Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
132,971
10,775,689,060
IssuesEvent
2019-11-03 15:54:51
dbeaver/dbeaver
https://api.github.com/repos/dbeaver/dbeaver
closed
NPE appears after changing active datasource [D8]
TestQuality bug xp:minor
#### System information: - Operating system (distribution) and version win 10 x64 - DBeaver version 6.2.4 #### Describe the problem you're observing: Npe appears when I switch in SQL editor from current DB (e.g. hana) to another (e.g. PostgreSQL) and back to first DB (hana). The issue has no affection for UI #### Steps to reproduce, if exist: 1. Execute some script 2. Change datasource to another (e.g. Hana - PostgreSQL) 3. Try to execute script - get error that item doesn't exist 4. Switch back to the previous DB - try to execute the script. - The script is successfully executed, but there is NPE in Error log [npe_changedatasource.log](https://github.com/dbeaver/dbeaver/files/3798444/npe_changedatasource.log) #### Include any warning/errors/backtraces from the logs <!-- Please, find the short guide how to find logs here: https://github.com/dbeaver/dbeaver/wiki/Log-files -->
1.0
NPE appears after changing active datasource [D8] - #### System information: - Operating system (distribution) and version win 10 x64 - DBeaver version 6.2.4 #### Describe the problem you're observing: Npe appears when I switch in SQL editor from current DB (e.g. hana) to another (e.g. PostgreSQL) and back to first DB (hana). The issue has no affection for UI #### Steps to reproduce, if exist: 1. Execute some script 2. Change datasource to another (e.g. Hana - PostgreSQL) 3. Try to execute script - get error that item doesn't exist 4. Switch back to the previous DB - try to execute the script. - The script is successfully executed, but there is NPE in Error log [npe_changedatasource.log](https://github.com/dbeaver/dbeaver/files/3798444/npe_changedatasource.log) #### Include any warning/errors/backtraces from the logs <!-- Please, find the short guide how to find logs here: https://github.com/dbeaver/dbeaver/wiki/Log-files -->
test
npe appears after changing active datasource system information operating system distribution and version win dbeaver version describe the problem you re observing npe appears when i switch in sql editor from current db e g hana to another e g postgresql and back to first db hana the issue has no affection for ui steps to reproduce if exist execute some script change datasource to another e g hana postgresql try to execute script get error that item doesn t exist switch back to the previous db try to execute the script the script is successfully executed but there is npe in error log include any warning errors backtraces from the logs
1
234,474
7,721,547,570
IssuesEvent
2018-05-24 06:01:51
FirecraftMC/FirecraftCore
https://api.github.com/repos/FirecraftMC/FirecraftCore
closed
Fully Implement Subranks
enhancement low-priority
NOTES: Staff will have limited donor permissions, however they will be able to use the commands/permissions from any donor rank they have bought
1.0
Fully Implement Subranks - NOTES: Staff will have limited donor permissions, however they will be able to use the commands/permissions from any donor rank they have bought
non_test
fully implement subranks notes staff will have limited donor permissions however they will be able to use the commands permissions from any donor rank they have bought
0
20,603
30,606,711,756
IssuesEvent
2023-07-23 04:52:14
mebjas/html5-qrcode
https://api.github.com/repos/mebjas/html5-qrcode
opened
Compatibility - [OS] [Browser] - [What is not working]
compatibility
**Describe the bug** A clear and concise description of what the bug is. - What is not working and what is expected. **Describe the browser:** - OS: [e.g. iOS, Android, MacOs or Windows] - Browser [e.g. chrome, safari, edge, firefox] - Version [e.g. 22] **Screenshots** If applicable, add screenshots to help explain your problem. **Additional context** Add any other context about the problem here.
True
Compatibility - [OS] [Browser] - [What is not working] - **Describe the bug** A clear and concise description of what the bug is. - What is not working and what is expected. **Describe the browser:** - OS: [e.g. iOS, Android, MacOs or Windows] - Browser [e.g. chrome, safari, edge, firefox] - Version [e.g. 22] **Screenshots** If applicable, add screenshots to help explain your problem. **Additional context** Add any other context about the problem here.
non_test
compatibility describe the bug a clear and concise description of what the bug is what is not working and what is expected describe the browser os browser version screenshots if applicable add screenshots to help explain your problem additional context add any other context about the problem here
0
5,992
2,800,962,290
IssuesEvent
2015-05-13 13:28:32
acemod/ACE3
https://api.github.com/repos/acemod/ACE3
closed
Game Crashing with Medical Modules
needs testing
ACE3 Version: 3.x.x **Mods:** * @cba_a3 * @ace3 **Placed ACE3 Modules:** * *Add the list of modules you have placed on the map. Use 'None' if the error occurs without using any modules.* Advanced Medical Settings Medical Settings (Tried with everything except BLUFOR tracking and the Set Medical modules earlier with same issues) **Description:** *Add a detailed description of the error. This makes it easier for us to fix the issue.* When loading into a mission the bar gets about 70% through and then the following error pops up: Bad version 68 in p3d file 'z\ace\addons\medical\data\bandage.p3d **Steps to reproduce:** * *Add the steps needed to reproduce the issue.* Im guessing run Arma with CBA and ACE3 place down the two modules about, set the difficulty to advanced and enable advanced wounds. **Where did the issue occur?** *A possible answer might be "Multiplayer", "Singleplayer"* Singleplayer **RPT log file:** *Add a link (pastebin.com) to the client or server RPT file.* http://pastebin.com/gcFZpkkW
1.0
Game Crashing with Medical Modules - ACE3 Version: 3.x.x **Mods:** * @cba_a3 * @ace3 **Placed ACE3 Modules:** * *Add the list of modules you have placed on the map. Use 'None' if the error occurs without using any modules.* Advanced Medical Settings Medical Settings (Tried with everything except BLUFOR tracking and the Set Medical modules earlier with same issues) **Description:** *Add a detailed description of the error. This makes it easier for us to fix the issue.* When loading into a mission the bar gets about 70% through and then the following error pops up: Bad version 68 in p3d file 'z\ace\addons\medical\data\bandage.p3d **Steps to reproduce:** * *Add the steps needed to reproduce the issue.* Im guessing run Arma with CBA and ACE3 place down the two modules about, set the difficulty to advanced and enable advanced wounds. **Where did the issue occur?** *A possible answer might be "Multiplayer", "Singleplayer"* Singleplayer **RPT log file:** *Add a link (pastebin.com) to the client or server RPT file.* http://pastebin.com/gcFZpkkW
test
game crashing with medical modules version x x mods cba placed modules add the list of modules you have placed on the map use none if the error occurs without using any modules advanced medical settings medical settings tried with everything except blufor tracking and the set medical modules earlier with same issues description add a detailed description of the error this makes it easier for us to fix the issue when loading into a mission the bar gets about through and then the following error pops up bad version in file z ace addons medical data bandage steps to reproduce add the steps needed to reproduce the issue im guessing run arma with cba and place down the two modules about set the difficulty to advanced and enable advanced wounds where did the issue occur a possible answer might be multiplayer singleplayer singleplayer rpt log file add a link pastebin com to the client or server rpt file
1
54,404
6,385,613,941
IssuesEvent
2017-08-03 09:00:52
SatelliteQE/robottelo
https://api.github.com/repos/SatelliteQE/robottelo
opened
Failure's on tests.foreman.ui.test_organization due to different pages for two screnarios
6.3 High test-failure
Same as described in #4592 but for Organization.
1.0
Failure's on tests.foreman.ui.test_organization due to different pages for two screnarios - Same as described in #4592 but for Organization.
test
failure s on tests foreman ui test organization due to different pages for two screnarios same as described in but for organization
1
297,411
25,729,159,087
IssuesEvent
2022-12-07 18:51:33
opensearch-project/OpenSearch
https://api.github.com/repos/opensearch-project/OpenSearch
closed
[BUG] org.opensearch.repositories.s3.RepositoryS3ClientYamlTestSuiteIT/test {yaml=repository_s3/20_repository_permanent_credentials/Snapshot and Restore with repository-s3 using permanent credentials} flaky
bug flaky-test
org.opensearch.repositories.s3.RepositoryS3ClientYamlTestSuiteIT/test {yaml=repository_s3/20_repository_permanent_credentials/Snapshot and Restore with repository-s3 using permanent credentials} https://build.ci.opensearch.org/job/gradle-check/6782/ https://build.ci.opensearch.org/job/gradle-check/6779/ https://build.ci.opensearch.org/job/gradle-check/6778/ https://build.ci.opensearch.org/job/gradle-check/6766/ https://build.ci.opensearch.org/job/gradle-check/6751/ https://build.ci.opensearch.org/job/gradle-check/6750/
1.0
[BUG] org.opensearch.repositories.s3.RepositoryS3ClientYamlTestSuiteIT/test {yaml=repository_s3/20_repository_permanent_credentials/Snapshot and Restore with repository-s3 using permanent credentials} flaky - org.opensearch.repositories.s3.RepositoryS3ClientYamlTestSuiteIT/test {yaml=repository_s3/20_repository_permanent_credentials/Snapshot and Restore with repository-s3 using permanent credentials} https://build.ci.opensearch.org/job/gradle-check/6782/ https://build.ci.opensearch.org/job/gradle-check/6779/ https://build.ci.opensearch.org/job/gradle-check/6778/ https://build.ci.opensearch.org/job/gradle-check/6766/ https://build.ci.opensearch.org/job/gradle-check/6751/ https://build.ci.opensearch.org/job/gradle-check/6750/
test
org opensearch repositories test yaml repository repository permanent credentials snapshot and restore with repository using permanent credentials flaky org opensearch repositories test yaml repository repository permanent credentials snapshot and restore with repository using permanent credentials
1
166,005
14,017,473,112
IssuesEvent
2020-10-29 15:41:44
tskit-dev/tskit
https://api.github.com/repos/tskit-dev/tskit
opened
Track documentation views as metric
Infrastructure and tools documentation
In this [blog post](https://blog.dask.org/2020/01/14/estimating-users) Matthew Rocklin talks about how they estimate the number of users Dask has. The number of unique IPs looking at the documentation seems like quite a good metric to me. Can we do something similar for tskit-dev related projects? (Since we're moving away from RTD?)
1.0
Track documentation views as metric - In this [blog post](https://blog.dask.org/2020/01/14/estimating-users) Matthew Rocklin talks about how they estimate the number of users Dask has. The number of unique IPs looking at the documentation seems like quite a good metric to me. Can we do something similar for tskit-dev related projects? (Since we're moving away from RTD?)
non_test
track documentation views as metric in this matthew rocklin talks about how they estimate the number of users dask has the number of unique ips looking at the documentation seems like quite a good metric to me can we do something similar for tskit dev related projects since we re moving away from rtd
0
20,100
4,495,115,771
IssuesEvent
2016-08-31 09:02:33
radare/radare2
https://api.github.com/repos/radare/radare2
closed
`afu` command undocumented
anal documentation
While looking how to resize a function that radare/anal incorrectly delimited, I accidentally stumbled on an `afu` command that is handled in cmd_anal_fcn, but it is not documented. What is it supposed to do?
1.0
`afu` command undocumented - While looking how to resize a function that radare/anal incorrectly delimited, I accidentally stumbled on an `afu` command that is handled in cmd_anal_fcn, but it is not documented. What is it supposed to do?
non_test
afu command undocumented while looking how to resize a function that radare anal incorrectly delimited i accidentally stumbled on an afu command that is handled in cmd anal fcn but it is not documented what is it supposed to do
0
333,502
29,669,509,681
IssuesEvent
2023-06-11 08:19:06
cse110-sp23-group23/Zoltar
https://api.github.com/repos/cse110-sp23-group23/Zoltar
opened
Mute button test
End to End Tests
Requirement Description: - Toggle mute button test Any key challenges: - Steps to Implementing: - visuals.test.js Other: -
1.0
Mute button test - Requirement Description: - Toggle mute button test Any key challenges: - Steps to Implementing: - visuals.test.js Other: -
test
mute button test requirement description toggle mute button test any key challenges steps to implementing visuals test js other
1
85,209
16,615,102,835
IssuesEvent
2021-06-02 15:44:24
WISE-Community/WISE-Client
https://api.github.com/repos/WISE-Community/WISE-Client
closed
Separate authoring-tool into node/project-authoring
2 points Code Quality
Separate authoring-tool.module.ts into node-authoring.module.ts and project-authoring.module.ts
1.0
Separate authoring-tool into node/project-authoring - Separate authoring-tool.module.ts into node-authoring.module.ts and project-authoring.module.ts
non_test
separate authoring tool into node project authoring separate authoring tool module ts into node authoring module ts and project authoring module ts
0
19,984
5,961,561,208
IssuesEvent
2017-05-29 17:59:34
ggez/ggez
https://api.github.com/repos/ggez/ggez
closed
Drawing is *horribly* slow on Windows
enhancement help wanted [CODE]
Like, what on Linux or Mac runs at 130 fps instead runs at 6 fps. Good example program: https://github.com/icefoxen/ggj2017 at least as of commit 851f83782d0e4dd05cb6633fd4a4e1ea9f2d1376 I blame SDL2. Still! Should get to the bottom of it one way or another. Tested on: * Intel core i5 laptop with NVidia 5100M GPU running Windows 10 * Intel core i5 laptop with NVidia 425M GPU running Windows 7
1.0
Drawing is *horribly* slow on Windows - Like, what on Linux or Mac runs at 130 fps instead runs at 6 fps. Good example program: https://github.com/icefoxen/ggj2017 at least as of commit 851f83782d0e4dd05cb6633fd4a4e1ea9f2d1376 I blame SDL2. Still! Should get to the bottom of it one way or another. Tested on: * Intel core i5 laptop with NVidia 5100M GPU running Windows 10 * Intel core i5 laptop with NVidia 425M GPU running Windows 7
non_test
drawing is horribly slow on windows like what on linux or mac runs at fps instead runs at fps good example program at least as of commit i blame still should get to the bottom of it one way or another tested on intel core laptop with nvidia gpu running windows intel core laptop with nvidia gpu running windows
0
11,186
3,185,520,441
IssuesEvent
2015-09-28 05:46:53
e-government-ua/i
https://api.github.com/repos/e-government-ua/i
closed
Настроить Томкат в Neatbeans
active test
https://github.com/e-government-ua/i/wiki/%D0%A3%D1%81%D1%82%D0%B0%D0%BD%D0%BE%D0%B2%D0%BA%D0%B0-%D0%B8-%D0%BD%D0%B0%D1%81%D1%82%D1%80%D0%BE%D0%B9%D0%BA%D0%B0-%D1%81%D0%B5%D1%80%D0%B2%D0%B5%D1%80%D0%B0-Tomcat-%28backend%29 На страничке "Установка и настройка сервера Tomcat (backend)" добавить описание в пункт 3 "Настроить Томкат в Neatbeans".
1.0
Настроить Томкат в Neatbeans - https://github.com/e-government-ua/i/wiki/%D0%A3%D1%81%D1%82%D0%B0%D0%BD%D0%BE%D0%B2%D0%BA%D0%B0-%D0%B8-%D0%BD%D0%B0%D1%81%D1%82%D1%80%D0%BE%D0%B9%D0%BA%D0%B0-%D1%81%D0%B5%D1%80%D0%B2%D0%B5%D1%80%D0%B0-Tomcat-%28backend%29 На страничке "Установка и настройка сервера Tomcat (backend)" добавить описание в пункт 3 "Настроить Томкат в Neatbeans".
test
настроить томкат в neatbeans на страничке установка и настройка сервера tomcat backend добавить описание в пункт настроить томкат в neatbeans
1
441,018
12,707,019,744
IssuesEvent
2020-06-23 08:13:36
luna/luna
https://api.github.com/repos/luna/luna
closed
Safe/Unsafe FFI Semantics
Category: Compiler Category: Semantics Category: Tooling Difficulty: Core Contributor Priority: High Type: Enhancement
Opening up some bikeshedding on the names for our safety annotations in Luna. The parser currently deals with the following: - `safe`: The FFI call can be executed on any thread. - `unsafe`: The FFI call must be executed on one thread. Chatting has also proposed `moveable` and `pinned` respectively, which I like a lot. I will implement whatever changes to the parser are necessary based on the results of the discussion here.
1.0
Safe/Unsafe FFI Semantics - Opening up some bikeshedding on the names for our safety annotations in Luna. The parser currently deals with the following: - `safe`: The FFI call can be executed on any thread. - `unsafe`: The FFI call must be executed on one thread. Chatting has also proposed `moveable` and `pinned` respectively, which I like a lot. I will implement whatever changes to the parser are necessary based on the results of the discussion here.
non_test
safe unsafe ffi semantics opening up some bikeshedding on the names for our safety annotations in luna the parser currently deals with the following safe the ffi call can be executed on any thread unsafe the ffi call must be executed on one thread chatting has also proposed moveable and pinned respectively which i like a lot i will implement whatever changes to the parser are necessary based on the results of the discussion here
0
3,422
2,671,941,248
IssuesEvent
2015-03-24 10:53:45
rssidlowski/Pollution_Source_Tracking
https://api.github.com/repos/rssidlowski/Pollution_Source_Tracking
closed
Zoom value after sample selected using identify feature
COBDev Ready for Testing enhancement moderate priority
From COB personnel: There appears to be a default zoom value that the map goes to after a sample is selected from the identify feature. In some cases where there are several clusters of separate sample points the map view will zoom out beyond the ability to distinguish the clusters. Is there a way to leave the zoom value as it was prior to sample selection? Before sample is selected: ![beforeidsamp](https://cloud.githubusercontent.com/assets/5039052/6491953/c3c972a6-c27e-11e4-8a49-ccc2a14b2c11.jpg) After selecting sample. ![afteridsamp](https://cloud.githubusercontent.com/assets/5039052/6491957/cb199cf2-c27e-11e4-83d5-64c68986733b.jpg)
1.0
Zoom value after sample selected using identify feature - From COB personnel: There appears to be a default zoom value that the map goes to after a sample is selected from the identify feature. In some cases where there are several clusters of separate sample points the map view will zoom out beyond the ability to distinguish the clusters. Is there a way to leave the zoom value as it was prior to sample selection? Before sample is selected: ![beforeidsamp](https://cloud.githubusercontent.com/assets/5039052/6491953/c3c972a6-c27e-11e4-8a49-ccc2a14b2c11.jpg) After selecting sample. ![afteridsamp](https://cloud.githubusercontent.com/assets/5039052/6491957/cb199cf2-c27e-11e4-83d5-64c68986733b.jpg)
test
zoom value after sample selected using identify feature from cob personnel there appears to be a default zoom value that the map goes to after a sample is selected from the identify feature in some cases where there are several clusters of separate sample points the map view will zoom out beyond the ability to distinguish the clusters is there a way to leave the zoom value as it was prior to sample selection before sample is selected after selecting sample
1
75,600
7,478,228,751
IssuesEvent
2018-04-04 10:54:05
FreeRDP/FreeRDP
https://api.github.com/repos/FreeRDP/FreeRDP
closed
Serial redirection reserved names `COM1-9`
client fixed-waiting-test wayland x11
hi, i've dowloaded sources from github and build deb with dpkg-buildpackage. Unfortunately i can't use option /serial with that build (if i omit it it works ok). here's my connection line: ```/opt/freerdp-nightly/bin/xfreerdp /serial:COM3,/dev/ttyS0 /printer:'laser' /clipboard /kbd:0x00000415 /load-balance-info:'TSV://MS Terminal Services Plugin.1.ATD' /cert-tofu /size:'90%' /auto-reconnect /v:XXXXXX /u:YYYY [16:06:28:457] [25852:25853] [INFO][com.freerdp.client.common.cmdline] - loading channelEx rdpdr [16:06:28:459] [25852:25853] [INFO][com.freerdp.client.common.cmdline] - loading channelEx rdpsnd [16:06:28:460] [25852:25853] [INFO][com.freerdp.client.common.cmdline] - loading channelEx cliprdr Password: [16:06:33:933] [25852:25853] [INFO][com.freerdp.gdi] - Local framebuffer format PIXEL_FORMAT_BGRX32 [16:06:33:935] [25852:25853] [INFO][com.freerdp.gdi] - Remote framebuffer format PIXEL_FORMAT_RGB16 [16:06:33:960] [25852:25853] [INFO][com.winpr.clipboard] - initialized POSIX local file subsystem [16:06:33:968] [25852:25858] [INFO][com.freerdp.channels.rdpdr.client] - Loading device service serial [COM3] (static) [16:06:33:970] [25852:25858] [ERROR][com.freerdp.channels.serial.client] - DefineCommDevice failed! [16:06:33:971] [25852:25858] [ERROR][com.freerdp.channels.rdpdr.client] - devman_load_device_service failed with error 1359! [16:06:33:973] [25852:25858] [ERROR][com.freerdp.channels.rdpdr.client] - rdpdr_process_connect failed with error 1359! [16:06:33:974] [25852:25853] [ERROR][com.freerdp.core] - rdpdr_virtual_channel_client_thread reported an error. Error was 1359 [16:06:33:975] [25852:25853] [INFO][com.freerdp.client.x11] - Network disconnect! [16:06:33:975] [25852:25853] [INFO][com.freerdp.client.x11] - Attempting reconnect (1 of 20) [16:06:33:196] [25852:25853] [ERROR][com.freerdp.core] - rdpdr_virtual_channel_client_thread reported an error. Error was 1359 [16:06:33:197] [25852:25853] [INFO][com.freerdp.client.x11] - Network disconnect! [16:06:33:198] [25852:25853] [INFO][com.freerdp.client.x11] - Attempting reconnect (1 of 20) [16:06:33:199] [25852:25861] [INFO][com.freerdp.channels.rdpdr.client] - Loading device service serial [COM3] (static) [16:06:33:200] [25852:25861] [ERROR][com.freerdp.channels.serial.client] - DefineCommDevice failed! [16:06:33:200] [25852:25861] [ERROR][com.freerdp.channels.rdpdr.client] - devman_load_device_service failed with error 1359! [16:06:33:201] [25852:25861] [ERROR][com.freerdp.channels.rdpdr.client] - rdpdr_process_connect failed with error 1359! ASAN:SIGSEGV ================================================================= ==25852==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000000 (pc 0x7fe4635318da bp 0x7fe451e80650 sp 0x7fe451e80640 T1) #0 0x7fe4635318d9 in MessagePipe_PostQuit (/opt/freerdp-nightly/bin/../lib/libwinpr2.so.2+0xc08d9) #1 0x7fe4643a63e3 (/opt/freerdp-nightly/bin/../lib/libfreerdp-client2.so.2+0xb93e3) #2 0x7fe463eed9a4 (/opt/freerdp-nightly/bin/../lib/libfreerdp2.so.2+0x1499a4) #3 0x7fe463f02ebf (/opt/freerdp-nightly/bin/../lib/libfreerdp2.so.2+0x15eebf) #4 0x443528 (/opt/freerdp-nightly/bin/xfreerdp+0x443528) #5 0x7fe463578a31 (/opt/freerdp-nightly/bin/../lib/libwinpr2.so.2+0x107a31) #6 0x7fe462e916b9 in start_thread (/lib/x86_64-linux-gnu/libpthread.so.0+0x76b9) #7 0x7fe4631ae41c in clone (/lib/x86_64-linux-gnu/libc.so.6+0x10741c) AddressSanitizer can not provide additional info. SUMMARY: AddressSanitizer: SEGV ??:0 MessagePipe_PostQuit Thread T1 created by T0 here: #0 0x7fe465a91253 in pthread_create (/usr/lib/x86_64-linux-gnu/libasan.so.2+0x36253) #1 0x7fe4635786c4 (/opt/freerdp-nightly/bin/../lib/libwinpr2.so.2+0x1076c4) #2 0x7fe463578ed7 in CreateThread (/opt/freerdp-nightly/bin/../lib/libwinpr2.so.2+0x107ed7) #3 0x43dc7f (/opt/freerdp-nightly/bin/xfreerdp+0x43dc7f) #4 0x407985 (/opt/freerdp-nightly/bin/xfreerdp+0x407985) #5 0x7fe4630c782f in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x2082f) ==25852==ABORTING```
1.0
Serial redirection reserved names `COM1-9` - hi, i've dowloaded sources from github and build deb with dpkg-buildpackage. Unfortunately i can't use option /serial with that build (if i omit it it works ok). here's my connection line: ```/opt/freerdp-nightly/bin/xfreerdp /serial:COM3,/dev/ttyS0 /printer:'laser' /clipboard /kbd:0x00000415 /load-balance-info:'TSV://MS Terminal Services Plugin.1.ATD' /cert-tofu /size:'90%' /auto-reconnect /v:XXXXXX /u:YYYY [16:06:28:457] [25852:25853] [INFO][com.freerdp.client.common.cmdline] - loading channelEx rdpdr [16:06:28:459] [25852:25853] [INFO][com.freerdp.client.common.cmdline] - loading channelEx rdpsnd [16:06:28:460] [25852:25853] [INFO][com.freerdp.client.common.cmdline] - loading channelEx cliprdr Password: [16:06:33:933] [25852:25853] [INFO][com.freerdp.gdi] - Local framebuffer format PIXEL_FORMAT_BGRX32 [16:06:33:935] [25852:25853] [INFO][com.freerdp.gdi] - Remote framebuffer format PIXEL_FORMAT_RGB16 [16:06:33:960] [25852:25853] [INFO][com.winpr.clipboard] - initialized POSIX local file subsystem [16:06:33:968] [25852:25858] [INFO][com.freerdp.channels.rdpdr.client] - Loading device service serial [COM3] (static) [16:06:33:970] [25852:25858] [ERROR][com.freerdp.channels.serial.client] - DefineCommDevice failed! [16:06:33:971] [25852:25858] [ERROR][com.freerdp.channels.rdpdr.client] - devman_load_device_service failed with error 1359! [16:06:33:973] [25852:25858] [ERROR][com.freerdp.channels.rdpdr.client] - rdpdr_process_connect failed with error 1359! [16:06:33:974] [25852:25853] [ERROR][com.freerdp.core] - rdpdr_virtual_channel_client_thread reported an error. Error was 1359 [16:06:33:975] [25852:25853] [INFO][com.freerdp.client.x11] - Network disconnect! [16:06:33:975] [25852:25853] [INFO][com.freerdp.client.x11] - Attempting reconnect (1 of 20) [16:06:33:196] [25852:25853] [ERROR][com.freerdp.core] - rdpdr_virtual_channel_client_thread reported an error. Error was 1359 [16:06:33:197] [25852:25853] [INFO][com.freerdp.client.x11] - Network disconnect! [16:06:33:198] [25852:25853] [INFO][com.freerdp.client.x11] - Attempting reconnect (1 of 20) [16:06:33:199] [25852:25861] [INFO][com.freerdp.channels.rdpdr.client] - Loading device service serial [COM3] (static) [16:06:33:200] [25852:25861] [ERROR][com.freerdp.channels.serial.client] - DefineCommDevice failed! [16:06:33:200] [25852:25861] [ERROR][com.freerdp.channels.rdpdr.client] - devman_load_device_service failed with error 1359! [16:06:33:201] [25852:25861] [ERROR][com.freerdp.channels.rdpdr.client] - rdpdr_process_connect failed with error 1359! ASAN:SIGSEGV ================================================================= ==25852==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000000 (pc 0x7fe4635318da bp 0x7fe451e80650 sp 0x7fe451e80640 T1) #0 0x7fe4635318d9 in MessagePipe_PostQuit (/opt/freerdp-nightly/bin/../lib/libwinpr2.so.2+0xc08d9) #1 0x7fe4643a63e3 (/opt/freerdp-nightly/bin/../lib/libfreerdp-client2.so.2+0xb93e3) #2 0x7fe463eed9a4 (/opt/freerdp-nightly/bin/../lib/libfreerdp2.so.2+0x1499a4) #3 0x7fe463f02ebf (/opt/freerdp-nightly/bin/../lib/libfreerdp2.so.2+0x15eebf) #4 0x443528 (/opt/freerdp-nightly/bin/xfreerdp+0x443528) #5 0x7fe463578a31 (/opt/freerdp-nightly/bin/../lib/libwinpr2.so.2+0x107a31) #6 0x7fe462e916b9 in start_thread (/lib/x86_64-linux-gnu/libpthread.so.0+0x76b9) #7 0x7fe4631ae41c in clone (/lib/x86_64-linux-gnu/libc.so.6+0x10741c) AddressSanitizer can not provide additional info. SUMMARY: AddressSanitizer: SEGV ??:0 MessagePipe_PostQuit Thread T1 created by T0 here: #0 0x7fe465a91253 in pthread_create (/usr/lib/x86_64-linux-gnu/libasan.so.2+0x36253) #1 0x7fe4635786c4 (/opt/freerdp-nightly/bin/../lib/libwinpr2.so.2+0x1076c4) #2 0x7fe463578ed7 in CreateThread (/opt/freerdp-nightly/bin/../lib/libwinpr2.so.2+0x107ed7) #3 0x43dc7f (/opt/freerdp-nightly/bin/xfreerdp+0x43dc7f) #4 0x407985 (/opt/freerdp-nightly/bin/xfreerdp+0x407985) #5 0x7fe4630c782f in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x2082f) ==25852==ABORTING```
test
serial redirection reserved names hi i ve dowloaded sources from github and build deb with dpkg buildpackage unfortunately i can t use option serial with that build if i omit it it works ok here s my connection line opt freerdp nightly bin xfreerdp serial dev printer laser clipboard kbd load balance info tsv ms terminal services plugin atd cert tofu size auto reconnect v xxxxxx u yyyy loading channelex rdpdr loading channelex rdpsnd loading channelex cliprdr password local framebuffer format pixel format remote framebuffer format pixel format initialized posix local file subsystem loading device service serial static definecommdevice failed devman load device service failed with error rdpdr process connect failed with error rdpdr virtual channel client thread reported an error error was network disconnect attempting reconnect of rdpdr virtual channel client thread reported an error error was network disconnect attempting reconnect of loading device service serial static definecommdevice failed devman load device service failed with error rdpdr process connect failed with error asan sigsegv error addresssanitizer segv on unknown address pc bp sp in messagepipe postquit opt freerdp nightly bin lib so opt freerdp nightly bin lib libfreerdp so opt freerdp nightly bin lib so opt freerdp nightly bin lib so opt freerdp nightly bin xfreerdp opt freerdp nightly bin lib so in start thread lib linux gnu libpthread so in clone lib linux gnu libc so addresssanitizer can not provide additional info summary addresssanitizer segv messagepipe postquit thread created by here in pthread create usr lib linux gnu libasan so opt freerdp nightly bin lib so in createthread opt freerdp nightly bin lib so opt freerdp nightly bin xfreerdp opt freerdp nightly bin xfreerdp in libc start main lib linux gnu libc so aborting
1
414,458
27,987,912,849
IssuesEvent
2023-03-26 22:11:49
Witiko/markdown
https://api.github.com/repos/Witiko/markdown
closed
Write TUGboat 44:1 article about attributes and attribute contexts
documentation
In Markdown 2.22.0, we will support attribute context renderers for [headings][1], [bracketed spans and fenced divs][2], and [links, images, code spans, and fenced code][3]. We should write a TUGboat article that introduces the concept of attributes and attribute contexts and show how coders can style attributes with the Markdown package. [1]: https://github.com/Witiko/markdown/issues/91 [2]: https://github.com/Witiko/markdown/issues/126 [3]: https://github.com/Witiko/markdown/issues/123 [4]: https://github.com/Witiko/markdown/issues/232
1.0
Write TUGboat 44:1 article about attributes and attribute contexts - In Markdown 2.22.0, we will support attribute context renderers for [headings][1], [bracketed spans and fenced divs][2], and [links, images, code spans, and fenced code][3]. We should write a TUGboat article that introduces the concept of attributes and attribute contexts and show how coders can style attributes with the Markdown package. [1]: https://github.com/Witiko/markdown/issues/91 [2]: https://github.com/Witiko/markdown/issues/126 [3]: https://github.com/Witiko/markdown/issues/123 [4]: https://github.com/Witiko/markdown/issues/232
non_test
write tugboat article about attributes and attribute contexts in markdown we will support attribute context renderers for and we should write a tugboat article that introduces the concept of attributes and attribute contexts and show how coders can style attributes with the markdown package
0
17,660
3,633,109,909
IssuesEvent
2016-02-11 13:15:32
AtomLinter/linter-javac
https://api.github.com/repos/AtomLinter/linter-javac
closed
spawn E2BIG when linting
bug untested
I am getting spawn E2BIG errors when saving .java files in Atom. It might have something to do with the number of files in the project folder, as open a window including only a single folder from the project does not produce the error. #### Steps * Make changes to an open .java file * Save file #### Expected: Atom updates linter display without error #### Actual: No linting, following error displayed: <img width="453" alt="screen shot 2015-11-19 at 1 55 28 pm" src="https://cloud.githubusercontent.com/assets/870130/11282529/6164fe5a-8ec5-11e5-9306-616b7a5f0ec0.png"> #### Setup: Atom 1.2.3 Mac OS X 10.11.1
1.0
spawn E2BIG when linting - I am getting spawn E2BIG errors when saving .java files in Atom. It might have something to do with the number of files in the project folder, as open a window including only a single folder from the project does not produce the error. #### Steps * Make changes to an open .java file * Save file #### Expected: Atom updates linter display without error #### Actual: No linting, following error displayed: <img width="453" alt="screen shot 2015-11-19 at 1 55 28 pm" src="https://cloud.githubusercontent.com/assets/870130/11282529/6164fe5a-8ec5-11e5-9306-616b7a5f0ec0.png"> #### Setup: Atom 1.2.3 Mac OS X 10.11.1
test
spawn when linting i am getting spawn errors when saving java files in atom it might have something to do with the number of files in the project folder as open a window including only a single folder from the project does not produce the error steps make changes to an open java file save file expected atom updates linter display without error actual no linting following error displayed img width alt screen shot at pm src setup atom mac os x
1
251,236
21,448,222,178
IssuesEvent
2022-04-25 08:43:43
hazelcast/hazelcast-csharp-client
https://api.github.com/repos/hazelcast/hazelcast-csharp-client
closed
DNS Host name test [API-1247]
Type: Test Failure Type: Backport Priority: Normal Jira State: Active
Implement and validate `com.hazelcast.client.ClientRegressionWithRealNetworkTest.testConnectWithDNSHostnames` Backport from https://github.com/hazelcast/hazelcast/pull/11368
1.0
DNS Host name test [API-1247] - Implement and validate `com.hazelcast.client.ClientRegressionWithRealNetworkTest.testConnectWithDNSHostnames` Backport from https://github.com/hazelcast/hazelcast/pull/11368
test
dns host name test implement and validate com hazelcast client clientregressionwithrealnetworktest testconnectwithdnshostnames backport from
1
64,047
26,598,039,628
IssuesEvent
2023-01-23 13:56:10
flexion/ef-cms
https://api.github.com/repos/flexion/ef-cms
closed
Court: Designated Service Person
Delay Ship Indefinitely - No (5) Service on Parties
As the Court, in order to comply with Rule 21(b)(1)(D)(2), I need to designate a service party for Petitioner and designate a service party for Respondent. ## Pre-Conditions: * Case has been created. ## Acceptance Criteria: * There must always be a designated service person for each party. * If petitioner is unrepresented, they are the designated service party * The first attorney on each side (Petitioner and Respondent) should be the default designated service person for that party * If petitioner has more than 1 counsel, a Court user needs to be able to change/designate which of Petitioner's counsel is the designated service person for Petitioner * If respondent has more than 1 counsel, a Court user needs to be able to change/designated which Respondent's counsel is the designated service person for Respondent
1.0
Court: Designated Service Person - As the Court, in order to comply with Rule 21(b)(1)(D)(2), I need to designate a service party for Petitioner and designate a service party for Respondent. ## Pre-Conditions: * Case has been created. ## Acceptance Criteria: * There must always be a designated service person for each party. * If petitioner is unrepresented, they are the designated service party * The first attorney on each side (Petitioner and Respondent) should be the default designated service person for that party * If petitioner has more than 1 counsel, a Court user needs to be able to change/designate which of Petitioner's counsel is the designated service person for Petitioner * If respondent has more than 1 counsel, a Court user needs to be able to change/designated which Respondent's counsel is the designated service person for Respondent
non_test
court designated service person as the court in order to comply with rule b d i need to designate a service party for petitioner and designate a service party for respondent pre conditions case has been created acceptance criteria there must always be a designated service person for each party if petitioner is unrepresented they are the designated service party the first attorney on each side petitioner and respondent should be the default designated service person for that party if petitioner has more than counsel a court user needs to be able to change designate which of petitioner s counsel is the designated service person for petitioner if respondent has more than counsel a court user needs to be able to change designated which respondent s counsel is the designated service person for respondent
0
56,562
3,080,253,843
IssuesEvent
2015-08-21 20:59:08
pavel-pimenov/flylinkdc-r5xx
https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx
closed
Падение при использовании мастера быстрой настройки.
bug imported Priority-High
_From [kotyar...@gmail.com](https://code.google.com/u/110049176879914219675/) on May 25, 2012 05:57:59_ r501 build 9474 Unhandled exception at 0x75a09673: Code 0xe06d7363 FlylinkDC++ r501 build 9474 startup on machine with: Number of processors: 2. Page size: 4096 Bytes. Processor type: x86. Memory config: There is 49 percent of memory in use. There are 2047 MB total of physical memory. There are 1036 MB free of physical memory. Running in Windows native (NT version 6.1). Текущее состояние системы: Memory config: There is 55 percent of memory in use. There are 2,00 ГБ total of physical memory. There are 912,99 МБ free of physical memory. Частота процессора: 2000,02 MHz на вин7, чистая установка, трижды вылетел при первой настройке на этапе определения айпишника _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=759_
1.0
Падение при использовании мастера быстрой настройки. - _From [kotyar...@gmail.com](https://code.google.com/u/110049176879914219675/) on May 25, 2012 05:57:59_ r501 build 9474 Unhandled exception at 0x75a09673: Code 0xe06d7363 FlylinkDC++ r501 build 9474 startup on machine with: Number of processors: 2. Page size: 4096 Bytes. Processor type: x86. Memory config: There is 49 percent of memory in use. There are 2047 MB total of physical memory. There are 1036 MB free of physical memory. Running in Windows native (NT version 6.1). Текущее состояние системы: Memory config: There is 55 percent of memory in use. There are 2,00 ГБ total of physical memory. There are 912,99 МБ free of physical memory. Частота процессора: 2000,02 MHz на вин7, чистая установка, трижды вылетел при первой настройке на этапе определения айпишника _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=759_
non_test
падение при использовании мастера быстрой настройки from on may build unhandled exception at code flylinkdc build startup on machine with number of processors page size bytes processor type memory config there is percent of memory in use there are mb total of physical memory there are mb free of physical memory running in windows native nt version текущее состояние системы memory config there is percent of memory in use there are гб total of physical memory there are мб free of physical memory частота процессора mhz на чистая установка трижды вылетел при первой настройке на этапе определения айпишника original issue
0
86,001
16,795,665,324
IssuesEvent
2021-06-16 02:52:41
GameDev-One/BattleofRhiannonNetwork
https://api.github.com/repos/GameDev-One/BattleofRhiannonNetwork
closed
Enemy does not react when being damaged
Art Code Design
The enemy does not react to being hit. Recommend having some sort of recoil and flash VFX to indicate that damage was taken.
1.0
Enemy does not react when being damaged - The enemy does not react to being hit. Recommend having some sort of recoil and flash VFX to indicate that damage was taken.
non_test
enemy does not react when being damaged the enemy does not react to being hit recommend having some sort of recoil and flash vfx to indicate that damage was taken
0
593,828
18,017,942,112
IssuesEvent
2021-09-16 15:46:27
pnxenopoulos/csgo
https://api.github.com/repos/pnxenopoulos/csgo
closed
Competitive Rank
Feature Request Low Priority
I would like to add the ranks of each player when i get the parsed json file. I asked the author of the demoinfocs-golang if this is implemented and he said that it is already there. [Code](https://pkg.go.dev/github.com/markus-wa/demoinfocs-golang/v2@v2.8.1/pkg/demoinfocs/events#RankUpdate) **Describe the solution you'd like** Maybe at the end of the array the relevant rank information for each player
1.0
Competitive Rank - I would like to add the ranks of each player when i get the parsed json file. I asked the author of the demoinfocs-golang if this is implemented and he said that it is already there. [Code](https://pkg.go.dev/github.com/markus-wa/demoinfocs-golang/v2@v2.8.1/pkg/demoinfocs/events#RankUpdate) **Describe the solution you'd like** Maybe at the end of the array the relevant rank information for each player
non_test
competitive rank i would like to add the ranks of each player when i get the parsed json file i asked the author of the demoinfocs golang if this is implemented and he said that it is already there describe the solution you d like maybe at the end of the array the relevant rank information for each player
0
714,665
24,569,654,396
IssuesEvent
2022-10-13 07:37:24
AlphaWallet/alpha-wallet-ios
https://api.github.com/repos/AlphaWallet/alpha-wallet-ios
opened
Can't connect using WalletConnect for minting with DevCon ticket attestation
High Priority
1. Visit the attestation URL in macOS Safari 2. In the same session, click "NFTs" in the top navigation bar 3. Click "Mint Arbitrum One NFT" 4. Dot click "Mint NFT" 5. Click "WALLET CONNECT" (it doesn't seem to matter which chain is enabled because the actionsheet to start the session doesn't appear at all for me) 6. Try to connect Don't click "Mint NFT" at any time :) It times out when I tried it a few times. Also have tried this: A. It works most of the time on Android AlphaWallet. Fails a few times B. It works most of the time on another wallet app. Fails a few times C. It works https://test.walletconnect.org the one time I tried (and accidentally mint 😭 a token on Polygon) D. AlphaWallet iOS works if I connect to http://example.walletconnect.org E. It doesn't work even if I comment out the timeout and wait
1.0
Can't connect using WalletConnect for minting with DevCon ticket attestation - 1. Visit the attestation URL in macOS Safari 2. In the same session, click "NFTs" in the top navigation bar 3. Click "Mint Arbitrum One NFT" 4. Dot click "Mint NFT" 5. Click "WALLET CONNECT" (it doesn't seem to matter which chain is enabled because the actionsheet to start the session doesn't appear at all for me) 6. Try to connect Don't click "Mint NFT" at any time :) It times out when I tried it a few times. Also have tried this: A. It works most of the time on Android AlphaWallet. Fails a few times B. It works most of the time on another wallet app. Fails a few times C. It works https://test.walletconnect.org the one time I tried (and accidentally mint 😭 a token on Polygon) D. AlphaWallet iOS works if I connect to http://example.walletconnect.org E. It doesn't work even if I comment out the timeout and wait
non_test
can t connect using walletconnect for minting with devcon ticket attestation visit the attestation url in macos safari in the same session click nfts in the top navigation bar click mint arbitrum one nft dot click mint nft click wallet connect it doesn t seem to matter which chain is enabled because the actionsheet to start the session doesn t appear at all for me try to connect don t click mint nft at any time it times out when i tried it a few times also have tried this a it works most of the time on android alphawallet fails a few times b it works most of the time on another wallet app fails a few times c it works the one time i tried and accidentally mint 😭 a token on polygon d alphawallet ios works if i connect to e it doesn t work even if i comment out the timeout and wait
0
46,624
24,631,476,104
IssuesEvent
2022-10-17 02:44:52
HypothesisWorks/hypothesis
https://api.github.com/repos/HypothesisWorks/hypothesis
opened
Efficient strategy for `st.text(...).filter(str.isidentifier)`
enhancement performance
As a follow-up to https://github.com/HypothesisWorks/hypothesis/issues/2693#issuecomment-823710924 and #3134, I'd like to return an efficient strategy for `st.text(...).filter(str.isidentifier)`. Adapting https://github.com/Zac-HD/hypothesmith/blob/85358991f8498db489569e81ac9dc9049c75773f/src/hypothesmith/syntactic.py#L39-L56 should make this pretty easy, even with the slight complication of incorporating the restrictions of the `alphabet=` strategy. We could optionally add similar support for the other `str.isX` methods, which mostly impose simpler constraints and could be satisfied more efficiently by replacing the `self.elements` strategy (e.g. `str.isspace` requires all-whitespace chars), and for some adding the filter on at the end (e.g. `str.isupper` is satisfied if all cased characters in the string are uppercase and there is at least one cased character)
True
Efficient strategy for `st.text(...).filter(str.isidentifier)` - As a follow-up to https://github.com/HypothesisWorks/hypothesis/issues/2693#issuecomment-823710924 and #3134, I'd like to return an efficient strategy for `st.text(...).filter(str.isidentifier)`. Adapting https://github.com/Zac-HD/hypothesmith/blob/85358991f8498db489569e81ac9dc9049c75773f/src/hypothesmith/syntactic.py#L39-L56 should make this pretty easy, even with the slight complication of incorporating the restrictions of the `alphabet=` strategy. We could optionally add similar support for the other `str.isX` methods, which mostly impose simpler constraints and could be satisfied more efficiently by replacing the `self.elements` strategy (e.g. `str.isspace` requires all-whitespace chars), and for some adding the filter on at the end (e.g. `str.isupper` is satisfied if all cased characters in the string are uppercase and there is at least one cased character)
non_test
efficient strategy for st text filter str isidentifier as a follow up to and i d like to return an efficient strategy for st text filter str isidentifier adapting should make this pretty easy even with the slight complication of incorporating the restrictions of the alphabet strategy we could optionally add similar support for the other str isx methods which mostly impose simpler constraints and could be satisfied more efficiently by replacing the self elements strategy e g str isspace requires all whitespace chars and for some adding the filter on at the end e g str isupper is satisfied if all cased characters in the string are uppercase and there is at least one cased character
0
512,208
14,890,271,512
IssuesEvent
2021-01-20 22:45:50
woocommerce/woocommerce-admin
https://api.github.com/repos/woocommerce/woocommerce-admin
opened
[1.9.0-rc.1] Home Screen Fatal
priority: critical
When starting the tasks list, an uninitialised `trackedCompletedTasks` prop is returned as `false` instead of an array causing a fatal error on the home screen. https://github.com/woocommerce/woocommerce-admin/blob/41970b8fcdd859a44aa5d15fe112e76e8c593e00/client/header/activity-panel/index.js#L312 <img width="1659" alt="Screen Shot 2021-01-21 at 11 42 20 AM" src="https://user-images.githubusercontent.com/1922453/105250046-f3232980-5bdd-11eb-9248-e4da49bcdad6.png"> Looks like this code was added in https://github.com/woocommerce/woocommerce-admin/pull/5826 cc @louwie17 ## To Reproduce * New site, click on skip onboarding details * Go to WooCommerce->home * Click on “Add my products” * Click on WooCommerce->home again * Click on “Add my products” again. * See blank screen
1.0
[1.9.0-rc.1] Home Screen Fatal - When starting the tasks list, an uninitialised `trackedCompletedTasks` prop is returned as `false` instead of an array causing a fatal error on the home screen. https://github.com/woocommerce/woocommerce-admin/blob/41970b8fcdd859a44aa5d15fe112e76e8c593e00/client/header/activity-panel/index.js#L312 <img width="1659" alt="Screen Shot 2021-01-21 at 11 42 20 AM" src="https://user-images.githubusercontent.com/1922453/105250046-f3232980-5bdd-11eb-9248-e4da49bcdad6.png"> Looks like this code was added in https://github.com/woocommerce/woocommerce-admin/pull/5826 cc @louwie17 ## To Reproduce * New site, click on skip onboarding details * Go to WooCommerce->home * Click on “Add my products” * Click on WooCommerce->home again * Click on “Add my products” again. * See blank screen
non_test
home screen fatal when starting the tasks list an uninitialised trackedcompletedtasks prop is returned as false instead of an array causing a fatal error on the home screen img width alt screen shot at am src looks like this code was added in cc to reproduce new site click on skip onboarding details go to woocommerce home click on “add my products” click on woocommerce home again click on “add my products” again see blank screen
0
139,666
11,275,602,444
IssuesEvent
2020-01-14 21:08:53
aliasrobotics/RVD
https://api.github.com/repos/aliasrobotics/RVD
opened
(error) Resource leak
bug cppcheck static analysis testing triage
```yaml { "system": "src/industrial_calibration/industrial_extrinsic_cal/src/nodes/wrist_cal_srv.cpp", "type": "bug", "exploitation": { "description": "", "exploitation-image": "", "exploitation-vector": "" }, "keywords": [ "cppcheck", "static analysis", "testing", "triage", "bug" ], "mitigation": { "description": "", "pull-request": "", "date-mitigation": "" }, "flaw": { "architectural-location": "N/A", "reported-by-relationship": "automatic", "phase": "testing", "reported-by": "Alias Robotics", "reproducibility": "always", "specificity": "N/A", "languages": "None", "detected-by-method": "testing static", "application": "N/A", "subsystem": "N/A", "trace": "", "package": "N/A", "reproduction": "See artifacts below (if available)", "date-detected": "2020-01-14 (21:08)", "detected-by": "Alias Robotics", "date-reported": "2020-01-14 (21:08)", "reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_ros_industrial/-/jobs/403099637/artifacts/download", "issue": "" }, "cve": "None", "description": "[src/industrial_calibration/industrial_extrinsic_cal/src/nodes/wrist_cal_srv.cpp:450]: (error) Resource leak: fp", "severity": { "cvss-vector": "", "rvss-score": 0, "rvss-vector": "", "cvss-score": 0, "severity-description": "" }, "title": "(error) Resource leak", "cwe": "None", "id": 1, "vendor": null, "links": "" } ```
1.0
(error) Resource leak - ```yaml { "system": "src/industrial_calibration/industrial_extrinsic_cal/src/nodes/wrist_cal_srv.cpp", "type": "bug", "exploitation": { "description": "", "exploitation-image": "", "exploitation-vector": "" }, "keywords": [ "cppcheck", "static analysis", "testing", "triage", "bug" ], "mitigation": { "description": "", "pull-request": "", "date-mitigation": "" }, "flaw": { "architectural-location": "N/A", "reported-by-relationship": "automatic", "phase": "testing", "reported-by": "Alias Robotics", "reproducibility": "always", "specificity": "N/A", "languages": "None", "detected-by-method": "testing static", "application": "N/A", "subsystem": "N/A", "trace": "", "package": "N/A", "reproduction": "See artifacts below (if available)", "date-detected": "2020-01-14 (21:08)", "detected-by": "Alias Robotics", "date-reported": "2020-01-14 (21:08)", "reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_ros_industrial/-/jobs/403099637/artifacts/download", "issue": "" }, "cve": "None", "description": "[src/industrial_calibration/industrial_extrinsic_cal/src/nodes/wrist_cal_srv.cpp:450]: (error) Resource leak: fp", "severity": { "cvss-vector": "", "rvss-score": 0, "rvss-vector": "", "cvss-score": 0, "severity-description": "" }, "title": "(error) Resource leak", "cwe": "None", "id": 1, "vendor": null, "links": "" } ```
test
error resource leak yaml system src industrial calibration industrial extrinsic cal src nodes wrist cal srv cpp type bug exploitation description exploitation image exploitation vector keywords cppcheck static analysis testing triage bug mitigation description pull request date mitigation flaw architectural location n a reported by relationship automatic phase testing reported by alias robotics reproducibility always specificity n a languages none detected by method testing static application n a subsystem n a trace package n a reproduction see artifacts below if available date detected detected by alias robotics date reported reproduction image gitlab com aliasrobotics offensive alurity pipelines active pipeline ros industrial jobs artifacts download issue cve none description error resource leak fp severity cvss vector rvss score rvss vector cvss score severity description title error resource leak cwe none id vendor null links
1
302,552
9,276,184,782
IssuesEvent
2019-03-20 01:46:50
evscott/Rambl
https://api.github.com/repos/evscott/Rambl
closed
Basic global styling
Low priority
Add basic global styles including: - header tags - p tags - links - background colours and Bootstrap colour override - generic flex wrapping container
1.0
Basic global styling - Add basic global styles including: - header tags - p tags - links - background colours and Bootstrap colour override - generic flex wrapping container
non_test
basic global styling add basic global styles including header tags p tags links background colours and bootstrap colour override generic flex wrapping container
0
148,669
13,243,593,871
IssuesEvent
2020-08-19 11:43:40
freenas/documentation
https://api.github.com/repos/freenas/documentation
closed
TrueNAS Core 12: Cloud Sync post-script
documentation
I'm missing some details for post-script. Is for example possible pass return code (or another vars) from task to this post-script? Source: https://www.truenas.com/docs/hub/tasks/scheduled/cloudsync/
1.0
TrueNAS Core 12: Cloud Sync post-script - I'm missing some details for post-script. Is for example possible pass return code (or another vars) from task to this post-script? Source: https://www.truenas.com/docs/hub/tasks/scheduled/cloudsync/
non_test
truenas core cloud sync post script i m missing some details for post script is for example possible pass return code or another vars from task to this post script source
0
225,071
17,791,753,399
IssuesEvent
2021-08-31 17:00:32
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
opened
FR: Record results of OpInfo reference tests and detect when numerics of an operator change
feature module: tests triaged
This would help identify the extent of changes to the operator set and require engineers appreciate the scope and impact of their change before committing it.
1.0
FR: Record results of OpInfo reference tests and detect when numerics of an operator change - This would help identify the extent of changes to the operator set and require engineers appreciate the scope and impact of their change before committing it.
test
fr record results of opinfo reference tests and detect when numerics of an operator change this would help identify the extent of changes to the operator set and require engineers appreciate the scope and impact of their change before committing it
1
273,884
8,554,221,030
IssuesEvent
2018-11-08 05:04:52
buttercup/buttercup-desktop
https://api.github.com/repos/buttercup/buttercup-desktop
closed
Cannot scroll at the column of entry
Effort: Medium Platform: Mac Priority: High Status: Available Type: Bug
OS: macOS 10.14 BC: 1.10.3 When the number of entries is greater than the one can be displayed in a window, mouse cannot scroll up/down the entries. But using trackpad is OK. ![2](https://user-images.githubusercontent.com/5687273/48061050-b0d9c180-e1f8-11e8-99f8-4abf6cf3cd98.jpg)
1.0
Cannot scroll at the column of entry - OS: macOS 10.14 BC: 1.10.3 When the number of entries is greater than the one can be displayed in a window, mouse cannot scroll up/down the entries. But using trackpad is OK. ![2](https://user-images.githubusercontent.com/5687273/48061050-b0d9c180-e1f8-11e8-99f8-4abf6cf3cd98.jpg)
non_test
cannot scroll at the column of entry os macos bc when the number of entries is greater than the one can be displayed in a window mouse cannot scroll up down the entries but using trackpad is ok
0
17,911
3,645,879,566
IssuesEvent
2016-02-15 16:23:52
backbee/backbee-standard
https://api.github.com/repos/backbee/backbee-standard
closed
[PAGE] ask user to validate modification before putting online a page that never has been validate from creation
bug To test
scenario Create a page Give a name and select a layout make modifications on this new page Go to the page tab and select "online" Save status In front office the page modifications are not made Solution : automatically validate new page creation and modifcation before putting online
1.0
[PAGE] ask user to validate modification before putting online a page that never has been validate from creation - scenario Create a page Give a name and select a layout make modifications on this new page Go to the page tab and select "online" Save status In front office the page modifications are not made Solution : automatically validate new page creation and modifcation before putting online
test
ask user to validate modification before putting online a page that never has been validate from creation scenario create a page give a name and select a layout make modifications on this new page go to the page tab and select online save status in front office the page modifications are not made solution automatically validate new page creation and modifcation before putting online
1
147,286
19,512,669,958
IssuesEvent
2021-12-29 02:56:27
ChoeMinji/deno-1.5.0
https://api.github.com/repos/ChoeMinji/deno-1.5.0
closed
CVE-2021-38191 (Medium) detected in tokio-0.2.22.crate - autoclosed
security vulnerability
## CVE-2021-38191 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tokio-0.2.22.crate</b></p></summary> <p>An event-driven, non-blocking I/O platform for writing asynchronous I/O backed applications. </p> <p>Library home page: <a href="https://crates.io/api/v1/crates/tokio/0.2.22/download">https://crates.io/api/v1/crates/tokio/0.2.22/download</a></p> <p> Dependency Hierarchy: - test_plugin-0.0.1 (Root Library) - test_util-0.1.0 - :x: **tokio-0.2.22.crate** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ChoeMinji/deno-1.5.0/commit/6bd9a93e55faf7abd43040d83fa5bb6fcbd55f5c">6bd9a93e55faf7abd43040d83fa5bb6fcbd55f5c</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the tokio crate before 1.8.1 for Rust. Upon a JoinHandle::abort, a Task may be dropped in the wrong thread. <p>Publish Date: 2021-08-08 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-38191>CVE-2021-38191</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://rustsec.org/advisories/RUSTSEC-2021-0072.html">https://rustsec.org/advisories/RUSTSEC-2021-0072.html</a></p> <p>Release Date: 2021-08-08</p> <p>Fix Resolution: tokio - 1.5.1,1.6.3,1.7.2, 1.8.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-38191 (Medium) detected in tokio-0.2.22.crate - autoclosed - ## CVE-2021-38191 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tokio-0.2.22.crate</b></p></summary> <p>An event-driven, non-blocking I/O platform for writing asynchronous I/O backed applications. </p> <p>Library home page: <a href="https://crates.io/api/v1/crates/tokio/0.2.22/download">https://crates.io/api/v1/crates/tokio/0.2.22/download</a></p> <p> Dependency Hierarchy: - test_plugin-0.0.1 (Root Library) - test_util-0.1.0 - :x: **tokio-0.2.22.crate** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ChoeMinji/deno-1.5.0/commit/6bd9a93e55faf7abd43040d83fa5bb6fcbd55f5c">6bd9a93e55faf7abd43040d83fa5bb6fcbd55f5c</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the tokio crate before 1.8.1 for Rust. Upon a JoinHandle::abort, a Task may be dropped in the wrong thread. <p>Publish Date: 2021-08-08 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-38191>CVE-2021-38191</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://rustsec.org/advisories/RUSTSEC-2021-0072.html">https://rustsec.org/advisories/RUSTSEC-2021-0072.html</a></p> <p>Release Date: 2021-08-08</p> <p>Fix Resolution: tokio - 1.5.1,1.6.3,1.7.2, 1.8.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in tokio crate autoclosed cve medium severity vulnerability vulnerable library tokio crate an event driven non blocking i o platform for writing asynchronous i o backed applications library home page a href dependency hierarchy test plugin root library test util x tokio crate vulnerable library found in head commit a href found in base branch master vulnerability details an issue was discovered in the tokio crate before for rust upon a joinhandle abort a task may be dropped in the wrong thread publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tokio step up your open source security game with whitesource
0
561,557
16,619,271,748
IssuesEvent
2021-06-02 21:18:01
googleapis/rules_gapic
https://api.github.com/repos/googleapis/rules_gapic
opened
feat: do not automatically generate gapic_library targets for protos that don't have a service
priority: p3 type: feature request
We currently generate gapic_library targets in the BUILD file for all protos, including protos that only contain messages (e.g. do not contain a service). We do not need to do that for protos that do not contain services. We could either parse the proto itself for a service, or look at the yaml file for APIs being referenced to determine whether gapic_library targets need to be included.
1.0
feat: do not automatically generate gapic_library targets for protos that don't have a service - We currently generate gapic_library targets in the BUILD file for all protos, including protos that only contain messages (e.g. do not contain a service). We do not need to do that for protos that do not contain services. We could either parse the proto itself for a service, or look at the yaml file for APIs being referenced to determine whether gapic_library targets need to be included.
non_test
feat do not automatically generate gapic library targets for protos that don t have a service we currently generate gapic library targets in the build file for all protos including protos that only contain messages e g do not contain a service we do not need to do that for protos that do not contain services we could either parse the proto itself for a service or look at the yaml file for apis being referenced to determine whether gapic library targets need to be included
0
165,099
20,574,159,560
IssuesEvent
2022-03-04 01:26:35
renfei/renfei-java-sdk
https://api.github.com/repos/renfei/renfei-java-sdk
opened
WS-2022-0089 (High) detected in nokogiri-1.11.5.gem
security vulnerability
## WS-2022-0089 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nokogiri-1.11.5.gem</b></p></summary> <p>Nokogiri (鋸) makes it easy and painless to work with XML and HTML from Ruby. It provides a sensible, easy-to-understand API for reading, writing, modifying, and querying documents. It is fast and standards-compliant by relying on native parsers like libxml2 (C) and xerces (Java). </p> <p>Library home page: <a href="https://rubygems.org/gems/nokogiri-1.11.5.gem">https://rubygems.org/gems/nokogiri-1.11.5.gem</a></p> <p> Dependency Hierarchy: - w3c_validators-1.3.5.gem (Root Library) - :x: **nokogiri-1.11.5.gem** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Nokogiri before version 1.13.2 is vulnerable. <p>Publish Date: 2022-03-01 <p>URL: <a href=https://github.com/sparklemotion/nokogiri/commit/472913378794b8cae21751b0777205e7c0606a95>WS-2022-0089</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sparklemotion/nokogiri/security/advisories/GHSA-fq42-c5rg-92c2">https://github.com/sparklemotion/nokogiri/security/advisories/GHSA-fq42-c5rg-92c2</a></p> <p>Release Date: 2022-03-01</p> <p>Fix Resolution: nokogiri - v1.13.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2022-0089 (High) detected in nokogiri-1.11.5.gem - ## WS-2022-0089 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nokogiri-1.11.5.gem</b></p></summary> <p>Nokogiri (鋸) makes it easy and painless to work with XML and HTML from Ruby. It provides a sensible, easy-to-understand API for reading, writing, modifying, and querying documents. It is fast and standards-compliant by relying on native parsers like libxml2 (C) and xerces (Java). </p> <p>Library home page: <a href="https://rubygems.org/gems/nokogiri-1.11.5.gem">https://rubygems.org/gems/nokogiri-1.11.5.gem</a></p> <p> Dependency Hierarchy: - w3c_validators-1.3.5.gem (Root Library) - :x: **nokogiri-1.11.5.gem** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Nokogiri before version 1.13.2 is vulnerable. <p>Publish Date: 2022-03-01 <p>URL: <a href=https://github.com/sparklemotion/nokogiri/commit/472913378794b8cae21751b0777205e7c0606a95>WS-2022-0089</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sparklemotion/nokogiri/security/advisories/GHSA-fq42-c5rg-92c2">https://github.com/sparklemotion/nokogiri/security/advisories/GHSA-fq42-c5rg-92c2</a></p> <p>Release Date: 2022-03-01</p> <p>Fix Resolution: nokogiri - v1.13.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
ws high detected in nokogiri gem ws high severity vulnerability vulnerable library nokogiri gem nokogiri 鋸 makes it easy and painless to work with xml and html from ruby it provides a sensible easy to understand api for reading writing modifying and querying documents it is fast and standards compliant by relying on native parsers like c and xerces java library home page a href dependency hierarchy validators gem root library x nokogiri gem vulnerable library found in base branch master vulnerability details nokogiri before version is vulnerable publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution nokogiri step up your open source security game with whitesource
0
308,835
26,634,439,034
IssuesEvent
2023-01-24 20:36:33
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Manual test run on macOS (Intel) for 1.47.x - Release #4
tests OS/macOS QA/Yes release-notes/exclude OS/Desktop
### Installer - [x] Check signature: - [x] If macOS, using x64 binary run `spctl --assess --verbose` for the installed version and make sure it returns `accepted` - [x] If macOS, using universal binary run `spctl --assess --verbose` for the installed version and make sure it returns `accepted` - [ ] If Windows right click on the `brave_installer-x64.exe` and go to Properties, go to the Digital Signatures tab and double click on the signature. Make sure it says "The digital signature is OK" in the popup window ### Widevine - [x] Verify `Widevine Notification` is shown when you visit Netflix for the first time - [x] Test that you can stream on Netflix on a fresh profile after installing Widevine - [x] Verify `Widevine Notification` is shown when you visit HBO Max for the first time - [x] Test that you can stream on HBO Max on a fresh profile after installing Widevine - [x] If macOS, run the above Widevine tests for both `x64` and `universal` builds ### Rewards - [x] Verify that none of the reward endpoints are being contacted when a user visits a media publisher (`youtube.com`, `reddit.com`, `twitter.com`, `github.com`) and hasn't interacted with rewards - [x] Verify that `rewards.brave.com`, `pcdn.brave.com`, `grant.rewards.brave.com` or `api.rewards.brave.com` are not being contacted - [x] Verify Rewards balance shows correct BAT and USD value - [x] Verify actions taken (claiming grant, tipping, auto-contribute) display in panel transactions list - [x] Verify when you click on the BR panel while on a site, the panel displays site specific information (site favicon, domain, attention %) - [x] Verify you are able to make one-time tip and they display in tips panel - [x] Verify you are able to make recurring tip and they display in tips panel - [x] Verify you can tip a verified publisher - [x] Verify you can tip a verified YouTube creator - [x] Verify you are able to perform a contribution - [x] Verify if you disable auto-contribute you are still able to tip regular sites and YouTube creators ### TLS Pinning - [x] Visit https://ssl-pinning.someblog.org/ and verify a pinning error is displayed - [x] Visit https://pinning-test.badssl.com/ and verify a pinning error is **not** displayed ## Update tests - [x] Verify visiting `brave://settings/help` triggers update check - [x] Verify once update is downloaded, prompts to `Relaunch` to install update #### Startup & Components - [x] Verify that Brave is only contacting `*.brave.com` endpoints on first launch using either `Charles Proxy`, `Fiddler`, `Wireshark` or `LittleSnitch` (or a similar application) - [x] Verify that opening a NTP doesn't trigger any outbound connections related to widgets without user interaction - [x] Delete Adblock folder from browser profile and restart browser. Visit `brave://components` and verify `Brave Ad Block Updater` downloads and update the component. Repeat for all Brave components ### Upgrade - [x] Make sure that data from the last version appears in the new version OK - [x] Ensure that `brave://version` lists the expected Brave & Chromium versions - [x] With data from the last version, verify that - [x] Bookmarks on the bookmark toolbar and bookmark folders can be opened - [x] Cookies are preserved - [x] Installed extensions are retained and work correctly - [x] Opened tabs can be reloaded - [x] Stored passwords are preserved - [x] Sync chain created in previous version is retained - [x] Social media blocking buttons changes are retained - [x] Rewards - [x] BAT balance is retained - [x] Auto-contribute list is retained - [x] Both Tips and Monthly Contributions are retained - [x] Panel transactions list is retained - [x] Changes to rewards settings are retained - [x] Ensure that Auto Contribute is not being enabled when upgrading to a new version if AC was disabled - [x] Ads - [x] Both `Estimated pending rewards` & `Ad notifications received this month` are retained - [x] Changes to ads settings are retained - [x] Ensure that ads are not being enabled when upgrading to a new version if they were disabled - [x] Ensure that ads are not disabled when upgrading to a new version if they were enabled
1.0
Manual test run on macOS (Intel) for 1.47.x - Release #4 - ### Installer - [x] Check signature: - [x] If macOS, using x64 binary run `spctl --assess --verbose` for the installed version and make sure it returns `accepted` - [x] If macOS, using universal binary run `spctl --assess --verbose` for the installed version and make sure it returns `accepted` - [ ] If Windows right click on the `brave_installer-x64.exe` and go to Properties, go to the Digital Signatures tab and double click on the signature. Make sure it says "The digital signature is OK" in the popup window ### Widevine - [x] Verify `Widevine Notification` is shown when you visit Netflix for the first time - [x] Test that you can stream on Netflix on a fresh profile after installing Widevine - [x] Verify `Widevine Notification` is shown when you visit HBO Max for the first time - [x] Test that you can stream on HBO Max on a fresh profile after installing Widevine - [x] If macOS, run the above Widevine tests for both `x64` and `universal` builds ### Rewards - [x] Verify that none of the reward endpoints are being contacted when a user visits a media publisher (`youtube.com`, `reddit.com`, `twitter.com`, `github.com`) and hasn't interacted with rewards - [x] Verify that `rewards.brave.com`, `pcdn.brave.com`, `grant.rewards.brave.com` or `api.rewards.brave.com` are not being contacted - [x] Verify Rewards balance shows correct BAT and USD value - [x] Verify actions taken (claiming grant, tipping, auto-contribute) display in panel transactions list - [x] Verify when you click on the BR panel while on a site, the panel displays site specific information (site favicon, domain, attention %) - [x] Verify you are able to make one-time tip and they display in tips panel - [x] Verify you are able to make recurring tip and they display in tips panel - [x] Verify you can tip a verified publisher - [x] Verify you can tip a verified YouTube creator - [x] Verify you are able to perform a contribution - [x] Verify if you disable auto-contribute you are still able to tip regular sites and YouTube creators ### TLS Pinning - [x] Visit https://ssl-pinning.someblog.org/ and verify a pinning error is displayed - [x] Visit https://pinning-test.badssl.com/ and verify a pinning error is **not** displayed ## Update tests - [x] Verify visiting `brave://settings/help` triggers update check - [x] Verify once update is downloaded, prompts to `Relaunch` to install update #### Startup & Components - [x] Verify that Brave is only contacting `*.brave.com` endpoints on first launch using either `Charles Proxy`, `Fiddler`, `Wireshark` or `LittleSnitch` (or a similar application) - [x] Verify that opening a NTP doesn't trigger any outbound connections related to widgets without user interaction - [x] Delete Adblock folder from browser profile and restart browser. Visit `brave://components` and verify `Brave Ad Block Updater` downloads and update the component. Repeat for all Brave components ### Upgrade - [x] Make sure that data from the last version appears in the new version OK - [x] Ensure that `brave://version` lists the expected Brave & Chromium versions - [x] With data from the last version, verify that - [x] Bookmarks on the bookmark toolbar and bookmark folders can be opened - [x] Cookies are preserved - [x] Installed extensions are retained and work correctly - [x] Opened tabs can be reloaded - [x] Stored passwords are preserved - [x] Sync chain created in previous version is retained - [x] Social media blocking buttons changes are retained - [x] Rewards - [x] BAT balance is retained - [x] Auto-contribute list is retained - [x] Both Tips and Monthly Contributions are retained - [x] Panel transactions list is retained - [x] Changes to rewards settings are retained - [x] Ensure that Auto Contribute is not being enabled when upgrading to a new version if AC was disabled - [x] Ads - [x] Both `Estimated pending rewards` & `Ad notifications received this month` are retained - [x] Changes to ads settings are retained - [x] Ensure that ads are not being enabled when upgrading to a new version if they were disabled - [x] Ensure that ads are not disabled when upgrading to a new version if they were enabled
test
manual test run on macos intel for x release installer check signature if macos using binary run spctl assess verbose for the installed version and make sure it returns accepted if macos using universal binary run spctl assess verbose for the installed version and make sure it returns accepted if windows right click on the brave installer exe and go to properties go to the digital signatures tab and double click on the signature make sure it says the digital signature is ok in the popup window widevine verify widevine notification is shown when you visit netflix for the first time test that you can stream on netflix on a fresh profile after installing widevine verify widevine notification is shown when you visit hbo max for the first time test that you can stream on hbo max on a fresh profile after installing widevine if macos run the above widevine tests for both and universal builds rewards verify that none of the reward endpoints are being contacted when a user visits a media publisher youtube com reddit com twitter com github com and hasn t interacted with rewards verify that rewards brave com pcdn brave com grant rewards brave com or api rewards brave com are not being contacted verify rewards balance shows correct bat and usd value verify actions taken claiming grant tipping auto contribute display in panel transactions list verify when you click on the br panel while on a site the panel displays site specific information site favicon domain attention verify you are able to make one time tip and they display in tips panel verify you are able to make recurring tip and they display in tips panel verify you can tip a verified publisher verify you can tip a verified youtube creator verify you are able to perform a contribution verify if you disable auto contribute you are still able to tip regular sites and youtube creators tls pinning visit and verify a pinning error is displayed visit and verify a pinning error is not displayed update tests verify visiting brave settings help triggers update check verify once update is downloaded prompts to relaunch to install update startup components verify that brave is only contacting brave com endpoints on first launch using either charles proxy fiddler wireshark or littlesnitch or a similar application verify that opening a ntp doesn t trigger any outbound connections related to widgets without user interaction delete adblock folder from browser profile and restart browser visit brave components and verify brave ad block updater downloads and update the component repeat for all brave components upgrade make sure that data from the last version appears in the new version ok ensure that brave version lists the expected brave chromium versions with data from the last version verify that bookmarks on the bookmark toolbar and bookmark folders can be opened cookies are preserved installed extensions are retained and work correctly opened tabs can be reloaded stored passwords are preserved sync chain created in previous version is retained social media blocking buttons changes are retained rewards bat balance is retained auto contribute list is retained both tips and monthly contributions are retained panel transactions list is retained changes to rewards settings are retained ensure that auto contribute is not being enabled when upgrading to a new version if ac was disabled ads both estimated pending rewards ad notifications received this month are retained changes to ads settings are retained ensure that ads are not being enabled when upgrading to a new version if they were disabled ensure that ads are not disabled when upgrading to a new version if they were enabled
1
297,440
25,731,600,818
IssuesEvent
2022-12-07 20:46:02
USEPA/haztrak
https://api.github.com/repos/USEPA/haztrak
closed
Update POC in services/api.ts
good first issue reactjs test
# 🐞 Bug Report The current source was converted from JS with `fetch` to TS with `axios` but the logic could us a serious makeover. convert to creating an instance of axios. Remove unnecessary parts. <!-- Please provide a clear and concise description, steps to reproduce, and any other information you believe may be useful. -->
1.0
Update POC in services/api.ts - # 🐞 Bug Report The current source was converted from JS with `fetch` to TS with `axios` but the logic could us a serious makeover. convert to creating an instance of axios. Remove unnecessary parts. <!-- Please provide a clear and concise description, steps to reproduce, and any other information you believe may be useful. -->
test
update poc in services api ts 🐞 bug report the current source was converted from js with fetch to ts with axios but the logic could us a serious makeover convert to creating an instance of axios remove unnecessary parts please provide a clear and concise description steps to reproduce and any other information you believe may be useful
1
296,298
25,543,126,672
IssuesEvent
2022-11-29 16:39:35
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
sql: TestSchemaUniqueColumnDropFailure failed
C-test-failure O-robot T-sql-schema branch-release-22.1
sql.TestSchemaUniqueColumnDropFailure [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=7747961&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=7747961&tab=artifacts#/) on release-22.1 @ [e12fd6a210d19ae33b6691f127e90ed660ac6035](https://github.com/cockroachdb/cockroach/commits/e12fd6a210d19ae33b6691f127e90ed660ac6035): ``` === RUN TestSchemaUniqueColumnDropFailure test_log_scope.go:79: test logs captured to: /artifacts/tmp/_tmp/233e6aa31ff7ad29c896f0479cb5d0c4/logTestSchemaUniqueColumnDropFailure3539421078 test_log_scope.go:80: use -show-logs to present logs inline schema_changer_test.go:2338: -- test log scope end -- schema_changer_test.go:2338: Leaked goroutine: goroutine 49292372 [IO wait]: internal/poll.runtime_pollWait(0x7f46753124e8, 0x72) GOROOT/src/runtime/netpoll.go:303 +0x85 internal/poll.(*pollDesc).wait(0xc013e4cd80, 0xc009a42620, 0x0) GOROOT/src/internal/poll/fd_poll_runtime.go:84 +0x32 internal/poll.(*pollDesc).waitRead(...) GOROOT/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0xc013e4cd80, {0xc009a42620, 0x1, 0x200}) GOROOT/src/internal/poll/fd_unix.go:167 +0x25a net.(*netFD).Read(0xc013e4cd80, {0xc009a42620, 0x7b1365, 0xc013e4cd80}) GOROOT/src/net/fd_posix.go:56 +0x29 net.(*conn).Read(0xc00b3f2448, {0xc009a42620, 0x7c5c0e, 0xc00b3f2448}) GOROOT/src/net/net.go:183 +0x45 io.ReadAtLeast({0x5e17d40, 0xc00b3f2448}, {0xc009a42620, 0x1, 0x200}, 0x1) GOROOT/src/io/io.go:328 +0x9a io.ReadFull(...) GOROOT/src/io/io.go:347 github.com/lib/pq.(*conn).ssl(0xc009a42600, 0xc000136020) github.com/lib/pq/external/com_github_lib_pq/conn.go:1108 +0x145 github.com/lib/pq.(*Connector).open(0xc01a37f158, {0x5eb5a78, 0xc000136020}) github.com/lib/pq/external/com_github_lib_pq/conn.go:361 +0x278 github.com/lib/pq.DialOpen({0x5e37788, 0xc01759d680}, {0xc00f26e000, 0xc000232250}) github.com/lib/pq/external/com_github_lib_pq/conn.go:328 +0x7b github.com/lib/pq.Open(...) github.com/lib/pq/external/com_github_lib_pq/conn.go:318 github.com/lib/pq.Driver.Open({}, {0xc00f26e000, 0x15c}) github.com/lib/pq/external/com_github_lib_pq/conn.go:56 +0x7a database/sql.dsnConnector.Connect(...) GOROOT/src/database/sql/sql.go:761 database/sql.(*DB).conn(0xc00ff16750, {0x5eb5a78, 0xc000136020}, 0x1) GOROOT/src/database/sql/sql.go:1364 +0x7ac database/sql.(*DB).exec(0xc0000ab718, {0x5eb5a78, 0xc000136020}, {0x4aac776, 0x20}, {0x0, 0x0, 0x0}, 0x0) GOROOT/src/database/sql/sql.go:1622 +0x5d database/sql.(*DB).ExecContext(0x10821c0, {0x5eb5a78, 0xc000136020}, {0x4aac776, 0x20}, {0x0, 0x0, 0x0}) GOROOT/src/database/sql/sql.go:1601 +0xed database/sql.(*DB).Exec(...) GOROOT/src/database/sql/sql.go:1618 github.com/cockroachdb/cockroach/pkg/sql_test.TestSchemaUniqueColumnDropFailure.func4() github.com/cockroachdb/cockroach/pkg/sql_test/pkg/sql/schema_changer_test.go:2312 +0x45 created by github.com/cockroachdb/cockroach/pkg/sql_test.TestSchemaUniqueColumnDropFailure github.com/cockroachdb/cockroach/pkg/sql_test/pkg/sql/schema_changer_test.go:2310 +0x685 --- FAIL: TestSchemaUniqueColumnDropFailure (6.62s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) Parameters in this failure: - TAGS=bazel,gss,deadlock </p> </details> /cc @cockroachdb/sql-schema <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestSchemaUniqueColumnDropFailure.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-21900
1.0
sql: TestSchemaUniqueColumnDropFailure failed - sql.TestSchemaUniqueColumnDropFailure [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=7747961&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=7747961&tab=artifacts#/) on release-22.1 @ [e12fd6a210d19ae33b6691f127e90ed660ac6035](https://github.com/cockroachdb/cockroach/commits/e12fd6a210d19ae33b6691f127e90ed660ac6035): ``` === RUN TestSchemaUniqueColumnDropFailure test_log_scope.go:79: test logs captured to: /artifacts/tmp/_tmp/233e6aa31ff7ad29c896f0479cb5d0c4/logTestSchemaUniqueColumnDropFailure3539421078 test_log_scope.go:80: use -show-logs to present logs inline schema_changer_test.go:2338: -- test log scope end -- schema_changer_test.go:2338: Leaked goroutine: goroutine 49292372 [IO wait]: internal/poll.runtime_pollWait(0x7f46753124e8, 0x72) GOROOT/src/runtime/netpoll.go:303 +0x85 internal/poll.(*pollDesc).wait(0xc013e4cd80, 0xc009a42620, 0x0) GOROOT/src/internal/poll/fd_poll_runtime.go:84 +0x32 internal/poll.(*pollDesc).waitRead(...) GOROOT/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0xc013e4cd80, {0xc009a42620, 0x1, 0x200}) GOROOT/src/internal/poll/fd_unix.go:167 +0x25a net.(*netFD).Read(0xc013e4cd80, {0xc009a42620, 0x7b1365, 0xc013e4cd80}) GOROOT/src/net/fd_posix.go:56 +0x29 net.(*conn).Read(0xc00b3f2448, {0xc009a42620, 0x7c5c0e, 0xc00b3f2448}) GOROOT/src/net/net.go:183 +0x45 io.ReadAtLeast({0x5e17d40, 0xc00b3f2448}, {0xc009a42620, 0x1, 0x200}, 0x1) GOROOT/src/io/io.go:328 +0x9a io.ReadFull(...) GOROOT/src/io/io.go:347 github.com/lib/pq.(*conn).ssl(0xc009a42600, 0xc000136020) github.com/lib/pq/external/com_github_lib_pq/conn.go:1108 +0x145 github.com/lib/pq.(*Connector).open(0xc01a37f158, {0x5eb5a78, 0xc000136020}) github.com/lib/pq/external/com_github_lib_pq/conn.go:361 +0x278 github.com/lib/pq.DialOpen({0x5e37788, 0xc01759d680}, {0xc00f26e000, 0xc000232250}) github.com/lib/pq/external/com_github_lib_pq/conn.go:328 +0x7b github.com/lib/pq.Open(...) github.com/lib/pq/external/com_github_lib_pq/conn.go:318 github.com/lib/pq.Driver.Open({}, {0xc00f26e000, 0x15c}) github.com/lib/pq/external/com_github_lib_pq/conn.go:56 +0x7a database/sql.dsnConnector.Connect(...) GOROOT/src/database/sql/sql.go:761 database/sql.(*DB).conn(0xc00ff16750, {0x5eb5a78, 0xc000136020}, 0x1) GOROOT/src/database/sql/sql.go:1364 +0x7ac database/sql.(*DB).exec(0xc0000ab718, {0x5eb5a78, 0xc000136020}, {0x4aac776, 0x20}, {0x0, 0x0, 0x0}, 0x0) GOROOT/src/database/sql/sql.go:1622 +0x5d database/sql.(*DB).ExecContext(0x10821c0, {0x5eb5a78, 0xc000136020}, {0x4aac776, 0x20}, {0x0, 0x0, 0x0}) GOROOT/src/database/sql/sql.go:1601 +0xed database/sql.(*DB).Exec(...) GOROOT/src/database/sql/sql.go:1618 github.com/cockroachdb/cockroach/pkg/sql_test.TestSchemaUniqueColumnDropFailure.func4() github.com/cockroachdb/cockroach/pkg/sql_test/pkg/sql/schema_changer_test.go:2312 +0x45 created by github.com/cockroachdb/cockroach/pkg/sql_test.TestSchemaUniqueColumnDropFailure github.com/cockroachdb/cockroach/pkg/sql_test/pkg/sql/schema_changer_test.go:2310 +0x685 --- FAIL: TestSchemaUniqueColumnDropFailure (6.62s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) Parameters in this failure: - TAGS=bazel,gss,deadlock </p> </details> /cc @cockroachdb/sql-schema <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestSchemaUniqueColumnDropFailure.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-21900
test
sql testschemauniquecolumndropfailure failed sql testschemauniquecolumndropfailure with on release run testschemauniquecolumndropfailure test log scope go test logs captured to artifacts tmp tmp test log scope go use show logs to present logs inline schema changer test go test log scope end schema changer test go leaked goroutine goroutine internal poll runtime pollwait goroot src runtime netpoll go internal poll polldesc wait goroot src internal poll fd poll runtime go internal poll polldesc waitread goroot src internal poll fd poll runtime go internal poll fd read goroot src internal poll fd unix go net netfd read goroot src net fd posix go net conn read goroot src net net go io readatleast goroot src io io go io readfull goroot src io io go github com lib pq conn ssl github com lib pq external com github lib pq conn go github com lib pq connector open github com lib pq external com github lib pq conn go github com lib pq dialopen github com lib pq external com github lib pq conn go github com lib pq open github com lib pq external com github lib pq conn go github com lib pq driver open github com lib pq external com github lib pq conn go database sql dsnconnector connect goroot src database sql sql go database sql db conn goroot src database sql sql go database sql db exec goroot src database sql sql go database sql db execcontext goroot src database sql sql go database sql db exec goroot src database sql sql go github com cockroachdb cockroach pkg sql test testschemauniquecolumndropfailure github com cockroachdb cockroach pkg sql test pkg sql schema changer test go created by github com cockroachdb cockroach pkg sql test testschemauniquecolumndropfailure github com cockroachdb cockroach pkg sql test pkg sql schema changer test go fail testschemauniquecolumndropfailure help see also parameters in this failure tags bazel gss deadlock cc cockroachdb sql schema jira issue crdb
1
288,405
24,903,187,949
IssuesEvent
2022-10-29 00:30:27
ValveSoftware/steam-for-linux
https://api.github.com/repos/ValveSoftware/steam-for-linux
closed
Dynamic collections allows scrolling way past its content
Steam client Need Retest
#### Your system information * Steam client version (build number or date): 1658944613 * Distribution (e.g. Ubuntu): Arch Linux * Opted into Steam client beta?: [Yes/No] No * Have you checked for system updates?: [Yes/No] Yes #### Please describe your issue in as much detail as possible: Dynamic collections of a certain size (seem to be more than 30 games) makes it so that you can scroll way past the end of the collection. This does not happen to static collections or even dynamic collections which had the issue that is snapshot to a static collection. I'm not sure if this happens on other Distros as I'm not able to set up an environment for testing at this moment. #### Steps for reproducing this issue: 1. Make sure you have a library of about 50+ games. 2. Create a **Dynamic Collection** 3. Set the snapshot filter to f.ex **Single player** (whichever filter yields the most results) 4. You can now scroll way past the end of the collection and the scrollbar barely moves. #### Screenshots: **Dynamic collection** ![2022-08-08_12-16](https://user-images.githubusercontent.com/554229/183396250-730127ce-ac7f-4e8f-81bf-af74869c6afa.png) ![2022-08-08_12-18](https://user-images.githubusercontent.com/554229/183396275-9a56ee15-27f3-4451-bd87-cce1572b66e5.png) **Static collection based on the steps of reproduction** ![2022-08-08_12-20](https://user-images.githubusercontent.com/554229/183396739-0aa03379-2eac-40dc-90bd-3ca87bdf3ba5.png) ![2022-08-08_12-20_1](https://user-images.githubusercontent.com/554229/183396746-aeae4c83-a643-4192-a5fa-5e48d32b8d50.png)
1.0
Dynamic collections allows scrolling way past its content - #### Your system information * Steam client version (build number or date): 1658944613 * Distribution (e.g. Ubuntu): Arch Linux * Opted into Steam client beta?: [Yes/No] No * Have you checked for system updates?: [Yes/No] Yes #### Please describe your issue in as much detail as possible: Dynamic collections of a certain size (seem to be more than 30 games) makes it so that you can scroll way past the end of the collection. This does not happen to static collections or even dynamic collections which had the issue that is snapshot to a static collection. I'm not sure if this happens on other Distros as I'm not able to set up an environment for testing at this moment. #### Steps for reproducing this issue: 1. Make sure you have a library of about 50+ games. 2. Create a **Dynamic Collection** 3. Set the snapshot filter to f.ex **Single player** (whichever filter yields the most results) 4. You can now scroll way past the end of the collection and the scrollbar barely moves. #### Screenshots: **Dynamic collection** ![2022-08-08_12-16](https://user-images.githubusercontent.com/554229/183396250-730127ce-ac7f-4e8f-81bf-af74869c6afa.png) ![2022-08-08_12-18](https://user-images.githubusercontent.com/554229/183396275-9a56ee15-27f3-4451-bd87-cce1572b66e5.png) **Static collection based on the steps of reproduction** ![2022-08-08_12-20](https://user-images.githubusercontent.com/554229/183396739-0aa03379-2eac-40dc-90bd-3ca87bdf3ba5.png) ![2022-08-08_12-20_1](https://user-images.githubusercontent.com/554229/183396746-aeae4c83-a643-4192-a5fa-5e48d32b8d50.png)
test
dynamic collections allows scrolling way past its content your system information steam client version build number or date distribution e g ubuntu arch linux opted into steam client beta no have you checked for system updates yes please describe your issue in as much detail as possible dynamic collections of a certain size seem to be more than games makes it so that you can scroll way past the end of the collection this does not happen to static collections or even dynamic collections which had the issue that is snapshot to a static collection i m not sure if this happens on other distros as i m not able to set up an environment for testing at this moment steps for reproducing this issue make sure you have a library of about games create a dynamic collection set the snapshot filter to f ex single player whichever filter yields the most results you can now scroll way past the end of the collection and the scrollbar barely moves screenshots dynamic collection static collection based on the steps of reproduction
1
186,907
14,426,868,284
IssuesEvent
2020-12-06 00:28:35
kalexmills/github-vet-tests-dec2020
https://api.github.com/repos/kalexmills/github-vet-tests-dec2020
closed
kubevirt/kubernetes-device-plugins: vendor/k8s.io/kubernetes/pkg/controller/deployment/util/deployment_util_test.go; 5 LoC
fresh test tiny vendored
Found a possible issue in [kubevirt/kubernetes-device-plugins](https://www.github.com/kubevirt/kubernetes-device-plugins) at [vendor/k8s.io/kubernetes/pkg/controller/deployment/util/deployment_util_test.go](https://github.com/kubevirt/kubernetes-device-plugins/blob/2439489f2cd0b3ddc00c5779dd5129680f0c2dcd/vendor/k8s.io/kubernetes/pkg/controller/deployment/util/deployment_util_test.go#L61-L65) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to rs at line 63 may start a goroutine [Click here to see the code in its original context.](https://github.com/kubevirt/kubernetes-device-plugins/blob/2439489f2cd0b3ddc00c5779dd5129680f0c2dcd/vendor/k8s.io/kubernetes/pkg/controller/deployment/util/deployment_util_test.go#L61-L65) <details> <summary>Click here to show the 5 line(s) of Go which triggered the analyzer.</summary> ```go for _, rs := range rsList.Items { if rs.Name == name { return true, &rs, nil } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 2439489f2cd0b3ddc00c5779dd5129680f0c2dcd
1.0
kubevirt/kubernetes-device-plugins: vendor/k8s.io/kubernetes/pkg/controller/deployment/util/deployment_util_test.go; 5 LoC - Found a possible issue in [kubevirt/kubernetes-device-plugins](https://www.github.com/kubevirt/kubernetes-device-plugins) at [vendor/k8s.io/kubernetes/pkg/controller/deployment/util/deployment_util_test.go](https://github.com/kubevirt/kubernetes-device-plugins/blob/2439489f2cd0b3ddc00c5779dd5129680f0c2dcd/vendor/k8s.io/kubernetes/pkg/controller/deployment/util/deployment_util_test.go#L61-L65) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to rs at line 63 may start a goroutine [Click here to see the code in its original context.](https://github.com/kubevirt/kubernetes-device-plugins/blob/2439489f2cd0b3ddc00c5779dd5129680f0c2dcd/vendor/k8s.io/kubernetes/pkg/controller/deployment/util/deployment_util_test.go#L61-L65) <details> <summary>Click here to show the 5 line(s) of Go which triggered the analyzer.</summary> ```go for _, rs := range rsList.Items { if rs.Name == name { return true, &rs, nil } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 2439489f2cd0b3ddc00c5779dd5129680f0c2dcd
test
kubevirt kubernetes device plugins vendor io kubernetes pkg controller deployment util deployment util test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to rs at line may start a goroutine click here to show the line s of go which triggered the analyzer go for rs range rslist items if rs name name return true rs nil leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
54,275
6,377,230,981
IssuesEvent
2017-08-02 09:31:59
xcat2/xcat-core
https://api.github.com/repos/xcat2/xcat-core
closed
[fvt] "pgsqlsetup -i " does not work well in rhels7.3 and rhels7.4
component:test priority:high sprint2 type:bug
``xcat version:`` Version 2.13.6 (git commit ea6498a638c31b1eb49167856e3fea6c0c164070, built Mon Jul 31 06:15:45 EDT 2017) ``OS:`` rhels 7.4 rc 1 ``pgsqlsetup -i `` does not work well in rhels7.3 and rhels7.4. The return code of "pgsqlsetup" is no-zore, but there is not error in the log.
1.0
[fvt] "pgsqlsetup -i " does not work well in rhels7.3 and rhels7.4 - ``xcat version:`` Version 2.13.6 (git commit ea6498a638c31b1eb49167856e3fea6c0c164070, built Mon Jul 31 06:15:45 EDT 2017) ``OS:`` rhels 7.4 rc 1 ``pgsqlsetup -i `` does not work well in rhels7.3 and rhels7.4. The return code of "pgsqlsetup" is no-zore, but there is not error in the log.
test
pgsqlsetup i does not work well in and xcat version version git commit built mon jul edt os rhels rc pgsqlsetup i does not work well in and the return code of pgsqlsetup is no zore but there is not error in the log
1
251,786
21,523,027,930
IssuesEvent
2022-04-28 15:43:21
damccorm/test-migration-target
https://api.github.com/repos/damccorm/test-migration-target
opened
Flink Tests affected by : Encountered unsupported logical type URN: int org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.UncheckedExecutionException
bug awaiting triage P2 testing
This is the list of tests affected by this error.  -- beam_LoadTests_Go_CoGBK_Flink_batch -- beam_LoadTests_Go_Combine_Flink_Batch -- beam_LoadTests_Go_GBK_Flink_Batch   The error  is Encountered unsupported logical type URN: int org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.UncheckedExecutionException **   Imported from Jira [BEAM-14173](https://issues.apache.org/jira/browse/BEAM-14173) Reported by: andoni.guzman.
1.0
Flink Tests affected by : Encountered unsupported logical type URN: int org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.UncheckedExecutionException - This is the list of tests affected by this error.  -- beam_LoadTests_Go_CoGBK_Flink_batch -- beam_LoadTests_Go_Combine_Flink_Batch -- beam_LoadTests_Go_GBK_Flink_Batch   The error  is Encountered unsupported logical type URN: int org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.UncheckedExecutionException **   Imported from Jira [BEAM-14173](https://issues.apache.org/jira/browse/BEAM-14173) Reported by: andoni.guzman.
test
flink tests affected by encountered unsupported logical type urn int org apache beam vendor guava jre com google common util concurrent uncheckedexecutionexception this is the list of tests affected by this error   beam loadtests go cogbk flink batch beam loadtests go combine flink batch beam loadtests go gbk flink batch   the error  is encountered unsupported logical type urn int org apache beam vendor guava jre com google common util concurrent uncheckedexecutionexception   imported from jira reported by andoni guzman
1
188,752
14,474,777,445
IssuesEvent
2020-12-10 00:03:13
kalexmills/github-vet-tests-dec2020
https://api.github.com/repos/kalexmills/github-vet-tests-dec2020
closed
a0x8o/kubernetes: staging/src/k8s.io/kube-apiextensions-server/test/integration/basic_test.go; 5 LoC
fresh test tiny
Found a possible issue in [a0x8o/kubernetes](https://www.github.com/a0x8o/kubernetes) at [staging/src/k8s.io/kube-apiextensions-server/test/integration/basic_test.go](https://github.com/a0x8o/kubernetes/blob/06e98f7a27a6d2f408ed1ce068bc95f00d1c467d/staging/src/k8s.io/kube-apiextensions-server/test/integration/basic_test.go#L460-L464) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to a at line 461 may start a goroutine [Click here to see the code in its original context.](https://github.com/a0x8o/kubernetes/blob/06e98f7a27a6d2f408ed1ce068bc95f00d1c467d/staging/src/k8s.io/kube-apiextensions-server/test/integration/basic_test.go#L460-L464) <details> <summary>Click here to show the 5 line(s) of Go which triggered the analyzer.</summary> ```go for _, a := range createdList.(*unstructured.UnstructuredList).Items { if e := instances[a.GetNamespace()]; !reflect.DeepEqual(e, &a) { t.Errorf("expected %v, got %v", e, a) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 06e98f7a27a6d2f408ed1ce068bc95f00d1c467d
1.0
a0x8o/kubernetes: staging/src/k8s.io/kube-apiextensions-server/test/integration/basic_test.go; 5 LoC - Found a possible issue in [a0x8o/kubernetes](https://www.github.com/a0x8o/kubernetes) at [staging/src/k8s.io/kube-apiextensions-server/test/integration/basic_test.go](https://github.com/a0x8o/kubernetes/blob/06e98f7a27a6d2f408ed1ce068bc95f00d1c467d/staging/src/k8s.io/kube-apiextensions-server/test/integration/basic_test.go#L460-L464) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to a at line 461 may start a goroutine [Click here to see the code in its original context.](https://github.com/a0x8o/kubernetes/blob/06e98f7a27a6d2f408ed1ce068bc95f00d1c467d/staging/src/k8s.io/kube-apiextensions-server/test/integration/basic_test.go#L460-L464) <details> <summary>Click here to show the 5 line(s) of Go which triggered the analyzer.</summary> ```go for _, a := range createdList.(*unstructured.UnstructuredList).Items { if e := instances[a.GetNamespace()]; !reflect.DeepEqual(e, &a) { t.Errorf("expected %v, got %v", e, a) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 06e98f7a27a6d2f408ed1ce068bc95f00d1c467d
test
kubernetes staging src io kube apiextensions server test integration basic test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to a at line may start a goroutine click here to show the line s of go which triggered the analyzer go for a range createdlist unstructured unstructuredlist items if e instances reflect deepequal e a t errorf expected v got v e a leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
153
2,551,578,204
IssuesEvent
2015-02-02 10:51:19
LinDA-tools/LindaWorkbench
https://api.github.com/repos/LinDA-tools/LindaWorkbench
closed
A user can transform the available datasets in different formats in order to be able to work on a common data format basis.
Requirement from Scenario Transformation tool
A user can transform the available datasets in different formats in order to be able to work on a common data format basis. scenario 4
1.0
A user can transform the available datasets in different formats in order to be able to work on a common data format basis. - A user can transform the available datasets in different formats in order to be able to work on a common data format basis. scenario 4
non_test
a user can transform the available datasets in different formats in order to be able to work on a common data format basis a user can transform the available datasets in different formats in order to be able to work on a common data format basis scenario
0
309,050
26,648,673,293
IssuesEvent
2023-01-25 12:01:58
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
opened
Fix view_tensor.test_view_tensor_sin
PyTorch Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3994356210/jobs/6851966030" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/3994356210/jobs/6851966030" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3994356210/jobs/6851966030" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/3994356210/jobs/6851966030" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_view_tensor.py::test_view_tensor_sin[cpu-ivy.functional.backends.torch-False-False]</summary> 2023-01-24T08:42:35.5769881Z E TypeError: len() of a 0-d tensor 2023-01-24T08:42:35.5775410Z E ivy.exceptions.IvyBackendException: torch: asarray: len() of a 0-d tensor </details>
1.0
Fix view_tensor.test_view_tensor_sin - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3994356210/jobs/6851966030" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/3994356210/jobs/6851966030" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3994356210/jobs/6851966030" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/3994356210/jobs/6851966030" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_view_tensor.py::test_view_tensor_sin[cpu-ivy.functional.backends.torch-False-False]</summary> 2023-01-24T08:42:35.5769881Z E TypeError: len() of a 0-d tensor 2023-01-24T08:42:35.5775410Z E ivy.exceptions.IvyBackendException: torch: asarray: len() of a 0-d tensor </details>
test
fix view tensor test view tensor sin tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test frontends test torch test view tensor py test view tensor sin e typeerror len of a d tensor e ivy exceptions ivybackendexception torch asarray len of a d tensor
1
28,082
4,363,684,240
IssuesEvent
2016-08-03 01:54:12
exercism/todo
https://api.github.com/repos/exercism/todo
closed
Extract shared inputs/outputs for exercise Triangle
shared-test-data
Goal: https://github.com/exercism/todo/issues/13 -------------- **Skillset:** This is mostly a matter of reading code (test suites) in several languages and creating some JSON. --------------- Create a file `triangle.json` containing the inputs and expected outputs for Triangle. Triangle has been implemented in the following languages: - https://github.com/exercism/xclojure/tree/master/triangle - https://github.com/exercism/xcoffeescript/tree/master/triangle - https://github.com/exercism/xcpp/tree/master/triangle - https://github.com/exercism/xcsharp/tree/master/triangle - https://github.com/exercism/xecmascript/tree/master/triangle - https://github.com/exercism/xelixir/tree/master/triangle - https://github.com/exercism/xfsharp/tree/master/triangle - https://github.com/exercism/xgo/tree/master/triangle - https://github.com/exercism/xhaskell/tree/master/triangle - https://github.com/exercism/xjava/tree/master/triangle - https://github.com/exercism/xjavascript/tree/master/triangle - https://github.com/exercism/xlisp/tree/master/triangle - https://github.com/exercism/xlua/tree/master/triangle - https://github.com/exercism/xperl5/tree/master/triangle - https://github.com/exercism/xpython/tree/master/triangle - https://github.com/exercism/xruby/tree/master/triangle - https://github.com/exercism/xscala/tree/master/triangle - https://github.com/exercism/xswift/tree/master/triangle See the following files for some examples: - https://github.com/exercism/x-common/blob/master/bob.json - https://github.com/exercism/x-common/blob/master/clock.json - https://github.com/exercism/x-common/blob/master/custom-set.json - https://github.com/exercism/x-common/blob/master/gigasecond.json - https://github.com/exercism/x-common/blob/master/hamming.json - https://github.com/exercism/x-common/blob/master/leap.json
1.0
Extract shared inputs/outputs for exercise Triangle - Goal: https://github.com/exercism/todo/issues/13 -------------- **Skillset:** This is mostly a matter of reading code (test suites) in several languages and creating some JSON. --------------- Create a file `triangle.json` containing the inputs and expected outputs for Triangle. Triangle has been implemented in the following languages: - https://github.com/exercism/xclojure/tree/master/triangle - https://github.com/exercism/xcoffeescript/tree/master/triangle - https://github.com/exercism/xcpp/tree/master/triangle - https://github.com/exercism/xcsharp/tree/master/triangle - https://github.com/exercism/xecmascript/tree/master/triangle - https://github.com/exercism/xelixir/tree/master/triangle - https://github.com/exercism/xfsharp/tree/master/triangle - https://github.com/exercism/xgo/tree/master/triangle - https://github.com/exercism/xhaskell/tree/master/triangle - https://github.com/exercism/xjava/tree/master/triangle - https://github.com/exercism/xjavascript/tree/master/triangle - https://github.com/exercism/xlisp/tree/master/triangle - https://github.com/exercism/xlua/tree/master/triangle - https://github.com/exercism/xperl5/tree/master/triangle - https://github.com/exercism/xpython/tree/master/triangle - https://github.com/exercism/xruby/tree/master/triangle - https://github.com/exercism/xscala/tree/master/triangle - https://github.com/exercism/xswift/tree/master/triangle See the following files for some examples: - https://github.com/exercism/x-common/blob/master/bob.json - https://github.com/exercism/x-common/blob/master/clock.json - https://github.com/exercism/x-common/blob/master/custom-set.json - https://github.com/exercism/x-common/blob/master/gigasecond.json - https://github.com/exercism/x-common/blob/master/hamming.json - https://github.com/exercism/x-common/blob/master/leap.json
test
extract shared inputs outputs for exercise triangle goal skillset this is mostly a matter of reading code test suites in several languages and creating some json create a file triangle json containing the inputs and expected outputs for triangle triangle has been implemented in the following languages see the following files for some examples
1
118,129
9,975,771,009
IssuesEvent
2019-07-09 13:46:53
xamarin/xamarin-macios
https://api.github.com/repos/xamarin/xamarin-macios
closed
[xcode11] Sample tests fail provisioning because bots don't have Xcode 11 (beta 1)
bug iOS macOS test-only-issue
https://dev.azure.com/xamarin/internal/_build/results?buildId=3048 The Azure DevOps bots don't have Xcode 11, so everything fails pretty much instantly. We can't expect the bots to have beta Xcode's, so we should implement support for provisioning those ourselves.
1.0
[xcode11] Sample tests fail provisioning because bots don't have Xcode 11 (beta 1) - https://dev.azure.com/xamarin/internal/_build/results?buildId=3048 The Azure DevOps bots don't have Xcode 11, so everything fails pretty much instantly. We can't expect the bots to have beta Xcode's, so we should implement support for provisioning those ourselves.
test
sample tests fail provisioning because bots don t have xcode beta the azure devops bots don t have xcode so everything fails pretty much instantly we can t expect the bots to have beta xcode s so we should implement support for provisioning those ourselves
1
7,032
2,872,057,944
IssuesEvent
2015-06-08 09:21:46
dpi/rng
https://api.github.com/repos/dpi/rng
opened
Dereference and delete related entities: Event entity/Event bundle
needs tests task
Deleting a __Event__ or __Event Bundle__: - Removes - __Registration__ entities - __Group__ entities - __Rule__ entities Renaming a __Event Bundle__: - Updates - Reference: __Registration__ -> __Event__ - Reference: __Rule__ -> __Event__
1.0
Dereference and delete related entities: Event entity/Event bundle - Deleting a __Event__ or __Event Bundle__: - Removes - __Registration__ entities - __Group__ entities - __Rule__ entities Renaming a __Event Bundle__: - Updates - Reference: __Registration__ -> __Event__ - Reference: __Rule__ -> __Event__
test
dereference and delete related entities event entity event bundle deleting a event or event bundle removes registration entities group entities rule entities renaming a event bundle updates reference registration event reference rule event
1
61,828
6,759,181,955
IssuesEvent
2017-10-24 16:16:01
sveneisenschmidt/yay
https://api.github.com/repos/sveneisenschmidt/yay
closed
(Test) Write tests for commands
help wanted test
- [x] Yay\Bundle\IntegrationBundle\Command\EnableCommand - [x] Yay\Bundle\IntegrationBundle\Command\DisableCommand - [x] Yay\Bundle\IntegrationBundle\Command\ValidateCommand - [x] Yay\Bundle\ApiBundle\Command\RecalculateCommand
1.0
(Test) Write tests for commands - - [x] Yay\Bundle\IntegrationBundle\Command\EnableCommand - [x] Yay\Bundle\IntegrationBundle\Command\DisableCommand - [x] Yay\Bundle\IntegrationBundle\Command\ValidateCommand - [x] Yay\Bundle\ApiBundle\Command\RecalculateCommand
test
test write tests for commands yay bundle integrationbundle command enablecommand yay bundle integrationbundle command disablecommand yay bundle integrationbundle command validatecommand yay bundle apibundle command recalculatecommand
1
325,393
27,874,115,236
IssuesEvent
2023-03-21 15:05:28
containers/podman-desktop
https://api.github.com/repos/containers/podman-desktop
closed
code coverage is not reported on files without tests
kind/enhancement area/tests
### Is your enhancement related to a problem? Please describe we have code coverage from the current files being tested but untested files are not reported in the report it's leading to invalid results ### Describe the solution you'd like Full code coverage report should report coverage for all the files, even those that have no associated tests ### Describe alternatives you've considered _No response_ ### Additional context _No response_
1.0
code coverage is not reported on files without tests - ### Is your enhancement related to a problem? Please describe we have code coverage from the current files being tested but untested files are not reported in the report it's leading to invalid results ### Describe the solution you'd like Full code coverage report should report coverage for all the files, even those that have no associated tests ### Describe alternatives you've considered _No response_ ### Additional context _No response_
test
code coverage is not reported on files without tests is your enhancement related to a problem please describe we have code coverage from the current files being tested but untested files are not reported in the report it s leading to invalid results describe the solution you d like full code coverage report should report coverage for all the files even those that have no associated tests describe alternatives you ve considered no response additional context no response
1
140,000
31,811,082,999
IssuesEvent
2023-09-13 16:52:59
mokkang/github-verademo
https://api.github.com/repos/mokkang/github-verademo
opened
Insertion of Sensitive Information Into Sent Data ('Information Leakage') [VID:202]
VeracodeFlaw: Low Veracode Policy Scan
**Filename:** UserController.java **Line:** 442 **CWE:** 201 (Insertion of Sensitive Information Into Sent Data ('Information Leakage')) <span> The application calls the showProfile() function, which will result in data being transferred out of the application (via the network or another medium). This data contains sensitive information. The potentially sensitive data originated from an earlier call to java.lang.System.getenv.</span> <span>Ensure that the transfer of the sensitive data is intended and that it does not violate application security policy. This flaw is categorized as low severity because it only impacts confidentiality, not integrity or availability. However, in the context of a mobile application, the significance of an information leak may be much greater, especially if misaligned with user expectations or data privacy policies.</span> <span>References: <a href="https://cwe.mitre.org/data/definitions/201.html">CWE</a> <a href="https://owasp.org/Top10/A05_2021-Security_Misconfiguration/">OWASP Security Misconfiguration</a> <a href="https://owasp.org/Top10/A02_2021-Cryptographic_Failures/">OWASP Cryptographic Failures</a></span>
2.0
Insertion of Sensitive Information Into Sent Data ('Information Leakage') [VID:202] - **Filename:** UserController.java **Line:** 442 **CWE:** 201 (Insertion of Sensitive Information Into Sent Data ('Information Leakage')) <span> The application calls the showProfile() function, which will result in data being transferred out of the application (via the network or another medium). This data contains sensitive information. The potentially sensitive data originated from an earlier call to java.lang.System.getenv.</span> <span>Ensure that the transfer of the sensitive data is intended and that it does not violate application security policy. This flaw is categorized as low severity because it only impacts confidentiality, not integrity or availability. However, in the context of a mobile application, the significance of an information leak may be much greater, especially if misaligned with user expectations or data privacy policies.</span> <span>References: <a href="https://cwe.mitre.org/data/definitions/201.html">CWE</a> <a href="https://owasp.org/Top10/A05_2021-Security_Misconfiguration/">OWASP Security Misconfiguration</a> <a href="https://owasp.org/Top10/A02_2021-Cryptographic_Failures/">OWASP Cryptographic Failures</a></span>
non_test
insertion of sensitive information into sent data information leakage filename usercontroller java line cwe insertion of sensitive information into sent data information leakage the application calls the showprofile function which will result in data being transferred out of the application via the network or another medium this data contains sensitive information the potentially sensitive data originated from an earlier call to java lang system getenv ensure that the transfer of the sensitive data is intended and that it does not violate application security policy this flaw is categorized as low severity because it only impacts confidentiality not integrity or availability however in the context of a mobile application the significance of an information leak may be much greater especially if misaligned with user expectations or data privacy policies references
0
345,969
30,856,945,410
IssuesEvent
2023-08-02 21:31:15
kevinbowen777/django-blog
https://api.github.com/repos/kevinbowen777/django-blog
closed
View tests for dynamic slugs failing
testing
Tests for `posts/tests/test_view.py` are generating errors for testing dynamic slugs: ``` > self.assertEqual( self.post.get_absolute_url(), f"/posts/2023/7/14/{self.post.slug}/" ) E AssertionError: '/posts/2023/8/2/a-good-title/' != '/posts/2023/7/14/a-good-title/' E - /posts/2023/8/2/a-good-title/ E ? ^ ^ E + /posts/2023/7/14/a-good-title/ E ? ^ ^^ posts/tests/test_views.py:45: AssertionError FAILED posts/tests/test_views.py::PostTests::test_get_absolute_url - AssertionError: '/posts/2023/8/2/a-good-title/' != '/posts/2023/7/14/a-good-title/' FAILED posts/tests/test_views.py::PostTests::test_post_detail_view - AssertionError: 404 != 200 Results (27.52s): 51 passed 2 failed - posts/tests/test_views.py:44 PostTests.test_get_absolute_url - posts/tests/test_views.py:49 PostTests.test_post_detail_view ```
1.0
View tests for dynamic slugs failing - Tests for `posts/tests/test_view.py` are generating errors for testing dynamic slugs: ``` > self.assertEqual( self.post.get_absolute_url(), f"/posts/2023/7/14/{self.post.slug}/" ) E AssertionError: '/posts/2023/8/2/a-good-title/' != '/posts/2023/7/14/a-good-title/' E - /posts/2023/8/2/a-good-title/ E ? ^ ^ E + /posts/2023/7/14/a-good-title/ E ? ^ ^^ posts/tests/test_views.py:45: AssertionError FAILED posts/tests/test_views.py::PostTests::test_get_absolute_url - AssertionError: '/posts/2023/8/2/a-good-title/' != '/posts/2023/7/14/a-good-title/' FAILED posts/tests/test_views.py::PostTests::test_post_detail_view - AssertionError: 404 != 200 Results (27.52s): 51 passed 2 failed - posts/tests/test_views.py:44 PostTests.test_get_absolute_url - posts/tests/test_views.py:49 PostTests.test_post_detail_view ```
test
view tests for dynamic slugs failing tests for posts tests test view py are generating errors for testing dynamic slugs self assertequal self post get absolute url f posts self post slug e assertionerror posts a good title posts a good title e posts a good title e e posts a good title e posts tests test views py assertionerror failed posts tests test views py posttests test get absolute url assertionerror posts a good title posts a good title failed posts tests test views py posttests test post detail view assertionerror results passed failed posts tests test views py posttests test get absolute url posts tests test views py posttests test post detail view
1
67,258
7,043,721,480
IssuesEvent
2017-12-31 11:40:38
spenmath/ScoutsEMSIssues
https://api.github.com/repos/spenmath/ScoutsEMSIssues
closed
Add event configuration for Staff Roles to DB
testing
Ref: https://scouts-ems.org.uk/staffreg.php Add Staff Roles to the database and give the event administrator the ability to manage the roles from the Event Menu
1.0
Add event configuration for Staff Roles to DB - Ref: https://scouts-ems.org.uk/staffreg.php Add Staff Roles to the database and give the event administrator the ability to manage the roles from the Event Menu
test
add event configuration for staff roles to db ref add staff roles to the database and give the event administrator the ability to manage the roles from the event menu
1
327,991
28,096,109,871
IssuesEvent
2023-03-30 15:54:02
woocommerce/woocommerce
https://api.github.com/repos/woocommerce/woocommerce
opened
Add new products Woo REST API tests to `api-core-tests`
type: task focus: api tests [team:Solaris]
### Describe the solution you'd like Add test for the following rest API endpoints to ensure there is coverage for all CRUD operations: GET /wp-json/wc/v3/products Documentation is [here](https://woocommerce.github.io/woocommerce-rest-api-docs/?javascript#list-all-products) Related to this https://github.com/woocommerce/woocommerce/issues/37497 ### Describe alternatives you've considered n/a - increasing test coverage ### Additional context _No response_
1.0
Add new products Woo REST API tests to `api-core-tests` - ### Describe the solution you'd like Add test for the following rest API endpoints to ensure there is coverage for all CRUD operations: GET /wp-json/wc/v3/products Documentation is [here](https://woocommerce.github.io/woocommerce-rest-api-docs/?javascript#list-all-products) Related to this https://github.com/woocommerce/woocommerce/issues/37497 ### Describe alternatives you've considered n/a - increasing test coverage ### Additional context _No response_
test
add new products woo rest api tests to api core tests describe the solution you d like add test for the following rest api endpoints to ensure there is coverage for all crud operations get wp json wc products documentation is related to this describe alternatives you ve considered n a increasing test coverage additional context no response
1
304,955
26,348,839,645
IssuesEvent
2023-01-11 01:45:32
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
opened
🥐🥐 Flaky E2E Test Detected! health-care-cerner.cypress.spec.js
authenticated-experience e2e-flaky-test
Your test spec `src/applications/personalization/dashboard/tests/e2e/health-care-cerner.cypress.spec.js` has been disallowed from running on `vets-website` due to flakiness. Please make corrections to the test spec and run it through CI to have it re-enabled. To see the details of the test failure, please visit: [Mochawesome Report](https://testing-tools-testing-reports.s3-us-gov-west-1.amazonaws.com/vets-website-cypress-stress-test-reports/7e10a994-1af4-48f1-8b0c-3844862080a8.html) Thanks, The QA Standards Team
1.0
🥐🥐 Flaky E2E Test Detected! health-care-cerner.cypress.spec.js - Your test spec `src/applications/personalization/dashboard/tests/e2e/health-care-cerner.cypress.spec.js` has been disallowed from running on `vets-website` due to flakiness. Please make corrections to the test spec and run it through CI to have it re-enabled. To see the details of the test failure, please visit: [Mochawesome Report](https://testing-tools-testing-reports.s3-us-gov-west-1.amazonaws.com/vets-website-cypress-stress-test-reports/7e10a994-1af4-48f1-8b0c-3844862080a8.html) Thanks, The QA Standards Team
test
🥐🥐 flaky test detected health care cerner cypress spec js your test spec src applications personalization dashboard tests health care cerner cypress spec js has been disallowed from running on vets website due to flakiness please make corrections to the test spec and run it through ci to have it re enabled to see the details of the test failure please visit thanks the qa standards team
1
832,320
32,077,495,608
IssuesEvent
2023-09-25 12:01:06
googleapis/google-cloud-ruby
https://api.github.com/repos/googleapis/google-cloud-ruby
closed
[Nightly CI Failures] Failures detected for google-cloud-binary_authorization
type: bug priority: p1 nightly failure
At 2023-09-09 09:39:30 UTC, detected failures in google-cloud-binary_authorization for: test. The CI logs can be found [here](https://github.com/googleapis/google-cloud-ruby/actions/runs/6129868852) report_key_0076e1fdc6162141d774c9374894f991
1.0
[Nightly CI Failures] Failures detected for google-cloud-binary_authorization - At 2023-09-09 09:39:30 UTC, detected failures in google-cloud-binary_authorization for: test. The CI logs can be found [here](https://github.com/googleapis/google-cloud-ruby/actions/runs/6129868852) report_key_0076e1fdc6162141d774c9374894f991
non_test
failures detected for google cloud binary authorization at utc detected failures in google cloud binary authorization for test the ci logs can be found report key
0
63,711
6,883,046,601
IssuesEvent
2017-11-21 07:48:06
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
opened
small-scale density test failure on 1.8 branch
kind/bug priority/failing-test sig/scalability
**Is this a BUG REPORT or FEATURE REQUEST?**: /kind bug @kubernetes/sig-scalability-bugs /priority failing-test /assign **What happened**: https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gci-gce-scalability-stable1/210 failed with: ``` I1120 18:51:35.941] [sig-scalability] Density I1120 18:51:35.941] /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scalability/framework.go:22 I1120 18:51:35.942] [Feature:Performance] should allow starting 30 pods per node using { ReplicationController} with 0 secrets, 0 configmaps and 0 daemons [It] I1120 18:51:35.942] /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scalability/density.go:783 I1120 18:51:35.942] I1120 18:51:35.942] Expected error: I1120 18:51:35.942] <*errors.errorString | 0xc4202b73d0>: { I1120 18:51:35.943] s: "timed out waiting for the condition", I1120 18:51:35.943] } I1120 18:51:35.943] timed out waiting for the condition I1120 18:51:35.943] not to have occurred I1120 18:51:35.944] I1120 18:51:35.944] /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scalability/density.go:777 ``` A little earlier, the logs read: ``` I1120 18:35:08.082] STEP: Scheduling additional Pods to measure startup latencies I1120 18:35:13.231] Nov 20 18:35:13.231: INFO: Found pod 'density-latency-pod-1' running [...] I1120 18:35:39.766] Nov 20 18:35:39.172: INFO: Found pod 'density-latency-pod-100' running I1120 18:36:07.884] STEP: Waiting for all Pods begin observed by the watch... I1120 18:36:10.178] Nov 20 18:35:39.172: INFO: 2 latency pods scheduled on e2e-scalability-1-6-minion-group-qxqh I1120 18:36:53.727] error dialing prow@35.196.238.140:22: 'ssh: handshake failed: read tcp 10.60.162.41:54680->35.196.238.140:22: i/o timeout', retrying I1120 18:37:07.096] Nov 20 18:37:04.576: INFO: 10% worst schedule latencies: [{density-latency-pod-62-nqzf6 e2e-scalability-1-6-minion-group-jpx3 0s} {density-latency-pod-66-c9w84 e2e-scalability-1-6-minion-group-5z1n 0s} {density-latency-pod-63-xvrdc e2e-scalability-1-6-minion-group-6g24 0s} {density-latency-pod-91-xnzdt e2e-scalability-1-6-minion-group-xf7k 0s} {density-latency-pod-13-wskj8 e2e-scalability-1-6-minion-group-8d0x 0s} {density-latency-pod-32-c7pv6 e2e-scalability-1-6-minion-group-9snm 0s} {density-latency-pod-44-jh2rc e2e-scalability-1-6-minion-group-xlkv 0s} {density-latency-pod-57-jdxhz e2e-scalability-1-6-minion-group-v449 0s} {density-latency-pod-94-77cl2 e2e-scalability-1-6-minion-group-wmg0 0s} {density-latency-pod-100-j7njs e2e-scalability-1-6-minion-group-sv0q 0s}] I1120 18:37:07.528] Nov 20 18:37:04.576: INFO: perc50: 0s, perc90: 0s, perc99: 0s I1120 18:37:07.666] Nov 20 18:37:04.576: INFO: 10% worst run-after-schedule latencies: [{density-latency-pod-62-nqzf6 e2e-scalability-1-6-minion-group-jpx3 1s} {density-latency-pod-66-c9w84 e2e-scalability-1-6-minion-group-5z1n 1s} {density-latency-pod-63-xvrdc e2e-scalability-1-6-minion-group-6g24 1s} {density-latency-pod-91-xnzdt e2e-scalability-1-6-minion-group-xf7k 1s} {density-latency-pod-13-wskj8 e2e-scalability-1-6-minion-group-8d0x 1s} {density-latency-pod-32-c7pv6 e2e-scalability-1-6-minion-group-9snm 1s} {density-latency-pod-44-jh2rc e2e-scalability-1-6-minion-group-xlkv 1s} {density-latency-pod-57-jdxhz e2e-scalability-1-6-minion-group-v449 1s} {density-latency-pod-94-77cl2 e2e-scalability-1-6-minion-group-wmg0 1s} {density-latency-pod-100-j7njs e2e-scalability-1-6-minion-group-sv0q 1s}] I1120 18:37:07.958] Nov 20 18:37:04.577: INFO: perc50: 1s, perc90: 1s, perc99: 1s I1120 18:37:08.229] Nov 20 18:37:04.577: INFO: 10% worst watch latencies: [{density-latency-pod-9-5m7xz e2e-scalability-1-6-minion-group-j4c7 1.755172589s} {density-latency-pod-84-w657q e2e-scalability-1-6-minion-group-017z 1.8380606s} {density-latency-pod-8-zrjmj e2e-scalability-1-6-minion-group-hrxk 1.838067735s} {density-latency-pod-56-bqc6k e2e-scalability-1-6-minion-group-kbc8 1.845675087s} {density-latency-pod-37-klfbp e2e-scalability-1-6-minion-group-qdsd 1.846620818s} {density-latency-pod-94-77cl2 e2e-scalability-1-6-minion-group-wmg0 1.961543168s} {density-latency-pod-93-dgxpm e2e-scalability-1-6-minion-group-s4c1 1.961553187s} {density-latency-pod-75-l5s27 e2e-scalability-1-6-minion-group-sgfp 2.110547217s} {density-latency-pod-95-hgzx5 e2e-scalability-1-6-minion-group-rrm1 2.633364677s} {density-latency-pod-92-zdz77 e2e-scalability-1-6-minion-group-vxlb 2.961550491s}] I1120 18:37:08.428] Nov 20 18:37:04.577: INFO: perc50: 1.104970952s, perc90: 1.737680884s, perc99: 2.633364677s I1120 18:37:08.840] Nov 20 18:37:04.577: INFO: 10% worst scheduled-to-end total latencies: [{density-latency-pod-86-p2rhc e2e-scalability-1-6-minion-group-2spc 2.641154569s} {density-latency-pod-55-mdvdf e2e-scalability-1-6-minion-group-x42j 2.652542108s} {density-latency-pod-67-gnfbd e2e-scalability-1-6-minion-group-x204 2.719568423s} {density-latency-pod-43-bsbls e2e-scalability-1-6-minion-group-zfx4 2.719678338s} {density-latency-pod-48-fhs6x e2e-scalability-1-6-minion-group-mdc1 2.730745699s} {density-latency-pod-8-zrjmj e2e-scalability-1-6-minion-group-hrxk 2.838067735s} {density-latency-pod-94-77cl2 e2e-scalability-1-6-minion-group-wmg0 2.961543168s} {density-latency-pod-92-zdz77 e2e-scalability-1-6-minion-group-vxlb 2.961550491s} {density-latency-pod-93-dgxpm e2e-scalability-1-6-minion-group-s4c1 2.961553187s} {density-latency-pod-95-hgzx5 e2e-scalability-1-6-minion-group-rrm1 3.633364677s}] I1120 18:37:08.996] Nov 20 18:37:04.577: INFO: perc50: 1.890265159s, perc90: 2.633388881s, perc99: 2.961553187s I1120 18:37:09.170] Nov 20 18:37:04.577: INFO: 10% worst e2e total latencies: [{density-latency-pod-86-p2rhc e2e-scalability-1-6-minion-group-2spc 2.641154569s} {density-latency-pod-55-mdvdf e2e-scalability-1-6-minion-group-x42j 2.652542108s} {density-latency-pod-67-gnfbd e2e-scalability-1-6-minion-group-x204 2.719568423s} {density-latency-pod-43-bsbls e2e-scalability-1-6-minion-group-zfx4 2.719678338s} {density-latency-pod-48-fhs6x e2e-scalability-1-6-minion-group-mdc1 2.730745699s} {density-latency-pod-8-zrjmj e2e-scalability-1-6-minion-group-hrxk 2.838067735s} {density-latency-pod-94-77cl2 e2e-scalability-1-6-minion-group-wmg0 2.961543168s} {density-latency-pod-92-zdz77 e2e-scalability-1-6-minion-group-vxlb 2.961550491s} {density-latency-pod-93-dgxpm e2e-scalability-1-6-minion-group-s4c1 2.961553187s} {density-latency-pod-95-hgzx5 e2e-scalability-1-6-minion-group-rrm1 3.633364677s}] I1120 18:37:09.595] Nov 20 18:37:04.577: INFO: perc50: 1.890265159s, perc90: 2.633388881s, perc99: 2.961553187s I1120 18:37:09.703] Nov 20 18:37:04.577: INFO: Approx throughput: 1651.3619009898246 pods/min I1120 18:37:09.931] STEP: Removing additional replication controllers I1120 18:37:10.139] STEP: deleting { ReplicationController} density-latency-pod-1 in namespace e2e-tests-density-30-1-2jts7, will wait for the garbage collector to delete the pods [...] I1120 18:37:25.349] STEP: deleting { ReplicationController} density-latency-pod-25 in namespace e2e-tests-density-30-1-2jts7, will wait for the garbage collector to delete the pods I1120 18:37:28.821] Nov 20 18:37:25.285: INFO: Error while trying to SSH to 104.196.71.126:22, skipping probe. Error: error creating session to prow@104.196.71.126:22: 'read tcp 10.60.162.41:36592->104.196.71.126:22: i/o timeout' I1120 18:37:56.998] error dialing prow@35.196.146.229:22: 'ssh: handshake failed: read tcp 10.60.162.41:52670->35.196.146.229:22: i/o timeout', retrying I1120 18:38:03.493] error dialing prow@35.196.134.58:22: 'ssh: handshake failed: read tcp 10.60.162.41:56588->35.196.134.58:22: i/o timeout', retrying I1120 18:38:10.249] error dialing prow@35.196.152.43:22: 'ssh: handshake failed: read tcp 10.60.162.41:50130->35.196.152.43:22: i/o timeout', retrying I1120 18:38:31.730] error dialing prow@35.196.10.128:22: 'ssh: handshake failed: read tcp 10.60.162.41:42388->35.196.10.128:22: i/o timeout', retrying I1120 18:38:48.126] Nov 20 18:38:47.768: INFO: Error while trying to SSH to 35.196.134.58:22, skipping probe. Error: error getting SSH client to prow@35.196.134.58:22: 'ssh: handshake failed: read tcp 10.60.162.41:56716->35.196.134.58:22: i/o timeout' I1120 18:38:50.602] Nov 20 18:38:50.597: INFO: Error while trying to SSH to 35.196.146.229:22, skipping probe. Error: error getting SSH client to prow@35.196.146.229:22: 'ssh: handshake failed: read tcp 10.60.162.41:52798->35.196.146.229:22: i/o timeout' I1120 18:39:18.638] error dialing prow@104.196.188.66:22: 'ssh: handshake failed: read tcp 10.60.162.41:48272->104.196.188.66:22: i/o timeout', retrying I1120 18:39:25.394] Nov 20 18:39:25.342: INFO: Error while trying to SSH to 35.196.152.43:22, skipping probe. Error: error getting SSH client to prow@35.196.152.43:22: 'ssh: handshake failed: read tcp 10.60.162.41:50242->35.196.152.43:22: i/o timeout' I1120 18:39:41.821] Nov 20 18:39:41.706: INFO: Error while trying to SSH to 35.196.10.128:22, skipping probe. Error: error getting SSH client to prow@35.196.10.128:22: 'ssh: handshake failed: read tcp 10.60.162.41:42518->35.196.10.128:22: i/o timeout' I1120 18:40:19.282] error dialing prow@35.196.199.87:22: 'ssh: handshake failed: read tcp 10.60.162.41:33490->35.196.199.87:22: i/o timeout', retrying I1120 18:40:19.886] error dialing prow@104.196.216.172:22: 'ssh: handshake failed: read tcp 10.60.162.41:39284->104.196.216.172:22: i/o timeout', retrying I1120 18:40:31.097] Nov 20 18:40:30.896: INFO: Error while trying to SSH to 104.196.188.66:22, skipping probe. Error: error getting SSH client to prow@104.196.188.66:22: 'ssh: handshake failed: read tcp 10.60.162.41:48440->104.196.188.66:22: i/o timeout' I1120 18:40:41.794] error dialing prow@35.196.174.144:22: 'ssh: handshake failed: read tcp 10.60.162.41:52262->35.196.174.144:22: i/o timeout', retrying I1120 18:43:04.790] Nov 20 18:42:58.627: INFO: Unexpected error occurred: timed out waiting for the condition I1120 18:43:09.789] Nov 20 18:42:58.627: INFO: Unexpected error occurred: timed out waiting for the condition I1120 18:43:11.040] STEP: deleting { ReplicationController} density-latency-pod-26 in namespace e2e-tests-density-30-1-2jts7, will wait for the garbage collector to delete the pods [...] I1120 18:45:06.875] Nov 20 18:45:06.875: INFO: Terminating { ReplicationController} density-latency-pod-99 pods took: 103.922056ms I1120 18:45:06.876] Nov 20 18:45:06.876: INFO: Terminating { ReplicationController} density-latency-pod-100 pods took: 102.482037ms I1120 18:45:18.075] STEP: Deleting created Collections I1120 18:45:18.075] STEP: Cleaning up only the { ReplicationController}, garbage collector will clean up the pods I1120 18:45:18.076] STEP: deleting { ReplicationController} density3000-0-2e4a4c8c-ce20-11e7-931f-0a580a3ca229 in namespace e2e-tests-density-30-1-2jts7, will wait for the garbage collector to delete the pods I1120 18:45:19.356] Nov 20 18:45:19.356: INFO: Deleting { ReplicationController} density3000-0-2e4a4c8c-ce20-11e7-931f-0a580a3ca229 took: 40.281025ms I1120 18:49:07.842] W1120 18:49:07.842619 5000 reflector.go:334] k8s.io/kubernetes/test/utils/pod_store.go:53: watch of *v1.Pod ended with: too old resource version: 75796 (87313) I1120 18:49:35.783] W1120 18:49:35.783341 5000 reflector.go:334] k8s.io/kubernetes/test/utils/pod_store.go:53: watch of *v1.Pod ended with: too old resource version: 75796 (88754) I1120 18:49:38.395] W1120 18:49:38.395045 5000 reflector.go:334] k8s.io/kubernetes/test/utils/pod_store.go:53: watch of *v1.Pod ended with: too old resource version: 75796 (88957) I1120 18:49:48.550] W1120 18:49:48.549348 5000 reflector.go:334] k8s.io/kubernetes/test/utils/pod_store.go:53: watch of *v1.Pod ended with: too old resource version: 75796 (89485) I1120 18:49:57.774] W1120 18:49:57.774640 5000 reflector.go:334] k8s.io/kubernetes/test/utils/pod_store.go:53: watch of *v1.Pod ended with: too old resource version: 75796 (89983) I1120 18:50:19.357] Nov 20 18:50:19.357: INFO: Terminating { ReplicationController} density3000-0-2e4a4c8c-ce20-11e7-931f-0a580a3ca229 pods took: 5m0.00090081s I1120 18:50:41.768] W1120 18:50:41.768353 5000 reflector.go:334] k8s.io/kubernetes/test/utils/pod_store.go:53: watch of *v1.Pod ended with: too old resource version: 75796 (91177) I1120 18:50:43.527] [AfterEach] [sig-scalability] Density I1120 18:50:43.527] /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scalability/density.go:357 I1120 18:50:43.529] [AfterEach] [sig-scalability] Density I1120 18:50:43.529] /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:135 I1120 18:50:43.530] STEP: Collecting events from namespace "e2e-tests-density-vr5m7". I1120 18:50:43.569] STEP: Found 0 events. ```
1.0
small-scale density test failure on 1.8 branch - **Is this a BUG REPORT or FEATURE REQUEST?**: /kind bug @kubernetes/sig-scalability-bugs /priority failing-test /assign **What happened**: https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gci-gce-scalability-stable1/210 failed with: ``` I1120 18:51:35.941] [sig-scalability] Density I1120 18:51:35.941] /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scalability/framework.go:22 I1120 18:51:35.942] [Feature:Performance] should allow starting 30 pods per node using { ReplicationController} with 0 secrets, 0 configmaps and 0 daemons [It] I1120 18:51:35.942] /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scalability/density.go:783 I1120 18:51:35.942] I1120 18:51:35.942] Expected error: I1120 18:51:35.942] <*errors.errorString | 0xc4202b73d0>: { I1120 18:51:35.943] s: "timed out waiting for the condition", I1120 18:51:35.943] } I1120 18:51:35.943] timed out waiting for the condition I1120 18:51:35.943] not to have occurred I1120 18:51:35.944] I1120 18:51:35.944] /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scalability/density.go:777 ``` A little earlier, the logs read: ``` I1120 18:35:08.082] STEP: Scheduling additional Pods to measure startup latencies I1120 18:35:13.231] Nov 20 18:35:13.231: INFO: Found pod 'density-latency-pod-1' running [...] I1120 18:35:39.766] Nov 20 18:35:39.172: INFO: Found pod 'density-latency-pod-100' running I1120 18:36:07.884] STEP: Waiting for all Pods begin observed by the watch... I1120 18:36:10.178] Nov 20 18:35:39.172: INFO: 2 latency pods scheduled on e2e-scalability-1-6-minion-group-qxqh I1120 18:36:53.727] error dialing prow@35.196.238.140:22: 'ssh: handshake failed: read tcp 10.60.162.41:54680->35.196.238.140:22: i/o timeout', retrying I1120 18:37:07.096] Nov 20 18:37:04.576: INFO: 10% worst schedule latencies: [{density-latency-pod-62-nqzf6 e2e-scalability-1-6-minion-group-jpx3 0s} {density-latency-pod-66-c9w84 e2e-scalability-1-6-minion-group-5z1n 0s} {density-latency-pod-63-xvrdc e2e-scalability-1-6-minion-group-6g24 0s} {density-latency-pod-91-xnzdt e2e-scalability-1-6-minion-group-xf7k 0s} {density-latency-pod-13-wskj8 e2e-scalability-1-6-minion-group-8d0x 0s} {density-latency-pod-32-c7pv6 e2e-scalability-1-6-minion-group-9snm 0s} {density-latency-pod-44-jh2rc e2e-scalability-1-6-minion-group-xlkv 0s} {density-latency-pod-57-jdxhz e2e-scalability-1-6-minion-group-v449 0s} {density-latency-pod-94-77cl2 e2e-scalability-1-6-minion-group-wmg0 0s} {density-latency-pod-100-j7njs e2e-scalability-1-6-minion-group-sv0q 0s}] I1120 18:37:07.528] Nov 20 18:37:04.576: INFO: perc50: 0s, perc90: 0s, perc99: 0s I1120 18:37:07.666] Nov 20 18:37:04.576: INFO: 10% worst run-after-schedule latencies: [{density-latency-pod-62-nqzf6 e2e-scalability-1-6-minion-group-jpx3 1s} {density-latency-pod-66-c9w84 e2e-scalability-1-6-minion-group-5z1n 1s} {density-latency-pod-63-xvrdc e2e-scalability-1-6-minion-group-6g24 1s} {density-latency-pod-91-xnzdt e2e-scalability-1-6-minion-group-xf7k 1s} {density-latency-pod-13-wskj8 e2e-scalability-1-6-minion-group-8d0x 1s} {density-latency-pod-32-c7pv6 e2e-scalability-1-6-minion-group-9snm 1s} {density-latency-pod-44-jh2rc e2e-scalability-1-6-minion-group-xlkv 1s} {density-latency-pod-57-jdxhz e2e-scalability-1-6-minion-group-v449 1s} {density-latency-pod-94-77cl2 e2e-scalability-1-6-minion-group-wmg0 1s} {density-latency-pod-100-j7njs e2e-scalability-1-6-minion-group-sv0q 1s}] I1120 18:37:07.958] Nov 20 18:37:04.577: INFO: perc50: 1s, perc90: 1s, perc99: 1s I1120 18:37:08.229] Nov 20 18:37:04.577: INFO: 10% worst watch latencies: [{density-latency-pod-9-5m7xz e2e-scalability-1-6-minion-group-j4c7 1.755172589s} {density-latency-pod-84-w657q e2e-scalability-1-6-minion-group-017z 1.8380606s} {density-latency-pod-8-zrjmj e2e-scalability-1-6-minion-group-hrxk 1.838067735s} {density-latency-pod-56-bqc6k e2e-scalability-1-6-minion-group-kbc8 1.845675087s} {density-latency-pod-37-klfbp e2e-scalability-1-6-minion-group-qdsd 1.846620818s} {density-latency-pod-94-77cl2 e2e-scalability-1-6-minion-group-wmg0 1.961543168s} {density-latency-pod-93-dgxpm e2e-scalability-1-6-minion-group-s4c1 1.961553187s} {density-latency-pod-75-l5s27 e2e-scalability-1-6-minion-group-sgfp 2.110547217s} {density-latency-pod-95-hgzx5 e2e-scalability-1-6-minion-group-rrm1 2.633364677s} {density-latency-pod-92-zdz77 e2e-scalability-1-6-minion-group-vxlb 2.961550491s}] I1120 18:37:08.428] Nov 20 18:37:04.577: INFO: perc50: 1.104970952s, perc90: 1.737680884s, perc99: 2.633364677s I1120 18:37:08.840] Nov 20 18:37:04.577: INFO: 10% worst scheduled-to-end total latencies: [{density-latency-pod-86-p2rhc e2e-scalability-1-6-minion-group-2spc 2.641154569s} {density-latency-pod-55-mdvdf e2e-scalability-1-6-minion-group-x42j 2.652542108s} {density-latency-pod-67-gnfbd e2e-scalability-1-6-minion-group-x204 2.719568423s} {density-latency-pod-43-bsbls e2e-scalability-1-6-minion-group-zfx4 2.719678338s} {density-latency-pod-48-fhs6x e2e-scalability-1-6-minion-group-mdc1 2.730745699s} {density-latency-pod-8-zrjmj e2e-scalability-1-6-minion-group-hrxk 2.838067735s} {density-latency-pod-94-77cl2 e2e-scalability-1-6-minion-group-wmg0 2.961543168s} {density-latency-pod-92-zdz77 e2e-scalability-1-6-minion-group-vxlb 2.961550491s} {density-latency-pod-93-dgxpm e2e-scalability-1-6-minion-group-s4c1 2.961553187s} {density-latency-pod-95-hgzx5 e2e-scalability-1-6-minion-group-rrm1 3.633364677s}] I1120 18:37:08.996] Nov 20 18:37:04.577: INFO: perc50: 1.890265159s, perc90: 2.633388881s, perc99: 2.961553187s I1120 18:37:09.170] Nov 20 18:37:04.577: INFO: 10% worst e2e total latencies: [{density-latency-pod-86-p2rhc e2e-scalability-1-6-minion-group-2spc 2.641154569s} {density-latency-pod-55-mdvdf e2e-scalability-1-6-minion-group-x42j 2.652542108s} {density-latency-pod-67-gnfbd e2e-scalability-1-6-minion-group-x204 2.719568423s} {density-latency-pod-43-bsbls e2e-scalability-1-6-minion-group-zfx4 2.719678338s} {density-latency-pod-48-fhs6x e2e-scalability-1-6-minion-group-mdc1 2.730745699s} {density-latency-pod-8-zrjmj e2e-scalability-1-6-minion-group-hrxk 2.838067735s} {density-latency-pod-94-77cl2 e2e-scalability-1-6-minion-group-wmg0 2.961543168s} {density-latency-pod-92-zdz77 e2e-scalability-1-6-minion-group-vxlb 2.961550491s} {density-latency-pod-93-dgxpm e2e-scalability-1-6-minion-group-s4c1 2.961553187s} {density-latency-pod-95-hgzx5 e2e-scalability-1-6-minion-group-rrm1 3.633364677s}] I1120 18:37:09.595] Nov 20 18:37:04.577: INFO: perc50: 1.890265159s, perc90: 2.633388881s, perc99: 2.961553187s I1120 18:37:09.703] Nov 20 18:37:04.577: INFO: Approx throughput: 1651.3619009898246 pods/min I1120 18:37:09.931] STEP: Removing additional replication controllers I1120 18:37:10.139] STEP: deleting { ReplicationController} density-latency-pod-1 in namespace e2e-tests-density-30-1-2jts7, will wait for the garbage collector to delete the pods [...] I1120 18:37:25.349] STEP: deleting { ReplicationController} density-latency-pod-25 in namespace e2e-tests-density-30-1-2jts7, will wait for the garbage collector to delete the pods I1120 18:37:28.821] Nov 20 18:37:25.285: INFO: Error while trying to SSH to 104.196.71.126:22, skipping probe. Error: error creating session to prow@104.196.71.126:22: 'read tcp 10.60.162.41:36592->104.196.71.126:22: i/o timeout' I1120 18:37:56.998] error dialing prow@35.196.146.229:22: 'ssh: handshake failed: read tcp 10.60.162.41:52670->35.196.146.229:22: i/o timeout', retrying I1120 18:38:03.493] error dialing prow@35.196.134.58:22: 'ssh: handshake failed: read tcp 10.60.162.41:56588->35.196.134.58:22: i/o timeout', retrying I1120 18:38:10.249] error dialing prow@35.196.152.43:22: 'ssh: handshake failed: read tcp 10.60.162.41:50130->35.196.152.43:22: i/o timeout', retrying I1120 18:38:31.730] error dialing prow@35.196.10.128:22: 'ssh: handshake failed: read tcp 10.60.162.41:42388->35.196.10.128:22: i/o timeout', retrying I1120 18:38:48.126] Nov 20 18:38:47.768: INFO: Error while trying to SSH to 35.196.134.58:22, skipping probe. Error: error getting SSH client to prow@35.196.134.58:22: 'ssh: handshake failed: read tcp 10.60.162.41:56716->35.196.134.58:22: i/o timeout' I1120 18:38:50.602] Nov 20 18:38:50.597: INFO: Error while trying to SSH to 35.196.146.229:22, skipping probe. Error: error getting SSH client to prow@35.196.146.229:22: 'ssh: handshake failed: read tcp 10.60.162.41:52798->35.196.146.229:22: i/o timeout' I1120 18:39:18.638] error dialing prow@104.196.188.66:22: 'ssh: handshake failed: read tcp 10.60.162.41:48272->104.196.188.66:22: i/o timeout', retrying I1120 18:39:25.394] Nov 20 18:39:25.342: INFO: Error while trying to SSH to 35.196.152.43:22, skipping probe. Error: error getting SSH client to prow@35.196.152.43:22: 'ssh: handshake failed: read tcp 10.60.162.41:50242->35.196.152.43:22: i/o timeout' I1120 18:39:41.821] Nov 20 18:39:41.706: INFO: Error while trying to SSH to 35.196.10.128:22, skipping probe. Error: error getting SSH client to prow@35.196.10.128:22: 'ssh: handshake failed: read tcp 10.60.162.41:42518->35.196.10.128:22: i/o timeout' I1120 18:40:19.282] error dialing prow@35.196.199.87:22: 'ssh: handshake failed: read tcp 10.60.162.41:33490->35.196.199.87:22: i/o timeout', retrying I1120 18:40:19.886] error dialing prow@104.196.216.172:22: 'ssh: handshake failed: read tcp 10.60.162.41:39284->104.196.216.172:22: i/o timeout', retrying I1120 18:40:31.097] Nov 20 18:40:30.896: INFO: Error while trying to SSH to 104.196.188.66:22, skipping probe. Error: error getting SSH client to prow@104.196.188.66:22: 'ssh: handshake failed: read tcp 10.60.162.41:48440->104.196.188.66:22: i/o timeout' I1120 18:40:41.794] error dialing prow@35.196.174.144:22: 'ssh: handshake failed: read tcp 10.60.162.41:52262->35.196.174.144:22: i/o timeout', retrying I1120 18:43:04.790] Nov 20 18:42:58.627: INFO: Unexpected error occurred: timed out waiting for the condition I1120 18:43:09.789] Nov 20 18:42:58.627: INFO: Unexpected error occurred: timed out waiting for the condition I1120 18:43:11.040] STEP: deleting { ReplicationController} density-latency-pod-26 in namespace e2e-tests-density-30-1-2jts7, will wait for the garbage collector to delete the pods [...] I1120 18:45:06.875] Nov 20 18:45:06.875: INFO: Terminating { ReplicationController} density-latency-pod-99 pods took: 103.922056ms I1120 18:45:06.876] Nov 20 18:45:06.876: INFO: Terminating { ReplicationController} density-latency-pod-100 pods took: 102.482037ms I1120 18:45:18.075] STEP: Deleting created Collections I1120 18:45:18.075] STEP: Cleaning up only the { ReplicationController}, garbage collector will clean up the pods I1120 18:45:18.076] STEP: deleting { ReplicationController} density3000-0-2e4a4c8c-ce20-11e7-931f-0a580a3ca229 in namespace e2e-tests-density-30-1-2jts7, will wait for the garbage collector to delete the pods I1120 18:45:19.356] Nov 20 18:45:19.356: INFO: Deleting { ReplicationController} density3000-0-2e4a4c8c-ce20-11e7-931f-0a580a3ca229 took: 40.281025ms I1120 18:49:07.842] W1120 18:49:07.842619 5000 reflector.go:334] k8s.io/kubernetes/test/utils/pod_store.go:53: watch of *v1.Pod ended with: too old resource version: 75796 (87313) I1120 18:49:35.783] W1120 18:49:35.783341 5000 reflector.go:334] k8s.io/kubernetes/test/utils/pod_store.go:53: watch of *v1.Pod ended with: too old resource version: 75796 (88754) I1120 18:49:38.395] W1120 18:49:38.395045 5000 reflector.go:334] k8s.io/kubernetes/test/utils/pod_store.go:53: watch of *v1.Pod ended with: too old resource version: 75796 (88957) I1120 18:49:48.550] W1120 18:49:48.549348 5000 reflector.go:334] k8s.io/kubernetes/test/utils/pod_store.go:53: watch of *v1.Pod ended with: too old resource version: 75796 (89485) I1120 18:49:57.774] W1120 18:49:57.774640 5000 reflector.go:334] k8s.io/kubernetes/test/utils/pod_store.go:53: watch of *v1.Pod ended with: too old resource version: 75796 (89983) I1120 18:50:19.357] Nov 20 18:50:19.357: INFO: Terminating { ReplicationController} density3000-0-2e4a4c8c-ce20-11e7-931f-0a580a3ca229 pods took: 5m0.00090081s I1120 18:50:41.768] W1120 18:50:41.768353 5000 reflector.go:334] k8s.io/kubernetes/test/utils/pod_store.go:53: watch of *v1.Pod ended with: too old resource version: 75796 (91177) I1120 18:50:43.527] [AfterEach] [sig-scalability] Density I1120 18:50:43.527] /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scalability/density.go:357 I1120 18:50:43.529] [AfterEach] [sig-scalability] Density I1120 18:50:43.529] /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:135 I1120 18:50:43.530] STEP: Collecting events from namespace "e2e-tests-density-vr5m7". I1120 18:50:43.569] STEP: Found 0 events. ```
test
small scale density test failure on branch is this a bug report or feature request kind bug kubernetes sig scalability bugs priority failing test assign what happened failed with density go src io kubernetes output dockerized go src io kubernetes test scalability framework go should allow starting pods per node using replicationcontroller with secrets configmaps and daemons go src io kubernetes output dockerized go src io kubernetes test scalability density go expected error s timed out waiting for the condition timed out waiting for the condition not to have occurred go src io kubernetes output dockerized go src io kubernetes test scalability density go a little earlier the logs read step scheduling additional pods to measure startup latencies nov info found pod density latency pod running nov info found pod density latency pod running step waiting for all pods begin observed by the watch nov info latency pods scheduled on scalability minion group qxqh error dialing prow ssh handshake failed read tcp i o timeout retrying nov info worst schedule latencies nov info nov info worst run after schedule latencies nov info nov info worst watch latencies nov info nov info worst scheduled to end total latencies nov info nov info worst total latencies nov info nov info approx throughput pods min step removing additional replication controllers step deleting replicationcontroller density latency pod in namespace tests density will wait for the garbage collector to delete the pods step deleting replicationcontroller density latency pod in namespace tests density will wait for the garbage collector to delete the pods nov info error while trying to ssh to skipping probe error error creating session to prow read tcp i o timeout error dialing prow ssh handshake failed read tcp i o timeout retrying error dialing prow ssh handshake failed read tcp i o timeout retrying error dialing prow ssh handshake failed read tcp i o timeout retrying error dialing prow ssh handshake failed read tcp i o timeout retrying nov info error while trying to ssh to skipping probe error error getting ssh client to prow ssh handshake failed read tcp i o timeout nov info error while trying to ssh to skipping probe error error getting ssh client to prow ssh handshake failed read tcp i o timeout error dialing prow ssh handshake failed read tcp i o timeout retrying nov info error while trying to ssh to skipping probe error error getting ssh client to prow ssh handshake failed read tcp i o timeout nov info error while trying to ssh to skipping probe error error getting ssh client to prow ssh handshake failed read tcp i o timeout error dialing prow ssh handshake failed read tcp i o timeout retrying error dialing prow ssh handshake failed read tcp i o timeout retrying nov info error while trying to ssh to skipping probe error error getting ssh client to prow ssh handshake failed read tcp i o timeout error dialing prow ssh handshake failed read tcp i o timeout retrying nov info unexpected error occurred timed out waiting for the condition nov info unexpected error occurred timed out waiting for the condition step deleting replicationcontroller density latency pod in namespace tests density will wait for the garbage collector to delete the pods nov info terminating replicationcontroller density latency pod pods took nov info terminating replicationcontroller density latency pod pods took step deleting created collections step cleaning up only the replicationcontroller garbage collector will clean up the pods step deleting replicationcontroller in namespace tests density will wait for the garbage collector to delete the pods nov info deleting replicationcontroller took reflector go io kubernetes test utils pod store go watch of pod ended with too old resource version reflector go io kubernetes test utils pod store go watch of pod ended with too old resource version reflector go io kubernetes test utils pod store go watch of pod ended with too old resource version reflector go io kubernetes test utils pod store go watch of pod ended with too old resource version reflector go io kubernetes test utils pod store go watch of pod ended with too old resource version nov info terminating replicationcontroller pods took reflector go io kubernetes test utils pod store go watch of pod ended with too old resource version density go src io kubernetes output dockerized go src io kubernetes test scalability density go density go src io kubernetes output dockerized go src io kubernetes test framework framework go step collecting events from namespace tests density step found events
1
138,686
11,211,578,612
IssuesEvent
2020-01-06 15:43:57
ayumi-cloud/oc-security-module
https://api.github.com/repos/ayumi-cloud/oc-security-module
closed
Add Maxthon browser to whitelist and block versions 1 and 2
Add to Blacklist Add to Whitelist Analytics FINSIHED Firewall Priority: Medium Testing - Passed enhancement
### Enhancement idea - [x] Add Maxthon browser to whitelist and block versions 1 and 2. - [x] Add Maxthon browser to analytics section.
1.0
Add Maxthon browser to whitelist and block versions 1 and 2 - ### Enhancement idea - [x] Add Maxthon browser to whitelist and block versions 1 and 2. - [x] Add Maxthon browser to analytics section.
test
add maxthon browser to whitelist and block versions and enhancement idea add maxthon browser to whitelist and block versions and add maxthon browser to analytics section
1
127,107
10,452,450,048
IssuesEvent
2019-09-19 14:42:35
ansible/awx
https://api.github.com/repos/ansible/awx
closed
Azure login icon not displayed
component:ui priority:medium qe:regression qe:visual state:needs_test type:bug
##### ISSUE TYPE - Bug Report ##### SUMMARY The Azure AD login icon is not displayed in login page, only a grey circle. ##### ENVIRONMENT * AWX version: 4.0.0 * AWX install method: kubernetes (microk8s) * Ansible version: 2.7.10 * Operating System: Ubuntu 18.04 * Web Browser: Firefox/Chrome ##### STEPS TO REPRODUCE Configured Azure AD Login; the login works fine, but the icon in the login page is a grey circle, instead of an icon. ##### EXPECTED RESULTS Show the Azure AD login icon. ##### ACTUAL RESULTS Only a grey circle is displayed. Screenshot here https://imgur.com/a/FAE3t6S
1.0
Azure login icon not displayed - ##### ISSUE TYPE - Bug Report ##### SUMMARY The Azure AD login icon is not displayed in login page, only a grey circle. ##### ENVIRONMENT * AWX version: 4.0.0 * AWX install method: kubernetes (microk8s) * Ansible version: 2.7.10 * Operating System: Ubuntu 18.04 * Web Browser: Firefox/Chrome ##### STEPS TO REPRODUCE Configured Azure AD Login; the login works fine, but the icon in the login page is a grey circle, instead of an icon. ##### EXPECTED RESULTS Show the Azure AD login icon. ##### ACTUAL RESULTS Only a grey circle is displayed. Screenshot here https://imgur.com/a/FAE3t6S
test
azure login icon not displayed issue type bug report summary the azure ad login icon is not displayed in login page only a grey circle environment awx version awx install method kubernetes ansible version operating system ubuntu web browser firefox chrome steps to reproduce configured azure ad login the login works fine but the icon in the login page is a grey circle instead of an icon expected results show the azure ad login icon actual results only a grey circle is displayed screenshot here
1
213,953
16,548,816,542
IssuesEvent
2021-05-28 05:39:54
shyaginuma/papers
https://api.github.com/repos/shyaginuma/papers
closed
Improving the Sensitivity of Online Controlled Experiments: Case Studies at Netflix
AB testing Paper
## Meta * Link: https://www.kdd.org/kdd2016/papers/files/adp0945-xieA.pdf * Author: Huizhi Xie, Netflix * Date: 2016 * Journal: KDD ## Motivation <!-- この文献に何を期待しているのか --> そもそも分散の小さなMetricsを使って評価をするということ以外に、どのようにしてSensitivityを高めることができるのかを知りたい。 ## Summary <!-- どんなものか、著者がやりたかったものは何か --> Variance Reductionの様々な手法*をNetflixのデータを使って比較している。数式から理論的にはこうなるはずと言う見立てて、結果を考察している。 * Simple Random Sampling: 通常のA/B割り当てと同じ * Stratified Sampling: A/Bの割り当て時に共変量のバランスが取れるように割り当てを行う * Post Stratification: A/Bの効果推定の際に共変量のバランスを考慮し補正を行う * CUPED: [3] の文献で詳細に説明されている。ちょっと一行では説明しづらい ## Contribution * Variance Reductionの手法を一通りまとめ、それらの関係性を俯瞰している * Stratified Samplingの実装方法を提案している点 * Variance Reductionの手法を実際のデータを用いて比較検討している点 ## Contents ## Evaluation ## Discussion > Hi’s are chosen to **minimize the loss of information** from summarizing the distribution of streaming hours using these thresholds. 離散化を行う際のビンの切り方、具体的にはどういう風に行っているのか気になる。 ## 仕事に活用できそうなところ * 新規/既存でセグメントを分けて評価を行う。新規の方がメトリクスが動きやすいとの記述あり。(2.1 Test Audience) ## Next to read * [6] C. A. Gomez-Uribe and N. Hunt. The Netflix recommender system: Algorithms, business value, and innovation. ACM Trans. on Management Information Systems, 6(4), December 2015. Netflixのレコメンデーションにおける過去の施策やその結果がまとまっているらしい * [3] A. Deng, Y. Xu, R. Kohavi, and T. Walker. Improving the sensitivity of online controlled experiments by utilizing pre-experiment data. In WSDM 13 Proceedings of the sixth ACM International Conference on Web Search and Data Mining, pages 123–132, 2013. CUPEDというvariance reduction手法の元論文
1.0
Improving the Sensitivity of Online Controlled Experiments: Case Studies at Netflix - ## Meta * Link: https://www.kdd.org/kdd2016/papers/files/adp0945-xieA.pdf * Author: Huizhi Xie, Netflix * Date: 2016 * Journal: KDD ## Motivation <!-- この文献に何を期待しているのか --> そもそも分散の小さなMetricsを使って評価をするということ以外に、どのようにしてSensitivityを高めることができるのかを知りたい。 ## Summary <!-- どんなものか、著者がやりたかったものは何か --> Variance Reductionの様々な手法*をNetflixのデータを使って比較している。数式から理論的にはこうなるはずと言う見立てて、結果を考察している。 * Simple Random Sampling: 通常のA/B割り当てと同じ * Stratified Sampling: A/Bの割り当て時に共変量のバランスが取れるように割り当てを行う * Post Stratification: A/Bの効果推定の際に共変量のバランスを考慮し補正を行う * CUPED: [3] の文献で詳細に説明されている。ちょっと一行では説明しづらい ## Contribution * Variance Reductionの手法を一通りまとめ、それらの関係性を俯瞰している * Stratified Samplingの実装方法を提案している点 * Variance Reductionの手法を実際のデータを用いて比較検討している点 ## Contents ## Evaluation ## Discussion > Hi’s are chosen to **minimize the loss of information** from summarizing the distribution of streaming hours using these thresholds. 離散化を行う際のビンの切り方、具体的にはどういう風に行っているのか気になる。 ## 仕事に活用できそうなところ * 新規/既存でセグメントを分けて評価を行う。新規の方がメトリクスが動きやすいとの記述あり。(2.1 Test Audience) ## Next to read * [6] C. A. Gomez-Uribe and N. Hunt. The Netflix recommender system: Algorithms, business value, and innovation. ACM Trans. on Management Information Systems, 6(4), December 2015. Netflixのレコメンデーションにおける過去の施策やその結果がまとまっているらしい * [3] A. Deng, Y. Xu, R. Kohavi, and T. Walker. Improving the sensitivity of online controlled experiments by utilizing pre-experiment data. In WSDM 13 Proceedings of the sixth ACM International Conference on Web Search and Data Mining, pages 123–132, 2013. CUPEDというvariance reduction手法の元論文
test
improving the sensitivity of online controlled experiments case studies at netflix meta link author huizhi xie netflix date journal kdd motivation そもそも分散の小さなmetricsを使って評価をするということ以外に、どのようにしてsensitivityを高めることができるのかを知りたい。 summary variance reductionの様々な手法 をnetflixのデータを使って比較している。数式から理論的にはこうなるはずと言う見立てて、結果を考察している。 simple random sampling 通常のa b割り当てと同じ stratified sampling a bの割り当て時に共変量のバランスが取れるように割り当てを行う post stratification a bの効果推定の際に共変量のバランスを考慮し補正を行う cuped の文献で詳細に説明されている。ちょっと一行では説明しづらい contribution variance reductionの手法を一通りまとめ、それらの関係性を俯瞰している stratified samplingの実装方法を提案している点 variance reductionの手法を実際のデータを用いて比較検討している点 contents evaluation discussion hi’s are chosen to minimize the loss of information from summarizing the distribution of streaming hours using these thresholds 離散化を行う際のビンの切り方、具体的にはどういう風に行っているのか気になる。 仕事に活用できそうなところ 新規 既存でセグメントを分けて評価を行う。新規の方がメトリクスが動きやすいとの記述あり。 test audience next to read c a gomez uribe and n hunt the netflix recommender system algorithms business value and innovation acm trans on management information systems december netflixのレコメンデーションにおける過去の施策やその結果がまとまっているらしい a deng y xu r kohavi and t walker improving the sensitivity of online controlled experiments by utilizing pre experiment data in wsdm proceedings of the sixth acm international conference on web search and data mining pages – cupedというvariance reduction手法の元論文
1
146,257
13,174,993,485
IssuesEvent
2020-08-12 00:09:37
nicholaschiasson/ngx_upstream_jdomain
https://api.github.com/repos/nicholaschiasson/ngx_upstream_jdomain
closed
Improve README.md
documentation good first issue help wanted
Document the project better. - [ ] Features: all supported directive attributes and what they do. - [ ] Development tools to use for local development: docker, [act](https://github.com/nektos/act), etc. - [ ] General instructions for local development: building, testing, running github actions locally, etc.
1.0
Improve README.md - Document the project better. - [ ] Features: all supported directive attributes and what they do. - [ ] Development tools to use for local development: docker, [act](https://github.com/nektos/act), etc. - [ ] General instructions for local development: building, testing, running github actions locally, etc.
non_test
improve readme md document the project better features all supported directive attributes and what they do development tools to use for local development docker etc general instructions for local development building testing running github actions locally etc
0
64,716
14,677,226,785
IssuesEvent
2020-12-30 22:35:03
GooseWSS/ksa
https://api.github.com/repos/GooseWSS/ksa
opened
CVE-2014-3578 (Medium) detected in spring-core-3.1.1.RELEASE.jar
security vulnerability
## CVE-2014-3578 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-core-3.1.1.RELEASE.jar</b></p></summary> <p>Spring Framework Parent</p> <p>Path to dependency file: ksa/ksa-core/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,ksa/ksa-web-root/ksa-web/target/ROOT/WEB-INF/lib/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar</p> <p> Dependency Hierarchy: - :x: **spring-core-3.1.1.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/GooseWSS/ksa/commit/92b4fc1a7755c8d454d53e9ae803447b86a9521a">92b4fc1a7755c8d454d53e9ae803447b86a9521a</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Directory traversal vulnerability in Pivotal Spring Framework 3.x before 3.2.9 and 4.0 before 4.0.5 allows remote attackers to read arbitrary files via a crafted URL. <p>Publish Date: 2015-02-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-3578>CVE-2014-3578</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2014-3578">https://nvd.nist.gov/vuln/detail/CVE-2014-3578</a></p> <p>Release Date: 2015-02-19</p> <p>Fix Resolution: 3.2.9,4.0.5</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework","packageName":"spring-core","packageVersion":"3.1.1.RELEASE","isTransitiveDependency":false,"dependencyTree":"org.springframework:spring-core:3.1.1.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.2.9,4.0.5"}],"vulnerabilityIdentifier":"CVE-2014-3578","vulnerabilityDetails":"Directory traversal vulnerability in Pivotal Spring Framework 3.x before 3.2.9 and 4.0 before 4.0.5 allows remote attackers to read arbitrary files via a crafted URL.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-3578","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> -->
True
CVE-2014-3578 (Medium) detected in spring-core-3.1.1.RELEASE.jar - ## CVE-2014-3578 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-core-3.1.1.RELEASE.jar</b></p></summary> <p>Spring Framework Parent</p> <p>Path to dependency file: ksa/ksa-core/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,ksa/ksa-web-root/ksa-web/target/ROOT/WEB-INF/lib/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-core/3.1.1.RELEASE/spring-core-3.1.1.RELEASE.jar</p> <p> Dependency Hierarchy: - :x: **spring-core-3.1.1.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/GooseWSS/ksa/commit/92b4fc1a7755c8d454d53e9ae803447b86a9521a">92b4fc1a7755c8d454d53e9ae803447b86a9521a</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Directory traversal vulnerability in Pivotal Spring Framework 3.x before 3.2.9 and 4.0 before 4.0.5 allows remote attackers to read arbitrary files via a crafted URL. <p>Publish Date: 2015-02-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-3578>CVE-2014-3578</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2014-3578">https://nvd.nist.gov/vuln/detail/CVE-2014-3578</a></p> <p>Release Date: 2015-02-19</p> <p>Fix Resolution: 3.2.9,4.0.5</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework","packageName":"spring-core","packageVersion":"3.1.1.RELEASE","isTransitiveDependency":false,"dependencyTree":"org.springframework:spring-core:3.1.1.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.2.9,4.0.5"}],"vulnerabilityIdentifier":"CVE-2014-3578","vulnerabilityDetails":"Directory traversal vulnerability in Pivotal Spring Framework 3.x before 3.2.9 and 4.0 before 4.0.5 allows remote attackers to read arbitrary files via a crafted URL.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-3578","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> -->
non_test
cve medium detected in spring core release jar cve medium severity vulnerability vulnerable library spring core release jar spring framework parent path to dependency file ksa ksa core pom xml path to vulnerable library canner repository org springframework spring core release spring core release jar home wss scanner repository org springframework spring core release spring core release jar home wss scanner repository org springframework spring core release spring core release jar home wss scanner repository org springframework spring core release spring core release jar home wss scanner repository org springframework spring core release spring core release jar home wss scanner repository org springframework spring core release spring core release jar home wss scanner repository org springframework spring core release spring core release jar home wss scanner repository org springframework spring core release spring core release jar home wss scanner repository org springframework spring core release spring core release jar home wss scanner repository org springframework spring core release spring core release jar home wss scanner repository org springframework spring core release spring core release jar home wss scanner repository org springframework spring core release spring core release jar home wss scanner repository org springframework spring core release spring core release jar home wss scanner repository org springframework spring core release spring core release jar home wss scanner repository org springframework spring core release spring core release jar home wss scanner repository org springframework spring core release spring core release jar home wss scanner repository org springframework spring core release spring core release jar ksa ksa web root ksa web target root web inf lib spring core release jar home wss scanner repository org springframework spring core release spring core release jar home wss scanner repository org springframework spring core release spring core release jar dependency hierarchy x spring core release jar vulnerable library found in head commit a href found in base branch master vulnerability details directory traversal vulnerability in pivotal spring framework x before and before allows remote attackers to read arbitrary files via a crafted url publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails directory traversal vulnerability in pivotal spring framework x before and before allows remote attackers to read arbitrary files via a crafted url vulnerabilityurl
0
141,225
11,404,723,774
IssuesEvent
2020-01-31 10:25:10
mozilla/addons-frontend
https://api.github.com/repos/mozilla/addons-frontend
closed
Error from FingersCrossedReporter when a test fails
component: testing priority: p2 qa: not needed
When running a test via Jest, if the test fails we now see an error being generated from the `FingersCrossedReporter` as opposed to the failure from the test. It looks like `getConsoleOutput` is no longer available from `jest-util`, and that is causing an error at https://github.com/mozilla/addons-frontend/blob/master/tests/jest-reporters/fingers-crossed.js#L23. Here's some example output: ``` FAIL tests/unit/amo/components/TestInstallWarning.js FAIL tests/unit/amo/components/TestInstallWarning.js ● Test suite failed to run TypeError: getConsoleOutput is not a function 21 | // prettier-ignore 22 | this.log( > 23 | ` ${TITLE_BULLET}Console\n\n${getConsoleOutput( | ^ 24 | config.cwd, 25 | !!this._globalConfig.verbose, 26 | consoleBuffer at FingersCrossedReporter.printTestFileHeader (tests/jest-reporters/fingers-crossed.js:23:40) at FingersCrossedReporter.onTestResult (node_modules/@jest/reporters/build/default_reporter.js:200:12) at ReporterDispatcher.onTestResult (node_modules/@jest/core/build/ReporterDispatcher.js:48:25) at onResult (node_modules/@jest/core/build/TestScheduler.js:193:30) Test Suites: 2 failed, 2 of 1 total Tests: 8 failed, 16 passed, 24 total Snapshots: 0 total Time: 3.641s Ran all test suites matching /installw/i. Flow: no errors 🌈 ✨ Eslint: no errors 💄 ✨ Watch Usage: Press w to show more. ``` I have confirmed that this regression occurred with the landing of https://github.com/mozilla/addons-frontend/commit/78f716e76b581a63aa8466105773b072252e6feb.
1.0
Error from FingersCrossedReporter when a test fails - When running a test via Jest, if the test fails we now see an error being generated from the `FingersCrossedReporter` as opposed to the failure from the test. It looks like `getConsoleOutput` is no longer available from `jest-util`, and that is causing an error at https://github.com/mozilla/addons-frontend/blob/master/tests/jest-reporters/fingers-crossed.js#L23. Here's some example output: ``` FAIL tests/unit/amo/components/TestInstallWarning.js FAIL tests/unit/amo/components/TestInstallWarning.js ● Test suite failed to run TypeError: getConsoleOutput is not a function 21 | // prettier-ignore 22 | this.log( > 23 | ` ${TITLE_BULLET}Console\n\n${getConsoleOutput( | ^ 24 | config.cwd, 25 | !!this._globalConfig.verbose, 26 | consoleBuffer at FingersCrossedReporter.printTestFileHeader (tests/jest-reporters/fingers-crossed.js:23:40) at FingersCrossedReporter.onTestResult (node_modules/@jest/reporters/build/default_reporter.js:200:12) at ReporterDispatcher.onTestResult (node_modules/@jest/core/build/ReporterDispatcher.js:48:25) at onResult (node_modules/@jest/core/build/TestScheduler.js:193:30) Test Suites: 2 failed, 2 of 1 total Tests: 8 failed, 16 passed, 24 total Snapshots: 0 total Time: 3.641s Ran all test suites matching /installw/i. Flow: no errors 🌈 ✨ Eslint: no errors 💄 ✨ Watch Usage: Press w to show more. ``` I have confirmed that this regression occurred with the landing of https://github.com/mozilla/addons-frontend/commit/78f716e76b581a63aa8466105773b072252e6feb.
test
error from fingerscrossedreporter when a test fails when running a test via jest if the test fails we now see an error being generated from the fingerscrossedreporter as opposed to the failure from the test it looks like getconsoleoutput is no longer available from jest util and that is causing an error at here s some example output fail tests unit amo components testinstallwarning js fail tests unit amo components testinstallwarning js ● test suite failed to run typeerror getconsoleoutput is not a function prettier ignore this log title bullet console n n getconsoleoutput config cwd this globalconfig verbose consolebuffer at fingerscrossedreporter printtestfileheader tests jest reporters fingers crossed js at fingerscrossedreporter ontestresult node modules jest reporters build default reporter js at reporterdispatcher ontestresult node modules jest core build reporterdispatcher js at onresult node modules jest core build testscheduler js test suites failed of total tests failed passed total snapshots total time ran all test suites matching installw i flow no errors 🌈 ✨ eslint no errors 💄 ✨ watch usage press w to show more i have confirmed that this regression occurred with the landing of
1
176,546
13,645,386,009
IssuesEvent
2020-09-25 20:41:16
namhyung/uftrace
https://api.github.com/repos/namhyung/uftrace
opened
Many kernel tests fail in aarch64
arch kernel tests
Many kernel tests fail in aarch64 as follows. ``` $ ./runtest.py kernel Start 15 tests with 8 worker Test case pg finstrument-fu ------------------------: O0 O1 O2 O3 Os O0 O1 O2 O3 Os 022 filter_kernel : OK OK OK OK OK OK OK OK OK OK 079 replay_kernel_D : NG NG NG NG NG NG NG NG NG NG 080 replay_kernel_D2 : OK OK OK OK OK OK OK OK OK OK 081 kernel_depth : NG NG NG NG NG NG NG NG NG NG 103 dump_kernel : NG NG NG NG NG NG NG NG NG NG 104 graph_kernel : OK NG NG OK NG NG NG NG NG NG 111 kernel_tid : NG NG NG NG NG NG NG NG NG NG 132 trigger_kernel : NG NG NG NG NG NG NG NG NG NG 137 kernel_tid_update : NG NG NG NG NG NG NG NG NG NG 138 kernel_dynamic : NG NG NG NG NG NG NG NG NG NG 139 kernel_dynamic2 : NG NG NG NG NG NG NG NG NG NG 143 recv_kernel : NG NG NG NG NG NG NG NG NG NG 148 event_kernel : OK OK OK OK OK OK OK OK OK OK 149 event_kernel2 : NG NG NG NG NG NG NG NG NG NG 174 replay_filter_kernel: NG NG NG NG NG NG NG NG NG NG ``` Some of the test looks as follows. ``` t081_kernel_depth: diff result of -pg -O0 --- expect 2020-09-25 20:31:32.946663505 +0000 +++ result 2020-09-25 20:31:32.946663505 +0000 @@ -4,2 +4,3 @@ do_sys_open(); + __exception_text_end(); } /* sys_openat */ @@ -10,2 +11,9 @@ } /* sys_close */ + __exception_text_end() { + handle_IPI(); + } /* __exception_text_end */ + do_notify_resume() { + task_work_run(); + mem_cgroup_handle_over_high(); + } /* do_notify_resume */ } /* fclose */ 081 kernel_depth : NG ``` It's not well filtered out with the following change. ```diff diff --git a/tests/t081_kernel_depth.py b/tests/t081_kernel_depth.py index 16227a9f..43036c8a 100644 --- a/tests/t081_kernel_depth.py +++ b/tests/t081_kernel_depth.py @@ -33,7 +33,9 @@ class TestCase(TestBase): def setup(self): self.option = '-k --kernel-depth=2 --match glob ' self.option += '-N exit_to_usermode_loop@kernel ' - self.option += '-N *do_page_fault@kernel' + self.option += '-N *do_page_fault@kernel ' + self.option += '-N __exception_text_end@kernel ' + self.option += '-N do_notify_resume@kernel ' def fixup(self, cflags, result): uname = os.uname() ``` It still shows the failure. ``` t081_kernel_depth: diff result of -pg -O0 --- expect 2020-09-25 20:40:46.735597411 +0000 +++ result 2020-09-25 20:40:46.735597411 +0000 @@ -3,3 +3,7 @@ sys_openat() { + __exception_text_end() { + handle_IPI(); + } /* __exception_text_end */ do_sys_open(); + __exception_text_end(); } /* sys_openat */ 081 kernel_depth : NG ```
1.0
Many kernel tests fail in aarch64 - Many kernel tests fail in aarch64 as follows. ``` $ ./runtest.py kernel Start 15 tests with 8 worker Test case pg finstrument-fu ------------------------: O0 O1 O2 O3 Os O0 O1 O2 O3 Os 022 filter_kernel : OK OK OK OK OK OK OK OK OK OK 079 replay_kernel_D : NG NG NG NG NG NG NG NG NG NG 080 replay_kernel_D2 : OK OK OK OK OK OK OK OK OK OK 081 kernel_depth : NG NG NG NG NG NG NG NG NG NG 103 dump_kernel : NG NG NG NG NG NG NG NG NG NG 104 graph_kernel : OK NG NG OK NG NG NG NG NG NG 111 kernel_tid : NG NG NG NG NG NG NG NG NG NG 132 trigger_kernel : NG NG NG NG NG NG NG NG NG NG 137 kernel_tid_update : NG NG NG NG NG NG NG NG NG NG 138 kernel_dynamic : NG NG NG NG NG NG NG NG NG NG 139 kernel_dynamic2 : NG NG NG NG NG NG NG NG NG NG 143 recv_kernel : NG NG NG NG NG NG NG NG NG NG 148 event_kernel : OK OK OK OK OK OK OK OK OK OK 149 event_kernel2 : NG NG NG NG NG NG NG NG NG NG 174 replay_filter_kernel: NG NG NG NG NG NG NG NG NG NG ``` Some of the test looks as follows. ``` t081_kernel_depth: diff result of -pg -O0 --- expect 2020-09-25 20:31:32.946663505 +0000 +++ result 2020-09-25 20:31:32.946663505 +0000 @@ -4,2 +4,3 @@ do_sys_open(); + __exception_text_end(); } /* sys_openat */ @@ -10,2 +11,9 @@ } /* sys_close */ + __exception_text_end() { + handle_IPI(); + } /* __exception_text_end */ + do_notify_resume() { + task_work_run(); + mem_cgroup_handle_over_high(); + } /* do_notify_resume */ } /* fclose */ 081 kernel_depth : NG ``` It's not well filtered out with the following change. ```diff diff --git a/tests/t081_kernel_depth.py b/tests/t081_kernel_depth.py index 16227a9f..43036c8a 100644 --- a/tests/t081_kernel_depth.py +++ b/tests/t081_kernel_depth.py @@ -33,7 +33,9 @@ class TestCase(TestBase): def setup(self): self.option = '-k --kernel-depth=2 --match glob ' self.option += '-N exit_to_usermode_loop@kernel ' - self.option += '-N *do_page_fault@kernel' + self.option += '-N *do_page_fault@kernel ' + self.option += '-N __exception_text_end@kernel ' + self.option += '-N do_notify_resume@kernel ' def fixup(self, cflags, result): uname = os.uname() ``` It still shows the failure. ``` t081_kernel_depth: diff result of -pg -O0 --- expect 2020-09-25 20:40:46.735597411 +0000 +++ result 2020-09-25 20:40:46.735597411 +0000 @@ -3,3 +3,7 @@ sys_openat() { + __exception_text_end() { + handle_IPI(); + } /* __exception_text_end */ do_sys_open(); + __exception_text_end(); } /* sys_openat */ 081 kernel_depth : NG ```
test
many kernel tests fail in many kernel tests fail in as follows runtest py kernel start tests with worker test case pg finstrument fu os os filter kernel ok ok ok ok ok ok ok ok ok ok replay kernel d ng ng ng ng ng ng ng ng ng ng replay kernel ok ok ok ok ok ok ok ok ok ok kernel depth ng ng ng ng ng ng ng ng ng ng dump kernel ng ng ng ng ng ng ng ng ng ng graph kernel ok ng ng ok ng ng ng ng ng ng kernel tid ng ng ng ng ng ng ng ng ng ng trigger kernel ng ng ng ng ng ng ng ng ng ng kernel tid update ng ng ng ng ng ng ng ng ng ng kernel dynamic ng ng ng ng ng ng ng ng ng ng kernel ng ng ng ng ng ng ng ng ng ng recv kernel ng ng ng ng ng ng ng ng ng ng event kernel ok ok ok ok ok ok ok ok ok ok event ng ng ng ng ng ng ng ng ng ng replay filter kernel ng ng ng ng ng ng ng ng ng ng some of the test looks as follows kernel depth diff result of pg expect result do sys open exception text end sys openat sys close exception text end handle ipi exception text end do notify resume task work run mem cgroup handle over high do notify resume fclose kernel depth ng it s not well filtered out with the following change diff diff git a tests kernel depth py b tests kernel depth py index a tests kernel depth py b tests kernel depth py class testcase testbase def setup self self option k kernel depth match glob self option n exit to usermode loop kernel self option n do page fault kernel self option n do page fault kernel self option n exception text end kernel self option n do notify resume kernel def fixup self cflags result uname os uname it still shows the failure kernel depth diff result of pg expect result sys openat exception text end handle ipi exception text end do sys open exception text end sys openat kernel depth ng
1
77,513
7,575,419,357
IssuesEvent
2018-04-24 01:35:30
pandas-dev/pandas
https://api.github.com/repos/pandas-dev/pandas
closed
TST: failure on s3 mocking
IO Parquet Testing
master seems to fail recently: https://travis-ci.org/pandas-dev/pandas/jobs/367585545 on ``pandas/tests/iotest_parquet.py -k s3``. works locally for me. very odd as this is mocked and shouldn't ever have a connection error. In particular xfailing: https://github.com/pandas-dev/pandas/commit/336fba7c0191444c3328009e6d4f9f5d00ee224b - [ ] ``test_read_csv_chunked_download`` in ``pandas.tests.io.test_network.py`` - [ ] ``test_s3_roundtrip`` for fp & pyarrow in ``pandas.testsio.test_parquet.py``
1.0
TST: failure on s3 mocking - master seems to fail recently: https://travis-ci.org/pandas-dev/pandas/jobs/367585545 on ``pandas/tests/iotest_parquet.py -k s3``. works locally for me. very odd as this is mocked and shouldn't ever have a connection error. In particular xfailing: https://github.com/pandas-dev/pandas/commit/336fba7c0191444c3328009e6d4f9f5d00ee224b - [ ] ``test_read_csv_chunked_download`` in ``pandas.tests.io.test_network.py`` - [ ] ``test_s3_roundtrip`` for fp & pyarrow in ``pandas.testsio.test_parquet.py``
test
tst failure on mocking master seems to fail recently on pandas tests iotest parquet py k works locally for me very odd as this is mocked and shouldn t ever have a connection error in particular xfailing test read csv chunked download in pandas tests io test network py test roundtrip for fp pyarrow in pandas testsio test parquet py
1
600,706
18,352,129,784
IssuesEvent
2021-10-08 13:44:33
aiidateam/aiida-quantumespresso
https://api.github.com/repos/aiidateam/aiida-quantumespresso
closed
Protocols: `overrides` should take precedence over `ElectronicType`
type/bug priority/critical topic/protocol
Since in the `PwBaseWorkChain` the `overrides` are merged with the inputs before adapting the inputs based on the `ElectronicType`, the latter will change the final inputs submitted to the builder. This is problematic in case of e.g. the `PdosWorkChain`, where `occupations` should be set to `tetrahedra`. Trying to run the `PdosWorkChain` with `ElectronicType.INSULATOR` will simply fail immediately due to the validation. It seems to me that the `overrides` should be absolute, i.e. merged with the inputs right before setting up the builder. However, since the `overrides` are also used to pass default protocol settings down the hierarchy, I'd have to double check if this causes problems anywhere.
1.0
Protocols: `overrides` should take precedence over `ElectronicType` - Since in the `PwBaseWorkChain` the `overrides` are merged with the inputs before adapting the inputs based on the `ElectronicType`, the latter will change the final inputs submitted to the builder. This is problematic in case of e.g. the `PdosWorkChain`, where `occupations` should be set to `tetrahedra`. Trying to run the `PdosWorkChain` with `ElectronicType.INSULATOR` will simply fail immediately due to the validation. It seems to me that the `overrides` should be absolute, i.e. merged with the inputs right before setting up the builder. However, since the `overrides` are also used to pass default protocol settings down the hierarchy, I'd have to double check if this causes problems anywhere.
non_test
protocols overrides should take precedence over electronictype since in the pwbaseworkchain the overrides are merged with the inputs before adapting the inputs based on the electronictype the latter will change the final inputs submitted to the builder this is problematic in case of e g the pdosworkchain where occupations should be set to tetrahedra trying to run the pdosworkchain with electronictype insulator will simply fail immediately due to the validation it seems to me that the overrides should be absolute i e merged with the inputs right before setting up the builder however since the overrides are also used to pass default protocol settings down the hierarchy i d have to double check if this causes problems anywhere
0
174,218
13,462,327,011
IssuesEvent
2020-09-09 15:55:37
phetsims/griddle
https://api.github.com/repos/phetsims/griddle
closed
CT object is not compatible with merge
type:automated-testing
``` griddle : fuzz : unbuilt https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/griddle_en.html?continuousTest=%7B%22test%22%3A%5B%22griddle%22%2C%22fuzz%22%2C%22unbuilt%22%5D%2C%22snapshotName%22%3A%22snapshot-1599640382837%22%2C%22timestamp%22%3A1599641397075%7D&brand=phet&ea&fuzz&memoryLimit=1000 Query: brand=phet&ea&fuzz&memoryLimit=1000 Uncaught Error: Assertion failed: object is not compatible with merge Error: Assertion failed: object is not compatible with merge at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/assert/js/assert.js:22:13) at assertIsMergeable (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/phet-core/js/merge.js:59:3) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/phet-core/js/merge.js:31:17 at arrayEach (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/sherpa/lib/lodash-4.17.4.js:537:11) at Function.forEach (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/sherpa/lib/lodash-4.17.4.js:9359:14) at merge (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/phet-core/js/merge.js:29:5) at new ScrollingChartNode (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/js/ScrollingChartNode.js:45:15) at new XYCursorPlot (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/js/XYCursorPlot.js:50:5) at Object.demoXYCursorPlot [as createNode] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/js/demo/GriddleDemoScreenView.js:489:20) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/sun/js/demo/DemosScreenView.js:129:24 id: Bayes Chrome Snapshot from 9/9/2020, 2:33:02 AM ---------------------------------- griddle : fuzz : unbuilt https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/griddle_en.html?continuousTest=%7B%22test%22%3A%5B%22griddle%22%2C%22fuzz%22%2C%22unbuilt%22%5D%2C%22snapshotName%22%3A%22snapshot-1599640382837%22%2C%22timestamp%22%3A1599646954172%7D&brand=phet&ea&fuzz&memoryLimit=1000 Query: brand=phet&ea&fuzz&memoryLimit=1000 Uncaught Error: Assertion failed: object is not compatible with merge Error: Assertion failed: object is not compatible with merge at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/assert/js/assert.js:22:13) at assertIsMergeable (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/phet-core/js/merge.js:59:3) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/phet-core/js/merge.js:31:17 at arrayEach (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/sherpa/lib/lodash-4.17.4.js:537:11) at Function.forEach (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/sherpa/lib/lodash-4.17.4.js:9359:14) at merge (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/phet-core/js/merge.js:29:5) at new ScrollingChartNode (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/js/ScrollingChartNode.js:45:15) at new XYCursorPlot (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/js/XYCursorPlot.js:50:5) at Object.demoXYCursorPlot [as createNode] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/js/demo/GriddleDemoScreenView.js:489:20) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/sun/js/demo/DemosScreenView.js:129:24 id: Bayes Chrome Snapshot from 9/9/2020, 2:33:02 AM ---------------------------------- griddle : xss-fuzz https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/griddle_en.html?continuousTest=%7B%22test%22%3A%5B%22griddle%22%2C%22xss-fuzz%22%5D%2C%22snapshotName%22%3A%22snapshot-1599640382837%22%2C%22timestamp%22%3A1599641390959%7D&brand=phet&ea&fuzz&stringTest=xss&memoryLimit=1000 Query: brand=phet&ea&fuzz&stringTest=xss&memoryLimit=1000 Uncaught Error: Assertion failed: object is not compatible with merge Error: Assertion failed: object is not compatible with merge at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/assert/js/assert.js:22:13) at assertIsMergeable (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/phet-core/js/merge.js:59:3) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/phet-core/js/merge.js:31:17 at arrayEach (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/sherpa/lib/lodash-4.17.4.js:537:11) at Function.forEach (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/sherpa/lib/lodash-4.17.4.js:9359:14) at merge (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/phet-core/js/merge.js:29:5) at new ScrollingChartNode (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/js/ScrollingChartNode.js:45:15) at new XYCursorPlot (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/js/XYCursorPlot.js:50:5) at Object.demoXYCursorPlot [as createNode] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/js/demo/GriddleDemoScreenView.js:489:20) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/sun/js/demo/DemosScreenView.js:129:24 id: Bayes Chrome Snapshot from 9/9/2020, 2:33:02 AM ```
1.0
CT object is not compatible with merge - ``` griddle : fuzz : unbuilt https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/griddle_en.html?continuousTest=%7B%22test%22%3A%5B%22griddle%22%2C%22fuzz%22%2C%22unbuilt%22%5D%2C%22snapshotName%22%3A%22snapshot-1599640382837%22%2C%22timestamp%22%3A1599641397075%7D&brand=phet&ea&fuzz&memoryLimit=1000 Query: brand=phet&ea&fuzz&memoryLimit=1000 Uncaught Error: Assertion failed: object is not compatible with merge Error: Assertion failed: object is not compatible with merge at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/assert/js/assert.js:22:13) at assertIsMergeable (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/phet-core/js/merge.js:59:3) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/phet-core/js/merge.js:31:17 at arrayEach (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/sherpa/lib/lodash-4.17.4.js:537:11) at Function.forEach (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/sherpa/lib/lodash-4.17.4.js:9359:14) at merge (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/phet-core/js/merge.js:29:5) at new ScrollingChartNode (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/js/ScrollingChartNode.js:45:15) at new XYCursorPlot (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/js/XYCursorPlot.js:50:5) at Object.demoXYCursorPlot [as createNode] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/js/demo/GriddleDemoScreenView.js:489:20) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/sun/js/demo/DemosScreenView.js:129:24 id: Bayes Chrome Snapshot from 9/9/2020, 2:33:02 AM ---------------------------------- griddle : fuzz : unbuilt https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/griddle_en.html?continuousTest=%7B%22test%22%3A%5B%22griddle%22%2C%22fuzz%22%2C%22unbuilt%22%5D%2C%22snapshotName%22%3A%22snapshot-1599640382837%22%2C%22timestamp%22%3A1599646954172%7D&brand=phet&ea&fuzz&memoryLimit=1000 Query: brand=phet&ea&fuzz&memoryLimit=1000 Uncaught Error: Assertion failed: object is not compatible with merge Error: Assertion failed: object is not compatible with merge at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/assert/js/assert.js:22:13) at assertIsMergeable (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/phet-core/js/merge.js:59:3) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/phet-core/js/merge.js:31:17 at arrayEach (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/sherpa/lib/lodash-4.17.4.js:537:11) at Function.forEach (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/sherpa/lib/lodash-4.17.4.js:9359:14) at merge (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/phet-core/js/merge.js:29:5) at new ScrollingChartNode (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/js/ScrollingChartNode.js:45:15) at new XYCursorPlot (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/js/XYCursorPlot.js:50:5) at Object.demoXYCursorPlot [as createNode] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/js/demo/GriddleDemoScreenView.js:489:20) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/sun/js/demo/DemosScreenView.js:129:24 id: Bayes Chrome Snapshot from 9/9/2020, 2:33:02 AM ---------------------------------- griddle : xss-fuzz https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/griddle_en.html?continuousTest=%7B%22test%22%3A%5B%22griddle%22%2C%22xss-fuzz%22%5D%2C%22snapshotName%22%3A%22snapshot-1599640382837%22%2C%22timestamp%22%3A1599641390959%7D&brand=phet&ea&fuzz&stringTest=xss&memoryLimit=1000 Query: brand=phet&ea&fuzz&stringTest=xss&memoryLimit=1000 Uncaught Error: Assertion failed: object is not compatible with merge Error: Assertion failed: object is not compatible with merge at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/assert/js/assert.js:22:13) at assertIsMergeable (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/phet-core/js/merge.js:59:3) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/phet-core/js/merge.js:31:17 at arrayEach (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/sherpa/lib/lodash-4.17.4.js:537:11) at Function.forEach (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/sherpa/lib/lodash-4.17.4.js:9359:14) at merge (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/phet-core/js/merge.js:29:5) at new ScrollingChartNode (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/js/ScrollingChartNode.js:45:15) at new XYCursorPlot (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/js/XYCursorPlot.js:50:5) at Object.demoXYCursorPlot [as createNode] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/griddle/js/demo/GriddleDemoScreenView.js:489:20) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1599640382837/sun/js/demo/DemosScreenView.js:129:24 id: Bayes Chrome Snapshot from 9/9/2020, 2:33:02 AM ```
test
ct object is not compatible with merge griddle fuzz unbuilt query brand phet ea fuzz memorylimit uncaught error assertion failed object is not compatible with merge error assertion failed object is not compatible with merge at window assertions assertfunction at assertismergeable at at arrayeach at function foreach at merge at new scrollingchartnode at new xycursorplot at object demoxycursorplot at id bayes chrome snapshot from am griddle fuzz unbuilt query brand phet ea fuzz memorylimit uncaught error assertion failed object is not compatible with merge error assertion failed object is not compatible with merge at window assertions assertfunction at assertismergeable at at arrayeach at function foreach at merge at new scrollingchartnode at new xycursorplot at object demoxycursorplot at id bayes chrome snapshot from am griddle xss fuzz query brand phet ea fuzz stringtest xss memorylimit uncaught error assertion failed object is not compatible with merge error assertion failed object is not compatible with merge at window assertions assertfunction at assertismergeable at at arrayeach at function foreach at merge at new scrollingchartnode at new xycursorplot at object demoxycursorplot at id bayes chrome snapshot from am
1
206,546
15,736,926,406
IssuesEvent
2021-03-30 01:45:37
kubernetes/minikube
https://api.github.com/repos/kubernetes/minikube
closed
Failing test: TestFunctional/DockerEnv on Windows Docker
co/docker-driver help wanted kind/failing-test os/windows priority/important-soon
https://storage.googleapis.com/minikube-builds/logs/10603/fb9f08b/Docker_Windows.html#fail_TestFunctional%2fparallel%2fDockerEnv This is due to how we run our windows test. We ssh from a linux machine into a windows machine, starting a cmd terminal that then executes a powershell script. `minikube docker-env` then detects the cmd env and gives us incorrect commands since we're actually in powershell.
1.0
Failing test: TestFunctional/DockerEnv on Windows Docker - https://storage.googleapis.com/minikube-builds/logs/10603/fb9f08b/Docker_Windows.html#fail_TestFunctional%2fparallel%2fDockerEnv This is due to how we run our windows test. We ssh from a linux machine into a windows machine, starting a cmd terminal that then executes a powershell script. `minikube docker-env` then detects the cmd env and gives us incorrect commands since we're actually in powershell.
test
failing test testfunctional dockerenv on windows docker this is due to how we run our windows test we ssh from a linux machine into a windows machine starting a cmd terminal that then executes a powershell script minikube docker env then detects the cmd env and gives us incorrect commands since we re actually in powershell
1
79,662
7,722,871,982
IssuesEvent
2018-05-24 10:34:37
openshiftio/openshift.io
https://api.github.com/repos/openshiftio/openshift.io
closed
Integration Test for Booster support: Clean up
area/tests team/analytics/wizard type/task
# Test case: Clean up A test case for the https://github.com/openshiftio/openshift.io/issues/2399#issuecomment-372520115 test suite. ## Steps after login: - [x] Reset user environment. - [x] Verify environment is cleaned.
1.0
Integration Test for Booster support: Clean up - # Test case: Clean up A test case for the https://github.com/openshiftio/openshift.io/issues/2399#issuecomment-372520115 test suite. ## Steps after login: - [x] Reset user environment. - [x] Verify environment is cleaned.
test
integration test for booster support clean up test case clean up a test case for the test suite steps after login reset user environment verify environment is cleaned
1
339,494
30,451,341,427
IssuesEvent
2023-07-16 10:36:52
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix tensor.test_torch_instance_hsplit
PyTorch Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5566951250/jobs/10168458957"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5566951250/jobs/10168458957"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5566951250/jobs/10168458957"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5566951250/jobs/10168458957"><img src=https://img.shields.io/badge/-failure-red></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5566951250/jobs/10168458957"><img src=https://img.shields.io/badge/-failure-red></a>
1.0
Fix tensor.test_torch_instance_hsplit - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5566951250/jobs/10168458957"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5566951250/jobs/10168458957"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5566951250/jobs/10168458957"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5566951250/jobs/10168458957"><img src=https://img.shields.io/badge/-failure-red></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5566951250/jobs/10168458957"><img src=https://img.shields.io/badge/-failure-red></a>
test
fix tensor test torch instance hsplit tensorflow a href src jax a href src numpy a href src torch a href src paddle a href src
1
107,573
9,217,949,819
IssuesEvent
2019-03-11 12:09:09
magento/magento-coding-standard
https://api.github.com/repos/magento/magento-coding-standard
closed
Change the bin directory
Event: MageTestFest2019 accepted bug good first issue
### Preconditions Have PHP CodeSniffer installed in the project. ### Steps to reproduce 1. Require `magento/magento-coding-standard` 2. Run `vendor/bin/phpcs -i` ### Expected result 1. `Magento` coding Standard is present in the list of installed standards. ### Actual result 1. `Magento` coding Standard is missing in the list of installed standards. ### How to Fix Remove composer `bin-dir`, fix `scripts` to run `vendor/bin/phpcs` and change `README.md`.
1.0
Change the bin directory - ### Preconditions Have PHP CodeSniffer installed in the project. ### Steps to reproduce 1. Require `magento/magento-coding-standard` 2. Run `vendor/bin/phpcs -i` ### Expected result 1. `Magento` coding Standard is present in the list of installed standards. ### Actual result 1. `Magento` coding Standard is missing in the list of installed standards. ### How to Fix Remove composer `bin-dir`, fix `scripts` to run `vendor/bin/phpcs` and change `README.md`.
test
change the bin directory preconditions have php codesniffer installed in the project steps to reproduce require magento magento coding standard run vendor bin phpcs i expected result magento coding standard is present in the list of installed standards actual result magento coding standard is missing in the list of installed standards how to fix remove composer bin dir fix scripts to run vendor bin phpcs and change readme md
1
172,445
27,281,390,895
IssuesEvent
2023-02-23 10:19:44
CMPUT301W23T33/QR-Quest
https://api.github.com/repos/CMPUT301W23T33/QR-Quest
closed
Draw screens (20, 21)
UI design
## Personal task - Draw 1 layout for: - Screen 20, 21 ## General notice - **Deadline**: 4 pm, Sunday 2/26/2023 (MST) - Later: Screen 10 (Hardest) - The screen amounts are based on the difficulty level that screen (fewer screens do not mean the workload is less) - Each newline represents a layout that needs to be drawn - Only the layout is needed, you do not need to implement the functions for any buttons (yet) - Top left screen is screen 1, you start counting from left to right in this picture: ![Image](https://user-images.githubusercontent.com/90273567/220757104-1b6f3479-4c18-41c7-b7a9-a18b8d6fca49.png)
1.0
Draw screens (20, 21) - ## Personal task - Draw 1 layout for: - Screen 20, 21 ## General notice - **Deadline**: 4 pm, Sunday 2/26/2023 (MST) - Later: Screen 10 (Hardest) - The screen amounts are based on the difficulty level that screen (fewer screens do not mean the workload is less) - Each newline represents a layout that needs to be drawn - Only the layout is needed, you do not need to implement the functions for any buttons (yet) - Top left screen is screen 1, you start counting from left to right in this picture: ![Image](https://user-images.githubusercontent.com/90273567/220757104-1b6f3479-4c18-41c7-b7a9-a18b8d6fca49.png)
non_test
draw screens personal task draw layout for screen general notice deadline pm sunday mst later screen hardest the screen amounts are based on the difficulty level that screen fewer screens do not mean the workload is less each newline represents a layout that needs to be drawn only the layout is needed you do not need to implement the functions for any buttons yet top left screen is screen you start counting from left to right in this picture
0
63,228
6,829,910,903
IssuesEvent
2017-11-09 03:07:41
subutai-io/tray
https://api.github.com/repos/subutai-io/tray
closed
[Linux] Tray crashed after some time
bug critical testing
Without any activity, I noticed that after a while (20-30 minutes after launch) the tray was no longer in the task tray. It had crashed. This is an Ubuntu Linux 17.04 system and the Debian package that was supplied with hotfixes were used. The P2P daemon was also installed.
1.0
[Linux] Tray crashed after some time - Without any activity, I noticed that after a while (20-30 minutes after launch) the tray was no longer in the task tray. It had crashed. This is an Ubuntu Linux 17.04 system and the Debian package that was supplied with hotfixes were used. The P2P daemon was also installed.
test
tray crashed after some time without any activity i noticed that after a while minutes after launch the tray was no longer in the task tray it had crashed this is an ubuntu linux system and the debian package that was supplied with hotfixes were used the daemon was also installed
1
553,238
16,360,706,609
IssuesEvent
2021-05-14 09:01:31
primefaces/primevue
https://api.github.com/repos/primefaces/primevue
closed
Add loading prop to Dropdown, Multiselect ...
enhancement priority - medium
Hi, I started using your components recently and i'm so happy that i did. This is such an amazing product, so thanks for that. Before using vue prime i used to work with semantic ui dropdown and it has a cool feature. If you add a `.loading` class to the dropdown it will show a little loading spinner ![image](https://user-images.githubusercontent.com/1449151/112835953-41e7b680-909a-11eb-8f05-bde6341ab55f.png) This allows for a nice feedback to the user that the app is loading something and it does not clutter the html with loader components. Are you open to the idea of adding a `loading` prop to the Dropdown, Multiselect and other components ? Maybe this could be my first contribution to this amazing project. Thanks
1.0
Add loading prop to Dropdown, Multiselect ... - Hi, I started using your components recently and i'm so happy that i did. This is such an amazing product, so thanks for that. Before using vue prime i used to work with semantic ui dropdown and it has a cool feature. If you add a `.loading` class to the dropdown it will show a little loading spinner ![image](https://user-images.githubusercontent.com/1449151/112835953-41e7b680-909a-11eb-8f05-bde6341ab55f.png) This allows for a nice feedback to the user that the app is loading something and it does not clutter the html with loader components. Are you open to the idea of adding a `loading` prop to the Dropdown, Multiselect and other components ? Maybe this could be my first contribution to this amazing project. Thanks
non_test
add loading prop to dropdown multiselect hi i started using your components recently and i m so happy that i did this is such an amazing product so thanks for that before using vue prime i used to work with semantic ui dropdown and it has a cool feature if you add a loading class to the dropdown it will show a little loading spinner this allows for a nice feedback to the user that the app is loading something and it does not clutter the html with loader components are you open to the idea of adding a loading prop to the dropdown multiselect and other components maybe this could be my first contribution to this amazing project thanks
0
252,776
21,630,957,104
IssuesEvent
2022-05-05 09:39:07
hotosm/galaxy-api
https://api.github.com/repos/hotosm/galaxy-api
closed
Optimize memory usage of Raw Data Postprocessing
type:enhancement status:testing
Since we will be getting large amount of rows back from Database . Currently We are grabbing those results , binding them to geojson and again compressing that file to zip file ( all in memory ) and As a result single request of approx 400-500mb of file size is eating approx 2-4gb of ram until we deliver compressed zip file to user. Memory issue could be a serious problem when we will be getting Multiple requests at a same time . Need to do more of research to overcome this approach cc: @ramyaragupathy _Originally posted by @itskshitiz321 in https://github.com/hotosm/galaxy-api/issues/130#issuecomment-1047888543_
1.0
Optimize memory usage of Raw Data Postprocessing - Since we will be getting large amount of rows back from Database . Currently We are grabbing those results , binding them to geojson and again compressing that file to zip file ( all in memory ) and As a result single request of approx 400-500mb of file size is eating approx 2-4gb of ram until we deliver compressed zip file to user. Memory issue could be a serious problem when we will be getting Multiple requests at a same time . Need to do more of research to overcome this approach cc: @ramyaragupathy _Originally posted by @itskshitiz321 in https://github.com/hotosm/galaxy-api/issues/130#issuecomment-1047888543_
test
optimize memory usage of raw data postprocessing since we will be getting large amount of rows back from database currently we are grabbing those results binding them to geojson and again compressing that file to zip file all in memory and as a result single request of approx of file size is eating approx of ram until we deliver compressed zip file to user memory issue could be a serious problem when we will be getting multiple requests at a same time need to do more of research to overcome this approach cc ramyaragupathy originally posted by in
1
331,283
28,808,214,183
IssuesEvent
2023-05-03 00:51:10
Azure/azure-sdk-for-net
https://api.github.com/repos/Azure/azure-sdk-for-net
closed
[FormRecognizer] Re-enable StartBuildModelFailsWithInvalidPrefix
Client Cognitive - Form Recognizer test-reliability
Re-enable `StartBuildModelFailsWithInvalidPrefix` test when issue is fixed.
1.0
[FormRecognizer] Re-enable StartBuildModelFailsWithInvalidPrefix - Re-enable `StartBuildModelFailsWithInvalidPrefix` test when issue is fixed.
test
re enable startbuildmodelfailswithinvalidprefix re enable startbuildmodelfailswithinvalidprefix test when issue is fixed
1
34,078
4,890,323,546
IssuesEvent
2016-11-18 13:30:41
Grumnir/IDEmm
https://api.github.com/repos/Grumnir/IDEmm
closed
Create testplan in Wiki
Projectmanagement Test
It would be a nice idea to have something lika a testplan in the wiki. Create one.
1.0
Create testplan in Wiki - It would be a nice idea to have something lika a testplan in the wiki. Create one.
test
create testplan in wiki it would be a nice idea to have something lika a testplan in the wiki create one
1
351,627
10,521,584,652
IssuesEvent
2019-09-30 06:31:50
alebor-5/D7024E
https://api.github.com/repos/alebor-5/D7024E
opened
Object expiration delay
Medium Priority enhancement
To prevent files from expiring, the node that originally uploaded each object sends a refresh command to the nodes having copies of it to prevent them from being deleted. In particular, the command resets the TTL for the refreshed data object without actually requesting it. As long as the uploading node can contact the storing nodes, the object in question should never expire. Completing this objective requires that U1 is also completed.
1.0
Object expiration delay - To prevent files from expiring, the node that originally uploaded each object sends a refresh command to the nodes having copies of it to prevent them from being deleted. In particular, the command resets the TTL for the refreshed data object without actually requesting it. As long as the uploading node can contact the storing nodes, the object in question should never expire. Completing this objective requires that U1 is also completed.
non_test
object expiration delay to prevent files from expiring the node that originally uploaded each object sends a refresh command to the nodes having copies of it to prevent them from being deleted in particular the command resets the ttl for the refreshed data object without actually requesting it as long as the uploading node can contact the storing nodes the object in question should never expire completing this objective requires that is also completed
0
226,065
7,498,243,849
IssuesEvent
2018-04-09 03:13:48
CSC480-18S/Red-Team
https://api.github.com/repos/CSC480-18S/Red-Team
closed
Start word position validation
(ノಠ益ಠ)ノ彡┻━┻ Priority: High Type: Bug
When playing a two letter word as first word on empty board, more often then not frontend receives a INVALID_PLACEMENT error
1.0
Start word position validation - When playing a two letter word as first word on empty board, more often then not frontend receives a INVALID_PLACEMENT error
non_test
start word position validation when playing a two letter word as first word on empty board more often then not frontend receives a invalid placement error
0
418,390
28,114,738,848
IssuesEvent
2023-03-31 09:52:35
tzixi/ped
https://api.github.com/repos/tzixi/ped
opened
UG Table of Contents flaw
type.DocumentationBug severity.VeryLow
Clicking on the links in the table of contents in UG does not bring the user to the specified part of the UG ![image.png](https://raw.githubusercontent.com/tzixi/ped/main/files/e561c565-7d66-4e90-bd56-d7af7c4406d0.png) <!--session: 1680252431172-d43e5087-6923-46f0-ba82-c411655cba8e--> <!--Version: Web v3.4.7-->
1.0
UG Table of Contents flaw - Clicking on the links in the table of contents in UG does not bring the user to the specified part of the UG ![image.png](https://raw.githubusercontent.com/tzixi/ped/main/files/e561c565-7d66-4e90-bd56-d7af7c4406d0.png) <!--session: 1680252431172-d43e5087-6923-46f0-ba82-c411655cba8e--> <!--Version: Web v3.4.7-->
non_test
ug table of contents flaw clicking on the links in the table of contents in ug does not bring the user to the specified part of the ug
0
301,608
26,078,356,553
IssuesEvent
2022-12-24 23:15:37
Zexlo/FNF-Sonic-EXE-psych-engine-port
https://api.github.com/repos/Zexlo/FNF-Sonic-EXE-psych-engine-port
closed
Triple Trouble charting, 5 key conversion to 4 key
bug testing needed still in progress
Triple Trouble is a 5 key chart. As far as it's known, there is no way to play 5 key charts in Psych Engine. This chart needs to be fixed to be playable with 4 keys.
1.0
Triple Trouble charting, 5 key conversion to 4 key - Triple Trouble is a 5 key chart. As far as it's known, there is no way to play 5 key charts in Psych Engine. This chart needs to be fixed to be playable with 4 keys.
test
triple trouble charting key conversion to key triple trouble is a key chart as far as it s known there is no way to play key charts in psych engine this chart needs to be fixed to be playable with keys
1
46,463
5,810,236,276
IssuesEvent
2017-05-04 15:02:31
dealii/dealii
https://api.github.com/repos/dealii/dealii
closed
warnings in vector_tools_project_qpmf.cc
Tests
I am getting the following warnings with gcc 4.6.3 and I couldn't immediately find what is wrong. Anybody got an idea? ``` In file included from /ssd/deal-git/include/deal.II/matrix_free/operators.h:30:0, from /ssd/deal-git/include/deal.II/numerics/vector_tools.templates.h:43, from /ssd/deal-git/source/numerics/vector_tools_project_qpmf.cc:17: /ssd/deal-git/include/deal.II/base/vectorization.h: In static member function ‘static void dealii::internal::FEEvaluationImpl<type, dim, fe_degree, n_q_points_1d, n_components, Number>::evaluate(const dealii::internal::MatrixFreeFunctions::ShapeInfo<Number>&, dealii::VectorizedArray<Number>**, dealii::VectorizedArray<Number>**, dealii::VectorizedArray<Number>* (*)[dim], dealii::VectorizedArray<Number>* (*)[((dim * (dim + 1)) / 2)], bool, bool, bool) [with dealii::internal::MatrixFreeFunctions::ElementType type = (dealii::internal::MatrixFreeFunctions::ElementType)3u, int dim = 2, int fe_degree = 1, int n_q_points_1d = 2, int n_components = 1, Number = double]’: /ssd/deal-git/include/deal.II/base/vectorization.h:1586:21: warning: ‘*((void*)(& temp1)+48).dealii::VectorizedArray<double>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+48).dealii::VectorizedArray<double>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1586:21: warning: ‘*((void*)(& temp1)+32).dealii::VectorizedArray<double>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+32).dealii::VectorizedArray<double>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1586:21: warning: ‘*((void*)(& temp1)+16).dealii::VectorizedArray<double>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+16).dealii::VectorizedArray<double>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1586:21: warning: ‘temp1.dealii::VectorizedArray<double>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘temp1.dealii::VectorizedArray<double>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h: In static member function ‘static void dealii::internal::FEEvaluationImpl<type, dim, fe_degree, n_q_points_1d, n_components, Number>::evaluate(const dealii::internal::MatrixFreeFunctions::ShapeInfo<Number>&, dealii::VectorizedArray<Number>**, dealii::VectorizedArray<Number>**, dealii::VectorizedArray<Number>* (*)[dim], dealii::VectorizedArray<Number>* (*)[((dim * (dim + 1)) / 2)], bool, bool, bool) [with dealii::internal::MatrixFreeFunctions::ElementType type = (dealii::internal::MatrixFreeFunctions::ElementType)3u, int dim = 2, int fe_degree = 1, int n_q_points_1d = 2, int n_components = 1, Number = float]’: /ssd/deal-git/include/deal.II/base/vectorization.h:1856:21: warning: ‘*((void*)(& temp1)+48).dealii::VectorizedArray<float>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+48).dealii::VectorizedArray<float>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1856:21: warning: ‘*((void*)(& temp1)+32).dealii::VectorizedArray<float>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+32).dealii::VectorizedArray<float>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1856:21: warning: ‘*((void*)(& temp1)+16).dealii::VectorizedArray<float>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+16).dealii::VectorizedArray<float>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1856:21: warning: ‘temp1.dealii::VectorizedArray<float>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘temp1.dealii::VectorizedArray<float>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h: In static member function ‘static void dealii::internal::FEEvaluationImpl<type, dim, fe_degree, n_q_points_1d, n_components, Number>::evaluate(const dealii::internal::MatrixFreeFunctions::ShapeInfo<Number>&, dealii::VectorizedArray<Number>**, dealii::VectorizedArray<Number>**, dealii::VectorizedArray<Number>* (*)[dim], dealii::VectorizedArray<Number>* (*)[((dim * (dim + 1)) / 2)], bool, bool, bool) [with dealii::internal::MatrixFreeFunctions::ElementType type = (dealii::internal::MatrixFreeFunctions::ElementType)1u, int dim = 2, int fe_degree = 1, int n_q_points_1d = 2, int n_components = 1, Number = float]’: /ssd/deal-git/include/deal.II/base/vectorization.h:1856:21: warning: ‘*((void*)(& temp1)+48).dealii::VectorizedArray<float>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+48).dealii::VectorizedArray<float>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1856:21: warning: ‘*((void*)(& temp1)+32).dealii::VectorizedArray<float>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+32).dealii::VectorizedArray<float>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1856:21: warning: ‘*((void*)(& temp1)+16).dealii::VectorizedArray<float>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+16).dealii::VectorizedArray<float>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1856:21: warning: ‘temp1.dealii::VectorizedArray<float>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘temp1.dealii::VectorizedArray<float>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h: In static member function ‘static void dealii::internal::FEEvaluationImpl<type, dim, fe_degree, n_q_points_1d, n_components, Number>::evaluate(const dealii::internal::MatrixFreeFunctions::ShapeInfo<Number>&, dealii::VectorizedArray<Number>**, dealii::VectorizedArray<Number>**, dealii::VectorizedArray<Number>* (*)[dim], dealii::VectorizedArray<Number>* (*)[((dim * (dim + 1)) / 2)], bool, bool, bool) [with dealii::internal::MatrixFreeFunctions::ElementType type = (dealii::internal::MatrixFreeFunctions::ElementType)1u, int dim = 2, int fe_degree = 1, int n_q_points_1d = 2, int n_components = 1, Number = double]’: /ssd/deal-git/include/deal.II/base/vectorization.h:1586:21: warning: ‘*((void*)(& temp1)+48).dealii::VectorizedArray<double>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+48).dealii::VectorizedArray<double>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1586:21: warning: ‘*((void*)(& temp1)+32).dealii::VectorizedArray<double>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+32).dealii::VectorizedArray<double>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1586:21: warning: ‘*((void*)(& temp1)+16).dealii::VectorizedArray<double>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+16).dealii::VectorizedArray<double>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1586:21: warning: ‘temp1.dealii::VectorizedArray<double>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘temp1.dealii::VectorizedArray<double>::data’ was declared here ```
1.0
warnings in vector_tools_project_qpmf.cc - I am getting the following warnings with gcc 4.6.3 and I couldn't immediately find what is wrong. Anybody got an idea? ``` In file included from /ssd/deal-git/include/deal.II/matrix_free/operators.h:30:0, from /ssd/deal-git/include/deal.II/numerics/vector_tools.templates.h:43, from /ssd/deal-git/source/numerics/vector_tools_project_qpmf.cc:17: /ssd/deal-git/include/deal.II/base/vectorization.h: In static member function ‘static void dealii::internal::FEEvaluationImpl<type, dim, fe_degree, n_q_points_1d, n_components, Number>::evaluate(const dealii::internal::MatrixFreeFunctions::ShapeInfo<Number>&, dealii::VectorizedArray<Number>**, dealii::VectorizedArray<Number>**, dealii::VectorizedArray<Number>* (*)[dim], dealii::VectorizedArray<Number>* (*)[((dim * (dim + 1)) / 2)], bool, bool, bool) [with dealii::internal::MatrixFreeFunctions::ElementType type = (dealii::internal::MatrixFreeFunctions::ElementType)3u, int dim = 2, int fe_degree = 1, int n_q_points_1d = 2, int n_components = 1, Number = double]’: /ssd/deal-git/include/deal.II/base/vectorization.h:1586:21: warning: ‘*((void*)(& temp1)+48).dealii::VectorizedArray<double>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+48).dealii::VectorizedArray<double>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1586:21: warning: ‘*((void*)(& temp1)+32).dealii::VectorizedArray<double>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+32).dealii::VectorizedArray<double>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1586:21: warning: ‘*((void*)(& temp1)+16).dealii::VectorizedArray<double>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+16).dealii::VectorizedArray<double>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1586:21: warning: ‘temp1.dealii::VectorizedArray<double>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘temp1.dealii::VectorizedArray<double>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h: In static member function ‘static void dealii::internal::FEEvaluationImpl<type, dim, fe_degree, n_q_points_1d, n_components, Number>::evaluate(const dealii::internal::MatrixFreeFunctions::ShapeInfo<Number>&, dealii::VectorizedArray<Number>**, dealii::VectorizedArray<Number>**, dealii::VectorizedArray<Number>* (*)[dim], dealii::VectorizedArray<Number>* (*)[((dim * (dim + 1)) / 2)], bool, bool, bool) [with dealii::internal::MatrixFreeFunctions::ElementType type = (dealii::internal::MatrixFreeFunctions::ElementType)3u, int dim = 2, int fe_degree = 1, int n_q_points_1d = 2, int n_components = 1, Number = float]’: /ssd/deal-git/include/deal.II/base/vectorization.h:1856:21: warning: ‘*((void*)(& temp1)+48).dealii::VectorizedArray<float>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+48).dealii::VectorizedArray<float>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1856:21: warning: ‘*((void*)(& temp1)+32).dealii::VectorizedArray<float>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+32).dealii::VectorizedArray<float>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1856:21: warning: ‘*((void*)(& temp1)+16).dealii::VectorizedArray<float>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+16).dealii::VectorizedArray<float>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1856:21: warning: ‘temp1.dealii::VectorizedArray<float>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘temp1.dealii::VectorizedArray<float>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h: In static member function ‘static void dealii::internal::FEEvaluationImpl<type, dim, fe_degree, n_q_points_1d, n_components, Number>::evaluate(const dealii::internal::MatrixFreeFunctions::ShapeInfo<Number>&, dealii::VectorizedArray<Number>**, dealii::VectorizedArray<Number>**, dealii::VectorizedArray<Number>* (*)[dim], dealii::VectorizedArray<Number>* (*)[((dim * (dim + 1)) / 2)], bool, bool, bool) [with dealii::internal::MatrixFreeFunctions::ElementType type = (dealii::internal::MatrixFreeFunctions::ElementType)1u, int dim = 2, int fe_degree = 1, int n_q_points_1d = 2, int n_components = 1, Number = float]’: /ssd/deal-git/include/deal.II/base/vectorization.h:1856:21: warning: ‘*((void*)(& temp1)+48).dealii::VectorizedArray<float>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+48).dealii::VectorizedArray<float>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1856:21: warning: ‘*((void*)(& temp1)+32).dealii::VectorizedArray<float>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+32).dealii::VectorizedArray<float>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1856:21: warning: ‘*((void*)(& temp1)+16).dealii::VectorizedArray<float>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+16).dealii::VectorizedArray<float>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1856:21: warning: ‘temp1.dealii::VectorizedArray<float>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘temp1.dealii::VectorizedArray<float>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h: In static member function ‘static void dealii::internal::FEEvaluationImpl<type, dim, fe_degree, n_q_points_1d, n_components, Number>::evaluate(const dealii::internal::MatrixFreeFunctions::ShapeInfo<Number>&, dealii::VectorizedArray<Number>**, dealii::VectorizedArray<Number>**, dealii::VectorizedArray<Number>* (*)[dim], dealii::VectorizedArray<Number>* (*)[((dim * (dim + 1)) / 2)], bool, bool, bool) [with dealii::internal::MatrixFreeFunctions::ElementType type = (dealii::internal::MatrixFreeFunctions::ElementType)1u, int dim = 2, int fe_degree = 1, int n_q_points_1d = 2, int n_components = 1, Number = double]’: /ssd/deal-git/include/deal.II/base/vectorization.h:1586:21: warning: ‘*((void*)(& temp1)+48).dealii::VectorizedArray<double>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+48).dealii::VectorizedArray<double>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1586:21: warning: ‘*((void*)(& temp1)+32).dealii::VectorizedArray<double>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+32).dealii::VectorizedArray<double>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1586:21: warning: ‘*((void*)(& temp1)+16).dealii::VectorizedArray<double>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘*((void*)(& temp1)+16).dealii::VectorizedArray<double>::data’ was declared here /ssd/deal-git/include/deal.II/base/vectorization.h:1586:21: warning: ‘temp1.dealii::VectorizedArray<double>::data’ may be used uninitialized in this function [-Wuninitialized] /ssd/deal-git/include/deal.II/matrix_free/fe_evaluation.h:6254:37: note: ‘temp1.dealii::VectorizedArray<double>::data’ was declared here ```
test
warnings in vector tools project qpmf cc i am getting the following warnings with gcc and i couldn t immediately find what is wrong anybody got an idea in file included from ssd deal git include deal ii matrix free operators h from ssd deal git include deal ii numerics vector tools templates h from ssd deal git source numerics vector tools project qpmf cc ssd deal git include deal ii base vectorization h in static member function ‘static void dealii internal feevaluationimpl evaluate const dealii internal matrixfreefunctions shapeinfo dealii vectorizedarray dealii vectorizedarray dealii vectorizedarray dealii vectorizedarray bool bool bool ’ ssd deal git include deal ii base vectorization h warning ‘ void dealii vectorizedarray data’ may be used uninitialized in this function ssd deal git include deal ii matrix free fe evaluation h note ‘ void dealii vectorizedarray data’ was declared here ssd deal git include deal ii base vectorization h warning ‘ void dealii vectorizedarray data’ may be used uninitialized in this function ssd deal git include deal ii matrix free fe evaluation h note ‘ void dealii vectorizedarray data’ was declared here ssd deal git include deal ii base vectorization h warning ‘ void dealii vectorizedarray data’ may be used uninitialized in this function ssd deal git include deal ii matrix free fe evaluation h note ‘ void dealii vectorizedarray data’ was declared here ssd deal git include deal ii base vectorization h warning ‘ dealii vectorizedarray data’ may be used uninitialized in this function ssd deal git include deal ii matrix free fe evaluation h note ‘ dealii vectorizedarray data’ was declared here ssd deal git include deal ii base vectorization h in static member function ‘static void dealii internal feevaluationimpl evaluate const dealii internal matrixfreefunctions shapeinfo dealii vectorizedarray dealii vectorizedarray dealii vectorizedarray dealii vectorizedarray bool bool bool ’ ssd deal git include deal ii base vectorization h warning ‘ void dealii vectorizedarray data’ may be used uninitialized in this function ssd deal git include deal ii matrix free fe evaluation h note ‘ void dealii vectorizedarray data’ was declared here ssd deal git include deal ii base vectorization h warning ‘ void dealii vectorizedarray data’ may be used uninitialized in this function ssd deal git include deal ii matrix free fe evaluation h note ‘ void dealii vectorizedarray data’ was declared here ssd deal git include deal ii base vectorization h warning ‘ void dealii vectorizedarray data’ may be used uninitialized in this function ssd deal git include deal ii matrix free fe evaluation h note ‘ void dealii vectorizedarray data’ was declared here ssd deal git include deal ii base vectorization h warning ‘ dealii vectorizedarray data’ may be used uninitialized in this function ssd deal git include deal ii matrix free fe evaluation h note ‘ dealii vectorizedarray data’ was declared here ssd deal git include deal ii base vectorization h in static member function ‘static void dealii internal feevaluationimpl evaluate const dealii internal matrixfreefunctions shapeinfo dealii vectorizedarray dealii vectorizedarray dealii vectorizedarray dealii vectorizedarray bool bool bool ’ ssd deal git include deal ii base vectorization h warning ‘ void dealii vectorizedarray data’ may be used uninitialized in this function ssd deal git include deal ii matrix free fe evaluation h note ‘ void dealii vectorizedarray data’ was declared here ssd deal git include deal ii base vectorization h warning ‘ void dealii vectorizedarray data’ may be used uninitialized in this function ssd deal git include deal ii matrix free fe evaluation h note ‘ void dealii vectorizedarray data’ was declared here ssd deal git include deal ii base vectorization h warning ‘ void dealii vectorizedarray data’ may be used uninitialized in this function ssd deal git include deal ii matrix free fe evaluation h note ‘ void dealii vectorizedarray data’ was declared here ssd deal git include deal ii base vectorization h warning ‘ dealii vectorizedarray data’ may be used uninitialized in this function ssd deal git include deal ii matrix free fe evaluation h note ‘ dealii vectorizedarray data’ was declared here ssd deal git include deal ii base vectorization h in static member function ‘static void dealii internal feevaluationimpl evaluate const dealii internal matrixfreefunctions shapeinfo dealii vectorizedarray dealii vectorizedarray dealii vectorizedarray dealii vectorizedarray bool bool bool ’ ssd deal git include deal ii base vectorization h warning ‘ void dealii vectorizedarray data’ may be used uninitialized in this function ssd deal git include deal ii matrix free fe evaluation h note ‘ void dealii vectorizedarray data’ was declared here ssd deal git include deal ii base vectorization h warning ‘ void dealii vectorizedarray data’ may be used uninitialized in this function ssd deal git include deal ii matrix free fe evaluation h note ‘ void dealii vectorizedarray data’ was declared here ssd deal git include deal ii base vectorization h warning ‘ void dealii vectorizedarray data’ may be used uninitialized in this function ssd deal git include deal ii matrix free fe evaluation h note ‘ void dealii vectorizedarray data’ was declared here ssd deal git include deal ii base vectorization h warning ‘ dealii vectorizedarray data’ may be used uninitialized in this function ssd deal git include deal ii matrix free fe evaluation h note ‘ dealii vectorizedarray data’ was declared here
1
95,405
27,497,310,816
IssuesEvent
2023-03-05 09:39:51
chaotic-aur/packages
https://api.github.com/repos/chaotic-aur/packages
closed
[Outdated] nvim-nightly
waiting:upstream-fix request:rebuild-pkg priority:low bug:PKGBUILD
### If available, link to the latest build [nvim-nightly.log](https://builds.garudalinux.org/repos/chaotic-aur/logs/nvim-nightly.log) ### Package name `nvim-nightly` ### Latest build `0.9.0.dev.1c47839-1` ### Latest version available `0.9.0.dev.1dd2424-1` ### Have you tested if the package builds in a clean chroot? - [ ] Yes ### More information The package is outdated, it uses [the 4 months old version](https://github.com/neovim/neovim/commit/1c478391ca7754bf5ecb4b76c29acfa9b4978393) instead of [the latest one](https://github.com/neovim/neovim/releases/tag/nightly). The AUR package is outdated too, flagged it out-of-date.
2.0
[Outdated] nvim-nightly - ### If available, link to the latest build [nvim-nightly.log](https://builds.garudalinux.org/repos/chaotic-aur/logs/nvim-nightly.log) ### Package name `nvim-nightly` ### Latest build `0.9.0.dev.1c47839-1` ### Latest version available `0.9.0.dev.1dd2424-1` ### Have you tested if the package builds in a clean chroot? - [ ] Yes ### More information The package is outdated, it uses [the 4 months old version](https://github.com/neovim/neovim/commit/1c478391ca7754bf5ecb4b76c29acfa9b4978393) instead of [the latest one](https://github.com/neovim/neovim/releases/tag/nightly). The AUR package is outdated too, flagged it out-of-date.
non_test
nvim nightly if available link to the latest build package name nvim nightly latest build dev latest version available dev have you tested if the package builds in a clean chroot yes more information the package is outdated it uses instead of the aur package is outdated too flagged it out of date
0
758,661
26,563,952,891
IssuesEvent
2023-01-20 18:16:33
bcgov/entity
https://api.github.com/repos/bcgov/entity
closed
Retain buttons when using smaller viewport
bug Priority2 Assets
When using a smaller viewport the action buttons do not display. Currently the only workaround is to use the zoom to shrink the size of your browser. ![image.png](https://images.zenhubusercontent.com/5ba90c7629a785544981da1c/6516033e-86bd-4f83-8b98-ca0c1384f189)
1.0
Retain buttons when using smaller viewport - When using a smaller viewport the action buttons do not display. Currently the only workaround is to use the zoom to shrink the size of your browser. ![image.png](https://images.zenhubusercontent.com/5ba90c7629a785544981da1c/6516033e-86bd-4f83-8b98-ca0c1384f189)
non_test
retain buttons when using smaller viewport when using a smaller viewport the action buttons do not display currently the only workaround is to use the zoom to shrink the size of your browser
0
66,497
27,498,052,452
IssuesEvent
2023-03-05 11:29:13
APIs-guru/openapi-directory
https://api.github.com/repos/APIs-guru/openapi-directory
closed
Add "Aviation Edge Routes API" API
needs-servicename stale
**Format**: OpenAPI 3.0 **Official**: YES **Url**: https://aviation-edge.com/airline-routes-database-and-api/ **Name**: Aviation Edge Routes API **Category**: transport **Logo**:
1.0
Add "Aviation Edge Routes API" API - **Format**: OpenAPI 3.0 **Official**: YES **Url**: https://aviation-edge.com/airline-routes-database-and-api/ **Name**: Aviation Edge Routes API **Category**: transport **Logo**:
non_test
add aviation edge routes api api format openapi official yes url name aviation edge routes api category transport logo
0
810,481
30,244,698,092
IssuesEvent
2023-07-06 15:38:02
quickwit-oss/quickwit
https://api.github.com/repos/quickwit-oss/quickwit
closed
Merge pipeline failed when index has too many splits
bug high-priority
**Describe the bug** quickwit-indexer got this error ``` 2023-06-26T07:47:49.057Z WARN spawn_merge_pipeline{index=xxx-index gen=0}: quickwit_metastore::metastore::retrying_metastore::retry: Request failed attempt_count=5 2023-06-26T07:47:49.057Z ERROR quickwit_indexing::actors::merge_pipeline: Error while spawning indexing pipeline, retrying after some time. error=Internal error: `Error, message length too large: found 50114113 bytes, the limit is: 4194304 bytes` Cause: ``. retry_count=0 retry_delay=2s ``` **Expected behavior** no error **Configuration:** quickwit version: v0.6.1 deploy with helm, default settings (with cpu/memory request/limit increased) ---- I found the same error(message length too large) when request `/api/v1/indexes/<index-id>/splits` on searcher, but successed on metastore. I found this error template in tonic's source, so I guess that error occurs on large grpc response of splits.
1.0
Merge pipeline failed when index has too many splits - **Describe the bug** quickwit-indexer got this error ``` 2023-06-26T07:47:49.057Z WARN spawn_merge_pipeline{index=xxx-index gen=0}: quickwit_metastore::metastore::retrying_metastore::retry: Request failed attempt_count=5 2023-06-26T07:47:49.057Z ERROR quickwit_indexing::actors::merge_pipeline: Error while spawning indexing pipeline, retrying after some time. error=Internal error: `Error, message length too large: found 50114113 bytes, the limit is: 4194304 bytes` Cause: ``. retry_count=0 retry_delay=2s ``` **Expected behavior** no error **Configuration:** quickwit version: v0.6.1 deploy with helm, default settings (with cpu/memory request/limit increased) ---- I found the same error(message length too large) when request `/api/v1/indexes/<index-id>/splits` on searcher, but successed on metastore. I found this error template in tonic's source, so I guess that error occurs on large grpc response of splits.
non_test
merge pipeline failed when index has too many splits describe the bug quickwit indexer got this error warn spawn merge pipeline index xxx index gen quickwit metastore metastore retrying metastore retry request failed attempt count error quickwit indexing actors merge pipeline error while spawning indexing pipeline retrying after some time error internal error error message length too large found bytes the limit is bytes cause retry count retry delay expected behavior no error configuration quickwit version deploy with helm default settings with cpu memory request limit increased i found the same error message length too large when request api indexes splits on searcher but successed on metastore i found this error template in tonic s source so i guess that error occurs on large grpc response of splits
0
85,264
15,736,675,909
IssuesEvent
2021-03-30 01:10:59
saurockSaurav/weather-information-api
https://api.github.com/repos/saurockSaurav/weather-information-api
opened
CVE-2019-17563 (High) detected in tomcat-embed-core-8.5.39.jar
security vulnerability
## CVE-2019-17563 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.5.39.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Path to dependency file: /weather-information-api/weather-rest-api-service/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.39/tomcat-embed-core-8.5.39.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.5.20.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-1.5.20.RELEASE.jar - :x: **tomcat-embed-core-8.5.39.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When using FORM authentication with Apache Tomcat 9.0.0.M1 to 9.0.29, 8.5.0 to 8.5.49 and 7.0.0 to 7.0.98 there was a narrow window where an attacker could perform a session fixation attack. The window was considered too narrow for an exploit to be practical but, erring on the side of caution, this issue has been treated as a security vulnerability. <p>Publish Date: 2019-12-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17563>CVE-2019-17563</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17563">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17563</a></p> <p>Release Date: 2019-12-23</p> <p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:7.0.99,8.5.50,9.0.30;org.apache.tomcat:tomcat-catalina:7.0.99,8.5.50,9.0.30</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-17563 (High) detected in tomcat-embed-core-8.5.39.jar - ## CVE-2019-17563 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.5.39.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Path to dependency file: /weather-information-api/weather-rest-api-service/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.39/tomcat-embed-core-8.5.39.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.5.20.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-1.5.20.RELEASE.jar - :x: **tomcat-embed-core-8.5.39.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When using FORM authentication with Apache Tomcat 9.0.0.M1 to 9.0.29, 8.5.0 to 8.5.49 and 7.0.0 to 7.0.98 there was a narrow window where an attacker could perform a session fixation attack. The window was considered too narrow for an exploit to be practical but, erring on the side of caution, this issue has been treated as a security vulnerability. <p>Publish Date: 2019-12-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17563>CVE-2019-17563</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17563">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17563</a></p> <p>Release Date: 2019-12-23</p> <p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:7.0.99,8.5.50,9.0.30;org.apache.tomcat:tomcat-catalina:7.0.99,8.5.50,9.0.30</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in tomcat embed core jar cve high severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation path to dependency file weather information api weather rest api service pom xml path to vulnerable library root repository org apache tomcat embed tomcat embed core tomcat embed core jar dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar x tomcat embed core jar vulnerable library vulnerability details when using form authentication with apache tomcat to to and to there was a narrow window where an attacker could perform a session fixation attack the window was considered too narrow for an exploit to be practical but erring on the side of caution this issue has been treated as a security vulnerability publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat embed tomcat embed core org apache tomcat tomcat catalina step up your open source security game with whitesource
0
531,265
15,443,772,087
IssuesEvent
2021-03-08 09:33:20
Laravel-Backpack/CRUD
https://api.github.com/repos/Laravel-Backpack/CRUD
closed
[Bug] CrudControllers with setFromDb() don't work with DBAL v3
Bug Priority: MUST triage urgent
# Bug report ### What I did A simple CRUD that uses `setFromDb()`, for example Countries with: ```php <?php use Illuminate\Database\Migrations\Migration; use Illuminate\Database\Schema\Blueprint; class CreateCountriesTable extends Migration { /** * Run the migrations. * * @return void */ public function up() { Schema::create('countries', function (Blueprint $table) { $table->increments('id'); $table->string('code', 2)->index(); $table->string('name', 75); }); } /** * Reverse the migrations. * * @return void */ public function down() { Schema::drop('countries'); } } ``` ### What I expected to happen The default CountryCrudController to work if you do nothing. So this to work: ```php <?php namespace App\Http\Controllers\Admin; use App\Http\Requests\CountryRequest as StoreRequest; // VALIDATION: change the requests to match your own file names if you need form validation use App\Http\Requests\CountryRequest as UpdateRequest; use Backpack\CRUD\app\Http\Controllers\CrudController; class CountryCrudController extends CrudController { use \Backpack\CRUD\app\Http\Controllers\Operations\ListOperation; use \Backpack\CRUD\app\Http\Controllers\Operations\CreateOperation; use \Backpack\CRUD\app\Http\Controllers\Operations\UpdateOperation; use \Backpack\CRUD\app\Http\Controllers\Operations\DeleteOperation; public function setup() { /* |-------------------------------------------------------------------------- | BASIC CRUD INFORMATION |-------------------------------------------------------------------------- */ $this->crud->setModel(\App\Models\Country::class); $this->crud->setRoute('admin/country'); $this->crud->setEntityNameStrings('country', 'countries'); /* |-------------------------------------------------------------------------- | BASIC CRUD INFORMATION |-------------------------------------------------------------------------- */ $this->crud->setFromDb(); } protected function setupCreateOperation() { $this->crud->setValidation(StoreRequest::class); } protected function setupUpdateOperation() { $this->crud->setValidation(UpdateRequest::class); } } ``` ### What happened `“Class 'Doctrine\DBAL\Driver\\PDOMySql\Driver' not found”` error was thrown when going on the list page, so the ListOperation is unusable. [Click here for stack trace](http://flareapp.io/share/xPQ6LxD7). Screenshot below: ![Screenshot 2021-03-05 at 20 24 46](https://user-images.githubusercontent.com/1032474/110157537-f3276380-7df0-11eb-8364-77ff51f25a7b.png) ### What I've already tried to fix it Gone through the [change history for `AutoSet.php`](https://github.com/Laravel-Backpack/CRUD/commits/master/src/app/Library/CrudPanel/Traits/AutoSet.php) but it doesn't _look like_ the last few changes are the one that pushed it over the edge... I used a panel that uses `sstFromDb()` about ~4 days ago, it was working, then I ran a `composer update`, and... now it's not working. The problem is - we're just asking everyone to run a `composer update` for security issues this week. And the version everyone's going to get... will break CRUDs that use `setFromDb()`. ### Backpack, Laravel, PHP, DB version When I run ```php artisan backpack:version``` the output is: ``` ### PHP VERSION: PHP 8.0.2 (cli) (built: Feb 4 2021 18:57:55) ( NTS ) Copyright (c) The PHP Group Zend Engine v4.0.2, Copyright (c) Zend Technologies with Zend OPcache v8.0.2, Copyright (c), by Zend Technologies ### LARAVEL VERSION: v7.30.4@9dd38140dc2924daa1a020a3d7a45f9ceff03df3 ### BACKPACK VERSION: 4.1.37@f829c37501d2f015bc5a4c93ef6dbc4533354ba8 ```
1.0
[Bug] CrudControllers with setFromDb() don't work with DBAL v3 - # Bug report ### What I did A simple CRUD that uses `setFromDb()`, for example Countries with: ```php <?php use Illuminate\Database\Migrations\Migration; use Illuminate\Database\Schema\Blueprint; class CreateCountriesTable extends Migration { /** * Run the migrations. * * @return void */ public function up() { Schema::create('countries', function (Blueprint $table) { $table->increments('id'); $table->string('code', 2)->index(); $table->string('name', 75); }); } /** * Reverse the migrations. * * @return void */ public function down() { Schema::drop('countries'); } } ``` ### What I expected to happen The default CountryCrudController to work if you do nothing. So this to work: ```php <?php namespace App\Http\Controllers\Admin; use App\Http\Requests\CountryRequest as StoreRequest; // VALIDATION: change the requests to match your own file names if you need form validation use App\Http\Requests\CountryRequest as UpdateRequest; use Backpack\CRUD\app\Http\Controllers\CrudController; class CountryCrudController extends CrudController { use \Backpack\CRUD\app\Http\Controllers\Operations\ListOperation; use \Backpack\CRUD\app\Http\Controllers\Operations\CreateOperation; use \Backpack\CRUD\app\Http\Controllers\Operations\UpdateOperation; use \Backpack\CRUD\app\Http\Controllers\Operations\DeleteOperation; public function setup() { /* |-------------------------------------------------------------------------- | BASIC CRUD INFORMATION |-------------------------------------------------------------------------- */ $this->crud->setModel(\App\Models\Country::class); $this->crud->setRoute('admin/country'); $this->crud->setEntityNameStrings('country', 'countries'); /* |-------------------------------------------------------------------------- | BASIC CRUD INFORMATION |-------------------------------------------------------------------------- */ $this->crud->setFromDb(); } protected function setupCreateOperation() { $this->crud->setValidation(StoreRequest::class); } protected function setupUpdateOperation() { $this->crud->setValidation(UpdateRequest::class); } } ``` ### What happened `“Class 'Doctrine\DBAL\Driver\\PDOMySql\Driver' not found”` error was thrown when going on the list page, so the ListOperation is unusable. [Click here for stack trace](http://flareapp.io/share/xPQ6LxD7). Screenshot below: ![Screenshot 2021-03-05 at 20 24 46](https://user-images.githubusercontent.com/1032474/110157537-f3276380-7df0-11eb-8364-77ff51f25a7b.png) ### What I've already tried to fix it Gone through the [change history for `AutoSet.php`](https://github.com/Laravel-Backpack/CRUD/commits/master/src/app/Library/CrudPanel/Traits/AutoSet.php) but it doesn't _look like_ the last few changes are the one that pushed it over the edge... I used a panel that uses `sstFromDb()` about ~4 days ago, it was working, then I ran a `composer update`, and... now it's not working. The problem is - we're just asking everyone to run a `composer update` for security issues this week. And the version everyone's going to get... will break CRUDs that use `setFromDb()`. ### Backpack, Laravel, PHP, DB version When I run ```php artisan backpack:version``` the output is: ``` ### PHP VERSION: PHP 8.0.2 (cli) (built: Feb 4 2021 18:57:55) ( NTS ) Copyright (c) The PHP Group Zend Engine v4.0.2, Copyright (c) Zend Technologies with Zend OPcache v8.0.2, Copyright (c), by Zend Technologies ### LARAVEL VERSION: v7.30.4@9dd38140dc2924daa1a020a3d7a45f9ceff03df3 ### BACKPACK VERSION: 4.1.37@f829c37501d2f015bc5a4c93ef6dbc4533354ba8 ```
non_test
crudcontrollers with setfromdb don t work with dbal bug report what i did a simple crud that uses setfromdb for example countries with php php use illuminate database migrations migration use illuminate database schema blueprint class createcountriestable extends migration run the migrations return void public function up schema create countries function blueprint table table increments id table string code index table string name reverse the migrations return void public function down schema drop countries what i expected to happen the default countrycrudcontroller to work if you do nothing so this to work php php namespace app http controllers admin use app http requests countryrequest as storerequest validation change the requests to match your own file names if you need form validation use app http requests countryrequest as updaterequest use backpack crud app http controllers crudcontroller class countrycrudcontroller extends crudcontroller use backpack crud app http controllers operations listoperation use backpack crud app http controllers operations createoperation use backpack crud app http controllers operations updateoperation use backpack crud app http controllers operations deleteoperation public function setup basic crud information this crud setmodel app models country class this crud setroute admin country this crud setentitynamestrings country countries basic crud information this crud setfromdb protected function setupcreateoperation this crud setvalidation storerequest class protected function setupupdateoperation this crud setvalidation updaterequest class what happened “class doctrine dbal driver pdomysql driver not found” error was thrown when going on the list page so the listoperation is unusable screenshot below what i ve already tried to fix it gone through the but it doesn t look like the last few changes are the one that pushed it over the edge i used a panel that uses sstfromdb about days ago it was working then i ran a composer update and now it s not working the problem is we re just asking everyone to run a composer update for security issues this week and the version everyone s going to get will break cruds that use setfromdb backpack laravel php db version when i run php artisan backpack version the output is php version php cli built feb nts copyright c the php group zend engine copyright c zend technologies with zend opcache copyright c by zend technologies laravel version backpack version
0
40,107
5,273,088,113
IssuesEvent
2017-02-06 14:46:25
ntop/ntopng
https://api.github.com/repos/ntop/ntopng
closed
Cant count correct traffic in Elasticsearch/Kibana
testing needed
Hi, I export flows with the ntopng free version to Elasticsearch like this: `-F="es;flows;ntopng-%Y.%m.%d;http://localhost:9200/_bulk;"` After dowloading a file I can see the correct amount of traffic in the ntopng WebGUI, but when I calculate the traffic for that connection in Kibana there is a much smaller amount of traffic. Is ntopng doing sampling when exporting flows to Elasticsearch. If yes - how can I disable this?
1.0
Cant count correct traffic in Elasticsearch/Kibana - Hi, I export flows with the ntopng free version to Elasticsearch like this: `-F="es;flows;ntopng-%Y.%m.%d;http://localhost:9200/_bulk;"` After dowloading a file I can see the correct amount of traffic in the ntopng WebGUI, but when I calculate the traffic for that connection in Kibana there is a much smaller amount of traffic. Is ntopng doing sampling when exporting flows to Elasticsearch. If yes - how can I disable this?
test
cant count correct traffic in elasticsearch kibana hi i export flows with the ntopng free version to elasticsearch like this f es flows ntopng y m d after dowloading a file i can see the correct amount of traffic in the ntopng webgui but when i calculate the traffic for that connection in kibana there is a much smaller amount of traffic is ntopng doing sampling when exporting flows to elasticsearch if yes how can i disable this
1
50,159
21,036,235,475
IssuesEvent
2022-03-31 08:07:27
elastic/beats
https://api.github.com/repos/elastic/beats
closed
improve logging by log separation for better troubleshooting purposes
enhancement libbeat Stalled Team:Services
**Describe the enhancement:** I would like to have a better segregated logging in filebeat, metricbeat,... per module/input/output **Describe a specific use case for the enhancement or feature:** Sometimes it happens that we have more complex filebeat configuration files with lots of enabled inputs and lots of enabled modules it can sometimes get very tricky even with enabled DEBUG logging to turn down certain issues where they are really located when it comes to logical configuration issues Looking away to logstash there are really helpful tools for troubleshooting this. 1. Since if I recall correctly 7.4 we always have a prefix in the logstash logs, that always log as well the pipeline where a certain event was happening. 2. You additionally have the possibility to segregate the logs per pipeline using setting `pipeline.separate_logs` https://www.elastic.co/guide/en/logstash/current/logstash-settings-file.html Now having this said, can't we have a similar functionality in the beats logs, that tell when something happened in an input/output/processor/module which one really was it? Can we separate these log files per input/output/module as well. I think that should help for such issues to take the burden a little bit away from troubleshooting knowing exactly where certain things are happening?
1.0
improve logging by log separation for better troubleshooting purposes - **Describe the enhancement:** I would like to have a better segregated logging in filebeat, metricbeat,... per module/input/output **Describe a specific use case for the enhancement or feature:** Sometimes it happens that we have more complex filebeat configuration files with lots of enabled inputs and lots of enabled modules it can sometimes get very tricky even with enabled DEBUG logging to turn down certain issues where they are really located when it comes to logical configuration issues Looking away to logstash there are really helpful tools for troubleshooting this. 1. Since if I recall correctly 7.4 we always have a prefix in the logstash logs, that always log as well the pipeline where a certain event was happening. 2. You additionally have the possibility to segregate the logs per pipeline using setting `pipeline.separate_logs` https://www.elastic.co/guide/en/logstash/current/logstash-settings-file.html Now having this said, can't we have a similar functionality in the beats logs, that tell when something happened in an input/output/processor/module which one really was it? Can we separate these log files per input/output/module as well. I think that should help for such issues to take the burden a little bit away from troubleshooting knowing exactly where certain things are happening?
non_test
improve logging by log separation for better troubleshooting purposes describe the enhancement i would like to have a better segregated logging in filebeat metricbeat per module input output describe a specific use case for the enhancement or feature sometimes it happens that we have more complex filebeat configuration files with lots of enabled inputs and lots of enabled modules it can sometimes get very tricky even with enabled debug logging to turn down certain issues where they are really located when it comes to logical configuration issues looking away to logstash there are really helpful tools for troubleshooting this since if i recall correctly we always have a prefix in the logstash logs that always log as well the pipeline where a certain event was happening you additionally have the possibility to segregate the logs per pipeline using setting pipeline separate logs now having this said can t we have a similar functionality in the beats logs that tell when something happened in an input output processor module which one really was it can we separate these log files per input output module as well i think that should help for such issues to take the burden a little bit away from troubleshooting knowing exactly where certain things are happening
0
111,800
24,197,586,782
IssuesEvent
2022-09-24 04:53:26
microsoft/vscode-cpptools
https://api.github.com/repos/microsoft/vscode-cpptools
closed
clang-format complaing about unknown key after I deleted the bundled executable?
Language Service more info needed Feature: Code Formatting
### Environment - OS and Version: Windows 10 - VS Code Version: 1.71.1 - C/C++ Extension Version: 1.12.4 - Other extensions you installed (and if the issue persists after disabling them): remote - ssh, clangd - If using SSH remote, specify OS of remote machine: Linux - A clear and concise description of what the bug is, including information about the workspace (i.e. is the workspace a single project or multiple projects, size of the project, etc). See bug summary below ### Bug Summary and Steps to Reproduce Bug Summary: When trying to format a .c file I see this error on clangd log: /home/user/.clang-format:61:1: error: unknown key 'SeparateDefinitionBlocks' SeparateDefinitionBlocks: Always But if I run the bundled clang-format manually from .vscode-server/extensions/ms-vscode.cpptools-1.12.4-linux-x64/LLVM/bin, it is able to format the same file without errors (I double checked it was reading same settings from .clang-format). The clang-format documentation for that tag says is supported on version 14, which is the one bundled. So I am puzzled ... Just for the sake of it, I deleted the clang-format executable from the mentioned path, expecting to get a different error (missing executable or something). To my surprise though, it continues to complain about the unknown key! I thought that error was coming from clang-format itself, but per my experiment is coming from some other place? Please clarify and advise how to fix this, Thanks. Steps to reproduce: 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error ### Expected behavior . ### Code sample and Logs ```shell . ``` ### Screenshots . ### Additional context _No response_
1.0
clang-format complaing about unknown key after I deleted the bundled executable? - ### Environment - OS and Version: Windows 10 - VS Code Version: 1.71.1 - C/C++ Extension Version: 1.12.4 - Other extensions you installed (and if the issue persists after disabling them): remote - ssh, clangd - If using SSH remote, specify OS of remote machine: Linux - A clear and concise description of what the bug is, including information about the workspace (i.e. is the workspace a single project or multiple projects, size of the project, etc). See bug summary below ### Bug Summary and Steps to Reproduce Bug Summary: When trying to format a .c file I see this error on clangd log: /home/user/.clang-format:61:1: error: unknown key 'SeparateDefinitionBlocks' SeparateDefinitionBlocks: Always But if I run the bundled clang-format manually from .vscode-server/extensions/ms-vscode.cpptools-1.12.4-linux-x64/LLVM/bin, it is able to format the same file without errors (I double checked it was reading same settings from .clang-format). The clang-format documentation for that tag says is supported on version 14, which is the one bundled. So I am puzzled ... Just for the sake of it, I deleted the clang-format executable from the mentioned path, expecting to get a different error (missing executable or something). To my surprise though, it continues to complain about the unknown key! I thought that error was coming from clang-format itself, but per my experiment is coming from some other place? Please clarify and advise how to fix this, Thanks. Steps to reproduce: 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error ### Expected behavior . ### Code sample and Logs ```shell . ``` ### Screenshots . ### Additional context _No response_
non_test
clang format complaing about unknown key after i deleted the bundled executable environment os and version windows vs code version c c extension version other extensions you installed and if the issue persists after disabling them remote ssh clangd if using ssh remote specify os of remote machine linux a clear and concise description of what the bug is including information about the workspace i e is the workspace a single project or multiple projects size of the project etc see bug summary below bug summary and steps to reproduce bug summary when trying to format a c file i see this error on clangd log home user clang format error unknown key separatedefinitionblocks separatedefinitionblocks always but if i run the bundled clang format manually from vscode server extensions ms vscode cpptools linux llvm bin it is able to format the same file without errors i double checked it was reading same settings from clang format the clang format documentation for that tag says is supported on version which is the one bundled so i am puzzled just for the sake of it i deleted the clang format executable from the mentioned path expecting to get a different error missing executable or something to my surprise though it continues to complain about the unknown key i thought that error was coming from clang format itself but per my experiment is coming from some other place please clarify and advise how to fix this thanks steps to reproduce go to click on scroll down to see error expected behavior code sample and logs shell screenshots additional context no response
0
28,193
5,437,580,230
IssuesEvent
2017-03-06 07:35:47
JamesRamm/longclaw
https://api.github.com/repos/JamesRamm/longclaw
opened
Document payment backends
documentation
Write documentation to describe: - Configuring payment backends - Available payment backends - Selecting a payment backend (in the settings file) - Additional settings needed for each backend (i.e. authorisation tokens) - Developing new backends - `BasePayment` class - Methods that need implementing (`create_payment`) - Philosophy
1.0
Document payment backends - Write documentation to describe: - Configuring payment backends - Available payment backends - Selecting a payment backend (in the settings file) - Additional settings needed for each backend (i.e. authorisation tokens) - Developing new backends - `BasePayment` class - Methods that need implementing (`create_payment`) - Philosophy
non_test
document payment backends write documentation to describe configuring payment backends available payment backends selecting a payment backend in the settings file additional settings needed for each backend i e authorisation tokens developing new backends basepayment class methods that need implementing create payment philosophy
0
179,080
14,690,287,299
IssuesEvent
2021-01-02 14:31:29
nanoy42/django-voting-app
https://api.github.com/repos/nanoy42/django-voting-app
closed
Order of documentation not clear
documentation
In the documentation, some parts of the installation cannot be done without the configuration, and the configuration page is displayed after the installation page.
1.0
Order of documentation not clear - In the documentation, some parts of the installation cannot be done without the configuration, and the configuration page is displayed after the installation page.
non_test
order of documentation not clear in the documentation some parts of the installation cannot be done without the configuration and the configuration page is displayed after the installation page
0
334,806
29,992,331,880
IssuesEvent
2023-06-26 00:03:29
flojoy-io/studio
https://api.github.com/repos/flojoy-io/studio
closed
Control Panel Unit Tests
unit test CI & test automation
- [ ] ControlComponentState - [ ] ControlComponent - [ ] PlotControlState - [ ] PlotControl - [ ] NodeReference - [ ] KnobCtrl
2.0
Control Panel Unit Tests - - [ ] ControlComponentState - [ ] ControlComponent - [ ] PlotControlState - [ ] PlotControl - [ ] NodeReference - [ ] KnobCtrl
test
control panel unit tests controlcomponentstate controlcomponent plotcontrolstate plotcontrol nodereference knobctrl
1
50,785
6,111,760,294
IssuesEvent
2017-06-21 17:47:04
Augugrumi/libris
https://api.github.com/repos/Augugrumi/libris
closed
Fix error 503 in GoogleImage unit-tests
bug image test
When running GoogleImage tests on Travis-CI sometimes a 503 error is thrown. We need to fix this issue.
1.0
Fix error 503 in GoogleImage unit-tests - When running GoogleImage tests on Travis-CI sometimes a 503 error is thrown. We need to fix this issue.
test
fix error in googleimage unit tests when running googleimage tests on travis ci sometimes a error is thrown we need to fix this issue
1
211,990
16,386,697,830
IssuesEvent
2021-05-17 11:25:48
packit/packit
https://api.github.com/repos/packit/packit
closed
F32 and EPEL8 tests fail in TF with `AttributeError: 'PosixPath' object has no attribute 'is_relative_to'`
testing
http://artifacts.dev.testing-farm.io/05d3ab42-b596-4f6d-ab05-fa649a8c7078/ is that a feature of a newer python? F33 and F34 pass
1.0
F32 and EPEL8 tests fail in TF with `AttributeError: 'PosixPath' object has no attribute 'is_relative_to'` - http://artifacts.dev.testing-farm.io/05d3ab42-b596-4f6d-ab05-fa649a8c7078/ is that a feature of a newer python? F33 and F34 pass
test
and tests fail in tf with attributeerror posixpath object has no attribute is relative to is that a feature of a newer python and pass
1
203,050
15,342,582,022
IssuesEvent
2021-02-27 16:48:17
Perl/perl5
https://api.github.com/repos/Perl/perl5
closed
pp_split: \s+ pattern does not hit RXf_WHITE branch
Bug affects-blead hasPatch hasTest type-regex
**Description** According to [split](https://perldoc.perl.org/functions/split) and [perlreapi](https://github.com/Perl/perl5/blob/blead/pod/perlreapi.pod) docs, the `RXf_WHITE `branch in `pp_split` is supposed to be used - rather than the regex engine - if either of two patterns is used for the split: * " " * \s+ In blead, the first pattern does trigger the use of `pp_split`'s `RXf_WHITE` branch, but the second does not. **Steps to Reproduce** By sticking a `PerlIO_stdoutf("")` statement at the start of each of the main `pp_split` branches, the following is seen to be handled by a regex branch, rather than the `RXf_WHITE` branch as advertised. `# ./perl -e 'my $perls = "Perl "x1_000_000; for ( split /\s+/, $perls ) {}'` Note: The `RXf_START_ONLY` and `RXf_NULL` branches were checked and still work as advertised. **Expected behavior** The sample code above uses the `RXf_WHITE` branch. The current behaviour is at odds with documented behaviour and may also represent a performance regression. **TODO** 1) Find out when this stopped working. I'll have a go at manually bisecting this, since it looks hard to automatically bisect. 2) Find out why it's broken. Not sure if it's the parser or regex engine at fault... 3) Fix it!
1.0
pp_split: \s+ pattern does not hit RXf_WHITE branch - **Description** According to [split](https://perldoc.perl.org/functions/split) and [perlreapi](https://github.com/Perl/perl5/blob/blead/pod/perlreapi.pod) docs, the `RXf_WHITE `branch in `pp_split` is supposed to be used - rather than the regex engine - if either of two patterns is used for the split: * " " * \s+ In blead, the first pattern does trigger the use of `pp_split`'s `RXf_WHITE` branch, but the second does not. **Steps to Reproduce** By sticking a `PerlIO_stdoutf("")` statement at the start of each of the main `pp_split` branches, the following is seen to be handled by a regex branch, rather than the `RXf_WHITE` branch as advertised. `# ./perl -e 'my $perls = "Perl "x1_000_000; for ( split /\s+/, $perls ) {}'` Note: The `RXf_START_ONLY` and `RXf_NULL` branches were checked and still work as advertised. **Expected behavior** The sample code above uses the `RXf_WHITE` branch. The current behaviour is at odds with documented behaviour and may also represent a performance regression. **TODO** 1) Find out when this stopped working. I'll have a go at manually bisecting this, since it looks hard to automatically bisect. 2) Find out why it's broken. Not sure if it's the parser or regex engine at fault... 3) Fix it!
test
pp split s pattern does not hit rxf white branch description according to and docs the rxf white branch in pp split is supposed to be used rather than the regex engine if either of two patterns is used for the split s in blead the first pattern does trigger the use of pp split s rxf white branch but the second does not steps to reproduce by sticking a perlio stdoutf statement at the start of each of the main pp split branches the following is seen to be handled by a regex branch rather than the rxf white branch as advertised perl e my perls perl for split s perls note the rxf start only and rxf null branches were checked and still work as advertised expected behavior the sample code above uses the rxf white branch the current behaviour is at odds with documented behaviour and may also represent a performance regression todo find out when this stopped working i ll have a go at manually bisecting this since it looks hard to automatically bisect find out why it s broken not sure if it s the parser or regex engine at fault fix it
1
197,933
6,966,152,367
IssuesEvent
2017-12-09 15:29:03
lucaslioli/payless-api
https://api.github.com/repos/lucaslioli/payless-api
opened
Create a word dictionary
future feature medium priority
To improve the user experience, becomes necessary to provide a better mechanism for searching the product names, also to return a precisely response to the user about what he searches for.
1.0
Create a word dictionary - To improve the user experience, becomes necessary to provide a better mechanism for searching the product names, also to return a precisely response to the user about what he searches for.
non_test
create a word dictionary to improve the user experience becomes necessary to provide a better mechanism for searching the product names also to return a precisely response to the user about what he searches for
0
48,370
13,068,468,024
IssuesEvent
2020-07-31 03:40:25
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
Remove Noise paremter injecto in Vuvuzela segment (Trac #2243)
Migrated from Trac combo simulation defect
As per Andrii Terliuk's presentation in the IceCube bi-weekly call (22 January 2019): [https://drive.google.com/drive/folders/1DAiEVsRibOBXlRKsWO6lelsIdWmG29K4] Migrated from https://code.icecube.wisc.edu/ticket/2243 ```json { "status": "closed", "changetime": "2019-02-27T20:09:10", "description": "As per Andrii Terliuk's presentation in the IceCube bi-weekly call (22 January 2019):\n\n[https://drive.google.com/drive/folders/1DAiEVsRibOBXlRKsWO6lelsIdWmG29K4]", "reporter": "juancarlos", "cc": "", "resolution": "fixed", "_ts": "1551298150937247", "component": "combo simulation", "summary": "Remove Noise paremter injecto in Vuvuzela segment", "priority": "major", "keywords": "Vuvuzela", "time": "2019-02-27T20:08:12", "milestone": "", "owner": "juancarlos", "type": "defect" } ```
1.0
Remove Noise paremter injecto in Vuvuzela segment (Trac #2243) - As per Andrii Terliuk's presentation in the IceCube bi-weekly call (22 January 2019): [https://drive.google.com/drive/folders/1DAiEVsRibOBXlRKsWO6lelsIdWmG29K4] Migrated from https://code.icecube.wisc.edu/ticket/2243 ```json { "status": "closed", "changetime": "2019-02-27T20:09:10", "description": "As per Andrii Terliuk's presentation in the IceCube bi-weekly call (22 January 2019):\n\n[https://drive.google.com/drive/folders/1DAiEVsRibOBXlRKsWO6lelsIdWmG29K4]", "reporter": "juancarlos", "cc": "", "resolution": "fixed", "_ts": "1551298150937247", "component": "combo simulation", "summary": "Remove Noise paremter injecto in Vuvuzela segment", "priority": "major", "keywords": "Vuvuzela", "time": "2019-02-27T20:08:12", "milestone": "", "owner": "juancarlos", "type": "defect" } ```
non_test
remove noise paremter injecto in vuvuzela segment trac as per andrii terliuk s presentation in the icecube bi weekly call january migrated from json status closed changetime description as per andrii terliuk s presentation in the icecube bi weekly call january n n reporter juancarlos cc resolution fixed ts component combo simulation summary remove noise paremter injecto in vuvuzela segment priority major keywords vuvuzela time milestone owner juancarlos type defect
0
193,400
22,216,147,141
IssuesEvent
2022-06-08 02:00:42
artsking/linux-4.1.15
https://api.github.com/repos/artsking/linux-4.1.15
reopened
CVE-2017-8797 (High) detected in linux-stable-rtv4.1.33
security vulnerability
## CVE-2017-8797 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/artsking/linux-4.1.15/commit/b1c15f7dc4cfe553aeed8332e46f285ee92b5756">b1c15f7dc4cfe553aeed8332e46f285ee92b5756</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The NFSv4 server in the Linux kernel before 4.11.3 does not properly validate the layout type when processing the NFSv4 pNFS GETDEVICEINFO or LAYOUTGET operand in a UDP packet from a remote attacker. This type value is uninitialized upon encountering certain error conditions. This value is used as an array index for dereferencing, which leads to an OOPS and eventually a DoS of knfsd and a soft-lockup of the whole system. <p>Publish Date: 2017-07-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-8797>CVE-2017-8797</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2017-8797">https://www.linuxkernelcves.com/cves/CVE-2017-8797</a></p> <p>Release Date: 2017-07-02</p> <p>Fix Resolution: v4.12-rc1,v4.11.3,v4.9.30</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-8797 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2017-8797 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/artsking/linux-4.1.15/commit/b1c15f7dc4cfe553aeed8332e46f285ee92b5756">b1c15f7dc4cfe553aeed8332e46f285ee92b5756</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The NFSv4 server in the Linux kernel before 4.11.3 does not properly validate the layout type when processing the NFSv4 pNFS GETDEVICEINFO or LAYOUTGET operand in a UDP packet from a remote attacker. This type value is uninitialized upon encountering certain error conditions. This value is used as an array index for dereferencing, which leads to an OOPS and eventually a DoS of knfsd and a soft-lockup of the whole system. <p>Publish Date: 2017-07-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-8797>CVE-2017-8797</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2017-8797">https://www.linuxkernelcves.com/cves/CVE-2017-8797</a></p> <p>Release Date: 2017-07-02</p> <p>Fix Resolution: v4.12-rc1,v4.11.3,v4.9.30</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details the server in the linux kernel before does not properly validate the layout type when processing the pnfs getdeviceinfo or layoutget operand in a udp packet from a remote attacker this type value is uninitialized upon encountering certain error conditions this value is used as an array index for dereferencing which leads to an oops and eventually a dos of knfsd and a soft lockup of the whole system publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
71,884
8,688,864,605
IssuesEvent
2018-12-03 17:07:19
CCBlueX/LiquidBounce1.8-Issues
https://api.github.com/repos/CCBlueX/LiquidBounce1.8-Issues
closed
Better Colorchanger
GUI (Design etc) Request
Something like this instead of having slider for red, green and blue. ![color](https://user-images.githubusercontent.com/29576723/34543060-5e7ac9d0-f0df-11e7-9298-57d7131901be.PNG)
1.0
Better Colorchanger - Something like this instead of having slider for red, green and blue. ![color](https://user-images.githubusercontent.com/29576723/34543060-5e7ac9d0-f0df-11e7-9298-57d7131901be.PNG)
non_test
better colorchanger something like this instead of having slider for red green and blue
0