Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
957
labels
stringlengths
4
795
body
stringlengths
1
259k
index
stringclasses
12 values
text_combine
stringlengths
96
259k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
852
2,648,610,319
IssuesEvent
2015-03-14 02:37:20
WPN-XM/WPN-XM
https://api.github.com/repos/WPN-XM/WPN-XM
closed
try to compile the innosetup scripts on ci server
build tools installation wizard
**Try to compile the InnoSetup installation wizards on a CI server/unix** - [x] travis-ci or - [ ] jenkins-ci - [x] use wine - [x] install innosetup into wine - we already have innosetup in /bin/innosetup - it's no problem to execute the compiler `wineconsole issc`or `wine issc` - [x] run the build file - fix possible windows path issues - maybe clone the build file for unix? - run build file with phing and ant at the same time - ant would be used for the parts relying on git. we dont need to install git into wine. **Build Artifact Handling** - [ ] upload build artifacts - to a nightly folder ? - http://docs.travis-ci.com/user/deployment/releases/ - `after_success: php upload.php` ? - $GITHUB_TOKEN - $TRAVIS_COMMIT - file to upload - release filename - $TRAVIS_BRANCH - [ ] add "do-release" task, to push a new release to github - [ ] add "add-artifact" task, to push a new artifact to a release tag on github - [ ] add "add-artifacts" task, to push all artifacts (executables) via add-artifact **Automatic Deployment** - [ ] official wpnxm server via sftp/scp - [ ] github releases (curl) - https://developer.github.com/v3/repos/releases/#upload-a-release-asset - https://developer.github.com/changes/2013-09-25-releases-api/ - **CURL POST** `$ curl -H "Authorization: token TOKEN" \ -H "Accept: application/vnd.github.manifold-preview" \ -H "Content-Type: application/zip" \ --data-binary @build/mac/package.zip \ "https://uploads.github.com/repos/hubot/singularity/releases/123/assets?name=1.0.0-mac.zip"` - [ ] sourceforge (sftp/scp/curl) - https://sourceforge.net/p/forge/community-docs/Using%20the%20Release%20API/ - https://sourceforge.net/p/forge/documentation/Release%20Files%20for%20Download/#scp
1.0
try to compile the innosetup scripts on ci server - **Try to compile the InnoSetup installation wizards on a CI server/unix** - [x] travis-ci or - [ ] jenkins-ci - [x] use wine - [x] install innosetup into wine - we already have innosetup in /bin/innosetup - it's no problem to execute the compiler `wineconsole issc`or `wine issc` - [x] run the build file - fix possible windows path issues - maybe clone the build file for unix? - run build file with phing and ant at the same time - ant would be used for the parts relying on git. we dont need to install git into wine. **Build Artifact Handling** - [ ] upload build artifacts - to a nightly folder ? - http://docs.travis-ci.com/user/deployment/releases/ - `after_success: php upload.php` ? - $GITHUB_TOKEN - $TRAVIS_COMMIT - file to upload - release filename - $TRAVIS_BRANCH - [ ] add "do-release" task, to push a new release to github - [ ] add "add-artifact" task, to push a new artifact to a release tag on github - [ ] add "add-artifacts" task, to push all artifacts (executables) via add-artifact **Automatic Deployment** - [ ] official wpnxm server via sftp/scp - [ ] github releases (curl) - https://developer.github.com/v3/repos/releases/#upload-a-release-asset - https://developer.github.com/changes/2013-09-25-releases-api/ - **CURL POST** `$ curl -H "Authorization: token TOKEN" \ -H "Accept: application/vnd.github.manifold-preview" \ -H "Content-Type: application/zip" \ --data-binary @build/mac/package.zip \ "https://uploads.github.com/repos/hubot/singularity/releases/123/assets?name=1.0.0-mac.zip"` - [ ] sourceforge (sftp/scp/curl) - https://sourceforge.net/p/forge/community-docs/Using%20the%20Release%20API/ - https://sourceforge.net/p/forge/documentation/Release%20Files%20for%20Download/#scp
non_priority
try to compile the innosetup scripts on ci server try to compile the innosetup installation wizards on a ci server unix travis ci or jenkins ci use wine install innosetup into wine we already have innosetup in bin innosetup it s no problem to execute the compiler wineconsole issc or wine issc run the build file fix possible windows path issues maybe clone the build file for unix run build file with phing and ant at the same time ant would be used for the parts relying on git we dont need to install git into wine build artifact handling upload build artifacts to a nightly folder after success php upload php github token travis commit file to upload release filename travis branch add do release task to push a new release to github add add artifact task to push a new artifact to a release tag on github add add artifacts task to push all artifacts executables via add artifact automatic deployment official wpnxm server via sftp scp github releases curl curl post curl h authorization token token h accept application vnd github manifold preview h content type application zip data binary build mac package zip sourceforge sftp scp curl
0
166,827
12,973,078,912
IssuesEvent
2020-07-21 13:31:55
LiskHQ/lisk-sdk
https://api.github.com/repos/LiskHQ/lisk-sdk
opened
Refactor synchronization unit tests
type: refactoring type: test
### Description Currently synchronization unit tests (fast chain switching and block syncrhonization) are using complex setup of stubs. If any of the implementation detail changes, it is very hard to update the tests. It should be cleaned up to be more modifiable and readable with less mock setups.
1.0
Refactor synchronization unit tests - ### Description Currently synchronization unit tests (fast chain switching and block syncrhonization) are using complex setup of stubs. If any of the implementation detail changes, it is very hard to update the tests. It should be cleaned up to be more modifiable and readable with less mock setups.
non_priority
refactor synchronization unit tests description currently synchronization unit tests fast chain switching and block syncrhonization are using complex setup of stubs if any of the implementation detail changes it is very hard to update the tests it should be cleaned up to be more modifiable and readable with less mock setups
0
22,046
11,462,903,789
IssuesEvent
2020-02-07 15:01:10
motss/app-datepicker
https://api.github.com/repos/motss/app-datepicker
closed
Recommended performance improvement
enhancement help-wanted help-welcome in-progress lit-element needs investigation performance
### Description The Google Chrome browser warns (console) about a performance hit with non-passive event listeners. It points to tracker.js. A recommendation for improvement is documented here: [passive-event-listeners](https://developers.google.com/web/tools/lighthouse/audits/passive-event-listeners) ### Modifications tested My apologies, I didn't have much time to look at this but to quickly test, and serve as an example I made the following updates to tracker.js: 1. in class Tracker, add the passive option to the event listeners. In practice the recommendation is to check if the passive option is support. In my case it wasn't necessary, hence: ``` _element.addEventListener('mousedown', this._down, {passive:true}); _element.addEventListener('touchstart', this._down, {passive:true}); _element.addEventListener('touchmove', this._move, {passive:true}); _element.addEventListener('touchend', this._up, {passive:true}); ``` 2. in _onDown(down), remove: `ev.preventDefault()` ### Result The Google warning went away and the element appeared to work OK. Hopefully this is useful to you.
True
Recommended performance improvement - ### Description The Google Chrome browser warns (console) about a performance hit with non-passive event listeners. It points to tracker.js. A recommendation for improvement is documented here: [passive-event-listeners](https://developers.google.com/web/tools/lighthouse/audits/passive-event-listeners) ### Modifications tested My apologies, I didn't have much time to look at this but to quickly test, and serve as an example I made the following updates to tracker.js: 1. in class Tracker, add the passive option to the event listeners. In practice the recommendation is to check if the passive option is support. In my case it wasn't necessary, hence: ``` _element.addEventListener('mousedown', this._down, {passive:true}); _element.addEventListener('touchstart', this._down, {passive:true}); _element.addEventListener('touchmove', this._move, {passive:true}); _element.addEventListener('touchend', this._up, {passive:true}); ``` 2. in _onDown(down), remove: `ev.preventDefault()` ### Result The Google warning went away and the element appeared to work OK. Hopefully this is useful to you.
non_priority
recommended performance improvement description the google chrome browser warns console about a performance hit with non passive event listeners it points to tracker js a recommendation for improvement is documented here modifications tested my apologies i didn t have much time to look at this but to quickly test and serve as an example i made the following updates to tracker js in class tracker add the passive option to the event listeners in practice the recommendation is to check if the passive option is support in my case it wasn t necessary hence element addeventlistener mousedown this down passive true element addeventlistener touchstart this down passive true element addeventlistener touchmove this move passive true element addeventlistener touchend this up passive true in ondown down remove ev preventdefault result the google warning went away and the element appeared to work ok hopefully this is useful to you
0
45,455
11,672,031,528
IssuesEvent
2020-03-04 05:19:31
spring-projects/spring-security
https://api.github.com/repos/spring-projects/spring-security
closed
Update to spring-build-conventions:0.0.31.RELEASE
in: build type: dependency-upgrade
<!-- For Security Vulnerabilities, please use https://pivotal.io/security#reporting --> ### Summary Update to spring-build-conventions:0.0.31.RELEASE This will fix gh-8025
1.0
Update to spring-build-conventions:0.0.31.RELEASE - <!-- For Security Vulnerabilities, please use https://pivotal.io/security#reporting --> ### Summary Update to spring-build-conventions:0.0.31.RELEASE This will fix gh-8025
non_priority
update to spring build conventions release for security vulnerabilities please use summary update to spring build conventions release this will fix gh
0
349,874
24,959,666,910
IssuesEvent
2022-11-01 14:36:15
AY2223S1-CS2103T-T14-3/tp
https://api.github.com/repos/AY2223S1-CS2103T-T14-3/tp
closed
[PE-D][Tester E] UG: Misleading birthdate format for 'add' command
documentation Must Fix
In the UG, for 'add' command, the tip given for birthdate is as follows: "Tip: Birthdate must be given in DD-MM-YYYY HHmm format!" This is not accurate, as the correct format should have been DD-MM-YYYY instead! Be careful! <!--session: 1666945098506-7b9a3a6b-9771-404c-be16-3420d7464ef5--> <!--Version: Web v3.4.4--> ------------- Labels: `type.DocumentationBug` `severity.High` original: eugenelmj/ped#9
1.0
[PE-D][Tester E] UG: Misleading birthdate format for 'add' command - In the UG, for 'add' command, the tip given for birthdate is as follows: "Tip: Birthdate must be given in DD-MM-YYYY HHmm format!" This is not accurate, as the correct format should have been DD-MM-YYYY instead! Be careful! <!--session: 1666945098506-7b9a3a6b-9771-404c-be16-3420d7464ef5--> <!--Version: Web v3.4.4--> ------------- Labels: `type.DocumentationBug` `severity.High` original: eugenelmj/ped#9
non_priority
ug misleading birthdate format for add command in the ug for add command the tip given for birthdate is as follows tip birthdate must be given in dd mm yyyy hhmm format this is not accurate as the correct format should have been dd mm yyyy instead be careful labels type documentationbug severity high original eugenelmj ped
0
38,357
8,461,339,707
IssuesEvent
2018-10-22 21:30:37
MicrosoftDocs/live-share
https://api.github.com/repos/MicrosoftDocs/live-share
closed
[VS Code] The user code is invalid or expired. Try signing in again.
area: identity and sign-in duplicate os: linux vscode
<!-- For Visual Studio problems/feedback, please use the "Report a Problem..." feature built into the tool. See https://aka.ms/vsls-vsproblem. For VS Code issues, attach verbose logs as follows: 1. Press F1 (or Ctrl-Shift-P), type "export logs" and run the "Live Share: Export Logs" command. 2. Drag and drop the zip to the issue on this screen and wait for it to upload before creating the issue. For feature requests, please include enough of this same info so we know if the request is tool or language/platform specific. --> ## Error: The user code is invalid or expired. Try signing in again. ## Steps to Reproduce: 1. 2. ||Version Data| |-:|:-| |**extensionName**|VSLS| |**extensionVersion**|0.3.788| |**protocolVersion**|2.2| |**applicationName**|VSCode| |**applicationVersion**|1.28.0| |**platformName**|Linux| |**platformVersion**|4.15.0-36-generic|
1.0
[VS Code] The user code is invalid or expired. Try signing in again. - <!-- For Visual Studio problems/feedback, please use the "Report a Problem..." feature built into the tool. See https://aka.ms/vsls-vsproblem. For VS Code issues, attach verbose logs as follows: 1. Press F1 (or Ctrl-Shift-P), type "export logs" and run the "Live Share: Export Logs" command. 2. Drag and drop the zip to the issue on this screen and wait for it to upload before creating the issue. For feature requests, please include enough of this same info so we know if the request is tool or language/platform specific. --> ## Error: The user code is invalid or expired. Try signing in again. ## Steps to Reproduce: 1. 2. ||Version Data| |-:|:-| |**extensionName**|VSLS| |**extensionVersion**|0.3.788| |**protocolVersion**|2.2| |**applicationName**|VSCode| |**applicationVersion**|1.28.0| |**platformName**|Linux| |**platformVersion**|4.15.0-36-generic|
non_priority
the user code is invalid or expired try signing in again for visual studio problems feedback please use the report a problem feature built into the tool see for vs code issues attach verbose logs as follows press or ctrl shift p type export logs and run the live share export logs command drag and drop the zip to the issue on this screen and wait for it to upload before creating the issue for feature requests please include enough of this same info so we know if the request is tool or language platform specific error the user code is invalid or expired try signing in again steps to reproduce version data extensionname vsls extensionversion protocolversion applicationname vscode applicationversion platformname linux platformversion generic
0
301,796
22,775,237,020
IssuesEvent
2022-07-08 13:53:26
brpocock/grizzards
https://api.github.com/repos/brpocock/grizzards
closed
Final map graphic for manual
documentation enhancement
At least to include the Demo area and hints of Anchor Village, Lost Mine, Spiral Woods, Port Lion, and Shadow Valley.
1.0
Final map graphic for manual - At least to include the Demo area and hints of Anchor Village, Lost Mine, Spiral Woods, Port Lion, and Shadow Valley.
non_priority
final map graphic for manual at least to include the demo area and hints of anchor village lost mine spiral woods port lion and shadow valley
0
22,877
20,405,588,530
IssuesEvent
2022-02-23 04:44:25
geneontology/go-site
https://api.github.com/repos/geneontology/go-site
closed
Get validate.py to run all GORULES again
bug (B: affects usability)
@pgaudet Observed in 2021-05-22 `snapshot` that most GORULES were not running. This appears due to https://github.com/biolink/ontobio/issues/564. In the short term, to bypass the wonky default, we can just explicitly add `--rule-set all` to this `validate.py` cmd: https://github.com/geneontology/go-site/blob/0f6c248b5aa2f253cfe3a9a736cb424f7851e91c/pipeline/Makefile#L89 Tagging @kltm
True
Get validate.py to run all GORULES again - @pgaudet Observed in 2021-05-22 `snapshot` that most GORULES were not running. This appears due to https://github.com/biolink/ontobio/issues/564. In the short term, to bypass the wonky default, we can just explicitly add `--rule-set all` to this `validate.py` cmd: https://github.com/geneontology/go-site/blob/0f6c248b5aa2f253cfe3a9a736cb424f7851e91c/pipeline/Makefile#L89 Tagging @kltm
non_priority
get validate py to run all gorules again pgaudet observed in snapshot that most gorules were not running this appears due to in the short term to bypass the wonky default we can just explicitly add rule set all to this validate py cmd tagging kltm
0
6,441
8,737,904,727
IssuesEvent
2018-12-12 00:30:36
scylladb/scylla
https://api.github.com/repos/scylladb/scylla
closed
Cassandra's parameter for compression chunk configuration not accepted by Scylla
CQL cassandra 2.2 compatibility n00b
Cassandra configures the chunk size of compression with the parameter `chunk_length_in_kb`. Our parameter is a bit different, and it was probably an overlook on our side: `chunk_length_kb` Because we now would have Scylla deployments using our parameter, best would be to accept both.
True
Cassandra's parameter for compression chunk configuration not accepted by Scylla - Cassandra configures the chunk size of compression with the parameter `chunk_length_in_kb`. Our parameter is a bit different, and it was probably an overlook on our side: `chunk_length_kb` Because we now would have Scylla deployments using our parameter, best would be to accept both.
non_priority
cassandra s parameter for compression chunk configuration not accepted by scylla cassandra configures the chunk size of compression with the parameter chunk length in kb our parameter is a bit different and it was probably an overlook on our side chunk length kb because we now would have scylla deployments using our parameter best would be to accept both
0
50,297
13,187,429,124
IssuesEvent
2020-08-13 03:23:08
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
I3ParticleVector pybindings missing bases (Trac #479)
Migrated from Trac dataclasses defect
I3ParticleVector pybindings are missing the I3FrameObject base, and maybe other things. We also have a similar pybinding by the name of I3VectorI3Particle that does work. Let's resolve this naming duplication. Also examine other I3Vector classes for similar problems. I3RecoPulseSeries was noted. <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/479 , reported by david.schultz and owned by olivas_</summary> <p> ```json { "status": "closed", "changetime": "2015-02-12T06:52:34", "description": "I3ParticleVector pybindings are missing the I3FrameObject base, and maybe other things.\n\nWe also have a similar pybinding by the name of I3VectorI3Particle that does work. Let's resolve this naming duplication.\n\nAlso examine other I3Vector classes for similar problems. I3RecoPulseSeries was noted.", "reporter": "david.schultz", "cc": "", "resolution": "fixed", "_ts": "1423723954189338", "component": "dataclasses", "summary": "I3ParticleVector pybindings missing bases", "priority": "normal", "keywords": "", "time": "2014-01-22T03:15:57", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
1.0
I3ParticleVector pybindings missing bases (Trac #479) - I3ParticleVector pybindings are missing the I3FrameObject base, and maybe other things. We also have a similar pybinding by the name of I3VectorI3Particle that does work. Let's resolve this naming duplication. Also examine other I3Vector classes for similar problems. I3RecoPulseSeries was noted. <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/479 , reported by david.schultz and owned by olivas_</summary> <p> ```json { "status": "closed", "changetime": "2015-02-12T06:52:34", "description": "I3ParticleVector pybindings are missing the I3FrameObject base, and maybe other things.\n\nWe also have a similar pybinding by the name of I3VectorI3Particle that does work. Let's resolve this naming duplication.\n\nAlso examine other I3Vector classes for similar problems. I3RecoPulseSeries was noted.", "reporter": "david.schultz", "cc": "", "resolution": "fixed", "_ts": "1423723954189338", "component": "dataclasses", "summary": "I3ParticleVector pybindings missing bases", "priority": "normal", "keywords": "", "time": "2014-01-22T03:15:57", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
non_priority
pybindings missing bases trac pybindings are missing the base and maybe other things we also have a similar pybinding by the name of that does work let s resolve this naming duplication also examine other classes for similar problems was noted migrated from reported by david schultz and owned by olivas json status closed changetime description pybindings are missing the base and maybe other things n nwe also have a similar pybinding by the name of that does work let s resolve this naming duplication n nalso examine other classes for similar problems was noted reporter david schultz cc resolution fixed ts component dataclasses summary pybindings missing bases priority normal keywords time milestone owner olivas type defect
0
226,504
17,353,857,320
IssuesEvent
2021-07-29 12:12:38
STooDs-tools/RSTooDs
https://api.github.com/repos/STooDs-tools/RSTooDs
opened
Improvements to vignettes
documentation
* All vignettes: set knit directory to working directory (probably: knitr::opts_knit$set(root.dir = getwd()) in setup chunk). * Vignette "covariate": plot the pmf of the estimated Bernoulli by consistency with the first "getting started" vignette. * Create new vignette "Bayesian Basics" reproducing existing teaching material with RSTooDs.
1.0
Improvements to vignettes - * All vignettes: set knit directory to working directory (probably: knitr::opts_knit$set(root.dir = getwd()) in setup chunk). * Vignette "covariate": plot the pmf of the estimated Bernoulli by consistency with the first "getting started" vignette. * Create new vignette "Bayesian Basics" reproducing existing teaching material with RSTooDs.
non_priority
improvements to vignettes all vignettes set knit directory to working directory probably knitr opts knit set root dir getwd in setup chunk vignette covariate plot the pmf of the estimated bernoulli by consistency with the first getting started vignette create new vignette bayesian basics reproducing existing teaching material with rstoods
0
337,347
24,536,223,210
IssuesEvent
2022-10-11 21:01:26
trimble-oss/dba-dash
https://api.github.com/repos/trimble-oss/dba-dash
closed
Performance Summary riddles and puzzels
documentation
Hi Today I was reviewing the "PerfMon Summary" information (Version used 2.22.8) and I was wondering about following points below. It is worth to mention, all the data presented in all the issues are real world data and they are collected from productive systems refelecting real scenarios. Starting from top to bottom (in the image): - We got a column called "MB/Sec". I couldn't **clearly** decipher what this values is. Maybe a better column header would help. There are also no Tool Tips which explain the columns. This even could be implemnted as an additional feature: Tool Tip which explains the column headers in all the table grids, application wide. - Column "Read Latency" shows a painted cell as warning. I was seraching on the server itself for a quite while and I couldn't find anything which would help me to identify and localize the issue reported **on the spot**. I then searched even more and eventually I discovered a way to get some values and responsibles for the latency. This was hard and to some extent cryptical but maybe there are better ways. Anyhow, this was my approach: 1. Switch to single server. 2. Click on "Performance" Tab 3. On the "IO Perfromance" chart click "View Table Summary" (Which is yet another Modal Window I discovered :D) and then review the reported values which by the way they aren't painted at this stage so if one isn't paying attention he/she will then miss the point and leave. The above just tried to demonstrate how cumbersome is to identify the reported problem in detail. I hope you would come up with a easier / better way to solve this case. - Columns "Read Latecy" and "Write Latecy" are being painted as green. I really tried to understand **why**. I couldn't come up with any logic. - There are some cells in "Critical Wait" column which are gray colored. **Why**? Value 0.00 can't be gray AND green at the same time! - In "Lock Wait" and "Total Wait" columns, I find values way high in compare to the other similar values but these cells aren't painted at all. **Why**? Aren't we using any color indicators in theses columns? Shouldn't we do that? - There are some cells in the "Read Latecy" with no values at all. **Why?** Isn't there any values existing or are they so small that could be ignored? If so, then why can't we show 0,0? I am sure each one of the above questions can be answered one way or another but the MAIN point here is: The end user needs to **work** with these data and try not to decipher them. One way or another everyone (with enough persistence) can get along but that is also not the point of having the values collected. Application should do the most job for the user (because it is the tool, it is hard to keep track of these values and using visual effect one can transfer messages which helps the process.) and the user mostly needs to focus on identifying, anlasying of the collected data. This of course implies, collected data need to be precise and relyable on one hand and the application should try to avoid ambiguity on the other hand. ![perfmon_summary](https://user-images.githubusercontent.com/40255900/193550637-c2db8c0e-7635-4a58-b7fe-c50f70fe7e43.png)
1.0
Performance Summary riddles and puzzels - Hi Today I was reviewing the "PerfMon Summary" information (Version used 2.22.8) and I was wondering about following points below. It is worth to mention, all the data presented in all the issues are real world data and they are collected from productive systems refelecting real scenarios. Starting from top to bottom (in the image): - We got a column called "MB/Sec". I couldn't **clearly** decipher what this values is. Maybe a better column header would help. There are also no Tool Tips which explain the columns. This even could be implemnted as an additional feature: Tool Tip which explains the column headers in all the table grids, application wide. - Column "Read Latency" shows a painted cell as warning. I was seraching on the server itself for a quite while and I couldn't find anything which would help me to identify and localize the issue reported **on the spot**. I then searched even more and eventually I discovered a way to get some values and responsibles for the latency. This was hard and to some extent cryptical but maybe there are better ways. Anyhow, this was my approach: 1. Switch to single server. 2. Click on "Performance" Tab 3. On the "IO Perfromance" chart click "View Table Summary" (Which is yet another Modal Window I discovered :D) and then review the reported values which by the way they aren't painted at this stage so if one isn't paying attention he/she will then miss the point and leave. The above just tried to demonstrate how cumbersome is to identify the reported problem in detail. I hope you would come up with a easier / better way to solve this case. - Columns "Read Latecy" and "Write Latecy" are being painted as green. I really tried to understand **why**. I couldn't come up with any logic. - There are some cells in "Critical Wait" column which are gray colored. **Why**? Value 0.00 can't be gray AND green at the same time! - In "Lock Wait" and "Total Wait" columns, I find values way high in compare to the other similar values but these cells aren't painted at all. **Why**? Aren't we using any color indicators in theses columns? Shouldn't we do that? - There are some cells in the "Read Latecy" with no values at all. **Why?** Isn't there any values existing or are they so small that could be ignored? If so, then why can't we show 0,0? I am sure each one of the above questions can be answered one way or another but the MAIN point here is: The end user needs to **work** with these data and try not to decipher them. One way or another everyone (with enough persistence) can get along but that is also not the point of having the values collected. Application should do the most job for the user (because it is the tool, it is hard to keep track of these values and using visual effect one can transfer messages which helps the process.) and the user mostly needs to focus on identifying, anlasying of the collected data. This of course implies, collected data need to be precise and relyable on one hand and the application should try to avoid ambiguity on the other hand. ![perfmon_summary](https://user-images.githubusercontent.com/40255900/193550637-c2db8c0e-7635-4a58-b7fe-c50f70fe7e43.png)
non_priority
performance summary riddles and puzzels hi today i was reviewing the perfmon summary information version used and i was wondering about following points below it is worth to mention all the data presented in all the issues are real world data and they are collected from productive systems refelecting real scenarios starting from top to bottom in the image we got a column called mb sec i couldn t clearly decipher what this values is maybe a better column header would help there are also no tool tips which explain the columns this even could be implemnted as an additional feature tool tip which explains the column headers in all the table grids application wide column read latency shows a painted cell as warning i was seraching on the server itself for a quite while and i couldn t find anything which would help me to identify and localize the issue reported on the spot i then searched even more and eventually i discovered a way to get some values and responsibles for the latency this was hard and to some extent cryptical but maybe there are better ways anyhow this was my approach switch to single server click on performance tab on the io perfromance chart click view table summary which is yet another modal window i discovered d and then review the reported values which by the way they aren t painted at this stage so if one isn t paying attention he she will then miss the point and leave the above just tried to demonstrate how cumbersome is to identify the reported problem in detail i hope you would come up with a easier better way to solve this case columns read latecy and write latecy are being painted as green i really tried to understand why i couldn t come up with any logic there are some cells in critical wait column which are gray colored why value can t be gray and green at the same time in lock wait and total wait columns i find values way high in compare to the other similar values but these cells aren t painted at all why aren t we using any color indicators in theses columns shouldn t we do that there are some cells in the read latecy with no values at all why isn t there any values existing or are they so small that could be ignored if so then why can t we show i am sure each one of the above questions can be answered one way or another but the main point here is the end user needs to work with these data and try not to decipher them one way or another everyone with enough persistence can get along but that is also not the point of having the values collected application should do the most job for the user because it is the tool it is hard to keep track of these values and using visual effect one can transfer messages which helps the process and the user mostly needs to focus on identifying anlasying of the collected data this of course implies collected data need to be precise and relyable on one hand and the application should try to avoid ambiguity on the other hand
0
84,915
10,421,114,040
IssuesEvent
2019-09-16 04:32:57
YugaByte/yugabyte-db
https://api.github.com/repos/YugaByte/yugabyte-db
closed
[docs] Fix copy button copying for prompts
area/documentation kind/improve-ux
With the recent change of default username from `postgres` to `yugabyte` in commit 713233a, some docs pages have started to migrate towards using `yugabyte` in their prompts, and that is the intended plan as detailed in issue #2319. Unfortunately, this breaks the smart copying functionality that previously worked when `postgres` was used in the prompts. Previously, the copy button would only copy commands and not the prompt `postgres#= `. Now, the copy button simply copies it all. Restore the previous functionality.
1.0
[docs] Fix copy button copying for prompts - With the recent change of default username from `postgres` to `yugabyte` in commit 713233a, some docs pages have started to migrate towards using `yugabyte` in their prompts, and that is the intended plan as detailed in issue #2319. Unfortunately, this breaks the smart copying functionality that previously worked when `postgres` was used in the prompts. Previously, the copy button would only copy commands and not the prompt `postgres#= `. Now, the copy button simply copies it all. Restore the previous functionality.
non_priority
fix copy button copying for prompts with the recent change of default username from postgres to yugabyte in commit some docs pages have started to migrate towards using yugabyte in their prompts and that is the intended plan as detailed in issue unfortunately this breaks the smart copying functionality that previously worked when postgres was used in the prompts previously the copy button would only copy commands and not the prompt postgres now the copy button simply copies it all restore the previous functionality
0
76,280
9,932,663,810
IssuesEvent
2019-07-02 10:21:00
microsoft/vcpkg
https://api.github.com/repos/microsoft/vcpkg
closed
CmakeSettings example uses absolute path ?
documentation
I'm pretty new to C++ and Windows dev ... from what I've seen CMakeSettings.json should be in git. The [example here](https://github.com/Microsoft/vcpkg/blob/master/docs/examples/installing-and-using-packages.md#cmake-toolchain-file) adds a path in ```D:\\src\\vcpkg\\....\\vcpkg.cmake```. Is there any way this could be somewhere under the user profile ? I can't expect other users of my code to have the same drive mappings as I do, even if I do move it from ```D:\```. Or should I add a copy vcpkg.cmake somewhere in my own source repo ?
1.0
CmakeSettings example uses absolute path ? - I'm pretty new to C++ and Windows dev ... from what I've seen CMakeSettings.json should be in git. The [example here](https://github.com/Microsoft/vcpkg/blob/master/docs/examples/installing-and-using-packages.md#cmake-toolchain-file) adds a path in ```D:\\src\\vcpkg\\....\\vcpkg.cmake```. Is there any way this could be somewhere under the user profile ? I can't expect other users of my code to have the same drive mappings as I do, even if I do move it from ```D:\```. Or should I add a copy vcpkg.cmake somewhere in my own source repo ?
non_priority
cmakesettings example uses absolute path i m pretty new to c and windows dev from what i ve seen cmakesettings json should be in git the adds a path in d src vcpkg vcpkg cmake is there any way this could be somewhere under the user profile i can t expect other users of my code to have the same drive mappings as i do even if i do move it from d or should i add a copy vcpkg cmake somewhere in my own source repo
0
134,001
18,403,273,437
IssuesEvent
2021-10-12 18:50:05
susanstwss/WebGoat
https://api.github.com/repos/susanstwss/WebGoat
opened
CVE-2020-26259 (Medium) detected in xstream-1.4.5.jar
security vulnerability
## CVE-2020-26259 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.5.jar</b></p></summary> <p>XStream is a serialization library from Java objects to XML and back.</p> <p>Path to dependency file: WebGoat/webgoat-server/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar,m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar</p> <p> Dependency Hierarchy: - :x: **xstream-1.4.5.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/susanstwss/WebGoat/commit/2b2a58a6bee86bd0e98aa367052cda5f99597abb">2b2a58a6bee86bd0e98aa367052cda5f99597abb</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.15, is vulnerable to an Arbitrary File Deletion on the local host when unmarshalling. The vulnerability may allow a remote attacker to delete arbitrary know files on the host as log as the executing process has sufficient rights only by manipulating the processed input stream. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.15. The reported vulnerability does not exist running Java 15 or higher. No user is affected, who followed the recommendation to setup XStream's Security Framework with a whitelist! Anyone relying on XStream's default blacklist can immediately switch to a whilelist for the allowed types to avoid the vulnerability. Users of XStream 1.4.14 or below who still want to use XStream default blacklist can use a workaround described in more detailed in the referenced advisories. <p>Publish Date: 2020-12-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26259>CVE-2020-26259</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-jfvx-7wrx-43fh">https://github.com/x-stream/xstream/security/advisories/GHSA-jfvx-7wrx-43fh</a></p> <p>Release Date: 2020-12-16</p> <p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.15</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.thoughtworks.xstream","packageName":"xstream","packageVersion":"1.4.5","packageFilePaths":["/webgoat-server/pom.xml","/webgoat-lessons/vulnerable-components/pom.xml","/webgoat-integration-tests/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.thoughtworks.xstream:xstream:1.4.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.thoughtworks.xstream:xstream:1.4.15"}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2020-26259","vulnerabilityDetails":"XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.15, is vulnerable to an Arbitrary File Deletion on the local host when unmarshalling. The vulnerability may allow a remote attacker to delete arbitrary know files on the host as log as the executing process has sufficient rights only by manipulating the processed input stream. If you rely on XStream\u0027s default blacklist of the Security Framework, you will have to use at least version 1.4.15. The reported vulnerability does not exist running Java 15 or higher. No user is affected, who followed the recommendation to setup XStream\u0027s Security Framework with a whitelist! Anyone relying on XStream\u0027s default blacklist can immediately switch to a whilelist for the allowed types to avoid the vulnerability. Users of XStream 1.4.14 or below who still want to use XStream default blacklist can use a workaround described in more detailed in the referenced advisories.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26259","cvss3Severity":"medium","cvss3Score":"6.8","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Changed","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-26259 (Medium) detected in xstream-1.4.5.jar - ## CVE-2020-26259 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.5.jar</b></p></summary> <p>XStream is a serialization library from Java objects to XML and back.</p> <p>Path to dependency file: WebGoat/webgoat-server/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar,m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar</p> <p> Dependency Hierarchy: - :x: **xstream-1.4.5.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/susanstwss/WebGoat/commit/2b2a58a6bee86bd0e98aa367052cda5f99597abb">2b2a58a6bee86bd0e98aa367052cda5f99597abb</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.15, is vulnerable to an Arbitrary File Deletion on the local host when unmarshalling. The vulnerability may allow a remote attacker to delete arbitrary know files on the host as log as the executing process has sufficient rights only by manipulating the processed input stream. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.15. The reported vulnerability does not exist running Java 15 or higher. No user is affected, who followed the recommendation to setup XStream's Security Framework with a whitelist! Anyone relying on XStream's default blacklist can immediately switch to a whilelist for the allowed types to avoid the vulnerability. Users of XStream 1.4.14 or below who still want to use XStream default blacklist can use a workaround described in more detailed in the referenced advisories. <p>Publish Date: 2020-12-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26259>CVE-2020-26259</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-jfvx-7wrx-43fh">https://github.com/x-stream/xstream/security/advisories/GHSA-jfvx-7wrx-43fh</a></p> <p>Release Date: 2020-12-16</p> <p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.15</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.thoughtworks.xstream","packageName":"xstream","packageVersion":"1.4.5","packageFilePaths":["/webgoat-server/pom.xml","/webgoat-lessons/vulnerable-components/pom.xml","/webgoat-integration-tests/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.thoughtworks.xstream:xstream:1.4.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.thoughtworks.xstream:xstream:1.4.15"}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2020-26259","vulnerabilityDetails":"XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.15, is vulnerable to an Arbitrary File Deletion on the local host when unmarshalling. The vulnerability may allow a remote attacker to delete arbitrary know files on the host as log as the executing process has sufficient rights only by manipulating the processed input stream. If you rely on XStream\u0027s default blacklist of the Security Framework, you will have to use at least version 1.4.15. The reported vulnerability does not exist running Java 15 or higher. No user is affected, who followed the recommendation to setup XStream\u0027s Security Framework with a whitelist! Anyone relying on XStream\u0027s default blacklist can immediately switch to a whilelist for the allowed types to avoid the vulnerability. Users of XStream 1.4.14 or below who still want to use XStream default blacklist can use a workaround described in more detailed in the referenced advisories.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26259","cvss3Severity":"medium","cvss3Score":"6.8","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Changed","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_priority
cve medium detected in xstream jar cve medium severity vulnerability vulnerable library xstream jar xstream is a serialization library from java objects to xml and back path to dependency file webgoat webgoat server pom xml path to vulnerable library home wss scanner repository com thoughtworks xstream xstream xstream jar repository com thoughtworks xstream xstream xstream jar home wss scanner repository com thoughtworks xstream xstream xstream jar dependency hierarchy x xstream jar vulnerable library found in head commit a href found in base branch develop vulnerability details xstream is a java library to serialize objects to xml and back again in xstream before version is vulnerable to an arbitrary file deletion on the local host when unmarshalling the vulnerability may allow a remote attacker to delete arbitrary know files on the host as log as the executing process has sufficient rights only by manipulating the processed input stream if you rely on xstream s default blacklist of the security framework you will have to use at least version the reported vulnerability does not exist running java or higher no user is affected who followed the recommendation to setup xstream s security framework with a whitelist anyone relying on xstream s default blacklist can immediately switch to a whilelist for the allowed types to avoid the vulnerability users of xstream or below who still want to use xstream default blacklist can use a workaround described in more detailed in the referenced advisories publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope changed impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com thoughtworks xstream xstream check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com thoughtworks xstream xstream isminimumfixversionavailable true minimumfixversion com thoughtworks xstream xstream basebranches vulnerabilityidentifier cve vulnerabilitydetails xstream is a java library to serialize objects to xml and back again in xstream before version is vulnerable to an arbitrary file deletion on the local host when unmarshalling the vulnerability may allow a remote attacker to delete arbitrary know files on the host as log as the executing process has sufficient rights only by manipulating the processed input stream if you rely on xstream default blacklist of the security framework you will have to use at least version the reported vulnerability does not exist running java or higher no user is affected who followed the recommendation to setup xstream security framework with a whitelist anyone relying on xstream default blacklist can immediately switch to a whilelist for the allowed types to avoid the vulnerability users of xstream or below who still want to use xstream default blacklist can use a workaround described in more detailed in the referenced advisories vulnerabilityurl
0
145,644
19,339,776,022
IssuesEvent
2021-12-15 02:09:53
madhans23/linux-4.15
https://api.github.com/repos/madhans23/linux-4.15
opened
CVE-2021-43975 (Medium) detected in linux-stagingv5.15
security vulnerability
## CVE-2021-43975 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stagingv5.15</b></p></summary> <p> <p>hwmon staging tree</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/groeck/linux-staging.git>https://git.kernel.org/pub/scm/linux/kernel/git/groeck/linux-staging.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_utils.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In the Linux kernel through 5.15.2, hw_atl_utils_fw_rpc_wait in drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_utils.c allows an attacker (who can introduce a crafted device) to trigger an out-of-bounds write via a crafted length value. <p>Publish Date: 2021-11-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43975>CVE-2021-43975</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-43975">https://www.linuxkernelcves.com/cves/CVE-2021-43975</a></p> <p>Release Date: 2021-11-17</p> <p>Fix Resolution: v5.16-rc2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-43975 (Medium) detected in linux-stagingv5.15 - ## CVE-2021-43975 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stagingv5.15</b></p></summary> <p> <p>hwmon staging tree</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/groeck/linux-staging.git>https://git.kernel.org/pub/scm/linux/kernel/git/groeck/linux-staging.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_utils.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In the Linux kernel through 5.15.2, hw_atl_utils_fw_rpc_wait in drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_utils.c allows an attacker (who can introduce a crafted device) to trigger an out-of-bounds write via a crafted length value. <p>Publish Date: 2021-11-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43975>CVE-2021-43975</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-43975">https://www.linuxkernelcves.com/cves/CVE-2021-43975</a></p> <p>Release Date: 2021-11-17</p> <p>Fix Resolution: v5.16-rc2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in linux cve medium severity vulnerability vulnerable library linux hwmon staging tree library home page a href found in base branch master vulnerable source files drivers net ethernet aquantia atlantic hw atl hw atl utils c vulnerability details in the linux kernel through hw atl utils fw rpc wait in drivers net ethernet aquantia atlantic hw atl hw atl utils c allows an attacker who can introduce a crafted device to trigger an out of bounds write via a crafted length value publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
13,237
8,380,412,972
IssuesEvent
2018-10-07 14:31:59
godotengine/godot
https://api.github.com/repos/godotengine/godot
closed
New property inspector: Scripts don't show filename
enhancement topic:editor usability
**Godot version:** ba134d4 **Issue description:** This is more of a personal nitpick. While I find the script thumb preview *marginally* useful, seeing a filename at a glance would be much more useful, or in a tooltip if must be. I'm assuming this is simply a feature in testing (for supporting property field previews which weren't feasible before?), but nevertheless it seems to be an issue.
True
New property inspector: Scripts don't show filename - **Godot version:** ba134d4 **Issue description:** This is more of a personal nitpick. While I find the script thumb preview *marginally* useful, seeing a filename at a glance would be much more useful, or in a tooltip if must be. I'm assuming this is simply a feature in testing (for supporting property field previews which weren't feasible before?), but nevertheless it seems to be an issue.
non_priority
new property inspector scripts don t show filename godot version issue description this is more of a personal nitpick while i find the script thumb preview marginally useful seeing a filename at a glance would be much more useful or in a tooltip if must be i m assuming this is simply a feature in testing for supporting property field previews which weren t feasible before but nevertheless it seems to be an issue
0
70,401
30,660,454,586
IssuesEvent
2023-07-25 14:41:59
hashicorp/terraform-provider-azurerm
https://api.github.com/repos/hashicorp/terraform-provider-azurerm
closed
azurerm_application_gateway Health Prode - Body should not be required field
bug waiting-response service/application-gateway v/3.x
### Is there an existing issue for this? - [X] I have searched the existing issues ### Community Note <!--- Please keep this note for the community ---> * Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform Version 1.3.7 ### AzureRM Provider Version 3.42.0 ### Affected Resource(s)/Data Source(s) azurerm_application_gateway ### Terraform Configuration Files ```hcl resource "azurerm_application_gateway" "network" { name = "example-appgateway" resource_group_name = azurerm_resource_group.example.name location = azurerm_resource_group.example.location sku { name = "Standard_Small" tier = "Standard" capacity = 2 } gateway_ip_configuration { name = "my-gateway-ip-configuration" subnet_id = azurerm_subnet.frontend.id } frontend_port { name = local.frontend_port_name port = 80 } frontend_ip_configuration { name = local.frontend_ip_configuration_name public_ip_address_id = azurerm_public_ip.example.id } backend_address_pool { name = local.backend_address_pool_name } backend_http_settings { name = local.http_setting_name cookie_based_affinity = "Disabled" path = "/path1/" port = 80 protocol = "Http" request_timeout = 60 } http_listener { name = local.listener_name frontend_ip_configuration_name = local.frontend_ip_configuration_name frontend_port_name = local.frontend_port_name protocol = "Http" } request_routing_rule { name = local.request_routing_rule_name rule_type = "Basic" http_listener_name = local.listener_name backend_address_pool_name = local.backend_address_pool_name backend_http_settings_name = local.http_setting_name } probe { name = "health-probe-default-https" pick_host_name_from_backend_http_settings = true path = "/" protocol = "Https" port = 443 interval = 30 timeout = 15 unhealthy_threshold = 3 match { body = "" status_code = ["200-399"] } } ``` ### Debug Output/Panic Output ```shell Error: Missing required argument │ │ on app-gw-prod\main.tf line 203, in resource "azurerm_application_gateway" "external-prod-centralus": │ 203: match { │ │ The argument "body" is required, but no definition was found. ``` ### Expected Behaviour Body should not be a required field. Only status_code should be required. Can workaround by adding _body = ""_ ### Actual Behaviour Failed TF build ### Steps to Reproduce terraform apply ### Important Factoids _No response_ ### References _No response_
1.0
azurerm_application_gateway Health Prode - Body should not be required field - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Community Note <!--- Please keep this note for the community ---> * Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform Version 1.3.7 ### AzureRM Provider Version 3.42.0 ### Affected Resource(s)/Data Source(s) azurerm_application_gateway ### Terraform Configuration Files ```hcl resource "azurerm_application_gateway" "network" { name = "example-appgateway" resource_group_name = azurerm_resource_group.example.name location = azurerm_resource_group.example.location sku { name = "Standard_Small" tier = "Standard" capacity = 2 } gateway_ip_configuration { name = "my-gateway-ip-configuration" subnet_id = azurerm_subnet.frontend.id } frontend_port { name = local.frontend_port_name port = 80 } frontend_ip_configuration { name = local.frontend_ip_configuration_name public_ip_address_id = azurerm_public_ip.example.id } backend_address_pool { name = local.backend_address_pool_name } backend_http_settings { name = local.http_setting_name cookie_based_affinity = "Disabled" path = "/path1/" port = 80 protocol = "Http" request_timeout = 60 } http_listener { name = local.listener_name frontend_ip_configuration_name = local.frontend_ip_configuration_name frontend_port_name = local.frontend_port_name protocol = "Http" } request_routing_rule { name = local.request_routing_rule_name rule_type = "Basic" http_listener_name = local.listener_name backend_address_pool_name = local.backend_address_pool_name backend_http_settings_name = local.http_setting_name } probe { name = "health-probe-default-https" pick_host_name_from_backend_http_settings = true path = "/" protocol = "Https" port = 443 interval = 30 timeout = 15 unhealthy_threshold = 3 match { body = "" status_code = ["200-399"] } } ``` ### Debug Output/Panic Output ```shell Error: Missing required argument │ │ on app-gw-prod\main.tf line 203, in resource "azurerm_application_gateway" "external-prod-centralus": │ 203: match { │ │ The argument "body" is required, but no definition was found. ``` ### Expected Behaviour Body should not be a required field. Only status_code should be required. Can workaround by adding _body = ""_ ### Actual Behaviour Failed TF build ### Steps to Reproduce terraform apply ### Important Factoids _No response_ ### References _No response_
non_priority
azurerm application gateway health prode body should not be required field is there an existing issue for this i have searched the existing issues community note please vote on this issue by adding a thumbsup to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform version azurerm provider version affected resource s data source s azurerm application gateway terraform configuration files hcl resource azurerm application gateway network name example appgateway resource group name azurerm resource group example name location azurerm resource group example location sku name standard small tier standard capacity gateway ip configuration name my gateway ip configuration subnet id azurerm subnet frontend id frontend port name local frontend port name port frontend ip configuration name local frontend ip configuration name public ip address id azurerm public ip example id backend address pool name local backend address pool name backend http settings name local http setting name cookie based affinity disabled path port protocol http request timeout http listener name local listener name frontend ip configuration name local frontend ip configuration name frontend port name local frontend port name protocol http request routing rule name local request routing rule name rule type basic http listener name local listener name backend address pool name local backend address pool name backend http settings name local http setting name probe name health probe default https pick host name from backend http settings true path protocol https port interval timeout unhealthy threshold match body status code debug output panic output shell error missing required argument │ │ on app gw prod main tf line in resource azurerm application gateway external prod centralus │ match │ │ the argument body is required but no definition was found expected behaviour body should not be a required field only status code should be required can workaround by adding body actual behaviour failed tf build steps to reproduce terraform apply important factoids no response references no response
0
267,835
23,323,111,440
IssuesEvent
2022-08-08 18:20:35
mozilla-mobile/mobile-test-eng
https://api.github.com/repos/mozilla-mobile/mobile-test-eng
opened
Create experimental Fenix cron for FTL improved emulator beta
waiting android infra:mobile infra:ui-test
This depends on if we're accepted into their beta (waiting label). Context: FTL has a beta sign-up for new emulators with improved speeds and they're looking for early adopters. Scope: In the same fashion of as `legacy-x86` cron, we can create a similar cron on schedule with a new Flank configuration targeting the new emulator name. It can run the full-suite.
1.0
Create experimental Fenix cron for FTL improved emulator beta - This depends on if we're accepted into their beta (waiting label). Context: FTL has a beta sign-up for new emulators with improved speeds and they're looking for early adopters. Scope: In the same fashion of as `legacy-x86` cron, we can create a similar cron on schedule with a new Flank configuration targeting the new emulator name. It can run the full-suite.
non_priority
create experimental fenix cron for ftl improved emulator beta this depends on if we re accepted into their beta waiting label context ftl has a beta sign up for new emulators with improved speeds and they re looking for early adopters scope in the same fashion of as legacy cron we can create a similar cron on schedule with a new flank configuration targeting the new emulator name it can run the full suite
0
96,737
16,164,449,356
IssuesEvent
2021-05-01 07:50:34
AlexRogalskiy/wotd
https://api.github.com/repos/AlexRogalskiy/wotd
opened
CVE-2020-11022 (Medium) detected in jquery-1.9.1.js, jquery-1.8.1.min.js
security vulnerability
## CVE-2020-11022 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.9.1.js</b>, <b>jquery-1.8.1.min.js</b></p></summary> <p> <details><summary><b>jquery-1.9.1.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.js</a></p> <p>Path to dependency file: wotd/node_modules/tinygradient/bower_components/tinycolor/index.html</p> <p>Path to vulnerable library: wotd/node_modules/tinygradient/bower_components/tinycolor/demo/jquery-1.9.1.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.9.1.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: wotd/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: wotd/node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/wotd/commit/f6150f55f68ef1a1f68c2ccaa5f94d7cc6a7c5fb">f6150f55f68ef1a1f68c2ccaa5f94d7cc6a7c5fb</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-11022 (Medium) detected in jquery-1.9.1.js, jquery-1.8.1.min.js - ## CVE-2020-11022 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.9.1.js</b>, <b>jquery-1.8.1.min.js</b></p></summary> <p> <details><summary><b>jquery-1.9.1.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.js</a></p> <p>Path to dependency file: wotd/node_modules/tinygradient/bower_components/tinycolor/index.html</p> <p>Path to vulnerable library: wotd/node_modules/tinygradient/bower_components/tinycolor/demo/jquery-1.9.1.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.9.1.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: wotd/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: wotd/node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/wotd/commit/f6150f55f68ef1a1f68c2ccaa5f94d7cc6a7c5fb">f6150f55f68ef1a1f68c2ccaa5f94d7cc6a7c5fb</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in jquery js jquery min js cve medium severity vulnerability vulnerable libraries jquery js jquery min js jquery js javascript library for dom operations library home page a href path to dependency file wotd node modules tinygradient bower components tinycolor index html path to vulnerable library wotd node modules tinygradient bower components tinycolor demo jquery js dependency hierarchy x jquery js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file wotd node modules redeyed examples browser index html path to vulnerable library wotd node modules redeyed examples browser index html dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
0
126,856
26,929,163,850
IssuesEvent
2023-02-07 15:41:42
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
PHPMail is rejected by the mail server with Error when using PHP 8.X
No Code Attached Yet Information Required
### PHPMail is rejected by the mail server with an "overlength date field" error when using PHP 8.X. Using PHP 7.4 does not produce the error Steps to reproduce the issue Set mail as PHPMail in Global Configuration and set PHP to 8.X goto "Forgot Logon" and try to get a new password. Email will be rejected ### Expected result ### Actual result ### System information (as much as possible) ### Additional comments
1.0
PHPMail is rejected by the mail server with Error when using PHP 8.X - ### PHPMail is rejected by the mail server with an "overlength date field" error when using PHP 8.X. Using PHP 7.4 does not produce the error Steps to reproduce the issue Set mail as PHPMail in Global Configuration and set PHP to 8.X goto "Forgot Logon" and try to get a new password. Email will be rejected ### Expected result ### Actual result ### System information (as much as possible) ### Additional comments
non_priority
phpmail is rejected by the mail server with error when using php x phpmail is rejected by the mail server with an overlength date field error when using php x using php does not produce the error steps to reproduce the issue set mail as phpmail in global configuration and set php to x goto forgot logon and try to get a new password email will be rejected expected result actual result system information as much as possible additional comments
0
129,011
10,559,756,337
IssuesEvent
2019-10-04 12:24:42
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
PartitionMigrationListenerTest.testMigrationStats_afterPartitionsLost_when_NO_MIGRATION
Module: Partitioning Source: Internal Team: Core Type: Test-Failure
http://jenkins.hazelcast.com/job/Hazelcast-pr-builder/3601/ ``` Expected: a value equal to or greater than <4> but: <3> was less than <4> Stacktrace java.lang.AssertionError: Expected: a value equal to or greater than <4> but: <3> was less than <4> at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) at org.junit.Assert.assertThat(Assert.java:956) at org.junit.Assert.assertThat(Assert.java:923) at com.hazelcast.partition.PartitionMigrationListenerTest$EventCollectingMigrationListener.lambda$awaitEventPacksComplete$0(PartitionMigrationListenerTest.java:445) at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:1308) at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:1410) at com.hazelcast.partition.PartitionMigrationListenerTest$EventCollectingMigrationListener.awaitEventPacksComplete(PartitionMigrationListenerTest.java:444) at com.hazelcast.partition.PartitionMigrationListenerTest$EventCollectingMigrationListener.ensureAndGetEventPacks(PartitionMigrationListenerTest.java:435) at com.hazelcast.partition.PartitionMigrationListenerTest.testMigrationStats_afterPartitionsLost_when_NO_MIGRATION(PartitionMigrationListenerTest.java:178) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:114) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:106) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.lang.Thread.run(Thread.java:748) ```
1.0
PartitionMigrationListenerTest.testMigrationStats_afterPartitionsLost_when_NO_MIGRATION - http://jenkins.hazelcast.com/job/Hazelcast-pr-builder/3601/ ``` Expected: a value equal to or greater than <4> but: <3> was less than <4> Stacktrace java.lang.AssertionError: Expected: a value equal to or greater than <4> but: <3> was less than <4> at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) at org.junit.Assert.assertThat(Assert.java:956) at org.junit.Assert.assertThat(Assert.java:923) at com.hazelcast.partition.PartitionMigrationListenerTest$EventCollectingMigrationListener.lambda$awaitEventPacksComplete$0(PartitionMigrationListenerTest.java:445) at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:1308) at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:1410) at com.hazelcast.partition.PartitionMigrationListenerTest$EventCollectingMigrationListener.awaitEventPacksComplete(PartitionMigrationListenerTest.java:444) at com.hazelcast.partition.PartitionMigrationListenerTest$EventCollectingMigrationListener.ensureAndGetEventPacks(PartitionMigrationListenerTest.java:435) at com.hazelcast.partition.PartitionMigrationListenerTest.testMigrationStats_afterPartitionsLost_when_NO_MIGRATION(PartitionMigrationListenerTest.java:178) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:114) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:106) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.lang.Thread.run(Thread.java:748) ```
non_priority
partitionmigrationlistenertest testmigrationstats afterpartitionslost when no migration expected a value equal to or greater than but was less than stacktrace java lang assertionerror expected a value equal to or greater than but was less than at org hamcrest matcherassert assertthat matcherassert java at org junit assert assertthat assert java at org junit assert assertthat assert java at com hazelcast partition partitionmigrationlistenertest eventcollectingmigrationlistener lambda awaiteventpackscomplete partitionmigrationlistenertest java at com hazelcast test hazelcasttestsupport asserttrueeventually hazelcasttestsupport java at com hazelcast test hazelcasttestsupport asserttrueeventually hazelcasttestsupport java at com hazelcast partition partitionmigrationlistenertest eventcollectingmigrationlistener awaiteventpackscomplete partitionmigrationlistenertest java at com hazelcast partition partitionmigrationlistenertest eventcollectingmigrationlistener ensureandgeteventpacks partitionmigrationlistenertest java at com hazelcast partition partitionmigrationlistenertest testmigrationstats afterpartitionslost when no migration partitionmigrationlistenertest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at java util concurrent futuretask run futuretask java at java lang thread run thread java
0
92,452
18,869,526,185
IssuesEvent
2021-11-13 00:37:03
nopSolutions/nopCommerce
https://api.github.com/repos/nopSolutions/nopCommerce
closed
Add javascript events
refactoring / source code
We started working on Google Analytics Enhanced ecommerce plugin for nopCommerce. Unfortunately nopCommerce is lacking any javascript events for any user actions. For example there is no way to know when a product has been added to the cart. It would be nice if you raise javascript events like "productAddedToCart" etc. for the actions that happen via Ajax. For example when you add a product via Ajax it would be nice to have something like this: $.event.trigger({ type: "nopAjaxCartProductAddedToCartEvent", productId: productId, quantity: addToCartProductQuantity }); This way we can subscribe to such events and submit the data to Google Anaytics for example. Unfortunately the Remove from Cart does not use Ajax, so there is no way this to be done for now. Source: http://www.nopcommerce.com/boards/t/38391/suggestions-for-the-upcoming-version-370.aspx?p=2#156114 also see http://www.nopcommerce.com/boards/t/43297/improvements-before-380-release.aspx#171728 also see http://www.nopcommerce.com/boards/t/39325/add-javascript-events-in-the-payment-flow.aspx P.S. We already have such event defined in \Views\Product_ProductAttributes.cshtml
1.0
Add javascript events - We started working on Google Analytics Enhanced ecommerce plugin for nopCommerce. Unfortunately nopCommerce is lacking any javascript events for any user actions. For example there is no way to know when a product has been added to the cart. It would be nice if you raise javascript events like "productAddedToCart" etc. for the actions that happen via Ajax. For example when you add a product via Ajax it would be nice to have something like this: $.event.trigger({ type: "nopAjaxCartProductAddedToCartEvent", productId: productId, quantity: addToCartProductQuantity }); This way we can subscribe to such events and submit the data to Google Anaytics for example. Unfortunately the Remove from Cart does not use Ajax, so there is no way this to be done for now. Source: http://www.nopcommerce.com/boards/t/38391/suggestions-for-the-upcoming-version-370.aspx?p=2#156114 also see http://www.nopcommerce.com/boards/t/43297/improvements-before-380-release.aspx#171728 also see http://www.nopcommerce.com/boards/t/39325/add-javascript-events-in-the-payment-flow.aspx P.S. We already have such event defined in \Views\Product_ProductAttributes.cshtml
non_priority
add javascript events we started working on google analytics enhanced ecommerce plugin for nopcommerce unfortunately nopcommerce is lacking any javascript events for any user actions for example there is no way to know when a product has been added to the cart it would be nice if you raise javascript events like productaddedtocart etc for the actions that happen via ajax for example when you add a product via ajax it would be nice to have something like this event trigger type nopajaxcartproductaddedtocartevent productid productid quantity addtocartproductquantity this way we can subscribe to such events and submit the data to google anaytics for example unfortunately the remove from cart does not use ajax so there is no way this to be done for now source also see also see p s we already have such event defined in views product productattributes cshtml
0
230,194
17,605,605,552
IssuesEvent
2021-08-17 16:39:37
microsoft/vscode-docker
https://api.github.com/repos/microsoft/vscode-docker
closed
SSH tunnel setup doesn't work in PowerShell
documentation
In ["Connect to remote Docker over SSH/Set up SSH Tunneling"](https://code.visualstudio.com/docs/containers/ssh#_set-up-ssh-tunneling) it is advised to run `sc config ssh-agent start=auto`, but it only works in old `cmd` terminal. PowerShell will complain: ``` > sc config ssh-agent start=auto Set-Content : A positional parameter cannot be found that accepts argument 'start=auto'. At line:1 char:1 ``` Please update the documentation with correct steps: ``` Set-Service ssh-agent -StartupType "Automatic" Start-Service ssh-agent ```
1.0
SSH tunnel setup doesn't work in PowerShell - In ["Connect to remote Docker over SSH/Set up SSH Tunneling"](https://code.visualstudio.com/docs/containers/ssh#_set-up-ssh-tunneling) it is advised to run `sc config ssh-agent start=auto`, but it only works in old `cmd` terminal. PowerShell will complain: ``` > sc config ssh-agent start=auto Set-Content : A positional parameter cannot be found that accepts argument 'start=auto'. At line:1 char:1 ``` Please update the documentation with correct steps: ``` Set-Service ssh-agent -StartupType "Automatic" Start-Service ssh-agent ```
non_priority
ssh tunnel setup doesn t work in powershell in it is advised to run sc config ssh agent start auto but it only works in old cmd terminal powershell will complain sc config ssh agent start auto set content a positional parameter cannot be found that accepts argument start auto at line char please update the documentation with correct steps set service ssh agent startuptype automatic start service ssh agent
0
53,529
13,261,827,449
IssuesEvent
2020-08-20 20:36:16
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
CVMFS - doc builds are choking on the lack of `napoleon` (Trac #1608)
Migrated from Trac defect infrastructure
david- could you add napoleon, or update sphinx? <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1608">https://code.icecube.wisc.edu/projects/icecube/ticket/1608</a>, reported by negaand owned by david.schultz</em></summary> <p> ```json { "status": "closed", "changetime": "2016-03-29T19:54:11", "_ts": "1459281251636355", "description": "david- could you add napoleon, or update sphinx?", "reporter": "nega", "cc": "", "resolution": "fixed", "time": "2016-03-28T21:20:33", "component": "infrastructure", "summary": "CVMFS - doc builds are choking on the lack of `napoleon`", "priority": "normal", "keywords": "cvmfs sphinx documentation", "milestone": "", "owner": "david.schultz", "type": "defect" } ``` </p> </details>
1.0
CVMFS - doc builds are choking on the lack of `napoleon` (Trac #1608) - david- could you add napoleon, or update sphinx? <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1608">https://code.icecube.wisc.edu/projects/icecube/ticket/1608</a>, reported by negaand owned by david.schultz</em></summary> <p> ```json { "status": "closed", "changetime": "2016-03-29T19:54:11", "_ts": "1459281251636355", "description": "david- could you add napoleon, or update sphinx?", "reporter": "nega", "cc": "", "resolution": "fixed", "time": "2016-03-28T21:20:33", "component": "infrastructure", "summary": "CVMFS - doc builds are choking on the lack of `napoleon`", "priority": "normal", "keywords": "cvmfs sphinx documentation", "milestone": "", "owner": "david.schultz", "type": "defect" } ``` </p> </details>
non_priority
cvmfs doc builds are choking on the lack of napoleon trac david could you add napoleon or update sphinx migrated from json status closed changetime ts description david could you add napoleon or update sphinx reporter nega cc resolution fixed time component infrastructure summary cvmfs doc builds are choking on the lack of napoleon priority normal keywords cvmfs sphinx documentation milestone owner david schultz type defect
0
10,472
4,074,264,334
IssuesEvent
2016-05-28 09:51:53
SleepyTrousers/EnderIO
https://api.github.com/repos/SleepyTrousers/EnderIO
closed
TODO: Rename changed GUI textures before release
Code Complete Will Implement
otherwise we'll get many reports like this: ![](https://cloud.githubusercontent.com/assets/8468011/11293615/7eec1bc0-8f29-11e5-8450-b67f3c155a59.png)
1.0
TODO: Rename changed GUI textures before release - otherwise we'll get many reports like this: ![](https://cloud.githubusercontent.com/assets/8468011/11293615/7eec1bc0-8f29-11e5-8450-b67f3c155a59.png)
non_priority
todo rename changed gui textures before release otherwise we ll get many reports like this
0
2,142
4,460,926,581
IssuesEvent
2016-08-24 02:11:11
diamondinrough/app
https://api.github.com/repos/diamondinrough/app
closed
Create basic hybrid app package
dev requirement
1. create the basic code base for client side app, upload it to github 2. connect to the server side to make sure the client side and server side integration works. Please document and share out the related staffs.
1.0
Create basic hybrid app package - 1. create the basic code base for client side app, upload it to github 2. connect to the server side to make sure the client side and server side integration works. Please document and share out the related staffs.
non_priority
create basic hybrid app package create the basic code base for client side app upload it to github connect to the server side to make sure the client side and server side integration works please document and share out the related staffs
0
101,748
16,528,033,422
IssuesEvent
2021-05-26 23:32:10
alpersonalwebsite/cards
https://api.github.com/repos/alpersonalwebsite/cards
opened
CVE-2020-28500 (Medium) detected in lodash-4.17.11.tgz
security vulnerability
## CVE-2020-28500 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.11.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p> <p>Path to dependency file: cards/package.json</p> <p>Path to vulnerable library: cards/node_modules/lodash</p> <p> Dependency Hierarchy: - react-native-0.62.0.tgz (Root Library) - cli-4.14.0.tgz - inquirer-3.3.0.tgz - :x: **lodash-4.17.11.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/alpersonalwebsite/cards/commit/959053f33fb963a3ec04cdb4be3c6f705f0312c6">959053f33fb963a3ec04cdb4be3c6f705f0312c6</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500>CVE-2020-28500</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution: lodash-4.17.21</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-28500 (Medium) detected in lodash-4.17.11.tgz - ## CVE-2020-28500 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.11.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p> <p>Path to dependency file: cards/package.json</p> <p>Path to vulnerable library: cards/node_modules/lodash</p> <p> Dependency Hierarchy: - react-native-0.62.0.tgz (Root Library) - cli-4.14.0.tgz - inquirer-3.3.0.tgz - :x: **lodash-4.17.11.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/alpersonalwebsite/cards/commit/959053f33fb963a3ec04cdb4be3c6f705f0312c6">959053f33fb963a3ec04cdb4be3c6f705f0312c6</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500>CVE-2020-28500</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution: lodash-4.17.21</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in lodash tgz cve medium severity vulnerability vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file cards package json path to vulnerable library cards node modules lodash dependency hierarchy react native tgz root library cli tgz inquirer tgz x lodash tgz vulnerable library found in head commit a href vulnerability details lodash versions prior to are vulnerable to regular expression denial of service redos via the tonumber trim and trimend functions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash step up your open source security game with whitesource
0
32,688
4,786,530,536
IssuesEvent
2016-10-29 13:42:13
infiniteautomation/ma-core-public
https://api.github.com/repos/infiniteautomation/ma-core-public
closed
Watchlist - Chart legend only displays names
Bug Ready for Testing
It was observed [in the forum](http://infiniteautomation.com/forum/topic/2441/watchlist-chart-only-shows-names-in-legend-without-devicenames) that somewhere in 2.5 or after the watchlist chart's legend ceased displaying the device name of the point in the legend, and only displays the name.
1.0
Watchlist - Chart legend only displays names - It was observed [in the forum](http://infiniteautomation.com/forum/topic/2441/watchlist-chart-only-shows-names-in-legend-without-devicenames) that somewhere in 2.5 or after the watchlist chart's legend ceased displaying the device name of the point in the legend, and only displays the name.
non_priority
watchlist chart legend only displays names it was observed that somewhere in or after the watchlist chart s legend ceased displaying the device name of the point in the legend and only displays the name
0
324,673
27,814,729,344
IssuesEvent
2023-03-18 14:48:51
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
kv/kvserver: TestReplicateQueueExpirationLeasesOnly failed
C-test-failure O-robot branch-master T-kv-replication
kv/kvserver.TestReplicateQueueExpirationLeasesOnly [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/9122901?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/9122901?buildTab=artifacts#/) on master @ [0a72a49eb6e4ce5823b53d02125441c3da2e94f5](https://github.com/cockroachdb/cockroach/commits/0a72a49eb6e4ce5823b53d02125441c3da2e94f5): ``` === RUN TestReplicateQueueExpirationLeasesOnly test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/33e1d369c27b9c01b2b6009c561815a3/logTestReplicateQueueExpirationLeasesOnly2320733262 test_log_scope.go:79: use -show-logs to present logs inline replicate_queue_test.go:2406: initial: epochLeases=316 expLeases=2 replicate_queue_test.go:2411: Error Trace: github.com/cockroachdb/cockroach/pkg/kv/kvserver_test/pkg/kv/kvserver/replicate_queue_test.go:2411 Error: Condition never satisfied Test: TestReplicateQueueExpirationLeasesOnly panic.go:522: -- test log scope end -- --- FAIL: TestReplicateQueueExpirationLeasesOnly (319.95s) ``` <p>Parameters: <code>TAGS=bazel,gss,deadlock</code> </p> <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/kv <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestReplicateQueueExpirationLeasesOnly.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-25608
1.0
kv/kvserver: TestReplicateQueueExpirationLeasesOnly failed - kv/kvserver.TestReplicateQueueExpirationLeasesOnly [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/9122901?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/9122901?buildTab=artifacts#/) on master @ [0a72a49eb6e4ce5823b53d02125441c3da2e94f5](https://github.com/cockroachdb/cockroach/commits/0a72a49eb6e4ce5823b53d02125441c3da2e94f5): ``` === RUN TestReplicateQueueExpirationLeasesOnly test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/33e1d369c27b9c01b2b6009c561815a3/logTestReplicateQueueExpirationLeasesOnly2320733262 test_log_scope.go:79: use -show-logs to present logs inline replicate_queue_test.go:2406: initial: epochLeases=316 expLeases=2 replicate_queue_test.go:2411: Error Trace: github.com/cockroachdb/cockroach/pkg/kv/kvserver_test/pkg/kv/kvserver/replicate_queue_test.go:2411 Error: Condition never satisfied Test: TestReplicateQueueExpirationLeasesOnly panic.go:522: -- test log scope end -- --- FAIL: TestReplicateQueueExpirationLeasesOnly (319.95s) ``` <p>Parameters: <code>TAGS=bazel,gss,deadlock</code> </p> <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/kv <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestReplicateQueueExpirationLeasesOnly.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-25608
non_priority
kv kvserver testreplicatequeueexpirationleasesonly failed kv kvserver testreplicatequeueexpirationleasesonly with on master run testreplicatequeueexpirationleasesonly test log scope go test logs captured to artifacts tmp tmp test log scope go use show logs to present logs inline replicate queue test go initial epochleases expleases replicate queue test go error trace github com cockroachdb cockroach pkg kv kvserver test pkg kv kvserver replicate queue test go error condition never satisfied test testreplicatequeueexpirationleasesonly panic go test log scope end fail testreplicatequeueexpirationleasesonly parameters tags bazel gss deadlock help see also cc cockroachdb kv jira issue crdb
0
62,301
7,569,918,375
IssuesEvent
2018-04-23 07:12:14
akvo/akvo-web
https://api.github.com/repos/akvo/akvo-web
closed
All RSR Project Updates feed - images not optimised
1 - Content 1 - Design 2 - Maintenance
I see the RSR update feed continues to render small thumbnail images at full resolution they were downloaded at (e.g. 1.5 MB pic for one tiny pic I saw just now). Is there any likelihood of a fix for this to serve images at much smaller file sizes? Aware that this page must perform very slowly in some of the countries in which Akvo is active (and where it is marketing / selling its tools). http://akvodev.org/seeithappen/all-rsr-project-updates
1.0
All RSR Project Updates feed - images not optimised - I see the RSR update feed continues to render small thumbnail images at full resolution they were downloaded at (e.g. 1.5 MB pic for one tiny pic I saw just now). Is there any likelihood of a fix for this to serve images at much smaller file sizes? Aware that this page must perform very slowly in some of the countries in which Akvo is active (and where it is marketing / selling its tools). http://akvodev.org/seeithappen/all-rsr-project-updates
non_priority
all rsr project updates feed images not optimised i see the rsr update feed continues to render small thumbnail images at full resolution they were downloaded at e g mb pic for one tiny pic i saw just now is there any likelihood of a fix for this to serve images at much smaller file sizes aware that this page must perform very slowly in some of the countries in which akvo is active and where it is marketing selling its tools
0
19,062
26,490,364,383
IssuesEvent
2023-01-17 22:00:43
storybookjs/storybook
https://api.github.com/repos/storybookjs/storybook
closed
Storybook fails to load story when generated code from @openapitools/openapi-generator-cli is used
question / support help wanted compatibility with other tools
**Describe the bug** Storybook fails to load a story as soon as generated code from @openapitools/openapi-generator-cli is used. The context is a NX monorepo but there is no issue when self-written code is used. We could not find what exactly breaks storybook. **To Reproduce** Steps to reproduce the behavior: I created a minimal example: https://github.com/mgansler/openapi-storybook 1. Clone Repo 2. Install dependencies (`yarn install`) 3. Start Storybook: `nx run react-app:storybook` 4. http://localhost:4400/?path=/story/app--my-app-story should show a story 5. Remove usage of generated code: remove line 9 from `app.tsx` https://github.com/mgansler/openapi-storybook/blob/38517464b39eb781b825bc6c0fc10bf923a4589f/apps/react-app/src/app/app.tsx#L9 6. Story shows up **Expected behavior** Story should show up when generated code is used. **Code snippets** https://github.com/mgansler/openapi-storybook **System** Environment Info: System: OS: macOS 10.15.7 CPU: (8) x64 Intel(R) Core(TM) i7-7820HQ CPU @ 2.90GHz Binaries: Node: 12.19.0 - ~/.nvm/versions/node/v12.19.0/bin/node Yarn: 1.22.10 - ~/.nvm/versions/node/v12.19.0/bin/yarn npm: 7.5.2 - ~/.nvm/versions/node/v12.19.0/bin/npm Browsers: Chrome: 89.0.4389.90 Firefox: 87.0 Safari: 14.0.3 **Additional context** I did not try the code "in production" in this example but in our project (which I cannot share) the App itself works without any issues. Only storybook fails to show the story.
True
Storybook fails to load story when generated code from @openapitools/openapi-generator-cli is used - **Describe the bug** Storybook fails to load a story as soon as generated code from @openapitools/openapi-generator-cli is used. The context is a NX monorepo but there is no issue when self-written code is used. We could not find what exactly breaks storybook. **To Reproduce** Steps to reproduce the behavior: I created a minimal example: https://github.com/mgansler/openapi-storybook 1. Clone Repo 2. Install dependencies (`yarn install`) 3. Start Storybook: `nx run react-app:storybook` 4. http://localhost:4400/?path=/story/app--my-app-story should show a story 5. Remove usage of generated code: remove line 9 from `app.tsx` https://github.com/mgansler/openapi-storybook/blob/38517464b39eb781b825bc6c0fc10bf923a4589f/apps/react-app/src/app/app.tsx#L9 6. Story shows up **Expected behavior** Story should show up when generated code is used. **Code snippets** https://github.com/mgansler/openapi-storybook **System** Environment Info: System: OS: macOS 10.15.7 CPU: (8) x64 Intel(R) Core(TM) i7-7820HQ CPU @ 2.90GHz Binaries: Node: 12.19.0 - ~/.nvm/versions/node/v12.19.0/bin/node Yarn: 1.22.10 - ~/.nvm/versions/node/v12.19.0/bin/yarn npm: 7.5.2 - ~/.nvm/versions/node/v12.19.0/bin/npm Browsers: Chrome: 89.0.4389.90 Firefox: 87.0 Safari: 14.0.3 **Additional context** I did not try the code "in production" in this example but in our project (which I cannot share) the App itself works without any issues. Only storybook fails to show the story.
non_priority
storybook fails to load story when generated code from openapitools openapi generator cli is used describe the bug storybook fails to load a story as soon as generated code from openapitools openapi generator cli is used the context is a nx monorepo but there is no issue when self written code is used we could not find what exactly breaks storybook to reproduce steps to reproduce the behavior i created a minimal example clone repo install dependencies yarn install start storybook nx run react app storybook should show a story remove usage of generated code remove line from app tsx story shows up expected behavior story should show up when generated code is used code snippets system environment info system os macos cpu intel r core tm cpu binaries node nvm versions node bin node yarn nvm versions node bin yarn npm nvm versions node bin npm browsers chrome firefox safari additional context i did not try the code in production in this example but in our project which i cannot share the app itself works without any issues only storybook fails to show the story
0
20,974
11,033,665,391
IssuesEvent
2019-12-06 23:55:15
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
opt: common 5-way virtual table join takes 45ms to plan
C-performance
The following simplified query is used frequently by ActiveRecord. It currently takes 30-45ms to plan, which feels excessive, especially since the query is run so frequently. I've simplified the query to remove a bunch of extra cruft that makes it harder to read. ``` root@:26257/activerecord_unittest> explain(opt) SELECT attr.attname, nsp.nspname, def.adbin FROM pg_class AS t JOIN pg_attribute AS attr ON t.oid = attrelid JOIN pg_attrdef AS def ON adrelid = attrelid AND adnum = attnum JOIN pg_constraint AS cons ON conrelid = adrelid AND adnum = conkey[1] JOIN pg_namespace AS nsp ON t.relnamespace = nsp.oid WHERE t.oid = 3; text +-----------------------------------------------------------------------------------+ project └── inner-join (hash) ├── virtual-scan pg_namespace ├── inner-join (hash) │ ├── project │ │ ├── virtual-scan pg_constraint │ │ └── projections │ │ └── conkey[1] │ ├── inner-join (hash) │ │ ├── select │ │ │ ├── virtual-scan pg_class@pg_class_oid_idx │ │ │ │ └── constraint: /1: [/3 - /3] │ │ │ └── filters │ │ │ └── pg_class.oid = 3 │ │ ├── inner-join (hash) │ │ │ ├── virtual-scan pg_attrdef │ │ │ ├── select │ │ │ │ ├── virtual-scan pg_attribute@pg_attribute_attrelid_idx │ │ │ │ │ └── constraint: /29: [/3 - /3] │ │ │ │ └── filters │ │ │ │ └── attrelid = 3 │ │ │ └── filters │ │ │ ├── adrelid = attrelid │ │ │ └── adnum = attnum │ │ └── filters │ │ └── pg_class.oid = attrelid │ └── filters │ ├── conrelid = adrelid │ └── adnum = column81 └── filters └── relnamespace = pg_namespace.oid (31 rows) Time: 50.133ms ``` Is there anything to be done about this? For comparison, Postgres plans the same query in 1-3ms.
True
opt: common 5-way virtual table join takes 45ms to plan - The following simplified query is used frequently by ActiveRecord. It currently takes 30-45ms to plan, which feels excessive, especially since the query is run so frequently. I've simplified the query to remove a bunch of extra cruft that makes it harder to read. ``` root@:26257/activerecord_unittest> explain(opt) SELECT attr.attname, nsp.nspname, def.adbin FROM pg_class AS t JOIN pg_attribute AS attr ON t.oid = attrelid JOIN pg_attrdef AS def ON adrelid = attrelid AND adnum = attnum JOIN pg_constraint AS cons ON conrelid = adrelid AND adnum = conkey[1] JOIN pg_namespace AS nsp ON t.relnamespace = nsp.oid WHERE t.oid = 3; text +-----------------------------------------------------------------------------------+ project └── inner-join (hash) ├── virtual-scan pg_namespace ├── inner-join (hash) │ ├── project │ │ ├── virtual-scan pg_constraint │ │ └── projections │ │ └── conkey[1] │ ├── inner-join (hash) │ │ ├── select │ │ │ ├── virtual-scan pg_class@pg_class_oid_idx │ │ │ │ └── constraint: /1: [/3 - /3] │ │ │ └── filters │ │ │ └── pg_class.oid = 3 │ │ ├── inner-join (hash) │ │ │ ├── virtual-scan pg_attrdef │ │ │ ├── select │ │ │ │ ├── virtual-scan pg_attribute@pg_attribute_attrelid_idx │ │ │ │ │ └── constraint: /29: [/3 - /3] │ │ │ │ └── filters │ │ │ │ └── attrelid = 3 │ │ │ └── filters │ │ │ ├── adrelid = attrelid │ │ │ └── adnum = attnum │ │ └── filters │ │ └── pg_class.oid = attrelid │ └── filters │ ├── conrelid = adrelid │ └── adnum = column81 └── filters └── relnamespace = pg_namespace.oid (31 rows) Time: 50.133ms ``` Is there anything to be done about this? For comparison, Postgres plans the same query in 1-3ms.
non_priority
opt common way virtual table join takes to plan the following simplified query is used frequently by activerecord it currently takes to plan which feels excessive especially since the query is run so frequently i ve simplified the query to remove a bunch of extra cruft that makes it harder to read root activerecord unittest explain opt select attr attname nsp nspname def adbin from pg class as t join pg attribute as attr on t oid attrelid join pg attrdef as def on adrelid attrelid and adnum attnum join pg constraint as cons on conrelid adrelid and adnum conkey join pg namespace as nsp on t relnamespace nsp oid where t oid text project └── inner join hash ├── virtual scan pg namespace ├── inner join hash │ ├── project │ │ ├── virtual scan pg constraint │ │ └── projections │ │ └── conkey │ ├── inner join hash │ │ ├── select │ │ │ ├── virtual scan pg class pg class oid idx │ │ │ │ └── constraint │ │ │ └── filters │ │ │ └── pg class oid │ │ ├── inner join hash │ │ │ ├── virtual scan pg attrdef │ │ │ ├── select │ │ │ │ ├── virtual scan pg attribute pg attribute attrelid idx │ │ │ │ │ └── constraint │ │ │ │ └── filters │ │ │ │ └── attrelid │ │ │ └── filters │ │ │ ├── adrelid attrelid │ │ │ └── adnum attnum │ │ └── filters │ │ └── pg class oid attrelid │ └── filters │ ├── conrelid adrelid │ └── adnum └── filters └── relnamespace pg namespace oid rows time is there anything to be done about this for comparison postgres plans the same query in
0
36,624
5,074,641,476
IssuesEvent
2016-12-27 15:25:32
MajkiIT/polish-ads-filter
https://api.github.com/repos/MajkiIT/polish-ads-filter
closed
300polityka.pl
reguły gotowe/testowanie reklama
Boxy po reklamach: ![image 019](https://cloud.githubusercontent.com/assets/9846948/21500645/3f84d80e-cc41-11e6-9bad-a1bf0164e30c.png) identyfikowane jako: ``` ###text-2 ``` ![image 020](https://cloud.githubusercontent.com/assets/9846948/21500660/7ad9860c-cc41-11e6-97e4-e4c1e3662c4e.png) ``` ###text-3 ```
1.0
300polityka.pl - Boxy po reklamach: ![image 019](https://cloud.githubusercontent.com/assets/9846948/21500645/3f84d80e-cc41-11e6-9bad-a1bf0164e30c.png) identyfikowane jako: ``` ###text-2 ``` ![image 020](https://cloud.githubusercontent.com/assets/9846948/21500660/7ad9860c-cc41-11e6-97e4-e4c1e3662c4e.png) ``` ###text-3 ```
non_priority
pl boxy po reklamach identyfikowane jako text text
0
201,409
15,802,273,604
IssuesEvent
2021-04-03 08:56:01
chenling1022/ped
https://api.github.com/repos/chenling1022/ped
opened
Brackets in documentation
severity.VeryLow type.DocumentationBug
No details provided. ![Screenshot 2021-04-03 at 4.54.43 PM.png](https://raw.githubusercontent.com/chenling1022/ped/main/files/2e28a1b1-378d-4817-a4b4-d33770706644.png) In command description, brackets are used, however, brackets are not used in the real command examples. i.e. view t/E [a/ACTIVITY_NAME] [date/DD-MM-YYYY] <!--session: 1617437413881-36cf2e4d-b336-4d21-91df-8378b3b68f93-->
1.0
Brackets in documentation - No details provided. ![Screenshot 2021-04-03 at 4.54.43 PM.png](https://raw.githubusercontent.com/chenling1022/ped/main/files/2e28a1b1-378d-4817-a4b4-d33770706644.png) In command description, brackets are used, however, brackets are not used in the real command examples. i.e. view t/E [a/ACTIVITY_NAME] [date/DD-MM-YYYY] <!--session: 1617437413881-36cf2e4d-b336-4d21-91df-8378b3b68f93-->
non_priority
brackets in documentation no details provided in command description brackets are used however brackets are not used in the real command examples i e view t e
0
34,096
6,289,965,642
IssuesEvent
2017-07-19 20:26:34
Naoghuman/lib-logger
https://api.github.com/repos/Naoghuman/lib-logger
closed
[JavaDoc] Use {@code xy} instead the tag <code> in JavaDoc.
documentation refactoring
[JavaDoc] Use {@code xy} instead the tag &lt;code&gt; in JavaDoc.
1.0
[JavaDoc] Use {@code xy} instead the tag <code> in JavaDoc. - [JavaDoc] Use {@code xy} instead the tag &lt;code&gt; in JavaDoc.
non_priority
use code xy instead the tag in javadoc use code xy instead the tag lt code gt in javadoc
0
73,802
7,359,079,826
IssuesEvent
2018-03-10 01:52:23
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
sql/logictest: TestParallel failed under stress
Robot test-failure
SHA: https://github.com/cockroachdb/cockroach/commits/ea5f8b22ba951799b76c84502aaab4456fe00a46 Parameters: ``` TAGS= GOFLAGS=-race ``` Stress build found a failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=552372&tab=buildLog ``` I180308 09:27:49.787154 4678 storage/replica_raftstorage.go:732 [n3,s3,r9/?:/Table/1{2-3}] applied preemptive snapshot in 8ms [clear=0ms batch=0ms entries=7ms commit=0ms] I180308 09:27:49.902388 835 testutils/testcluster/testcluster.go:534 [n1,s1] has 1 underreplicated ranges W180308 09:27:49.965287 1035 storage/store.go:3866 [n1,s1] handle raft ready: 1.0s [processed=0] I180308 09:27:49.992899 4679 storage/replica.go:3209 [n1,s1,r3/1:/System/NodeLiveness{-Max}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n4,s4):2 (n2,s2):3] next=4 I180308 09:27:49.998836 4659 storage/replica_command.go:1760 [replicate,n2,s2,r9/2:/Table/1{2-3}] change replicas (ADD_REPLICA (n3,s3):4): read existing descriptor r9:/Table/1{2-3} [(n1,s1):1, (n2,s2):2, (n5,s5):3, next=4] W180308 09:27:50.033177 1806 storage/node_liveness.go:501 [n3,hb] slow heartbeat took 1.1s I180308 09:27:50.226728 835 testutils/testcluster/testcluster.go:534 [n1,s1] has 1 underreplicated ranges I180308 09:27:50.327811 4695 storage/replica_raftstorage.go:520 [replicate,n1,s1,r10/1:/Table/1{3-4}] generated preemptive snapshot ae1bb9ed at index 153 I180308 09:27:50.408223 4615 storage/store.go:2350 [replicaGC,n1,s1,r16/1:/Table/{19-20}] removing replica I180308 09:27:50.482741 4695 storage/store.go:3745 [replicate,n1,s1,r10/1:/Table/1{3-4}] streamed snapshot to (n5,s5):?: kv pairs: 407, log entries: 51, rate-limit: 2.0 MiB/sec, 154ms I180308 09:27:50.524357 4615 storage/replica.go:836 [replicaGC,n1,s1,r16/1:/Table/{19-20}] removed 11 (0+11) keys in 40ms [clear=0ms commit=40ms] I180308 09:27:50.738648 4697 storage/replica_raftstorage.go:726 [n5,s5,r10/?:{-}] applying preemptive snapshot at index 153 (id=ae1bb9ed, encoded size=81088, 1 rocksdb batches, 51 log entries) I180308 09:27:50.786926 4653 storage/replica.go:3209 [n4,s4,r18/2:/Table/2{1-2}] proposing ADD_REPLICA((n3,s3):4): updated=[(n1,s1):1 (n4,s4):2 (n5,s5):3 (n3,s3):4] next=5 I180308 09:27:50.817425 1748 gossip/gossip.go:1306 [n3] node has connected to cluster via gossip I180308 09:27:50.818926 1748 storage/stores.go:350 [n3] wrote 4 node addresses to persistent storage I180308 09:27:50.857137 4684 storage/replica.go:3209 [n3,s3,r13/2:/Table/1{6-7}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n4,s4):4 (n3,s3):2 (n5,s5):3] next=5 I180308 09:27:51.086023 4697 storage/replica_raftstorage.go:732 [n5,s5,r10/?:/Table/1{3-4}] applied preemptive snapshot in 347ms [clear=0ms batch=0ms entries=187ms commit=1ms] I180308 09:27:51.243460 4695 storage/replica_command.go:1760 [replicate,n1,s1,r10/1:/Table/1{3-4}] change replicas (ADD_REPLICA (n5,s5):4): read existing descriptor r10:/Table/1{3-4} [(n1,s1):1, (n3,s3):2, (n2,s2):3, next=4] I180308 09:27:51.381230 4761 storage/replica_command.go:1760 [replicate,n4,s4,r18/2:/Table/2{1-2}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r18:/Table/2{1-2} [(n1,s1):1, (n4,s4):2, (n5,s5):3, (n3,s3):4, next=5] I180308 09:27:51.509691 4735 storage/replica_raftstorage.go:520 [replicate,n3,s3,r6/2:/{System/tse-Table/System…}] generated preemptive snapshot 27aebd8b at index 30 I180308 09:27:51.563486 835 sql/logictest/parallel_test.go:213 Setting range split size: 32768 I180308 09:27:51.601539 4735 storage/store.go:3745 [replicate,n3,s3,r6/2:/{System/tse-Table/System…}] streamed snapshot to (n2,s2):?: kv pairs: 14, log entries: 20, rate-limit: 2.0 MiB/sec, 91ms I180308 09:27:51.619862 1901 storage/store.go:3479 [n1,s1,r13/1:/Table/1{6-7}] added to replica GC queue (peer suggestion) I180308 09:27:51.630538 4789 storage/replica_raftstorage.go:726 [n2,s2,r6/?:{-}] applying preemptive snapshot at index 30 (id=27aebd8b, encoded size=8130, 1 rocksdb batches, 20 log entries) I180308 09:27:51.728873 4789 storage/replica_raftstorage.go:732 [n2,s2,r6/?:/{System/tse-Table/System…}] applied preemptive snapshot in 98ms [clear=0ms batch=0ms entries=94ms commit=3ms] I180308 09:27:51.807803 4735 storage/replica_command.go:1760 [replicate,n3,s3,r6/2:/{System/tse-Table/System…}] change replicas (ADD_REPLICA (n2,s2):4): read existing descriptor r6:/{System/tse-Table/SystemConfigSpan/Start} [(n1,s1):1, (n3,s3):2, (n4,s4):3, next=4] I180308 09:27:51.828942 4767 storage/replica_raftstorage.go:520 [raftsnapshot,n4,s4,r18/2:/Table/2{1-2}] generated Raft snapshot df1ffd91 at index 31 I180308 09:27:51.834992 4767 storage/store.go:3745 [raftsnapshot,n4,s4,r18/2:/Table/2{1-2}] streamed snapshot to (n3,s3):4: kv pairs: 18, log entries: 21, rate-limit: 8.0 MiB/sec, 5ms I180308 09:27:51.837164 4834 storage/replica_raftstorage.go:726 [n3,s3,r18/4:{-}] applying Raft snapshot at index 31 (id=df1ffd91, encoded size=7716, 1 rocksdb batches, 21 log entries) I180308 09:27:51.842351 4834 storage/replica_raftstorage.go:732 [n3,s3,r18/4:/Table/2{1-2}] applied Raft snapshot in 5ms [clear=0ms batch=0ms entries=3ms commit=0ms] I180308 09:27:52.277701 4836 storage/store.go:2350 [replicaGC,n5,s5,r10/?:/Table/1{3-4}] removing replica I180308 09:27:52.302073 4671 storage/store.go:2350 [replicaGC,n1,s1,r13/1:/Table/1{6-7}] removing replica I180308 09:27:52.431403 4836 storage/replica.go:836 [replicaGC,n5,s5,r10/?:/Table/1{3-4}] removed 389 (379+10) keys in 153ms [clear=0ms commit=153ms] I180308 09:27:52.549535 4671 storage/replica.go:836 [replicaGC,n1,s1,r13/1:/Table/1{6-7}] removed 12 (0+12) keys in 219ms [clear=0ms commit=219ms] I180308 09:27:52.628009 4816 storage/replica.go:3209 [n2,s2,r9/2:/Table/1{2-3}] proposing ADD_REPLICA((n3,s3):4): updated=[(n1,s1):1 (n2,s2):2 (n5,s5):3 (n3,s3):4] next=5 I180308 09:27:52.699806 1055 server/status/runtime.go:219 [n1] runtime stats: 1.1 GiB RSS, 1117 goroutines, 76 MiB/16 MiB/119 MiB GO alloc/idle/total, 154 MiB/186 MiB CGO alloc/total, 862.81cgo/sec, 0.95/0.07 %(u/s)time, 0.05 %gc (5x) I180308 09:27:52.842796 4890 storage/replica_command.go:1760 [replicate,n2,s2,r9/2:/Table/1{2-3}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r9:/Table/1{2-3} [(n1,s1):1, (n2,s2):2, (n5,s5):3, (n3,s3):4, next=5] I180308 09:27:53.053227 4909 storage/replica.go:3209 [n3,s3,r6/2:/{System/tse-Table/System…}] proposing ADD_REPLICA((n2,s2):4): updated=[(n1,s1):1 (n3,s3):2 (n4,s4):3 (n2,s2):4] next=5 I180308 09:27:53.204274 4783 storage/replica_command.go:1760 [replicate,n3,s3,r6/2:/{System/tse-Table/System…}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r6:/{System/tse-Table/SystemConfigSpan/Start} [(n1,s1):1, (n3,s3):2, (n4,s4):3, (n2,s2):4, next=5] I180308 09:27:53.385647 4933 storage/replica.go:3209 [n4,s4,r18/2:/Table/2{1-2}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n3,s3):4 (n4,s4):2 (n5,s5):3] next=5 I180308 09:27:53.654566 835 sql/logictest/parallel_test.go:227 Creating database I180308 09:27:53.714521 1518 gossip/gossip.go:1306 [n2] node has connected to cluster via gossip I180308 09:27:53.812099 1518 storage/stores.go:350 [n2] wrote 4 node addresses to persistent storage I180308 09:27:53.949703 2370 storage/store.go:3479 [n1,s1,r18/1:/Table/2{1-2}] added to replica GC queue (peer suggestion) I180308 09:27:53.950132 2370 storage/store.go:3479 [n1,s1,r18/1:/Table/2{1-2}] added to replica GC queue (peer suggestion) I180308 09:27:54.187036 4918 storage/replica_command.go:838 [split,n1,s1,r5/1:/System/ts{d-e}] initiating a split of this range at key /System/tsd/cr.node.gossip.infos.sent/4/10s/2018-03-08T09:00:00Z [r21] I180308 09:27:54.490089 4921 storage/replica_command.go:838 [split,n1,s1,r4/1:/System/{NodeLive…-tsd}] initiating a split of this range at key /System/StatusNode/2 [r22] I180308 09:27:54.641807 4982 storage/replica_command.go:838 [split,n1,s1,r10/1:/Table/1{3-4}] initiating a split of this range at key /Table/13/1/2018-03-08T09:27:29.777287Z/329091808619495425 [r23] I180308 09:27:54.767000 4695 storage/replica.go:3209 [n1,s1,r10/1:/Table/1{3-4}] proposing ADD_REPLICA((n5,s5):4): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3 (n5,s5):4] next=5 I180308 09:27:54.924274 4936 storage/store.go:2350 [replicaGC,n1,s1,r18/1:/Table/2{1-2}] removing replica I180308 09:27:55.227444 4936 storage/replica.go:836 [replicaGC,n1,s1,r18/1:/Table/2{1-2}] removed 11 (0+11) keys in 51ms [clear=0ms commit=50ms] W180308 09:27:55.283305 4861 storage/replica.go:3016 [n1,s1,r10/1:/Table/1{3-4}] context canceled before proposing: 1 HeartbeatTxn I180308 09:27:55.373122 5042 storage/replica_command.go:838 [split,n1,s1,r10/1:/Table/1{3-4}] initiating a split of this range at key /Table/13/1/2018-03-08T09:27:29.777287Z/329091808619495425 [r24] I180308 09:27:55.591922 5003 storage/replica.go:3209 [n2,s2,r9/2:/Table/1{2-3}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n3,s3):4 (n2,s2):2 (n5,s5):3] next=5 I180308 09:27:55.659550 1544 server/status/runtime.go:219 [n2] runtime stats: 1.1 GiB RSS, 1126 goroutines, 57 MiB/35 MiB/120 MiB GO alloc/idle/total, 155 MiB/187 MiB CGO alloc/total, 798.50cgo/sec, 0.95/0.07 %(u/s)time, 0.03 %gc (5x) I180308 09:27:56.074863 2121 storage/store.go:3479 [n1,s1,r9/1:/Table/1{2-3}] added to replica GC queue (peer suggestion) I180308 09:27:56.304086 4874 sql/event_log.go:121 [n1,client=127.0.0.1:43858,user=root] Event: "create_database", target: 50, info: {DatabaseName:test Statement:CREATE DATABASE test User:root} I180308 09:27:56.307831 2121 storage/store.go:3479 [n1,s1,r9/1:/Table/1{2-3}] added to replica GC queue (peer suggestion) I180308 09:27:56.399786 1635 storage/replica.go:3277 [n2,s2,r2/3:/System/{-NodeLive…}] not quiescing: 1 pending commands I180308 09:27:56.412847 2121 storage/store.go:3479 [n1,s1,r9/1:/Table/1{2-3}] added to replica GC queue (peer suggestion) I180308 09:27:56.413411 2121 storage/store.go:3479 [n1,s1,r9/1:/Table/1{2-3}] added to replica GC queue (peer suggestion) I180308 09:27:56.429913 5110 storage/store.go:2350 [replicaGC,n1,s1,r9/1:/Table/1{2-3}] removing replica I180308 09:27:56.448755 1797 server/status/runtime.go:219 [n3] runtime stats: 1.1 GiB RSS, 1135 goroutines, 66 MiB/26 MiB/120 MiB GO alloc/idle/total, 156 MiB/187 MiB CGO alloc/total, 773.55cgo/sec, 0.94/0.07 %(u/s)time, 0.03 %gc (5x) I180308 09:27:56.688465 5110 storage/replica.go:836 [replicaGC,n1,s1,r9/1:/Table/1{2-3}] removed 71 (60+11) keys in 257ms [clear=0ms commit=257ms] I180308 09:27:57.234316 5009 storage/replica_command.go:838 [split,n4,s4,r20/2:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r41] I180308 09:27:57.715312 5007 storage/replica_raftstorage.go:520 [raftsnapshot,n1,s1,r10/1:/Table/1{3-4}] generated Raft snapshot 729bfb72 at index 173 W180308 09:27:57.734210 982 storage/store.go:3866 [n1,s1] handle raft ready: 0.6s [processed=0] I180308 09:27:57.771678 2044 server/status/runtime.go:219 [n4] runtime stats: 1.2 GiB RSS, 1138 goroutines, 51 MiB/40 MiB/120 MiB GO alloc/idle/total, 156 MiB/187 MiB CGO alloc/total, 717.70cgo/sec, 0.95/0.07 %(u/s)time, 0.03 %gc (5x) I180308 09:27:57.853130 5007 storage/store.go:3745 [raftsnapshot,n1,s1,r10/1:/Table/1{3-4}] streamed snapshot to (n5,s5):4: kv pairs: 456, log entries: 71, rate-limit: 8.0 MiB/sec, 104ms I180308 09:27:57.904274 5196 storage/replica_raftstorage.go:726 [n5,s5,r10/4:{-}] applying Raft snapshot at index 173 (id=729bfb72, encoded size=103834, 1 rocksdb batches, 71 log entries) W180308 09:27:58.307548 1553 storage/node_liveness.go:501 [n2,hb] slow heartbeat took 1.2s I180308 09:27:58.890124 5216 storage/replica.go:3209 [n3,s3,r6/2:/{System/tse-Table/System…}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n2,s2):4 (n3,s3):2 (n4,s4):3] next=5 I180308 09:27:59.234979 2484 server/status/runtime.go:219 [n5] runtime stats: 1.2 GiB RSS, 1139 goroutines, 78 MiB/15 MiB/120 MiB GO alloc/idle/total, 156 MiB/188 MiB CGO alloc/total, 720.50cgo/sec, 0.94/0.06 %(u/s)time, 0.03 %gc (5x) I180308 09:27:59.534800 1901 storage/store.go:3479 [n1,s1,r6/1:/{System/tse-Table/System…}] added to replica GC queue (peer suggestion) I180308 09:27:59.617241 1901 storage/store.go:3479 [n1,s1,r6/1:/{System/tse-Table/System…}] added to replica GC queue (peer suggestion) W180308 09:27:59.658775 1806 storage/node_liveness.go:501 [n3,hb] slow heartbeat took 1.7s I180308 09:27:59.883988 835 sql/logictest/parallel_test.go:236 Test setup done I180308 09:27:59.884248 835 sql/logictest/parallel_test.go:159 testdata/parallel_test/subquery_retry_multinode/test.yaml: run list 0: 0:setup W180308 09:28:00.194433 1096 storage/node_liveness.go:501 [n1,hb] slow heartbeat took 1.4s I180308 09:28:00.297625 5240 storage/store.go:2350 [replicaGC,n1,s1,r6/1:/{System/tse-Table/System…}] removing replica I180308 09:28:00.329573 1028 storage/replica_proposal.go:201 [n1,s1,r4/1:/System/{NodeLive…-tsd}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0 I180308 09:28:00.379562 5248 storage/replica_command.go:838 [split,n1,s1,r22/1:/System/{StatusNo…-tsd}] initiating a split of this range at key /System/StatusNode/3 [r25] I180308 09:28:00.448234 5240 storage/replica.go:836 [replicaGC,n1,s1,r6/1:/{System/tse-Table/System…}] removed 12 (0+12) keys in 91ms [clear=0ms commit=91ms] W180308 09:28:00.707751 2147 storage/node_liveness.go:501 [n4,hb] slow heartbeat took 1.4s W180308 09:28:00.731795 5352 storage/replica_backpressure.go:135 [n1,s1,r5/1:/System/ts{d-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.security.certificate.expiration.ca/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.security.certificate.expiration.node/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.connections.incoming/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.connections.refused/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.bytes.received/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.bytes.sent/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.infos.received/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.infos.sent/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.connections.outgoing/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.batches/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.batches.partial/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.rpc.sent/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.rpc.sent.local/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.rpc.sent.nextreplicaerror/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.errors.notleaseholder/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.requests.slow.distsender/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.aborts/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/1/10s/2018-03-08T09:00:00Z,/Min), ... 405 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/1/10s/2018-03-08T09:00:00Z,/Min) I180308 09:28:01.257194 983 storage/replica_proposal.go:201 [n1,s1,r10/1:/Table/1{3-4}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.325864565,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.325864565,0 I180308 09:28:01.263525 5297 storage/replica_command.go:838 [split,n1,s1,r24/1:/Table/1{3/1/2018…-4}] initiating a split of this range at key /Table/13/1/2018-03-08T09:27:40.666166Z/329091846156877826 [r26] I180308 09:28:01.312145 5196 storage/replica_raftstorage.go:732 [n5,s5,r10/4:/Table/1{3-4}] applied Raft snapshot in 3408ms [clear=0ms batch=0ms entries=3374ms commit=21ms] I180308 09:28:01.454527 2220 storage/replica_proposal.go:201 [n4,s4,r20/2:/{Table/23-Max}] new range lease repl=(n4,s4):2 seq=3 start=1520501251.624682287,1 epo=1 pro=1520501255.402825609,0 following repl=(n4,s4):2 seq=3 start=1520501251.624682287,1 epo=1 pro=1520501255.402825609,0 I180308 09:28:01.482106 1005 storage/replica_proposal.go:201 [n1,s1,r5/1:/System/ts{d-e}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0 W180308 09:28:01.575951 5351 storage/replica.go:3016 [n1,s1,r10/1:/Table/13{-/1/2018-…}] context canceled before proposing: 1 HeartbeatTxn I180308 09:28:01.658992 5386 storage/replica_command.go:838 [split,n1,s1,r21/1:/System/ts{d/cr.no…-e}] initiating a split of this range at key /System/tsd/cr.node.sql.bytesin/3/10s/2018-03-08T09:00:00Z [r27] I180308 09:28:01.678252 5296 storage/replica_command.go:1760 [replicate,n1,s1,r10/1:/Table/13{-/1/2018-…}] change replicas (REMOVE_REPLICA (n3,s3):2): read existing descriptor r10:/Table/13{-/1/2018-03-08T09:27:29.777287Z/329091808619495425} [(n1,s1):1, (n3,s3):2, (n2,s2):3, (n5,s5):4, next=5] I180308 09:28:01.932018 5423 storage/replica_raftstorage.go:520 [replicate,n4,s4,r41/2:/{Table/50-Max}] generated preemptive snapshot e7efcdd4 at index 10 I180308 09:28:01.954883 5491 storage/replica_raftstorage.go:520 [replicate,n3,s3,r17/2:/Table/2{0-1}] generated preemptive snapshot b7e6a194 at index 36 E180308 09:28:02.175447 5423 storage/queue.go:778 [replicate,n4,s4,r41/2:/{Table/50-Max}] raft status not initialized I180308 09:28:02.188640 4874 sql/event_log.go:121 [n1,client=127.0.0.1:43858,user=root] Event: "create_table", target: 51, info: {TableName:test.public.t Statement:CREATE TABLE t (k INT, filler STRING) User:root} --- progress: testdata/parallel_test/subquery_retry_multinode/setup: 1 statements/queries I180308 09:28:02.231950 5276 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/setup: 1 statements/queries I180308 09:28:02.309261 5491 storage/store.go:3745 [replicate,n3,s3,r17/2:/Table/2{0-1}] streamed snapshot to (n1,s1):?: kv pairs: 16, log entries: 26, rate-limit: 2.0 MiB/sec, 97ms I180308 09:28:02.727426 5515 storage/replica_command.go:838 [split,n4,s4,r41/2:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r42] I180308 09:28:02.767793 1979 gossip/gossip.go:1306 [n4] node has connected to cluster via gossip I180308 09:28:02.791521 1055 server/status/runtime.go:219 [n1] runtime stats: 1.2 GiB RSS, 1166 goroutines, 65 MiB/26 MiB/120 MiB GO alloc/idle/total, 157 MiB/187 MiB CGO alloc/total, 1065.63cgo/sec, 0.94/0.07 %(u/s)time, 0.02 %gc (5x) I180308 09:28:03.002229 5486 storage/replica_raftstorage.go:726 [n1,s1,r17/?:{-}] applying preemptive snapshot at index 36 (id=b7e6a194, encoded size=8224, 1 rocksdb batches, 26 log entries) I180308 09:28:03.005301 1979 storage/stores.go:350 [n4] wrote 4 node addresses to persistent storage I180308 09:28:03.154898 5486 storage/replica_raftstorage.go:732 [n1,s1,r17/?:/Table/2{0-1}] applied preemptive snapshot in 152ms [clear=0ms batch=0ms entries=127ms commit=12ms] W180308 09:28:03.190672 5455 storage/replica_backpressure.go:135 [n2,s2,r21/2:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.security.certificate.expiration.ca/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.security.certificate.expiration.node/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.requests.slow.distsender/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.aborts/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), ... 379 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/2/10s/2018-03-08T09:00:00Z,/Min) I180308 09:28:03.513617 5491 storage/replica_command.go:1760 [replicate,n3,s3,r17/2:/Table/2{0-1}] change replicas (ADD_REPLICA (n1,s1):5): read existing descriptor r17:/Table/2{0-1} [(n5,s5):4, (n3,s3):2, (n2,s2):3, next=5] W180308 09:28:04.034472 5607 storage/replica_backpressure.go:135 [n1,s1,r21/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.security.certificate.expiration.ca/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.security.certificate.expiration.node/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.infos.sent/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.requests.slow.distsender/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.aborts/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/4/10s/2018-03-08T09:00:00Z,/Min), ... 380 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/4/10s/2018-03-08T09:00:00Z,/Min) W180308 09:28:04.459926 5676 storage/replica_backpressure.go:135 [n1,s1,r21/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.security.certificate.expiration.ca/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.security.certificate.expiration.node/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.infos.sent/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.requests.slow.distsender/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.aborts/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/5/10s/2018-03-08T09:00:00Z,/Min), ... 380 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/5/10s/2018-03-08T09:00:00Z,/Min) --- progress: testdata/parallel_test/subquery_retry_multinode/setup: 2 statements/queries I180308 09:28:04.488626 5276 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/setup: 2 statements/queries --- done: testdata/parallel_test/subquery_retry_multinode/setup with config : 2 tests, 0 failures I180308 09:28:04.488940 5276 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/setup with config : 2 tests, 0 failures I180308 09:28:04.494592 835 sql/logictest/parallel_test.go:159 testdata/parallel_test/subquery_retry_multinode/test.yaml: run list 1: 0:txn, 1:txn, 2:txn, 3:txn, 4:txn I180308 09:28:04.692538 5642 storage/replica.go:3209 [n1,s1,r10/1:/Table/13{-/1/2018-…}] proposing REMOVE_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n5,s5):4 (n2,s2):3] next=5 I180308 09:28:05.177727 1003 storage/replica_proposal.go:201 [n1,s1,r22/1:/System/{StatusNo…-tsd}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0 I180308 09:28:05.184170 1972 storage/store.go:3479 [n3,s3,r10/2:/Table/13{-/1/2018-…}] added to replica GC queue (peer suggestion) I180308 09:28:05.275031 5772 storage/replica_command.go:838 [split,n1,s1,r25/1:/System/{StatusNo…-tsd}] initiating a split of this range at key /System/StatusNode/4 [r28] I180308 09:28:05.629470 1544 server/status/runtime.go:219 [n2] runtime stats: 1.2 GiB RSS, 1171 goroutines, 87 MiB/11 MiB/125 MiB GO alloc/idle/total, 157 MiB/189 MiB CGO alloc/total, 1173.83cgo/sec, 0.94/0.07 %(u/s)time, 0.02 %gc (4x) I180308 09:28:05.634410 1827 storage/replica_proposal.go:201 [n3,s3,r4/2:/System/{NodeLive…-StatusNo…}] new range lease repl=(n3,s3):2 seq=4 start=1520501285.016867674,0 epo=1 pro=1520501285.016882538,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0 W180308 09:28:06.031522 5844 storage/replica.go:3016 [n1,s1,r22/1:/System/StatusNode/{2-3}] context canceled before proposing: 1 HeartbeatTxn --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries I180308 09:28:06.562121 5731 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures I180308 09:28:06.562430 5731 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures I180308 09:28:06.630349 1797 server/status/runtime.go:219 [n3] runtime stats: 1.2 GiB RSS, 1178 goroutines, 97 MiB/5.9 MiB/129 MiB GO alloc/idle/total, 158 MiB/189 MiB CGO alloc/total, 1136.76cgo/sec, 0.94/0.07 %(u/s)time, 0.03 %gc (5x) W180308 09:28:06.862124 2493 storage/node_liveness.go:501 [n5,hb] slow heartbeat took 1.7s W180308 09:28:07.341217 1553 storage/node_liveness.go:501 [n2,hb] slow heartbeat took 1.2s I180308 09:28:07.775014 2044 server/status/runtime.go:219 [n4] runtime stats: 1.2 GiB RSS, 1174 goroutines, 75 MiB/24 MiB/129 MiB GO alloc/idle/total, 157 MiB/189 MiB CGO alloc/total, 1173.62cgo/sec, 0.94/0.07 %(u/s)time, 0.03 %gc (4x) I180308 09:28:07.899393 993 storage/replica_proposal.go:201 [n1,s1,r21/1:/System/ts{d/cr.no…-e}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0 I180308 09:28:08.092040 6071 storage/replica_command.go:838 [split,n1,s1,r27/1:/System/ts{d/cr.no…-e}] initiating a split of this range at key /System/tsd/cr.node.sql.distsql.service.latency-p99.999/1/10s/2018-03-08T09:00:00Z [r29] W180308 09:28:09.182946 1850 storage/store.go:3866 [n3,s3] handle raft ready: 0.7s [processed=0] I180308 09:28:09.228630 2484 server/status/runtime.go:219 [n5] runtime stats: 1.3 GiB RSS, 1216 goroutines, 96 MiB/9.2 MiB/132 MiB GO alloc/idle/total, 160 MiB/191 MiB CGO alloc/total, 1213.47cgo/sec, 0.94/0.07 %(u/s)time, 0.03 %gc (5x) W180308 09:28:09.431040 987 storage/store.go:3866 [n1,s1] handle raft ready: 0.6s [processed=0] W180308 09:28:09.702153 1029 storage/store.go:3866 [n1,s1] handle raft ready: 0.5s [processed=0] I180308 09:28:09.884861 6241 storage/replica_raftstorage.go:520 [replicate,n1,s1,r5/1:/System/tsd{-/cr.nod…}] generated preemptive snapshot c879f910 at index 68 W180308 09:28:09.920664 1096 storage/node_liveness.go:501 [n1,hb] slow heartbeat took 2.2s I180308 09:28:09.926471 6051 storage/store.go:2350 [replicaGC,n1,s1,r17/?:/Table/2{0-1}] removing replica I180308 09:28:10.164342 6051 storage/replica.go:836 [replicaGC,n1,s1,r17/?:/Table/2{0-1}] removed 11 (0+11) keys in 177ms [clear=0ms commit=177ms] I180308 09:28:10.208745 6080 storage/replica_raftstorage.go:520 [replicate,n2,s2,r2/3:/System/{-NodeLive…}] generated preemptive snapshot c8cdd480 at index 34 I180308 09:28:10.482965 6080 storage/store.go:3745 [replicate,n2,s2,r2/3:/System/{-NodeLive…}] streamed snapshot to (n5,s5):?: kv pairs: 13, log entries: 24, rate-limit: 2.0 MiB/sec, 273ms I180308 09:28:10.627902 6241 storage/store.go:3745 [replicate,n1,s1,r5/1:/System/tsd{-/cr.nod…}] streamed snapshot to (n4,s4):?: kv pairs: 143, log entries: 13, rate-limit: 2.0 MiB/sec, 677ms I180308 09:28:10.693165 6287 storage/replica_raftstorage.go:726 [n5,s5,r2/?:{-}] applying preemptive snapshot at index 34 (id=c8cdd480, encoded size=56977, 1 rocksdb batches, 24 log entries) I180308 09:28:10.790504 6345 storage/replica_raftstorage.go:726 [n4,s4,r5/?:{-}] applying preemptive snapshot at index 68 (id=c879f910, encoded size=36756, 1 rocksdb batches, 13 log entries) W180308 09:28:10.795289 6380 storage/replica_backpressure.go:135 [n1,s1,r27/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.txn.aborts/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p90/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p75/3/10s/2018-03-08T09:00:00Z,/Min), ... 355 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/3/10s/2018-03-08T09:00:00Z,/Min) W180308 09:28:10.836796 1826 storage/store.go:3866 [n3,s3] handle raft ready: 0.5s [processed=0] W180308 09:28:10.963543 2147 storage/node_liveness.go:501 [n4,hb] slow heartbeat took 2.7s W180308 09:28:11.405336 6410 storage/replica_backpressure.go:135 [n1,s1,r27/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.txn.aborts/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p90/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p75/2/10s/2018-03-08T09:00:00Z,/Min), ... 354 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/2/10s/2018-03-08T09:00:00Z,/Min) I180308 09:28:11.779388 987 storage/replica_proposal.go:201 [n1,s1,r24/1:/Table/1{3/1/2018…-4}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.325864565,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.325864565,0 W180308 09:28:11.899345 2493 storage/node_liveness.go:501 [n5,hb] slow heartbeat took 2.2s W180308 09:28:12.551481 1553 storage/node_liveness.go:501 [n2,hb] slow heartbeat took 1.9s I180308 09:28:12.571934 6345 storage/replica_raftstorage.go:732 [n4,s4,r5/?:/System/tsd{-/cr.nod…}] applied preemptive snapshot in 1781ms [clear=0ms batch=0ms entries=1729ms commit=33ms] I180308 09:28:12.630942 6241 storage/replica_command.go:1760 [replicate,n1,s1,r5/1:/System/tsd{-/cr.nod…}] change replicas (ADD_REPLICA (n4,s4):4): read existing descriptor r5:/System/tsd{-/cr.node.gossip.infos.sent/4/10s/2018-03-08T09:00:00Z} [(n1,s1):1, (n2,s2):2, (n5,s5):3, next=4] W180308 09:28:12.654684 1806 storage/node_liveness.go:501 [n3,hb] slow heartbeat took 1.2s I180308 09:28:12.667081 1055 server/status/runtime.go:219 [n1] runtime stats: 1.3 GiB RSS, 1182 goroutines, 78 MiB/23 MiB/132 MiB GO alloc/idle/total, 160 MiB/194 MiB CGO alloc/total, 934.67cgo/sec, 0.92/0.07 %(u/s)time, 0.02 %gc (4x) I180308 09:28:13.552826 6287 storage/replica_raftstorage.go:732 [n5,s5,r2/?:/System/{-NodeLive…}] applied preemptive snapshot in 2833ms [clear=0ms batch=0ms entries=2609ms commit=164ms] I180308 09:28:13.769924 6080 storage/replica_command.go:1760 [replicate,n2,s2,r2/3:/System/{-NodeLive…}] change replicas (ADD_REPLICA (n5,s5):4): read existing descriptor r2:/System/{-NodeLiveness} [(n1,s1):1, (n3,s3):2, (n2,s2):3, next=4] W180308 09:28:13.950338 2147 storage/node_liveness.go:501 [n4,hb] slow heartbeat took 1.2s I180308 09:28:14.047030 6643 storage/replica.go:3209 [n3,s3,r17/2:/Table/2{0-1}] proposing ADD_REPLICA((n1,s1):5): updated=[(n5,s5):4 (n3,s3):2 (n2,s2):3 (n1,s1):5] next=6 --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries I180308 09:28:14.991532 5735 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures I180308 09:28:14.991818 5735 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures W180308 09:28:15.210739 2493 storage/node_liveness.go:501 [n5,hb] slow heartbeat took 1.1s I180308 09:28:15.459366 2539 storage/replica_proposal.go:201 [n5,s5,r17/4:/Table/2{0-1}] new range lease repl=(n5,s5):4 seq=4 start=1520501294.765769961,0 epo=1 pro=1520501294.765803960,0 following repl=(n3,s3):2 seq=3 start=1520501251.624682287,1 epo=1 pro=1520501257.825367317,0 I180308 09:28:15.554730 6653 storage/replica_raftstorage.go:520 [raftsnapshot,n3,s3,r17/2:/Table/2{0-1}] generated Raft snapshot 901a1124 at index 49 I180308 09:28:15.697368 1544 server/status/runtime.go:219 [n2] runtime stats: 1.3 GiB RSS, 1181 goroutines, 73 MiB/35 MiB/138 MiB GO alloc/idle/total, 160 MiB/195 MiB CGO alloc/total, 840.48cgo/sec, 0.90/0.07 %(u/s)time, 0.02 %gc (4x) I180308 09:28:15.764467 6653 storage/store.go:3745 [raftsnapshot,n3,s3,r17/2:/Table/2{0-1}] streamed snapshot to (n1,s1):5: kv pairs: 18, log entries: 39, rate-limit: 8.0 MiB/sec, 61ms I180308 09:28:15.778517 6602 storage/replica_raftstorage.go:726 [n1,s1,r17/5:{-}] applying Raft snapshot at index 49 (id=901a1124, encoded size=12107, 1 rocksdb batches, 39 log entries) I180308 09:28:16.177052 6521 storage/store.go:2350 [replicaGC,n4,s4,r5/?:/System/tsd{-/cr.nod…}] removing replica I180308 09:28:16.294600 6521 storage/replica.go:836 [replicaGC,n4,s4,r5/?:/System/tsd{-/cr.nod…}] removed 139 (128+11) keys in 117ms [clear=0ms commit=116ms] W180308 09:28:16.347616 1553 storage/node_liveness.go:501 [n2,hb] slow heartbeat took 1.3s I180308 09:28:16.372076 994 storage/replica_proposal.go:201 [n1,s1,r25/1:/System/{StatusNo…-tsd}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0 I180308 09:28:16.421976 5631 storage/store.go:2350 [replicaGC,n3,s3,r10/2:/Table/13{-/1/2018-…}] removing replica I180308 09:28:16.665522 1797 server/status/runtime.go:219 [n3] runtime stats: 1.3 GiB RSS, 1171 goroutines, 88 MiB/20 MiB/138 MiB GO alloc/idle/total, 162 MiB/196 MiB CGO alloc/total, 884.78cgo/sec, 0.89/0.07 %(u/s)time, 0.02 %gc (3x) I180308 09:28:16.681548 6668 storage/replica_command.go:838 [split,n1,s1,r28/1:/System/{StatusNo…-tsd}] initiating a split of this range at key /System/StatusNode/5 [r30] I180308 09:28:16.801652 6602 storage/replica_raftstorage.go:732 [n1,s1,r17/5:/Table/2{0-1}] applied Raft snapshot in 1003ms [clear=2ms batch=0ms entries=991ms commit=8ms] I180308 09:28:16.806835 5631 storage/replica.go:836 [replicaGC,n3,s3,r10/2:/Table/13{-/1/2018-…}] removed 181 (169+12) keys in 384ms [clear=0ms commit=384ms] W180308 09:28:17.412329 1806 storage/node_liveness.go:501 [n3,hb] slow heartbeat took 1.5s I180308 09:28:17.506774 6629 storage/replica_consistency.go:129 [replica consistency checker,n1,s1,r21/1:/System/tsd/cr.node.{go…-sq…}] triggering stats recomputation to resolve delta of {ContainsEstimates:true LastUpdateNanos:1520501288989219219 IntentAge:0 GCBytesAge:0 LiveBytes:4092 LiveCount:0 KeyBytes:0 KeyCount:0 ValBytes:4092 ValCount:0 IntentBytes:0 IntentCount:0 SysBytes:0 SysCount:0} I180308 09:28:17.510115 2214 storage/replica_proposal.go:201 [n4,s4,r41/2:/{Table/50-Max}] new range lease repl=(n4,s4):2 seq=3 start=1520501251.624682287,1 epo=1 pro=1520501255.402825609,0 following repl=(n4,s4):2 seq=3 start=1520501251.624682287,1 epo=1 pro=1520501255.402825609,0 I180308 09:28:17.785167 2044 server/status/runtime.go:219 [n4] runtime stats: 1.3 GiB RSS, 1172 goroutines, 62 MiB/44 MiB/138 MiB GO alloc/idle/total, 162 MiB/195 MiB CGO alloc/total, 981.39cgo/sec, 0.90/0.07 %(u/s)time, 0.02 %gc (4x) I180308 09:28:17.789384 6779 storage/replica.go:3209 [n2,s2,r2/3:/System/{-NodeLive…}] proposing ADD_REPLICA((n5,s5):4): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3 (n5,s5):4] next=5 W180308 09:28:17.792842 6603 storage/replica.go:3016 [n4,s4,r41/2:/Table/5{0-1}] context canceled before proposing: 1 HeartbeatTxn W180308 09:28:18.718662 1096 storage/node_liveness.go:501 [n1,hb] slow heartbeat took 2.0s I180308 09:28:18.788824 6931 storage/replica_command.go:1760 [replicate,n5,s5,r17/4:/Table/2{0-1}] change replicas (REMOVE_REPLICA (n1,s1):5): read existing descriptor r17:/Table/2{0-1} [(n5,s5):4, (n3,s3):2, (n2,s2):3, (n1,s1):5, next=6] I180308 09:28:19.053142 1010 storage/replica_proposal.go:201 [n1,s1,r27/1:/System/ts{d/cr.no…-e}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0 E180308 09:28:19.095691 5732 sql/logictest/logic.go:2124 testdata/parallel_test/subquery_retry_multinode/txn:9: expected success, but found pq: TransactionStatusError: does not exist --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 0 tests, 1 failures I180308 09:28:19.153442 5732 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 0 tests, 1 failures E180308 09:28:19.153715 5732 sql/logictest/parallel_test.go:84 error processing testdata/parallel_test/subquery_retry_multinode/txn: pq: TransactionStatusError: does not exist I180308 09:28:19.146805 2484 server/status/runtime.go:219 [n5] runtime stats: 1.3 GiB RSS, 1162 goroutines, 95 MiB/13 MiB/138 MiB GO alloc/idle/total, 161 MiB/196 MiB CGO alloc/total, 1029.88cgo/sec, 0.90/0.08 %(u/s)time, 0.02 %gc (3x) W180308 09:28:19.277944 2147 storage/node_liveness.go:501 [n4,hb] slow heartbeat took 2.0s I180308 09:28:19.513328 7000 storage/replica_command.go:838 [split,n1,s1,r29/1:/System/ts{d/cr.no…-e}] initiating a split of this range at key /System/tsd/cr.node.sql.mem.admin.session.max-p99.9/1/10s/2018-03-08T09:00:00Z [r31] I180308 09:28:19.527756 6832 storage/replica.go:3209 [n1,s1,r5/1:/System/tsd{-/cr.nod…}] proposing ADD_REPLICA((n4,s4):4): updated=[(n1,s1):1 (n2,s2):2 (n5,s5):3 (n4,s4):4] next=5 I180308 09:28:20.040589 7039 storage/replica_command.go:1760 [replicate,n1,s1,r24/1:/Table/13/1/2018-03-08T09:27:…] change replicas (REMOVE_REPLICA (n2,s2):3): read existing descriptor r24:/Table/13/1/2018-03-08T09:27:{29.777287Z/329091808619495425-40.666166Z/329091846156877826} [(n1,s1):1, (n3,s3):2, (n2,s2):3, (n5,s5):4, next=5] W180308 09:28:20.196189 7047 storage/replica_backpressure.go:135 [n1,s1,r29/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.txn.aborts/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p90/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p75/1/10s/2018-03-08T09:00:00Z,/Min), ... 330 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/1/10s/2018-03-08T09:00:00Z,/Min) I180308 09:28:20.209629 7061 storage/replica_raftstorage.go:520 [raftsnapshot,n1,s1,r5/1:/System/tsd{-/cr.nod…}] generated Raft snapshot a58aa6bc at index 74 I180308 09:28:20.318346 7061 storage/store.go:3745 [raftsnapshot,n1,s1,r5/1:/System/tsd{-/cr.nod…}] streamed snapshot to (n4,s4):4: kv pairs: 145, log entries: 19, rate-limit: 8.0 MiB/sec, 103ms I180308 09:28:20.321961 7122 storage/replica_raftstorage.go:726 [n4,s4,r5/4:{-}] applying Raft snapshot at index 74 (id=a58aa6bc, encoded size=39489, 1 rocksdb batches, 19 log entries) --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries I180308 09:28:20.675686 5734 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures I180308 09:28:20.676049 5734 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures W180308 09:28:20.841923 7144 storage/replica_backpressure.go:135 [n1,s1,r29/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.txn.aborts/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p90/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p75/3/10s/2018-03-08T09:00:00Z,/Min), ... 330 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/3/10s/2018-03-08T09:00:00Z,/Min) W180308 09:28:21.168305 2493 storage/node_liveness.go:501 [n5,hb] slow heartbeat took 2.5s W180308 09:28:21.335806 1028 storage/store.go:3866 [n1,s1] handle raft ready: 0.6s [processed=0] W180308 09:28:21.372717 7073 storage/replica_backpressure.go:135 [n5,s5,r29/3:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.txn.aborts/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p90/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p75/2/10s/2018-03-08T09:00:00Z,/Min), ... 330 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/2/10s/2018-03-08T09:00:00Z,/Min) W180308 09:28:21.566361 1553 storage/node_liveness.go:501 [n2,hb] slow heartbeat took 2.0s I180308 09:28:21.934342 7122 storage/replica_raftstorage.go:732 [n4,s4,r5/4:/System/tsd{-/cr.nod…}] applied Raft snapshot in 1612ms [clear=0ms batch=0ms entries=1609ms commit=2ms] --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries I180308 09:28:22.172925 5733 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures I180308 09:28:22.173248 5733 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures I180308 09:28:22.173462 835 sql/logictest/parallel_test.go:159 testdata/parallel_test/subquery_retry_multinode/test.yaml: run list 2: 0:final W180308 09:28:22.273947 1096 storage/node_liveness.go:501 [n1,hb] slow heartbeat took 1.1s W180308 09:28:22.317366 1806 storage/node_liveness.go:501 [n3,hb] slow heartbeat took 1.9s --- done: testdata/parallel_test/subquery_retry_multinode/final with config : 2 tests, 0 failures I180308 09:28:22.631831 7227 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/final with config : 2 tests, 0 failures I180308 09:28:22.633272 7200 util/stop/stopper.go:473 quiescing; tasks left: 1 ts.poller: poll I180308 09:28:22.636235 7199 util/stop/stopper.go:473 quiescing; tasks left: 1 ts.poller: poll W180308 09:28:22.636697 3779 storage/raft_transport.go:465 [n3] raft transport stream to node 4 failed: rpc error: code = Unavailable desc = transport is closing I180308 09:28:22.637631 7198 util/stop/stopper.go:473 quiescing; tasks left: 6 node.Node: batch 3 [async] kv.TxnCoordSender: heartbeat loop 2 [async] storage.split: processing replica 1 ts.poller: poll 1 [async] transport racer 1 [async] storage.replicate: processing replica W180308 09:28:22.638338 2119 storage/raft_transport.go:465 [n1] raft transport stream to node 2 failed: EOF W180308 09:28:22.645863 7000 internal/client/txn.go:531 [split,n1,s1,r29/1:/System/ts{d/cr.no…-e}] failure aborting transaction: node unavailable; try another peer; abort caused by: node unavailable; try another peer W180308 09:28:22.658964 1047 storage/store.go:1480 [n1,s1,r7/1:/Table/{SystemCon…-11}] could not gossip system config: node unavailable; try another peer W180308 09:28:22.659676 2336 storage/raft_transport.go:465 [n1] raft transport stream to node 4 failed: rpc error: code = Unavailable desc = transport is closing I180308 09:28:22.663545 7198 util/stop/stopper.go:473 quiescing; tasks left: 3 node.Node: batch 3 [async] kv.TxnCoordSender: heartbeat loop 1 ts.poller: poll 1 [async] transport racer 1 [async] storage.split: processing replica 1 [async] storage.replicate: processing replica W180308 09:28:22.666411 6668 internal/client/txn.go:531 [split,n1,s1,r28/1:/System/{StatusNo…-tsd}] failure aborting transaction: node unavailable; try another peer; abort caused by: result is ambiguous (server shutdown) I180308 09:28:22.669226 7198 util/stop/stopper.go:473 quiescing; tasks left: 3 [async] kv.TxnCoordSender: heartbeat loop 1 ts.poller: poll 1 node.Node: batch 1 [async] transport racer 1 [async] storage.split: processing replica 1 [async] storage.replicate: processing replica W180308 09:28:22.670449 1545 ts/db.go:154 [n2,ts-poll] error writing time series data: the batch experienced mixed success and failure W180308 09:28:22.679032 1545 ts/db.go:157 [n2,ts-poll] node unavailable; try another peer I180308 09:28:22.679527 7198 util/stop/stopper.go:473 quiescing; tasks left: 3 [async] kv.TxnCoordSender: heartbeat loop 1 ts.poller: poll 1 node.Node: batch 1 [async] transport racer 1 [async] storage.replicate: processing replica I180308 09:28:22.705284 1055 server/status/runtime.go:219 [n1] runtime stats: 1.3 GiB RSS, 1118 goroutines, 67 MiB/42 MiB/140 MiB GO alloc/idle/total, 158 MiB/197 MiB CGO alloc/total, 881.83cgo/sec, 0.89/0.09 %(u/s)time, 0.03 %gc (4x) W180308 09:28:22.706902 1048 storage/store.go:1480 [n1,s1,r3/1:/System/NodeLiveness{-Max}] could not gossip node liveness: node unavailable; try another peer I180308 09:28:22.707664 7201 util/stop/stopper.go:473 quiescing; tasks left: 1 ts.poller: poll I180308 09:28:22.710501 7330 util/stop/stopper.go:473 quiescing; tasks left: 1 ts.poller: poll 1 [async] storage.replicate: processing replica 1 [async] kv.TxnCoordSender: heartbeat loop W180308 09:28:22.711733 1047 storage/store.go:1480 [n1,s1,r7/1:/Table/{SystemCon…-11}] could not gossip system config: node unavailable; try another peer W180308 09:28:22.716304 1046 gossip/infostore.go:303 [n1] node unavailable; try another peer W180308 09:28:22.716688 1046 storage/store.go:1480 [n1,s1,r1/1:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer I180308 09:28:22.717708 1053 gossip/gossip.go:487 [n1] gossip status (ok, 5 nodes) gossip client (0/3 cur/max conns) gossip server (3/3 cur/max conns, infos 2303/150 sent/received, bytes 774296B/46012B sent/received) 3: 127.0.0.1:46629 (56s) 4: 127.0.0.1:38315 (55s) 2: 127.0.0.1:37987 (57s) W180308 09:28:22.720545 1053 gossip/infostore.go:303 [n1] node unavailable; try another peer W180308 09:28:22.722669 2373 storage/raft_transport.go:465 [n4] raft transport stream to node 1 failed: rpc error: code = Unavailable desc = transport is closing W180308 09:28:22.727955 2485 ts/db.go:154 [n5,ts-poll] error writing time series data: the batch experienced mixed success and failure W180308 09:28:22.728252 2485 ts/db.go:157 [n5,ts-poll] node unavailable; try another peer I180308 09:28:22.728512 7330 util/stop/stopper.go:473 quiescing; tasks left: 1 [async] storage.replicate: processing replica 1 [async] kv.TxnCoordSender: heartbeat loop W180308 09:28:22.732195 2770 storage/raft_transport.go:465 [n1] raft transport stream to node 5 failed: rpc error: code = Unavailable desc = transport is closing W180308 09:28:22.737226 4045 storage/raft_transport.go:465 [n2] raft transport stream to node 3 failed: EOF W180308 09:28:22.742157 1899 storage/raft_transport.go:465 [n1] raft transport stream to node 3 failed: rpc error: code = Unavailable desc = transport is closing W180308 09:28:22.747260 4042 storage/raft_transport.go:465 [n3] raft transport stream to node 2 failed: rpc error: code = FailedPrecondition desc = grpc: the client connection is closing W180308 09:28:22.760150 1970 storage/raft_transport.go:465 [n3] raft transport stream to node 1 failed: rpc error: code = Unavailable desc = transport is closing W180308 09:28:22.763583 2045 ts/db.go:154 [n4,ts-poll] error writing time series data: the batch experienced mixed success and failure W180308 09:28:22.770300 2045 ts/db.go:157 [n4,ts-poll] node unavailable; try another peer I180308 09:28:22.763970 750 kv/transport_race.go:67 transport race promotion: ran 45 iterations on up to 2482 requests W180308 09:28:22.764917 3208 storage/raft_transport.go:465 [n4] raft transport stream to node 5 failed: rpc error: code = FailedPrecondition desc = grpc: the client connection is closing I180308 09:28:22.772524 7198 util/stop/stopper.go:473 quiescing; tasks left: 3 [async] kv.TxnCoordSender: heartbeat loop 1 ts.poller: poll 1 node.Node: batch 1 [async] storage.replicate: processing replica W180308 09:28:22.775756 1518 gossip/gossip.go:1292 [n2] no incoming or outgoing connections W180308 09:28:22.776260 6938 kv/txn_coord_sender.go:1036 [split,s1,r28/1:/System/{StatusNo…-tsd},n1,txn=fcc1998d] node unavailable; try another peer W180308 09:28:22.777253 3848 storage/raft_transport.go:465 [n4] raft transport stream to node 2 failed: rpc error: code = Unavailable desc = transport is closing W180308 09:28:22.792605 2178 storage/raft_transport.go:465 [n2] raft transport stream to node 1 failed: rpc error: code = Unavailable desc = transport is closing W180308 09:28:22.793349 3922 storage/raft_transport.go:465 [n2] raft transport stream to node 4 failed: rpc error: code = Unavailable desc = transport is closing W180308 09:28:22.803019 1056 ts/db.go:154 [n1,ts-poll] error writing time series data: the batch experienced mixed success and failure W180308 09:28:22.812175 1056 ts/db.go:157 [n1,ts-poll] node unavailable; try another peer W180308 09:28:22.803505 1046 gossip/infostore.go:303 [n1] node unavailable; try another peer W180308 09:28:22.812587 1046 storage/store.go:1480 [n1,s1,r1/1:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer W180308 09:28:22.807111 1048 storage/store.go:1480 [n1,s1,r3/1:/System/NodeLiveness{-Max}] could not gossip node liveness: node unavailable; try another peer W180308 09:28:22.807396 1057 server/node.go:776 [n1,summaries] error recording status summaries: node unavailable; try another peer W180308 09:28:22.808727 7039 internal/client/txn.go:531 [replicate,n1,s1,r24/1:/Table/13/1/2018-03-08T09:27:…] failure aborting transaction: node unavailable; try another peer; abort caused by: node unavailable; try another peer E180308 09:28:22.813785 7039 storage/queue.go:778 [replicate,n1,s1,r24/1:/Table/13/1/2018-03-08T09:27:…] change replicas of r24 failed: node unavailable; try another peer W180308 09:28:22.809209 3173 storage/raft_transport.go:465 [n2] raft transport stream to node 5 failed: rpc error: code = Unavailable desc = transport is closing W180308 09:28:22.815172 7191 kv/txn_coord_sender.go:1036 [split,s1,r29/1:/System/ts{d/cr.no…-e},n1,txn=db46a2ed] node unavailable; try another peer W180308 09:28:22.818081 6931 internal/client/txn.go:531 [replicate,n5,s5,r17/4:/Table/2{0-1}] failure aborting transaction: node unavailable; try another peer; abort caused by: node unavailable; try another peer W180308 09:28:22.820510 2743 storage/raft_transport.go:465 [n5] raft transport stream to node 1 failed: EOF W180308 09:28:22.824887 2238 gossip/gossip.go:1292 [n5] no incoming or outgoing connections W180308 09:28:22.825931 1047 storage/store.go:1480 [n1,s1,r7/1:/Table/{SystemCon…-11}] could not gossip system config: node unavailable; try another peer W180308 09:28:22.835973 3199 storage/raft_transport.go:465 [n5] raft transport stream to node 4 failed: EOF W180308 09:28:22.837525 3335 storage/raft_transport.go:465 [n5] raft transport stream to node 3 failed: EOF W180308 09:28:22.842099 1748 gossip/gossip.go:1292 [n3] no incoming or outgoing connections W180308 09:28:22.842893 3245 storage/raft_transport.go:465 [n3] raft transport stream to node 5 failed: rpc error: code = FailedPrecondition desc = grpc: the client connection is closing W180308 09:28:22.845802 3832 storage/raft_transport.go:465 [n4] raft transport stream to node 3 failed: rpc error: code = Unavailable desc = transport is closing W180308 09:28:22.853934 1798 ts/db.go:154 [n3,ts-poll] error writing time series data: the batch experienced mixed success and failure W180308 09:28:22.855078 1979 gossip/gossip.go:1292 [n4] no incoming or outgoing connections W180308 09:28:22.856870 3102 storage/raft_transport.go:465 [n5] raft transport stream to node 2 failed: EOF W180308 09:28:22.890039 1798 ts/db.go:157 [n3,ts-poll] node unavailable; try another peer I180308 09:28:23.130857 6 sql/logictest/parallel_test.go:266 1 out of 2 parallel tests failed ```
1.0
sql/logictest: TestParallel failed under stress - SHA: https://github.com/cockroachdb/cockroach/commits/ea5f8b22ba951799b76c84502aaab4456fe00a46 Parameters: ``` TAGS= GOFLAGS=-race ``` Stress build found a failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=552372&tab=buildLog ``` I180308 09:27:49.787154 4678 storage/replica_raftstorage.go:732 [n3,s3,r9/?:/Table/1{2-3}] applied preemptive snapshot in 8ms [clear=0ms batch=0ms entries=7ms commit=0ms] I180308 09:27:49.902388 835 testutils/testcluster/testcluster.go:534 [n1,s1] has 1 underreplicated ranges W180308 09:27:49.965287 1035 storage/store.go:3866 [n1,s1] handle raft ready: 1.0s [processed=0] I180308 09:27:49.992899 4679 storage/replica.go:3209 [n1,s1,r3/1:/System/NodeLiveness{-Max}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n4,s4):2 (n2,s2):3] next=4 I180308 09:27:49.998836 4659 storage/replica_command.go:1760 [replicate,n2,s2,r9/2:/Table/1{2-3}] change replicas (ADD_REPLICA (n3,s3):4): read existing descriptor r9:/Table/1{2-3} [(n1,s1):1, (n2,s2):2, (n5,s5):3, next=4] W180308 09:27:50.033177 1806 storage/node_liveness.go:501 [n3,hb] slow heartbeat took 1.1s I180308 09:27:50.226728 835 testutils/testcluster/testcluster.go:534 [n1,s1] has 1 underreplicated ranges I180308 09:27:50.327811 4695 storage/replica_raftstorage.go:520 [replicate,n1,s1,r10/1:/Table/1{3-4}] generated preemptive snapshot ae1bb9ed at index 153 I180308 09:27:50.408223 4615 storage/store.go:2350 [replicaGC,n1,s1,r16/1:/Table/{19-20}] removing replica I180308 09:27:50.482741 4695 storage/store.go:3745 [replicate,n1,s1,r10/1:/Table/1{3-4}] streamed snapshot to (n5,s5):?: kv pairs: 407, log entries: 51, rate-limit: 2.0 MiB/sec, 154ms I180308 09:27:50.524357 4615 storage/replica.go:836 [replicaGC,n1,s1,r16/1:/Table/{19-20}] removed 11 (0+11) keys in 40ms [clear=0ms commit=40ms] I180308 09:27:50.738648 4697 storage/replica_raftstorage.go:726 [n5,s5,r10/?:{-}] applying preemptive snapshot at index 153 (id=ae1bb9ed, encoded size=81088, 1 rocksdb batches, 51 log entries) I180308 09:27:50.786926 4653 storage/replica.go:3209 [n4,s4,r18/2:/Table/2{1-2}] proposing ADD_REPLICA((n3,s3):4): updated=[(n1,s1):1 (n4,s4):2 (n5,s5):3 (n3,s3):4] next=5 I180308 09:27:50.817425 1748 gossip/gossip.go:1306 [n3] node has connected to cluster via gossip I180308 09:27:50.818926 1748 storage/stores.go:350 [n3] wrote 4 node addresses to persistent storage I180308 09:27:50.857137 4684 storage/replica.go:3209 [n3,s3,r13/2:/Table/1{6-7}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n4,s4):4 (n3,s3):2 (n5,s5):3] next=5 I180308 09:27:51.086023 4697 storage/replica_raftstorage.go:732 [n5,s5,r10/?:/Table/1{3-4}] applied preemptive snapshot in 347ms [clear=0ms batch=0ms entries=187ms commit=1ms] I180308 09:27:51.243460 4695 storage/replica_command.go:1760 [replicate,n1,s1,r10/1:/Table/1{3-4}] change replicas (ADD_REPLICA (n5,s5):4): read existing descriptor r10:/Table/1{3-4} [(n1,s1):1, (n3,s3):2, (n2,s2):3, next=4] I180308 09:27:51.381230 4761 storage/replica_command.go:1760 [replicate,n4,s4,r18/2:/Table/2{1-2}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r18:/Table/2{1-2} [(n1,s1):1, (n4,s4):2, (n5,s5):3, (n3,s3):4, next=5] I180308 09:27:51.509691 4735 storage/replica_raftstorage.go:520 [replicate,n3,s3,r6/2:/{System/tse-Table/System…}] generated preemptive snapshot 27aebd8b at index 30 I180308 09:27:51.563486 835 sql/logictest/parallel_test.go:213 Setting range split size: 32768 I180308 09:27:51.601539 4735 storage/store.go:3745 [replicate,n3,s3,r6/2:/{System/tse-Table/System…}] streamed snapshot to (n2,s2):?: kv pairs: 14, log entries: 20, rate-limit: 2.0 MiB/sec, 91ms I180308 09:27:51.619862 1901 storage/store.go:3479 [n1,s1,r13/1:/Table/1{6-7}] added to replica GC queue (peer suggestion) I180308 09:27:51.630538 4789 storage/replica_raftstorage.go:726 [n2,s2,r6/?:{-}] applying preemptive snapshot at index 30 (id=27aebd8b, encoded size=8130, 1 rocksdb batches, 20 log entries) I180308 09:27:51.728873 4789 storage/replica_raftstorage.go:732 [n2,s2,r6/?:/{System/tse-Table/System…}] applied preemptive snapshot in 98ms [clear=0ms batch=0ms entries=94ms commit=3ms] I180308 09:27:51.807803 4735 storage/replica_command.go:1760 [replicate,n3,s3,r6/2:/{System/tse-Table/System…}] change replicas (ADD_REPLICA (n2,s2):4): read existing descriptor r6:/{System/tse-Table/SystemConfigSpan/Start} [(n1,s1):1, (n3,s3):2, (n4,s4):3, next=4] I180308 09:27:51.828942 4767 storage/replica_raftstorage.go:520 [raftsnapshot,n4,s4,r18/2:/Table/2{1-2}] generated Raft snapshot df1ffd91 at index 31 I180308 09:27:51.834992 4767 storage/store.go:3745 [raftsnapshot,n4,s4,r18/2:/Table/2{1-2}] streamed snapshot to (n3,s3):4: kv pairs: 18, log entries: 21, rate-limit: 8.0 MiB/sec, 5ms I180308 09:27:51.837164 4834 storage/replica_raftstorage.go:726 [n3,s3,r18/4:{-}] applying Raft snapshot at index 31 (id=df1ffd91, encoded size=7716, 1 rocksdb batches, 21 log entries) I180308 09:27:51.842351 4834 storage/replica_raftstorage.go:732 [n3,s3,r18/4:/Table/2{1-2}] applied Raft snapshot in 5ms [clear=0ms batch=0ms entries=3ms commit=0ms] I180308 09:27:52.277701 4836 storage/store.go:2350 [replicaGC,n5,s5,r10/?:/Table/1{3-4}] removing replica I180308 09:27:52.302073 4671 storage/store.go:2350 [replicaGC,n1,s1,r13/1:/Table/1{6-7}] removing replica I180308 09:27:52.431403 4836 storage/replica.go:836 [replicaGC,n5,s5,r10/?:/Table/1{3-4}] removed 389 (379+10) keys in 153ms [clear=0ms commit=153ms] I180308 09:27:52.549535 4671 storage/replica.go:836 [replicaGC,n1,s1,r13/1:/Table/1{6-7}] removed 12 (0+12) keys in 219ms [clear=0ms commit=219ms] I180308 09:27:52.628009 4816 storage/replica.go:3209 [n2,s2,r9/2:/Table/1{2-3}] proposing ADD_REPLICA((n3,s3):4): updated=[(n1,s1):1 (n2,s2):2 (n5,s5):3 (n3,s3):4] next=5 I180308 09:27:52.699806 1055 server/status/runtime.go:219 [n1] runtime stats: 1.1 GiB RSS, 1117 goroutines, 76 MiB/16 MiB/119 MiB GO alloc/idle/total, 154 MiB/186 MiB CGO alloc/total, 862.81cgo/sec, 0.95/0.07 %(u/s)time, 0.05 %gc (5x) I180308 09:27:52.842796 4890 storage/replica_command.go:1760 [replicate,n2,s2,r9/2:/Table/1{2-3}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r9:/Table/1{2-3} [(n1,s1):1, (n2,s2):2, (n5,s5):3, (n3,s3):4, next=5] I180308 09:27:53.053227 4909 storage/replica.go:3209 [n3,s3,r6/2:/{System/tse-Table/System…}] proposing ADD_REPLICA((n2,s2):4): updated=[(n1,s1):1 (n3,s3):2 (n4,s4):3 (n2,s2):4] next=5 I180308 09:27:53.204274 4783 storage/replica_command.go:1760 [replicate,n3,s3,r6/2:/{System/tse-Table/System…}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r6:/{System/tse-Table/SystemConfigSpan/Start} [(n1,s1):1, (n3,s3):2, (n4,s4):3, (n2,s2):4, next=5] I180308 09:27:53.385647 4933 storage/replica.go:3209 [n4,s4,r18/2:/Table/2{1-2}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n3,s3):4 (n4,s4):2 (n5,s5):3] next=5 I180308 09:27:53.654566 835 sql/logictest/parallel_test.go:227 Creating database I180308 09:27:53.714521 1518 gossip/gossip.go:1306 [n2] node has connected to cluster via gossip I180308 09:27:53.812099 1518 storage/stores.go:350 [n2] wrote 4 node addresses to persistent storage I180308 09:27:53.949703 2370 storage/store.go:3479 [n1,s1,r18/1:/Table/2{1-2}] added to replica GC queue (peer suggestion) I180308 09:27:53.950132 2370 storage/store.go:3479 [n1,s1,r18/1:/Table/2{1-2}] added to replica GC queue (peer suggestion) I180308 09:27:54.187036 4918 storage/replica_command.go:838 [split,n1,s1,r5/1:/System/ts{d-e}] initiating a split of this range at key /System/tsd/cr.node.gossip.infos.sent/4/10s/2018-03-08T09:00:00Z [r21] I180308 09:27:54.490089 4921 storage/replica_command.go:838 [split,n1,s1,r4/1:/System/{NodeLive…-tsd}] initiating a split of this range at key /System/StatusNode/2 [r22] I180308 09:27:54.641807 4982 storage/replica_command.go:838 [split,n1,s1,r10/1:/Table/1{3-4}] initiating a split of this range at key /Table/13/1/2018-03-08T09:27:29.777287Z/329091808619495425 [r23] I180308 09:27:54.767000 4695 storage/replica.go:3209 [n1,s1,r10/1:/Table/1{3-4}] proposing ADD_REPLICA((n5,s5):4): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3 (n5,s5):4] next=5 I180308 09:27:54.924274 4936 storage/store.go:2350 [replicaGC,n1,s1,r18/1:/Table/2{1-2}] removing replica I180308 09:27:55.227444 4936 storage/replica.go:836 [replicaGC,n1,s1,r18/1:/Table/2{1-2}] removed 11 (0+11) keys in 51ms [clear=0ms commit=50ms] W180308 09:27:55.283305 4861 storage/replica.go:3016 [n1,s1,r10/1:/Table/1{3-4}] context canceled before proposing: 1 HeartbeatTxn I180308 09:27:55.373122 5042 storage/replica_command.go:838 [split,n1,s1,r10/1:/Table/1{3-4}] initiating a split of this range at key /Table/13/1/2018-03-08T09:27:29.777287Z/329091808619495425 [r24] I180308 09:27:55.591922 5003 storage/replica.go:3209 [n2,s2,r9/2:/Table/1{2-3}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n3,s3):4 (n2,s2):2 (n5,s5):3] next=5 I180308 09:27:55.659550 1544 server/status/runtime.go:219 [n2] runtime stats: 1.1 GiB RSS, 1126 goroutines, 57 MiB/35 MiB/120 MiB GO alloc/idle/total, 155 MiB/187 MiB CGO alloc/total, 798.50cgo/sec, 0.95/0.07 %(u/s)time, 0.03 %gc (5x) I180308 09:27:56.074863 2121 storage/store.go:3479 [n1,s1,r9/1:/Table/1{2-3}] added to replica GC queue (peer suggestion) I180308 09:27:56.304086 4874 sql/event_log.go:121 [n1,client=127.0.0.1:43858,user=root] Event: "create_database", target: 50, info: {DatabaseName:test Statement:CREATE DATABASE test User:root} I180308 09:27:56.307831 2121 storage/store.go:3479 [n1,s1,r9/1:/Table/1{2-3}] added to replica GC queue (peer suggestion) I180308 09:27:56.399786 1635 storage/replica.go:3277 [n2,s2,r2/3:/System/{-NodeLive…}] not quiescing: 1 pending commands I180308 09:27:56.412847 2121 storage/store.go:3479 [n1,s1,r9/1:/Table/1{2-3}] added to replica GC queue (peer suggestion) I180308 09:27:56.413411 2121 storage/store.go:3479 [n1,s1,r9/1:/Table/1{2-3}] added to replica GC queue (peer suggestion) I180308 09:27:56.429913 5110 storage/store.go:2350 [replicaGC,n1,s1,r9/1:/Table/1{2-3}] removing replica I180308 09:27:56.448755 1797 server/status/runtime.go:219 [n3] runtime stats: 1.1 GiB RSS, 1135 goroutines, 66 MiB/26 MiB/120 MiB GO alloc/idle/total, 156 MiB/187 MiB CGO alloc/total, 773.55cgo/sec, 0.94/0.07 %(u/s)time, 0.03 %gc (5x) I180308 09:27:56.688465 5110 storage/replica.go:836 [replicaGC,n1,s1,r9/1:/Table/1{2-3}] removed 71 (60+11) keys in 257ms [clear=0ms commit=257ms] I180308 09:27:57.234316 5009 storage/replica_command.go:838 [split,n4,s4,r20/2:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r41] I180308 09:27:57.715312 5007 storage/replica_raftstorage.go:520 [raftsnapshot,n1,s1,r10/1:/Table/1{3-4}] generated Raft snapshot 729bfb72 at index 173 W180308 09:27:57.734210 982 storage/store.go:3866 [n1,s1] handle raft ready: 0.6s [processed=0] I180308 09:27:57.771678 2044 server/status/runtime.go:219 [n4] runtime stats: 1.2 GiB RSS, 1138 goroutines, 51 MiB/40 MiB/120 MiB GO alloc/idle/total, 156 MiB/187 MiB CGO alloc/total, 717.70cgo/sec, 0.95/0.07 %(u/s)time, 0.03 %gc (5x) I180308 09:27:57.853130 5007 storage/store.go:3745 [raftsnapshot,n1,s1,r10/1:/Table/1{3-4}] streamed snapshot to (n5,s5):4: kv pairs: 456, log entries: 71, rate-limit: 8.0 MiB/sec, 104ms I180308 09:27:57.904274 5196 storage/replica_raftstorage.go:726 [n5,s5,r10/4:{-}] applying Raft snapshot at index 173 (id=729bfb72, encoded size=103834, 1 rocksdb batches, 71 log entries) W180308 09:27:58.307548 1553 storage/node_liveness.go:501 [n2,hb] slow heartbeat took 1.2s I180308 09:27:58.890124 5216 storage/replica.go:3209 [n3,s3,r6/2:/{System/tse-Table/System…}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n2,s2):4 (n3,s3):2 (n4,s4):3] next=5 I180308 09:27:59.234979 2484 server/status/runtime.go:219 [n5] runtime stats: 1.2 GiB RSS, 1139 goroutines, 78 MiB/15 MiB/120 MiB GO alloc/idle/total, 156 MiB/188 MiB CGO alloc/total, 720.50cgo/sec, 0.94/0.06 %(u/s)time, 0.03 %gc (5x) I180308 09:27:59.534800 1901 storage/store.go:3479 [n1,s1,r6/1:/{System/tse-Table/System…}] added to replica GC queue (peer suggestion) I180308 09:27:59.617241 1901 storage/store.go:3479 [n1,s1,r6/1:/{System/tse-Table/System…}] added to replica GC queue (peer suggestion) W180308 09:27:59.658775 1806 storage/node_liveness.go:501 [n3,hb] slow heartbeat took 1.7s I180308 09:27:59.883988 835 sql/logictest/parallel_test.go:236 Test setup done I180308 09:27:59.884248 835 sql/logictest/parallel_test.go:159 testdata/parallel_test/subquery_retry_multinode/test.yaml: run list 0: 0:setup W180308 09:28:00.194433 1096 storage/node_liveness.go:501 [n1,hb] slow heartbeat took 1.4s I180308 09:28:00.297625 5240 storage/store.go:2350 [replicaGC,n1,s1,r6/1:/{System/tse-Table/System…}] removing replica I180308 09:28:00.329573 1028 storage/replica_proposal.go:201 [n1,s1,r4/1:/System/{NodeLive…-tsd}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0 I180308 09:28:00.379562 5248 storage/replica_command.go:838 [split,n1,s1,r22/1:/System/{StatusNo…-tsd}] initiating a split of this range at key /System/StatusNode/3 [r25] I180308 09:28:00.448234 5240 storage/replica.go:836 [replicaGC,n1,s1,r6/1:/{System/tse-Table/System…}] removed 12 (0+12) keys in 91ms [clear=0ms commit=91ms] W180308 09:28:00.707751 2147 storage/node_liveness.go:501 [n4,hb] slow heartbeat took 1.4s W180308 09:28:00.731795 5352 storage/replica_backpressure.go:135 [n1,s1,r5/1:/System/ts{d-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.security.certificate.expiration.ca/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.security.certificate.expiration.node/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.connections.incoming/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.connections.refused/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.bytes.received/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.bytes.sent/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.infos.received/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.infos.sent/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.connections.outgoing/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.batches/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.batches.partial/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.rpc.sent/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.rpc.sent.local/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.rpc.sent.nextreplicaerror/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.errors.notleaseholder/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.requests.slow.distsender/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.aborts/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/1/10s/2018-03-08T09:00:00Z,/Min), ... 405 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/1/10s/2018-03-08T09:00:00Z,/Min) I180308 09:28:01.257194 983 storage/replica_proposal.go:201 [n1,s1,r10/1:/Table/1{3-4}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.325864565,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.325864565,0 I180308 09:28:01.263525 5297 storage/replica_command.go:838 [split,n1,s1,r24/1:/Table/1{3/1/2018…-4}] initiating a split of this range at key /Table/13/1/2018-03-08T09:27:40.666166Z/329091846156877826 [r26] I180308 09:28:01.312145 5196 storage/replica_raftstorage.go:732 [n5,s5,r10/4:/Table/1{3-4}] applied Raft snapshot in 3408ms [clear=0ms batch=0ms entries=3374ms commit=21ms] I180308 09:28:01.454527 2220 storage/replica_proposal.go:201 [n4,s4,r20/2:/{Table/23-Max}] new range lease repl=(n4,s4):2 seq=3 start=1520501251.624682287,1 epo=1 pro=1520501255.402825609,0 following repl=(n4,s4):2 seq=3 start=1520501251.624682287,1 epo=1 pro=1520501255.402825609,0 I180308 09:28:01.482106 1005 storage/replica_proposal.go:201 [n1,s1,r5/1:/System/ts{d-e}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0 W180308 09:28:01.575951 5351 storage/replica.go:3016 [n1,s1,r10/1:/Table/13{-/1/2018-…}] context canceled before proposing: 1 HeartbeatTxn I180308 09:28:01.658992 5386 storage/replica_command.go:838 [split,n1,s1,r21/1:/System/ts{d/cr.no…-e}] initiating a split of this range at key /System/tsd/cr.node.sql.bytesin/3/10s/2018-03-08T09:00:00Z [r27] I180308 09:28:01.678252 5296 storage/replica_command.go:1760 [replicate,n1,s1,r10/1:/Table/13{-/1/2018-…}] change replicas (REMOVE_REPLICA (n3,s3):2): read existing descriptor r10:/Table/13{-/1/2018-03-08T09:27:29.777287Z/329091808619495425} [(n1,s1):1, (n3,s3):2, (n2,s2):3, (n5,s5):4, next=5] I180308 09:28:01.932018 5423 storage/replica_raftstorage.go:520 [replicate,n4,s4,r41/2:/{Table/50-Max}] generated preemptive snapshot e7efcdd4 at index 10 I180308 09:28:01.954883 5491 storage/replica_raftstorage.go:520 [replicate,n3,s3,r17/2:/Table/2{0-1}] generated preemptive snapshot b7e6a194 at index 36 E180308 09:28:02.175447 5423 storage/queue.go:778 [replicate,n4,s4,r41/2:/{Table/50-Max}] raft status not initialized I180308 09:28:02.188640 4874 sql/event_log.go:121 [n1,client=127.0.0.1:43858,user=root] Event: "create_table", target: 51, info: {TableName:test.public.t Statement:CREATE TABLE t (k INT, filler STRING) User:root} --- progress: testdata/parallel_test/subquery_retry_multinode/setup: 1 statements/queries I180308 09:28:02.231950 5276 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/setup: 1 statements/queries I180308 09:28:02.309261 5491 storage/store.go:3745 [replicate,n3,s3,r17/2:/Table/2{0-1}] streamed snapshot to (n1,s1):?: kv pairs: 16, log entries: 26, rate-limit: 2.0 MiB/sec, 97ms I180308 09:28:02.727426 5515 storage/replica_command.go:838 [split,n4,s4,r41/2:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r42] I180308 09:28:02.767793 1979 gossip/gossip.go:1306 [n4] node has connected to cluster via gossip I180308 09:28:02.791521 1055 server/status/runtime.go:219 [n1] runtime stats: 1.2 GiB RSS, 1166 goroutines, 65 MiB/26 MiB/120 MiB GO alloc/idle/total, 157 MiB/187 MiB CGO alloc/total, 1065.63cgo/sec, 0.94/0.07 %(u/s)time, 0.02 %gc (5x) I180308 09:28:03.002229 5486 storage/replica_raftstorage.go:726 [n1,s1,r17/?:{-}] applying preemptive snapshot at index 36 (id=b7e6a194, encoded size=8224, 1 rocksdb batches, 26 log entries) I180308 09:28:03.005301 1979 storage/stores.go:350 [n4] wrote 4 node addresses to persistent storage I180308 09:28:03.154898 5486 storage/replica_raftstorage.go:732 [n1,s1,r17/?:/Table/2{0-1}] applied preemptive snapshot in 152ms [clear=0ms batch=0ms entries=127ms commit=12ms] W180308 09:28:03.190672 5455 storage/replica_backpressure.go:135 [n2,s2,r21/2:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.security.certificate.expiration.ca/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.security.certificate.expiration.node/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.requests.slow.distsender/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.aborts/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), ... 379 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/2/10s/2018-03-08T09:00:00Z,/Min) I180308 09:28:03.513617 5491 storage/replica_command.go:1760 [replicate,n3,s3,r17/2:/Table/2{0-1}] change replicas (ADD_REPLICA (n1,s1):5): read existing descriptor r17:/Table/2{0-1} [(n5,s5):4, (n3,s3):2, (n2,s2):3, next=5] W180308 09:28:04.034472 5607 storage/replica_backpressure.go:135 [n1,s1,r21/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.security.certificate.expiration.ca/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.security.certificate.expiration.node/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.infos.sent/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.requests.slow.distsender/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.aborts/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/4/10s/2018-03-08T09:00:00Z,/Min), ... 380 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/4/10s/2018-03-08T09:00:00Z,/Min) W180308 09:28:04.459926 5676 storage/replica_backpressure.go:135 [n1,s1,r21/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.security.certificate.expiration.ca/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.security.certificate.expiration.node/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.infos.sent/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.requests.slow.distsender/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.aborts/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/5/10s/2018-03-08T09:00:00Z,/Min), ... 380 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/5/10s/2018-03-08T09:00:00Z,/Min) --- progress: testdata/parallel_test/subquery_retry_multinode/setup: 2 statements/queries I180308 09:28:04.488626 5276 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/setup: 2 statements/queries --- done: testdata/parallel_test/subquery_retry_multinode/setup with config : 2 tests, 0 failures I180308 09:28:04.488940 5276 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/setup with config : 2 tests, 0 failures I180308 09:28:04.494592 835 sql/logictest/parallel_test.go:159 testdata/parallel_test/subquery_retry_multinode/test.yaml: run list 1: 0:txn, 1:txn, 2:txn, 3:txn, 4:txn I180308 09:28:04.692538 5642 storage/replica.go:3209 [n1,s1,r10/1:/Table/13{-/1/2018-…}] proposing REMOVE_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n5,s5):4 (n2,s2):3] next=5 I180308 09:28:05.177727 1003 storage/replica_proposal.go:201 [n1,s1,r22/1:/System/{StatusNo…-tsd}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0 I180308 09:28:05.184170 1972 storage/store.go:3479 [n3,s3,r10/2:/Table/13{-/1/2018-…}] added to replica GC queue (peer suggestion) I180308 09:28:05.275031 5772 storage/replica_command.go:838 [split,n1,s1,r25/1:/System/{StatusNo…-tsd}] initiating a split of this range at key /System/StatusNode/4 [r28] I180308 09:28:05.629470 1544 server/status/runtime.go:219 [n2] runtime stats: 1.2 GiB RSS, 1171 goroutines, 87 MiB/11 MiB/125 MiB GO alloc/idle/total, 157 MiB/189 MiB CGO alloc/total, 1173.83cgo/sec, 0.94/0.07 %(u/s)time, 0.02 %gc (4x) I180308 09:28:05.634410 1827 storage/replica_proposal.go:201 [n3,s3,r4/2:/System/{NodeLive…-StatusNo…}] new range lease repl=(n3,s3):2 seq=4 start=1520501285.016867674,0 epo=1 pro=1520501285.016882538,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0 W180308 09:28:06.031522 5844 storage/replica.go:3016 [n1,s1,r22/1:/System/StatusNode/{2-3}] context canceled before proposing: 1 HeartbeatTxn --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries I180308 09:28:06.562121 5731 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures I180308 09:28:06.562430 5731 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures I180308 09:28:06.630349 1797 server/status/runtime.go:219 [n3] runtime stats: 1.2 GiB RSS, 1178 goroutines, 97 MiB/5.9 MiB/129 MiB GO alloc/idle/total, 158 MiB/189 MiB CGO alloc/total, 1136.76cgo/sec, 0.94/0.07 %(u/s)time, 0.03 %gc (5x) W180308 09:28:06.862124 2493 storage/node_liveness.go:501 [n5,hb] slow heartbeat took 1.7s W180308 09:28:07.341217 1553 storage/node_liveness.go:501 [n2,hb] slow heartbeat took 1.2s I180308 09:28:07.775014 2044 server/status/runtime.go:219 [n4] runtime stats: 1.2 GiB RSS, 1174 goroutines, 75 MiB/24 MiB/129 MiB GO alloc/idle/total, 157 MiB/189 MiB CGO alloc/total, 1173.62cgo/sec, 0.94/0.07 %(u/s)time, 0.03 %gc (4x) I180308 09:28:07.899393 993 storage/replica_proposal.go:201 [n1,s1,r21/1:/System/ts{d/cr.no…-e}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0 I180308 09:28:08.092040 6071 storage/replica_command.go:838 [split,n1,s1,r27/1:/System/ts{d/cr.no…-e}] initiating a split of this range at key /System/tsd/cr.node.sql.distsql.service.latency-p99.999/1/10s/2018-03-08T09:00:00Z [r29] W180308 09:28:09.182946 1850 storage/store.go:3866 [n3,s3] handle raft ready: 0.7s [processed=0] I180308 09:28:09.228630 2484 server/status/runtime.go:219 [n5] runtime stats: 1.3 GiB RSS, 1216 goroutines, 96 MiB/9.2 MiB/132 MiB GO alloc/idle/total, 160 MiB/191 MiB CGO alloc/total, 1213.47cgo/sec, 0.94/0.07 %(u/s)time, 0.03 %gc (5x) W180308 09:28:09.431040 987 storage/store.go:3866 [n1,s1] handle raft ready: 0.6s [processed=0] W180308 09:28:09.702153 1029 storage/store.go:3866 [n1,s1] handle raft ready: 0.5s [processed=0] I180308 09:28:09.884861 6241 storage/replica_raftstorage.go:520 [replicate,n1,s1,r5/1:/System/tsd{-/cr.nod…}] generated preemptive snapshot c879f910 at index 68 W180308 09:28:09.920664 1096 storage/node_liveness.go:501 [n1,hb] slow heartbeat took 2.2s I180308 09:28:09.926471 6051 storage/store.go:2350 [replicaGC,n1,s1,r17/?:/Table/2{0-1}] removing replica I180308 09:28:10.164342 6051 storage/replica.go:836 [replicaGC,n1,s1,r17/?:/Table/2{0-1}] removed 11 (0+11) keys in 177ms [clear=0ms commit=177ms] I180308 09:28:10.208745 6080 storage/replica_raftstorage.go:520 [replicate,n2,s2,r2/3:/System/{-NodeLive…}] generated preemptive snapshot c8cdd480 at index 34 I180308 09:28:10.482965 6080 storage/store.go:3745 [replicate,n2,s2,r2/3:/System/{-NodeLive…}] streamed snapshot to (n5,s5):?: kv pairs: 13, log entries: 24, rate-limit: 2.0 MiB/sec, 273ms I180308 09:28:10.627902 6241 storage/store.go:3745 [replicate,n1,s1,r5/1:/System/tsd{-/cr.nod…}] streamed snapshot to (n4,s4):?: kv pairs: 143, log entries: 13, rate-limit: 2.0 MiB/sec, 677ms I180308 09:28:10.693165 6287 storage/replica_raftstorage.go:726 [n5,s5,r2/?:{-}] applying preemptive snapshot at index 34 (id=c8cdd480, encoded size=56977, 1 rocksdb batches, 24 log entries) I180308 09:28:10.790504 6345 storage/replica_raftstorage.go:726 [n4,s4,r5/?:{-}] applying preemptive snapshot at index 68 (id=c879f910, encoded size=36756, 1 rocksdb batches, 13 log entries) W180308 09:28:10.795289 6380 storage/replica_backpressure.go:135 [n1,s1,r27/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.txn.aborts/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p90/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p75/3/10s/2018-03-08T09:00:00Z,/Min), ... 355 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/3/10s/2018-03-08T09:00:00Z,/Min) W180308 09:28:10.836796 1826 storage/store.go:3866 [n3,s3] handle raft ready: 0.5s [processed=0] W180308 09:28:10.963543 2147 storage/node_liveness.go:501 [n4,hb] slow heartbeat took 2.7s W180308 09:28:11.405336 6410 storage/replica_backpressure.go:135 [n1,s1,r27/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.txn.aborts/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p90/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p75/2/10s/2018-03-08T09:00:00Z,/Min), ... 354 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/2/10s/2018-03-08T09:00:00Z,/Min) I180308 09:28:11.779388 987 storage/replica_proposal.go:201 [n1,s1,r24/1:/Table/1{3/1/2018…-4}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.325864565,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.325864565,0 W180308 09:28:11.899345 2493 storage/node_liveness.go:501 [n5,hb] slow heartbeat took 2.2s W180308 09:28:12.551481 1553 storage/node_liveness.go:501 [n2,hb] slow heartbeat took 1.9s I180308 09:28:12.571934 6345 storage/replica_raftstorage.go:732 [n4,s4,r5/?:/System/tsd{-/cr.nod…}] applied preemptive snapshot in 1781ms [clear=0ms batch=0ms entries=1729ms commit=33ms] I180308 09:28:12.630942 6241 storage/replica_command.go:1760 [replicate,n1,s1,r5/1:/System/tsd{-/cr.nod…}] change replicas (ADD_REPLICA (n4,s4):4): read existing descriptor r5:/System/tsd{-/cr.node.gossip.infos.sent/4/10s/2018-03-08T09:00:00Z} [(n1,s1):1, (n2,s2):2, (n5,s5):3, next=4] W180308 09:28:12.654684 1806 storage/node_liveness.go:501 [n3,hb] slow heartbeat took 1.2s I180308 09:28:12.667081 1055 server/status/runtime.go:219 [n1] runtime stats: 1.3 GiB RSS, 1182 goroutines, 78 MiB/23 MiB/132 MiB GO alloc/idle/total, 160 MiB/194 MiB CGO alloc/total, 934.67cgo/sec, 0.92/0.07 %(u/s)time, 0.02 %gc (4x) I180308 09:28:13.552826 6287 storage/replica_raftstorage.go:732 [n5,s5,r2/?:/System/{-NodeLive…}] applied preemptive snapshot in 2833ms [clear=0ms batch=0ms entries=2609ms commit=164ms] I180308 09:28:13.769924 6080 storage/replica_command.go:1760 [replicate,n2,s2,r2/3:/System/{-NodeLive…}] change replicas (ADD_REPLICA (n5,s5):4): read existing descriptor r2:/System/{-NodeLiveness} [(n1,s1):1, (n3,s3):2, (n2,s2):3, next=4] W180308 09:28:13.950338 2147 storage/node_liveness.go:501 [n4,hb] slow heartbeat took 1.2s I180308 09:28:14.047030 6643 storage/replica.go:3209 [n3,s3,r17/2:/Table/2{0-1}] proposing ADD_REPLICA((n1,s1):5): updated=[(n5,s5):4 (n3,s3):2 (n2,s2):3 (n1,s1):5] next=6 --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries I180308 09:28:14.991532 5735 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures I180308 09:28:14.991818 5735 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures W180308 09:28:15.210739 2493 storage/node_liveness.go:501 [n5,hb] slow heartbeat took 1.1s I180308 09:28:15.459366 2539 storage/replica_proposal.go:201 [n5,s5,r17/4:/Table/2{0-1}] new range lease repl=(n5,s5):4 seq=4 start=1520501294.765769961,0 epo=1 pro=1520501294.765803960,0 following repl=(n3,s3):2 seq=3 start=1520501251.624682287,1 epo=1 pro=1520501257.825367317,0 I180308 09:28:15.554730 6653 storage/replica_raftstorage.go:520 [raftsnapshot,n3,s3,r17/2:/Table/2{0-1}] generated Raft snapshot 901a1124 at index 49 I180308 09:28:15.697368 1544 server/status/runtime.go:219 [n2] runtime stats: 1.3 GiB RSS, 1181 goroutines, 73 MiB/35 MiB/138 MiB GO alloc/idle/total, 160 MiB/195 MiB CGO alloc/total, 840.48cgo/sec, 0.90/0.07 %(u/s)time, 0.02 %gc (4x) I180308 09:28:15.764467 6653 storage/store.go:3745 [raftsnapshot,n3,s3,r17/2:/Table/2{0-1}] streamed snapshot to (n1,s1):5: kv pairs: 18, log entries: 39, rate-limit: 8.0 MiB/sec, 61ms I180308 09:28:15.778517 6602 storage/replica_raftstorage.go:726 [n1,s1,r17/5:{-}] applying Raft snapshot at index 49 (id=901a1124, encoded size=12107, 1 rocksdb batches, 39 log entries) I180308 09:28:16.177052 6521 storage/store.go:2350 [replicaGC,n4,s4,r5/?:/System/tsd{-/cr.nod…}] removing replica I180308 09:28:16.294600 6521 storage/replica.go:836 [replicaGC,n4,s4,r5/?:/System/tsd{-/cr.nod…}] removed 139 (128+11) keys in 117ms [clear=0ms commit=116ms] W180308 09:28:16.347616 1553 storage/node_liveness.go:501 [n2,hb] slow heartbeat took 1.3s I180308 09:28:16.372076 994 storage/replica_proposal.go:201 [n1,s1,r25/1:/System/{StatusNo…-tsd}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0 I180308 09:28:16.421976 5631 storage/store.go:2350 [replicaGC,n3,s3,r10/2:/Table/13{-/1/2018-…}] removing replica I180308 09:28:16.665522 1797 server/status/runtime.go:219 [n3] runtime stats: 1.3 GiB RSS, 1171 goroutines, 88 MiB/20 MiB/138 MiB GO alloc/idle/total, 162 MiB/196 MiB CGO alloc/total, 884.78cgo/sec, 0.89/0.07 %(u/s)time, 0.02 %gc (3x) I180308 09:28:16.681548 6668 storage/replica_command.go:838 [split,n1,s1,r28/1:/System/{StatusNo…-tsd}] initiating a split of this range at key /System/StatusNode/5 [r30] I180308 09:28:16.801652 6602 storage/replica_raftstorage.go:732 [n1,s1,r17/5:/Table/2{0-1}] applied Raft snapshot in 1003ms [clear=2ms batch=0ms entries=991ms commit=8ms] I180308 09:28:16.806835 5631 storage/replica.go:836 [replicaGC,n3,s3,r10/2:/Table/13{-/1/2018-…}] removed 181 (169+12) keys in 384ms [clear=0ms commit=384ms] W180308 09:28:17.412329 1806 storage/node_liveness.go:501 [n3,hb] slow heartbeat took 1.5s I180308 09:28:17.506774 6629 storage/replica_consistency.go:129 [replica consistency checker,n1,s1,r21/1:/System/tsd/cr.node.{go…-sq…}] triggering stats recomputation to resolve delta of {ContainsEstimates:true LastUpdateNanos:1520501288989219219 IntentAge:0 GCBytesAge:0 LiveBytes:4092 LiveCount:0 KeyBytes:0 KeyCount:0 ValBytes:4092 ValCount:0 IntentBytes:0 IntentCount:0 SysBytes:0 SysCount:0} I180308 09:28:17.510115 2214 storage/replica_proposal.go:201 [n4,s4,r41/2:/{Table/50-Max}] new range lease repl=(n4,s4):2 seq=3 start=1520501251.624682287,1 epo=1 pro=1520501255.402825609,0 following repl=(n4,s4):2 seq=3 start=1520501251.624682287,1 epo=1 pro=1520501255.402825609,0 I180308 09:28:17.785167 2044 server/status/runtime.go:219 [n4] runtime stats: 1.3 GiB RSS, 1172 goroutines, 62 MiB/44 MiB/138 MiB GO alloc/idle/total, 162 MiB/195 MiB CGO alloc/total, 981.39cgo/sec, 0.90/0.07 %(u/s)time, 0.02 %gc (4x) I180308 09:28:17.789384 6779 storage/replica.go:3209 [n2,s2,r2/3:/System/{-NodeLive…}] proposing ADD_REPLICA((n5,s5):4): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3 (n5,s5):4] next=5 W180308 09:28:17.792842 6603 storage/replica.go:3016 [n4,s4,r41/2:/Table/5{0-1}] context canceled before proposing: 1 HeartbeatTxn W180308 09:28:18.718662 1096 storage/node_liveness.go:501 [n1,hb] slow heartbeat took 2.0s I180308 09:28:18.788824 6931 storage/replica_command.go:1760 [replicate,n5,s5,r17/4:/Table/2{0-1}] change replicas (REMOVE_REPLICA (n1,s1):5): read existing descriptor r17:/Table/2{0-1} [(n5,s5):4, (n3,s3):2, (n2,s2):3, (n1,s1):5, next=6] I180308 09:28:19.053142 1010 storage/replica_proposal.go:201 [n1,s1,r27/1:/System/ts{d/cr.no…-e}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0 E180308 09:28:19.095691 5732 sql/logictest/logic.go:2124 testdata/parallel_test/subquery_retry_multinode/txn:9: expected success, but found pq: TransactionStatusError: does not exist --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 0 tests, 1 failures I180308 09:28:19.153442 5732 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 0 tests, 1 failures E180308 09:28:19.153715 5732 sql/logictest/parallel_test.go:84 error processing testdata/parallel_test/subquery_retry_multinode/txn: pq: TransactionStatusError: does not exist I180308 09:28:19.146805 2484 server/status/runtime.go:219 [n5] runtime stats: 1.3 GiB RSS, 1162 goroutines, 95 MiB/13 MiB/138 MiB GO alloc/idle/total, 161 MiB/196 MiB CGO alloc/total, 1029.88cgo/sec, 0.90/0.08 %(u/s)time, 0.02 %gc (3x) W180308 09:28:19.277944 2147 storage/node_liveness.go:501 [n4,hb] slow heartbeat took 2.0s I180308 09:28:19.513328 7000 storage/replica_command.go:838 [split,n1,s1,r29/1:/System/ts{d/cr.no…-e}] initiating a split of this range at key /System/tsd/cr.node.sql.mem.admin.session.max-p99.9/1/10s/2018-03-08T09:00:00Z [r31] I180308 09:28:19.527756 6832 storage/replica.go:3209 [n1,s1,r5/1:/System/tsd{-/cr.nod…}] proposing ADD_REPLICA((n4,s4):4): updated=[(n1,s1):1 (n2,s2):2 (n5,s5):3 (n4,s4):4] next=5 I180308 09:28:20.040589 7039 storage/replica_command.go:1760 [replicate,n1,s1,r24/1:/Table/13/1/2018-03-08T09:27:…] change replicas (REMOVE_REPLICA (n2,s2):3): read existing descriptor r24:/Table/13/1/2018-03-08T09:27:{29.777287Z/329091808619495425-40.666166Z/329091846156877826} [(n1,s1):1, (n3,s3):2, (n2,s2):3, (n5,s5):4, next=5] W180308 09:28:20.196189 7047 storage/replica_backpressure.go:135 [n1,s1,r29/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.txn.aborts/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p90/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p75/1/10s/2018-03-08T09:00:00Z,/Min), ... 330 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/1/10s/2018-03-08T09:00:00Z,/Min) I180308 09:28:20.209629 7061 storage/replica_raftstorage.go:520 [raftsnapshot,n1,s1,r5/1:/System/tsd{-/cr.nod…}] generated Raft snapshot a58aa6bc at index 74 I180308 09:28:20.318346 7061 storage/store.go:3745 [raftsnapshot,n1,s1,r5/1:/System/tsd{-/cr.nod…}] streamed snapshot to (n4,s4):4: kv pairs: 145, log entries: 19, rate-limit: 8.0 MiB/sec, 103ms I180308 09:28:20.321961 7122 storage/replica_raftstorage.go:726 [n4,s4,r5/4:{-}] applying Raft snapshot at index 74 (id=a58aa6bc, encoded size=39489, 1 rocksdb batches, 19 log entries) --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries I180308 09:28:20.675686 5734 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures I180308 09:28:20.676049 5734 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures W180308 09:28:20.841923 7144 storage/replica_backpressure.go:135 [n1,s1,r29/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.txn.aborts/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p90/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p75/3/10s/2018-03-08T09:00:00Z,/Min), ... 330 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/3/10s/2018-03-08T09:00:00Z,/Min) W180308 09:28:21.168305 2493 storage/node_liveness.go:501 [n5,hb] slow heartbeat took 2.5s W180308 09:28:21.335806 1028 storage/store.go:3866 [n1,s1] handle raft ready: 0.6s [processed=0] W180308 09:28:21.372717 7073 storage/replica_backpressure.go:135 [n5,s5,r29/3:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.txn.aborts/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p90/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p75/2/10s/2018-03-08T09:00:00Z,/Min), ... 330 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/2/10s/2018-03-08T09:00:00Z,/Min) W180308 09:28:21.566361 1553 storage/node_liveness.go:501 [n2,hb] slow heartbeat took 2.0s I180308 09:28:21.934342 7122 storage/replica_raftstorage.go:732 [n4,s4,r5/4:/System/tsd{-/cr.nod…}] applied Raft snapshot in 1612ms [clear=0ms batch=0ms entries=1609ms commit=2ms] --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries I180308 09:28:22.172925 5733 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures I180308 09:28:22.173248 5733 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures I180308 09:28:22.173462 835 sql/logictest/parallel_test.go:159 testdata/parallel_test/subquery_retry_multinode/test.yaml: run list 2: 0:final W180308 09:28:22.273947 1096 storage/node_liveness.go:501 [n1,hb] slow heartbeat took 1.1s W180308 09:28:22.317366 1806 storage/node_liveness.go:501 [n3,hb] slow heartbeat took 1.9s --- done: testdata/parallel_test/subquery_retry_multinode/final with config : 2 tests, 0 failures I180308 09:28:22.631831 7227 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/final with config : 2 tests, 0 failures I180308 09:28:22.633272 7200 util/stop/stopper.go:473 quiescing; tasks left: 1 ts.poller: poll I180308 09:28:22.636235 7199 util/stop/stopper.go:473 quiescing; tasks left: 1 ts.poller: poll W180308 09:28:22.636697 3779 storage/raft_transport.go:465 [n3] raft transport stream to node 4 failed: rpc error: code = Unavailable desc = transport is closing I180308 09:28:22.637631 7198 util/stop/stopper.go:473 quiescing; tasks left: 6 node.Node: batch 3 [async] kv.TxnCoordSender: heartbeat loop 2 [async] storage.split: processing replica 1 ts.poller: poll 1 [async] transport racer 1 [async] storage.replicate: processing replica W180308 09:28:22.638338 2119 storage/raft_transport.go:465 [n1] raft transport stream to node 2 failed: EOF W180308 09:28:22.645863 7000 internal/client/txn.go:531 [split,n1,s1,r29/1:/System/ts{d/cr.no…-e}] failure aborting transaction: node unavailable; try another peer; abort caused by: node unavailable; try another peer W180308 09:28:22.658964 1047 storage/store.go:1480 [n1,s1,r7/1:/Table/{SystemCon…-11}] could not gossip system config: node unavailable; try another peer W180308 09:28:22.659676 2336 storage/raft_transport.go:465 [n1] raft transport stream to node 4 failed: rpc error: code = Unavailable desc = transport is closing I180308 09:28:22.663545 7198 util/stop/stopper.go:473 quiescing; tasks left: 3 node.Node: batch 3 [async] kv.TxnCoordSender: heartbeat loop 1 ts.poller: poll 1 [async] transport racer 1 [async] storage.split: processing replica 1 [async] storage.replicate: processing replica W180308 09:28:22.666411 6668 internal/client/txn.go:531 [split,n1,s1,r28/1:/System/{StatusNo…-tsd}] failure aborting transaction: node unavailable; try another peer; abort caused by: result is ambiguous (server shutdown) I180308 09:28:22.669226 7198 util/stop/stopper.go:473 quiescing; tasks left: 3 [async] kv.TxnCoordSender: heartbeat loop 1 ts.poller: poll 1 node.Node: batch 1 [async] transport racer 1 [async] storage.split: processing replica 1 [async] storage.replicate: processing replica W180308 09:28:22.670449 1545 ts/db.go:154 [n2,ts-poll] error writing time series data: the batch experienced mixed success and failure W180308 09:28:22.679032 1545 ts/db.go:157 [n2,ts-poll] node unavailable; try another peer I180308 09:28:22.679527 7198 util/stop/stopper.go:473 quiescing; tasks left: 3 [async] kv.TxnCoordSender: heartbeat loop 1 ts.poller: poll 1 node.Node: batch 1 [async] transport racer 1 [async] storage.replicate: processing replica I180308 09:28:22.705284 1055 server/status/runtime.go:219 [n1] runtime stats: 1.3 GiB RSS, 1118 goroutines, 67 MiB/42 MiB/140 MiB GO alloc/idle/total, 158 MiB/197 MiB CGO alloc/total, 881.83cgo/sec, 0.89/0.09 %(u/s)time, 0.03 %gc (4x) W180308 09:28:22.706902 1048 storage/store.go:1480 [n1,s1,r3/1:/System/NodeLiveness{-Max}] could not gossip node liveness: node unavailable; try another peer I180308 09:28:22.707664 7201 util/stop/stopper.go:473 quiescing; tasks left: 1 ts.poller: poll I180308 09:28:22.710501 7330 util/stop/stopper.go:473 quiescing; tasks left: 1 ts.poller: poll 1 [async] storage.replicate: processing replica 1 [async] kv.TxnCoordSender: heartbeat loop W180308 09:28:22.711733 1047 storage/store.go:1480 [n1,s1,r7/1:/Table/{SystemCon…-11}] could not gossip system config: node unavailable; try another peer W180308 09:28:22.716304 1046 gossip/infostore.go:303 [n1] node unavailable; try another peer W180308 09:28:22.716688 1046 storage/store.go:1480 [n1,s1,r1/1:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer I180308 09:28:22.717708 1053 gossip/gossip.go:487 [n1] gossip status (ok, 5 nodes) gossip client (0/3 cur/max conns) gossip server (3/3 cur/max conns, infos 2303/150 sent/received, bytes 774296B/46012B sent/received) 3: 127.0.0.1:46629 (56s) 4: 127.0.0.1:38315 (55s) 2: 127.0.0.1:37987 (57s) W180308 09:28:22.720545 1053 gossip/infostore.go:303 [n1] node unavailable; try another peer W180308 09:28:22.722669 2373 storage/raft_transport.go:465 [n4] raft transport stream to node 1 failed: rpc error: code = Unavailable desc = transport is closing W180308 09:28:22.727955 2485 ts/db.go:154 [n5,ts-poll] error writing time series data: the batch experienced mixed success and failure W180308 09:28:22.728252 2485 ts/db.go:157 [n5,ts-poll] node unavailable; try another peer I180308 09:28:22.728512 7330 util/stop/stopper.go:473 quiescing; tasks left: 1 [async] storage.replicate: processing replica 1 [async] kv.TxnCoordSender: heartbeat loop W180308 09:28:22.732195 2770 storage/raft_transport.go:465 [n1] raft transport stream to node 5 failed: rpc error: code = Unavailable desc = transport is closing W180308 09:28:22.737226 4045 storage/raft_transport.go:465 [n2] raft transport stream to node 3 failed: EOF W180308 09:28:22.742157 1899 storage/raft_transport.go:465 [n1] raft transport stream to node 3 failed: rpc error: code = Unavailable desc = transport is closing W180308 09:28:22.747260 4042 storage/raft_transport.go:465 [n3] raft transport stream to node 2 failed: rpc error: code = FailedPrecondition desc = grpc: the client connection is closing W180308 09:28:22.760150 1970 storage/raft_transport.go:465 [n3] raft transport stream to node 1 failed: rpc error: code = Unavailable desc = transport is closing W180308 09:28:22.763583 2045 ts/db.go:154 [n4,ts-poll] error writing time series data: the batch experienced mixed success and failure W180308 09:28:22.770300 2045 ts/db.go:157 [n4,ts-poll] node unavailable; try another peer I180308 09:28:22.763970 750 kv/transport_race.go:67 transport race promotion: ran 45 iterations on up to 2482 requests W180308 09:28:22.764917 3208 storage/raft_transport.go:465 [n4] raft transport stream to node 5 failed: rpc error: code = FailedPrecondition desc = grpc: the client connection is closing I180308 09:28:22.772524 7198 util/stop/stopper.go:473 quiescing; tasks left: 3 [async] kv.TxnCoordSender: heartbeat loop 1 ts.poller: poll 1 node.Node: batch 1 [async] storage.replicate: processing replica W180308 09:28:22.775756 1518 gossip/gossip.go:1292 [n2] no incoming or outgoing connections W180308 09:28:22.776260 6938 kv/txn_coord_sender.go:1036 [split,s1,r28/1:/System/{StatusNo…-tsd},n1,txn=fcc1998d] node unavailable; try another peer W180308 09:28:22.777253 3848 storage/raft_transport.go:465 [n4] raft transport stream to node 2 failed: rpc error: code = Unavailable desc = transport is closing W180308 09:28:22.792605 2178 storage/raft_transport.go:465 [n2] raft transport stream to node 1 failed: rpc error: code = Unavailable desc = transport is closing W180308 09:28:22.793349 3922 storage/raft_transport.go:465 [n2] raft transport stream to node 4 failed: rpc error: code = Unavailable desc = transport is closing W180308 09:28:22.803019 1056 ts/db.go:154 [n1,ts-poll] error writing time series data: the batch experienced mixed success and failure W180308 09:28:22.812175 1056 ts/db.go:157 [n1,ts-poll] node unavailable; try another peer W180308 09:28:22.803505 1046 gossip/infostore.go:303 [n1] node unavailable; try another peer W180308 09:28:22.812587 1046 storage/store.go:1480 [n1,s1,r1/1:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer W180308 09:28:22.807111 1048 storage/store.go:1480 [n1,s1,r3/1:/System/NodeLiveness{-Max}] could not gossip node liveness: node unavailable; try another peer W180308 09:28:22.807396 1057 server/node.go:776 [n1,summaries] error recording status summaries: node unavailable; try another peer W180308 09:28:22.808727 7039 internal/client/txn.go:531 [replicate,n1,s1,r24/1:/Table/13/1/2018-03-08T09:27:…] failure aborting transaction: node unavailable; try another peer; abort caused by: node unavailable; try another peer E180308 09:28:22.813785 7039 storage/queue.go:778 [replicate,n1,s1,r24/1:/Table/13/1/2018-03-08T09:27:…] change replicas of r24 failed: node unavailable; try another peer W180308 09:28:22.809209 3173 storage/raft_transport.go:465 [n2] raft transport stream to node 5 failed: rpc error: code = Unavailable desc = transport is closing W180308 09:28:22.815172 7191 kv/txn_coord_sender.go:1036 [split,s1,r29/1:/System/ts{d/cr.no…-e},n1,txn=db46a2ed] node unavailable; try another peer W180308 09:28:22.818081 6931 internal/client/txn.go:531 [replicate,n5,s5,r17/4:/Table/2{0-1}] failure aborting transaction: node unavailable; try another peer; abort caused by: node unavailable; try another peer W180308 09:28:22.820510 2743 storage/raft_transport.go:465 [n5] raft transport stream to node 1 failed: EOF W180308 09:28:22.824887 2238 gossip/gossip.go:1292 [n5] no incoming or outgoing connections W180308 09:28:22.825931 1047 storage/store.go:1480 [n1,s1,r7/1:/Table/{SystemCon…-11}] could not gossip system config: node unavailable; try another peer W180308 09:28:22.835973 3199 storage/raft_transport.go:465 [n5] raft transport stream to node 4 failed: EOF W180308 09:28:22.837525 3335 storage/raft_transport.go:465 [n5] raft transport stream to node 3 failed: EOF W180308 09:28:22.842099 1748 gossip/gossip.go:1292 [n3] no incoming or outgoing connections W180308 09:28:22.842893 3245 storage/raft_transport.go:465 [n3] raft transport stream to node 5 failed: rpc error: code = FailedPrecondition desc = grpc: the client connection is closing W180308 09:28:22.845802 3832 storage/raft_transport.go:465 [n4] raft transport stream to node 3 failed: rpc error: code = Unavailable desc = transport is closing W180308 09:28:22.853934 1798 ts/db.go:154 [n3,ts-poll] error writing time series data: the batch experienced mixed success and failure W180308 09:28:22.855078 1979 gossip/gossip.go:1292 [n4] no incoming or outgoing connections W180308 09:28:22.856870 3102 storage/raft_transport.go:465 [n5] raft transport stream to node 2 failed: EOF W180308 09:28:22.890039 1798 ts/db.go:157 [n3,ts-poll] node unavailable; try another peer I180308 09:28:23.130857 6 sql/logictest/parallel_test.go:266 1 out of 2 parallel tests failed ```
non_priority
sql logictest testparallel failed under stress sha parameters tags goflags race stress build found a failed test storage replica raftstorage go applied preemptive snapshot in testutils testcluster testcluster go has underreplicated ranges storage store go handle raft ready storage replica go proposing add replica updated next storage replica command go change replicas add replica read existing descriptor table storage node liveness go slow heartbeat took testutils testcluster testcluster go has underreplicated ranges storage replica raftstorage go generated preemptive snapshot at index storage store go removing replica storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica go removed keys in storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica go proposing add replica updated next gossip gossip go node has connected to cluster via gossip storage stores go wrote node addresses to persistent storage storage replica go proposing remove replica updated next storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica command go change replicas remove replica read existing descriptor table storage replica raftstorage go generated preemptive snapshot at index sql logictest parallel test go setting range split size storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage store go added to replica gc queue peer suggestion storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor system tse table systemconfigspan start storage replica raftstorage go generated raft snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying raft snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied raft snapshot in storage store go removing replica storage store go removing replica storage replica go removed keys in storage replica go removed keys in storage replica go proposing add replica updated next server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage replica command go change replicas remove replica read existing descriptor table storage replica go proposing add replica updated next storage replica command go change replicas remove replica read existing descriptor system tse table systemconfigspan start storage replica go proposing remove replica updated next sql logictest parallel test go creating database gossip gossip go node has connected to cluster via gossip storage stores go wrote node addresses to persistent storage storage store go added to replica gc queue peer suggestion storage store go added to replica gc queue peer suggestion storage replica command go initiating a split of this range at key system tsd cr node gossip infos sent storage replica command go initiating a split of this range at key system statusnode storage replica command go initiating a split of this range at key table storage replica go proposing add replica updated next storage store go removing replica storage replica go removed keys in storage replica go context canceled before proposing heartbeattxn storage replica command go initiating a split of this range at key table storage replica go proposing remove replica updated next server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage store go added to replica gc queue peer suggestion sql event log go event create database target info databasename test statement create database test user root storage store go added to replica gc queue peer suggestion storage replica go not quiescing pending commands storage store go added to replica gc queue peer suggestion storage store go added to replica gc queue peer suggestion storage store go removing replica server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage replica go removed keys in storage replica command go initiating a split of this range at key table storage replica raftstorage go generated raft snapshot at index storage store go handle raft ready server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying raft snapshot at index id encoded size rocksdb batches log entries storage node liveness go slow heartbeat took storage replica go proposing remove replica updated next server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage store go added to replica gc queue peer suggestion storage store go added to replica gc queue peer suggestion storage node liveness go slow heartbeat took sql logictest parallel test go test setup done sql logictest parallel test go testdata parallel test subquery retry multinode test yaml run list setup storage node liveness go slow heartbeat took storage store go removing replica storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica command go initiating a split of this range at key system statusnode storage replica go removed keys in storage node liveness go slow heartbeat took storage replica backpressure go applying backpressure to limit range growth on batch merge system tsd cr node security certificate expiration ca min merge system tsd cr node security certificate expiration node min merge system tsd cr node gossip connections incoming min merge system tsd cr node gossip connections refused min merge system tsd cr node gossip bytes received min merge system tsd cr node gossip bytes sent min merge system tsd cr node gossip infos received min merge system tsd cr node gossip infos sent min merge system tsd cr node gossip connections outgoing min merge system tsd cr node distsender batches min merge system tsd cr node distsender batches partial min merge system tsd cr node distsender rpc sent min merge system tsd cr node distsender rpc sent local min merge system tsd cr node distsender rpc sent nextreplicaerror min merge system tsd cr node distsender errors notleaseholder min merge system tsd cr node requests slow distsender min merge system tsd cr node txn aborts min merge system tsd cr node txn commits min merge system tsd cr node txn min merge system tsd cr node txn autoretries min skipped merge system tsd cr store queue replicate removereplica min merge system tsd cr store queue replicate removedeadreplica min merge system tsd cr store queue replicate rebalancereplica min merge system tsd cr store queue replicate transferlease min merge system tsd cr store queue replicagc removereplica min storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica command go initiating a split of this range at key table storage replica raftstorage go applied raft snapshot in storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica go context canceled before proposing heartbeattxn storage replica command go initiating a split of this range at key system tsd cr node sql bytesin storage replica command go change replicas remove replica read existing descriptor table storage replica raftstorage go generated preemptive snapshot at index storage replica raftstorage go generated preemptive snapshot at index storage queue go raft status not initialized sql event log go event create table target info tablename test public t statement create table t k int filler string user root progress testdata parallel test subquery retry multinode setup statements queries sql logictest logic go progress testdata parallel test subquery retry multinode setup statements queries storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica command go initiating a split of this range at key table gossip gossip go node has connected to cluster via gossip server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage stores go wrote node addresses to persistent storage storage replica raftstorage go applied preemptive snapshot in storage replica backpressure go applying backpressure to limit range growth on batch merge system tsd cr node security certificate expiration ca min merge system tsd cr node security certificate expiration node min merge system tsd cr node requests slow distsender min merge system tsd cr node txn aborts min merge system tsd cr node txn commits min merge system tsd cr node txn min merge system tsd cr node txn autoretries min merge system tsd cr node txn abandons min merge system tsd cr node txn durations max min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn restarts max min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min skipped merge system tsd cr store queue replicate removereplica min merge system tsd cr store queue replicate removedeadreplica min merge system tsd cr store queue replicate rebalancereplica min merge system tsd cr store queue replicate transferlease min merge system tsd cr store queue replicagc removereplica min storage replica command go change replicas add replica read existing descriptor table storage replica backpressure go applying backpressure to limit range growth on batch merge system tsd cr node security certificate expiration ca min merge system tsd cr node security certificate expiration node min merge system tsd cr node gossip infos sent min merge system tsd cr node requests slow distsender min merge system tsd cr node txn aborts min merge system tsd cr node txn commits min merge system tsd cr node txn min merge system tsd cr node txn autoretries min merge system tsd cr node txn abandons min merge system tsd cr node txn durations max min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn restarts max min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min skipped merge system tsd cr store queue replicate removereplica min merge system tsd cr store queue replicate removedeadreplica min merge system tsd cr store queue replicate rebalancereplica min merge system tsd cr store queue replicate transferlease min merge system tsd cr store queue replicagc removereplica min storage replica backpressure go applying backpressure to limit range growth on batch merge system tsd cr node security certificate expiration ca min merge system tsd cr node security certificate expiration node min merge system tsd cr node gossip infos sent min merge system tsd cr node requests slow distsender min merge system tsd cr node txn aborts min merge system tsd cr node txn commits min merge system tsd cr node txn min merge system tsd cr node txn autoretries min merge system tsd cr node txn abandons min merge system tsd cr node txn durations max min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn restarts max min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min skipped merge system tsd cr store queue replicate removereplica min merge system tsd cr store queue replicate removedeadreplica min merge system tsd cr store queue replicate rebalancereplica min merge system tsd cr store queue replicate transferlease min merge system tsd cr store queue replicagc removereplica min progress testdata parallel test subquery retry multinode setup statements queries sql logictest logic go progress testdata parallel test subquery retry multinode setup statements queries done testdata parallel test subquery retry multinode setup with config tests failures sql logictest logic go done testdata parallel test subquery retry multinode setup with config tests failures sql logictest parallel test go testdata parallel test subquery retry multinode test yaml run list txn txn txn txn txn storage replica go proposing remove replica updated next storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage store go added to replica gc queue peer suggestion storage replica command go initiating a split of this range at key system statusnode server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica go context canceled before proposing heartbeattxn progress testdata parallel test subquery retry multinode txn statements queries sql logictest logic go progress testdata parallel test subquery retry multinode txn statements queries done testdata parallel test subquery retry multinode txn with config tests failures sql logictest logic go done testdata parallel test subquery retry multinode txn with config tests failures server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage node liveness go slow heartbeat took storage node liveness go slow heartbeat took server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica command go initiating a split of this range at key system tsd cr node sql distsql service latency storage store go handle raft ready server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage store go handle raft ready storage store go handle raft ready storage replica raftstorage go generated preemptive snapshot at index storage node liveness go slow heartbeat took storage store go removing replica storage replica go removed keys in storage replica raftstorage go generated preemptive snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica backpressure go applying backpressure to limit range growth on batch merge system tsd cr node txn aborts min merge system tsd cr node txn commits min merge system tsd cr node txn min merge system tsd cr node txn autoretries min merge system tsd cr node txn abandons min merge system tsd cr node txn durations max min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn restarts max min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min skipped merge system tsd cr store queue replicate removereplica min merge system tsd cr store queue replicate removedeadreplica min merge system tsd cr store queue replicate rebalancereplica min merge system tsd cr store queue replicate transferlease min merge system tsd cr store queue replicagc removereplica min storage store go handle raft ready storage node liveness go slow heartbeat took storage replica backpressure go applying backpressure to limit range growth on batch merge system tsd cr node txn aborts min merge system tsd cr node txn commits min merge system tsd cr node txn min merge system tsd cr node txn autoretries min merge system tsd cr node txn abandons min merge system tsd cr node txn durations max min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn restarts max min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min skipped merge system tsd cr store queue replicate removereplica min merge system tsd cr store queue replicate removedeadreplica min merge system tsd cr store queue replicate rebalancereplica min merge system tsd cr store queue replicate transferlease min merge system tsd cr store queue replicagc removereplica min storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage node liveness go slow heartbeat took storage node liveness go slow heartbeat took storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor system tsd cr node gossip infos sent storage node liveness go slow heartbeat took server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor system nodeliveness storage node liveness go slow heartbeat took storage replica go proposing add replica updated next progress testdata parallel test subquery retry multinode txn statements queries sql logictest logic go progress testdata parallel test subquery retry multinode txn statements queries done testdata parallel test subquery retry multinode txn with config tests failures sql logictest logic go done testdata parallel test subquery retry multinode txn with config tests failures storage node liveness go slow heartbeat took storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica raftstorage go generated raft snapshot at index server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying raft snapshot at index id encoded size rocksdb batches log entries storage store go removing replica storage replica go removed keys in storage node liveness go slow heartbeat took storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage store go removing replica server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage replica command go initiating a split of this range at key system statusnode storage replica raftstorage go applied raft snapshot in storage replica go removed keys in storage node liveness go slow heartbeat took storage replica consistency go triggering stats recomputation to resolve delta of containsestimates true lastupdatenanos intentage gcbytesage livebytes livecount keybytes keycount valbytes valcount intentbytes intentcount sysbytes syscount storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage replica go proposing add replica updated next storage replica go context canceled before proposing heartbeattxn storage node liveness go slow heartbeat took storage replica command go change replicas remove replica read existing descriptor table storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro sql logictest logic go testdata parallel test subquery retry multinode txn expected success but found pq transactionstatuserror does not exist done testdata parallel test subquery retry multinode txn with config tests failures sql logictest logic go done testdata parallel test subquery retry multinode txn with config tests failures sql logictest parallel test go error processing testdata parallel test subquery retry multinode txn pq transactionstatuserror does not exist server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage node liveness go slow heartbeat took storage replica command go initiating a split of this range at key system tsd cr node sql mem admin session max storage replica go proposing add replica updated next storage replica command go change replicas remove replica read existing descriptor table storage replica backpressure go applying backpressure to limit range growth on batch merge system tsd cr node txn aborts min merge system tsd cr node txn commits min merge system tsd cr node txn min merge system tsd cr node txn autoretries min merge system tsd cr node txn abandons min merge system tsd cr node txn durations max min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn restarts max min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min skipped merge system tsd cr store queue replicate removereplica min merge system tsd cr store queue replicate removedeadreplica min merge system tsd cr store queue replicate rebalancereplica min merge system tsd cr store queue replicate transferlease min merge system tsd cr store queue replicagc removereplica min storage replica raftstorage go generated raft snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying raft snapshot at index id encoded size rocksdb batches log entries progress testdata parallel test subquery retry multinode txn statements queries sql logictest logic go progress testdata parallel test subquery retry multinode txn statements queries done testdata parallel test subquery retry multinode txn with config tests failures sql logictest logic go done testdata parallel test subquery retry multinode txn with config tests failures storage replica backpressure go applying backpressure to limit range growth on batch merge system tsd cr node txn aborts min merge system tsd cr node txn commits min merge system tsd cr node txn min merge system tsd cr node txn autoretries min merge system tsd cr node txn abandons min merge system tsd cr node txn durations max min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn restarts max min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min skipped merge system tsd cr store queue replicate removereplica min merge system tsd cr store queue replicate removedeadreplica min merge system tsd cr store queue replicate rebalancereplica min merge system tsd cr store queue replicate transferlease min merge system tsd cr store queue replicagc removereplica min storage node liveness go slow heartbeat took storage store go handle raft ready storage replica backpressure go applying backpressure to limit range growth on batch merge system tsd cr node txn aborts min merge system tsd cr node txn commits min merge system tsd cr node txn min merge system tsd cr node txn autoretries min merge system tsd cr node txn abandons min merge system tsd cr node txn durations max min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn restarts max min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min skipped merge system tsd cr store queue replicate removereplica min merge system tsd cr store queue replicate removedeadreplica min merge system tsd cr store queue replicate rebalancereplica min merge system tsd cr store queue replicate transferlease min merge system tsd cr store queue replicagc removereplica min storage node liveness go slow heartbeat took storage replica raftstorage go applied raft snapshot in progress testdata parallel test subquery retry multinode txn statements queries sql logictest logic go progress testdata parallel test subquery retry multinode txn statements queries done testdata parallel test subquery retry multinode txn with config tests failures sql logictest logic go done testdata parallel test subquery retry multinode txn with config tests failures sql logictest parallel test go testdata parallel test subquery retry multinode test yaml run list final storage node liveness go slow heartbeat took storage node liveness go slow heartbeat took done testdata parallel test subquery retry multinode final with config tests failures sql logictest logic go done testdata parallel test subquery retry multinode final with config tests failures util stop stopper go quiescing tasks left ts poller poll util stop stopper go quiescing tasks left ts poller poll storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing util stop stopper go quiescing tasks left node node batch kv txncoordsender heartbeat loop storage split processing replica ts poller poll transport racer storage replicate processing replica storage raft transport go raft transport stream to node failed eof internal client txn go failure aborting transaction node unavailable try another peer abort caused by node unavailable try another peer storage store go could not gossip system config node unavailable try another peer storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing util stop stopper go quiescing tasks left node node batch kv txncoordsender heartbeat loop ts poller poll transport racer storage split processing replica storage replicate processing replica internal client txn go failure aborting transaction node unavailable try another peer abort caused by result is ambiguous server shutdown util stop stopper go quiescing tasks left kv txncoordsender heartbeat loop ts poller poll node node batch transport racer storage split processing replica storage replicate processing replica ts db go error writing time series data the batch experienced mixed success and failure ts db go node unavailable try another peer util stop stopper go quiescing tasks left kv txncoordsender heartbeat loop ts poller poll node node batch transport racer storage replicate processing replica server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage store go could not gossip node liveness node unavailable try another peer util stop stopper go quiescing tasks left ts poller poll util stop stopper go quiescing tasks left ts poller poll storage replicate processing replica kv txncoordsender heartbeat loop storage store go could not gossip system config node unavailable try another peer gossip infostore go node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer gossip gossip go gossip status ok nodes gossip client cur max conns gossip server cur max conns infos sent received bytes sent received gossip infostore go node unavailable try another peer storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing ts db go error writing time series data the batch experienced mixed success and failure ts db go node unavailable try another peer util stop stopper go quiescing tasks left storage replicate processing replica kv txncoordsender heartbeat loop storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing storage raft transport go raft transport stream to node failed eof storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing storage raft transport go raft transport stream to node failed rpc error code failedprecondition desc grpc the client connection is closing storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing ts db go error writing time series data the batch experienced mixed success and failure ts db go node unavailable try another peer kv transport race go transport race promotion ran iterations on up to requests storage raft transport go raft transport stream to node failed rpc error code failedprecondition desc grpc the client connection is closing util stop stopper go quiescing tasks left kv txncoordsender heartbeat loop ts poller poll node node batch storage replicate processing replica gossip gossip go no incoming or outgoing connections kv txn coord sender go node unavailable try another peer storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing ts db go error writing time series data the batch experienced mixed success and failure ts db go node unavailable try another peer gossip infostore go node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip node liveness node unavailable try another peer server node go error recording status summaries node unavailable try another peer internal client txn go failure aborting transaction node unavailable try another peer abort caused by node unavailable try another peer storage queue go change replicas of failed node unavailable try another peer storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing kv txn coord sender go node unavailable try another peer internal client txn go failure aborting transaction node unavailable try another peer abort caused by node unavailable try another peer storage raft transport go raft transport stream to node failed eof gossip gossip go no incoming or outgoing connections storage store go could not gossip system config node unavailable try another peer storage raft transport go raft transport stream to node failed eof storage raft transport go raft transport stream to node failed eof gossip gossip go no incoming or outgoing connections storage raft transport go raft transport stream to node failed rpc error code failedprecondition desc grpc the client connection is closing storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing ts db go error writing time series data the batch experienced mixed success and failure gossip gossip go no incoming or outgoing connections storage raft transport go raft transport stream to node failed eof ts db go node unavailable try another peer sql logictest parallel test go out of parallel tests failed
0
296,575
25,560,204,902
IssuesEvent
2022-11-30 10:11:30
saleor/saleor-dashboard
https://api.github.com/repos/saleor/saleor-dashboard
closed
Enable "noImplicitThis" in tsconfig
maitenance static tests
Enable rule `noImplicitThis` in tsconfig. If the rule needs massive changes in the code, please create separate smaller issues to solve it gradually.
1.0
Enable "noImplicitThis" in tsconfig - Enable rule `noImplicitThis` in tsconfig. If the rule needs massive changes in the code, please create separate smaller issues to solve it gradually.
non_priority
enable noimplicitthis in tsconfig enable rule noimplicitthis in tsconfig if the rule needs massive changes in the code please create separate smaller issues to solve it gradually
0
84,908
10,420,980,834
IssuesEvent
2019-09-16 03:48:18
tarantool/tarantool-java
https://api.github.com/repos/tarantool/tarantool-java
closed
List and briefly describe client and cluster client configuration options
documentation
I mean TarantoolClientConfig and TarantoolClusterClientConfig fields. It is okay to just write it in README.
1.0
List and briefly describe client and cluster client configuration options - I mean TarantoolClientConfig and TarantoolClusterClientConfig fields. It is okay to just write it in README.
non_priority
list and briefly describe client and cluster client configuration options i mean tarantoolclientconfig and tarantoolclusterclientconfig fields it is okay to just write it in readme
0
10,927
9,194,954,949
IssuesEvent
2019-03-07 00:21:24
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Role driven permissions structure
assigned-to-author azure-analysis-services/svc in-progress product-question triaged
Hi, We have a tabular model on Azure SSAS and we are generating PowerBI reports based on that model. We want to build a dynamic steering of the access of cube data based on which user logs in. By that I mean that depending on which report the user runs that a different slice of the cube is performed(with the help of dax filters). I have created different roles in the cube but I noticed that because it is the effective name that is used to pull the data from the cube, the rights of the two different roles are combined when creating the power bi report. I decided then to create two different Azure AD groups(email security) and use the groups to differentiate the data pull from the cube. The group emails were added to both power bi app and the cube roles. The result however was again the same. When one powerbi report is created, mapped to only one of the Azure AD groups and respective cube roles, data from both roles where the user belongs to are still defining the cube slice. Is there any way for one user to get different slices of a cube by being assigned two different roles(with different dax filters) in the cube itself? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 3f0e1195-b6a2-d789-a950-4454798f265e * Version Independent ID: d8ea7415-3b97-5985-44e7-1b89bf1196b6 * Content: [Manage database roles and users in Azure Analysis Services](https://docs.microsoft.com/en-us/azure/analysis-services/analysis-services-database-users#feedback) * Content Source: [articles/analysis-services/analysis-services-database-users.md](https://github.com/Microsoft/azure-docs/blob/master/articles/analysis-services/analysis-services-database-users.md) * Service: **azure-analysis-services** * GitHub Login: @Minewiskan * Microsoft Alias: **owend**
1.0
Role driven permissions structure - Hi, We have a tabular model on Azure SSAS and we are generating PowerBI reports based on that model. We want to build a dynamic steering of the access of cube data based on which user logs in. By that I mean that depending on which report the user runs that a different slice of the cube is performed(with the help of dax filters). I have created different roles in the cube but I noticed that because it is the effective name that is used to pull the data from the cube, the rights of the two different roles are combined when creating the power bi report. I decided then to create two different Azure AD groups(email security) and use the groups to differentiate the data pull from the cube. The group emails were added to both power bi app and the cube roles. The result however was again the same. When one powerbi report is created, mapped to only one of the Azure AD groups and respective cube roles, data from both roles where the user belongs to are still defining the cube slice. Is there any way for one user to get different slices of a cube by being assigned two different roles(with different dax filters) in the cube itself? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 3f0e1195-b6a2-d789-a950-4454798f265e * Version Independent ID: d8ea7415-3b97-5985-44e7-1b89bf1196b6 * Content: [Manage database roles and users in Azure Analysis Services](https://docs.microsoft.com/en-us/azure/analysis-services/analysis-services-database-users#feedback) * Content Source: [articles/analysis-services/analysis-services-database-users.md](https://github.com/Microsoft/azure-docs/blob/master/articles/analysis-services/analysis-services-database-users.md) * Service: **azure-analysis-services** * GitHub Login: @Minewiskan * Microsoft Alias: **owend**
non_priority
role driven permissions structure hi we have a tabular model on azure ssas and we are generating powerbi reports based on that model we want to build a dynamic steering of the access of cube data based on which user logs in by that i mean that depending on which report the user runs that a different slice of the cube is performed with the help of dax filters i have created different roles in the cube but i noticed that because it is the effective name that is used to pull the data from the cube the rights of the two different roles are combined when creating the power bi report i decided then to create two different azure ad groups email security and use the groups to differentiate the data pull from the cube the group emails were added to both power bi app and the cube roles the result however was again the same when one powerbi report is created mapped to only one of the azure ad groups and respective cube roles data from both roles where the user belongs to are still defining the cube slice is there any way for one user to get different slices of a cube by being assigned two different roles with different dax filters in the cube itself document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service azure analysis services github login minewiskan microsoft alias owend
0
49,753
13,187,264,423
IssuesEvent
2020-08-13 02:51:56
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
[stochastics] remove minuit dependency (Trac #2007)
Incomplete Migration Migrated from Trac combo reconstruction defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2007">https://code.icecube.wisc.edu/ticket/2007</a>, reported by kjmeagher and owned by jgonzalez</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:14:55", "description": "either replace it with the implementation of minuit in lilliput or use the gulliver framework", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "_ts": "1550067295757382", "component": "combo reconstruction", "summary": "[stochastics] remove minuit dependency", "priority": "normal", "keywords": "", "time": "2017-05-09T17:07:35", "milestone": "", "owner": "jgonzalez", "type": "defect" } ``` </p> </details>
1.0
[stochastics] remove minuit dependency (Trac #2007) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2007">https://code.icecube.wisc.edu/ticket/2007</a>, reported by kjmeagher and owned by jgonzalez</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:14:55", "description": "either replace it with the implementation of minuit in lilliput or use the gulliver framework", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "_ts": "1550067295757382", "component": "combo reconstruction", "summary": "[stochastics] remove minuit dependency", "priority": "normal", "keywords": "", "time": "2017-05-09T17:07:35", "milestone": "", "owner": "jgonzalez", "type": "defect" } ``` </p> </details>
non_priority
remove minuit dependency trac migrated from json status closed changetime description either replace it with the implementation of minuit in lilliput or use the gulliver framework reporter kjmeagher cc resolution fixed ts component combo reconstruction summary remove minuit dependency priority normal keywords time milestone owner jgonzalez type defect
0
105,394
13,183,295,712
IssuesEvent
2020-08-12 17:15:19
tektoncd/experimental
https://api.github.com/repos/tektoncd/experimental
closed
Add 'Repository' and 'Branch' filters to PipelineRuns
design-help-wanted webhooks
This issue is part of `experimental` epic #37 but concerns development changes in `dashboard`. The webhooks extension needs to be able to filter pipeline runs by repository and branch: it makes sense to surface this capability to dashboard users. Under Dashboard > PipelineRuns, users should be able to filter the pipeline runs shown by 'git repository' and 'branch'. These can be passed as query params to the /pipelineruns REST API. We should apply labels to pipeline runs so as to be able to filter by these labels when obtaining pipeline runs via the Go client.
1.0
Add 'Repository' and 'Branch' filters to PipelineRuns - This issue is part of `experimental` epic #37 but concerns development changes in `dashboard`. The webhooks extension needs to be able to filter pipeline runs by repository and branch: it makes sense to surface this capability to dashboard users. Under Dashboard > PipelineRuns, users should be able to filter the pipeline runs shown by 'git repository' and 'branch'. These can be passed as query params to the /pipelineruns REST API. We should apply labels to pipeline runs so as to be able to filter by these labels when obtaining pipeline runs via the Go client.
non_priority
add repository and branch filters to pipelineruns this issue is part of experimental epic but concerns development changes in dashboard the webhooks extension needs to be able to filter pipeline runs by repository and branch it makes sense to surface this capability to dashboard users under dashboard pipelineruns users should be able to filter the pipeline runs shown by git repository and branch these can be passed as query params to the pipelineruns rest api we should apply labels to pipeline runs so as to be able to filter by these labels when obtaining pipeline runs via the go client
0
207,221
23,434,919,563
IssuesEvent
2022-08-15 08:44:44
Gal-Doron/Baragon-36
https://api.github.com/repos/Gal-Doron/Baragon-36
opened
dropwizard-jetty-1.3.12.jar: 3 vulnerabilities (highest severity is: 5.3)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>dropwizard-jetty-1.3.12.jar</b></p></summary> <p></p> <p>Path to dependency file: /BaragonService/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/Baragon-36/commit/3335ef04e9449f11036516e41533318fc21bd8a3">3335ef04e9449f11036516e41533318fc21bd8a3</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2020-27223](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27223) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | jetty-http-9.4.18.v20190429.jar | Transitive | 2.0.0-rc0+test8 | &#9989; | | [CVE-2021-28169](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28169) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | jetty-http-9.4.18.v20190429.jar | Transitive | 2.0.0-rc0+test8 | &#9989; | | [CVE-2022-2047](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2047) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 2.7 | jetty-http-9.4.18.v20190429.jar | Transitive | 2.0.0-rc0+test8 | &#9989; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-27223</summary> ### Vulnerable Library - <b>jetty-http-9.4.18.v20190429.jar</b></p> <p>The Eclipse Jetty Project</p> <p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p> <p>Path to dependency file: /BaragonData/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar</p> <p> Dependency Hierarchy: - dropwizard-jetty-1.3.12.jar (Root Library) - :x: **jetty-http-9.4.18.v20190429.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/Baragon-36/commit/3335ef04e9449f11036516e41533318fc21bd8a3">3335ef04e9449f11036516e41533318fc21bd8a3</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Eclipse Jetty 9.4.6.v20170531 to 9.4.36.v20210114 (inclusive), 10.0.0, and 11.0.0 when Jetty handles a request containing multiple Accept headers with a large number of “quality” (i.e. q) parameters, the server may enter a denial of service (DoS) state due to high CPU usage processing those quality values, resulting in minutes of CPU time exhausted processing those quality values. <p>Publish Date: 2021-02-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27223>CVE-2020-27223</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/eclipse/jetty.project/security/advisories/GHSA-m394-8rww-3jr7">https://github.com/eclipse/jetty.project/security/advisories/GHSA-m394-8rww-3jr7</a></p> <p>Release Date: 2021-02-26</p> <p>Fix Resolution (org.eclipse.jetty:jetty-http): 9.4.37.v20210219</p> <p>Direct dependency fix Resolution (io.dropwizard:dropwizard-jetty): 2.0.0-rc0+test8</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-28169</summary> ### Vulnerable Library - <b>jetty-http-9.4.18.v20190429.jar</b></p> <p>The Eclipse Jetty Project</p> <p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p> <p>Path to dependency file: /BaragonData/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar</p> <p> Dependency Hierarchy: - dropwizard-jetty-1.3.12.jar (Root Library) - :x: **jetty-http-9.4.18.v20190429.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/Baragon-36/commit/3335ef04e9449f11036516e41533318fc21bd8a3">3335ef04e9449f11036516e41533318fc21bd8a3</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> For Eclipse Jetty versions <= 9.4.40, <= 10.0.2, <= 11.0.2, it is possible for requests to the ConcatServlet with a doubly encoded path to access protected resources within the WEB-INF directory. For example a request to `/concat?/%2557EB-INF/web.xml` can retrieve the web.xml file. This can reveal sensitive information regarding the implementation of a web application. <p>Publish Date: 2021-06-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28169>CVE-2021-28169</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/eclipse/jetty.project/security/advisories/GHSA-gwcr-j4wh-j3cq">https://github.com/eclipse/jetty.project/security/advisories/GHSA-gwcr-j4wh-j3cq</a></p> <p>Release Date: 2021-06-09</p> <p>Fix Resolution (org.eclipse.jetty:jetty-http): 9.4.41.v20210516</p> <p>Direct dependency fix Resolution (io.dropwizard:dropwizard-jetty): 2.0.0-rc0+test8</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> CVE-2022-2047</summary> ### Vulnerable Library - <b>jetty-http-9.4.18.v20190429.jar</b></p> <p>The Eclipse Jetty Project</p> <p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p> <p>Path to dependency file: /BaragonData/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar</p> <p> Dependency Hierarchy: - dropwizard-jetty-1.3.12.jar (Root Library) - :x: **jetty-http-9.4.18.v20190429.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/Baragon-36/commit/3335ef04e9449f11036516e41533318fc21bd8a3">3335ef04e9449f11036516e41533318fc21bd8a3</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Eclipse Jetty versions 9.4.0 thru 9.4.46, and 10.0.0 thru 10.0.9, and 11.0.0 thru 11.0.9 versions, the parsing of the authority segment of an http scheme URI, the Jetty HttpURI class improperly detects an invalid input as a hostname. This can lead to failures in a Proxy scenario. <p>Publish Date: 2022-07-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2047>CVE-2022-2047</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>2.7</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/eclipse/jetty.project/security/advisories/GHSA-cj7v-27pg-wf7q">https://github.com/eclipse/jetty.project/security/advisories/GHSA-cj7v-27pg-wf7q</a></p> <p>Release Date: 2022-07-07</p> <p>Fix Resolution (org.eclipse.jetty:jetty-http): 9.4.47.v20220610</p> <p>Direct dependency fix Resolution (io.dropwizard:dropwizard-jetty): 2.0.0-rc0+test8</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details> *** <p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
True
dropwizard-jetty-1.3.12.jar: 3 vulnerabilities (highest severity is: 5.3) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>dropwizard-jetty-1.3.12.jar</b></p></summary> <p></p> <p>Path to dependency file: /BaragonService/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/Baragon-36/commit/3335ef04e9449f11036516e41533318fc21bd8a3">3335ef04e9449f11036516e41533318fc21bd8a3</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2020-27223](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27223) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | jetty-http-9.4.18.v20190429.jar | Transitive | 2.0.0-rc0+test8 | &#9989; | | [CVE-2021-28169](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28169) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | jetty-http-9.4.18.v20190429.jar | Transitive | 2.0.0-rc0+test8 | &#9989; | | [CVE-2022-2047](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2047) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 2.7 | jetty-http-9.4.18.v20190429.jar | Transitive | 2.0.0-rc0+test8 | &#9989; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-27223</summary> ### Vulnerable Library - <b>jetty-http-9.4.18.v20190429.jar</b></p> <p>The Eclipse Jetty Project</p> <p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p> <p>Path to dependency file: /BaragonData/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar</p> <p> Dependency Hierarchy: - dropwizard-jetty-1.3.12.jar (Root Library) - :x: **jetty-http-9.4.18.v20190429.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/Baragon-36/commit/3335ef04e9449f11036516e41533318fc21bd8a3">3335ef04e9449f11036516e41533318fc21bd8a3</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Eclipse Jetty 9.4.6.v20170531 to 9.4.36.v20210114 (inclusive), 10.0.0, and 11.0.0 when Jetty handles a request containing multiple Accept headers with a large number of “quality” (i.e. q) parameters, the server may enter a denial of service (DoS) state due to high CPU usage processing those quality values, resulting in minutes of CPU time exhausted processing those quality values. <p>Publish Date: 2021-02-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27223>CVE-2020-27223</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/eclipse/jetty.project/security/advisories/GHSA-m394-8rww-3jr7">https://github.com/eclipse/jetty.project/security/advisories/GHSA-m394-8rww-3jr7</a></p> <p>Release Date: 2021-02-26</p> <p>Fix Resolution (org.eclipse.jetty:jetty-http): 9.4.37.v20210219</p> <p>Direct dependency fix Resolution (io.dropwizard:dropwizard-jetty): 2.0.0-rc0+test8</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-28169</summary> ### Vulnerable Library - <b>jetty-http-9.4.18.v20190429.jar</b></p> <p>The Eclipse Jetty Project</p> <p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p> <p>Path to dependency file: /BaragonData/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar</p> <p> Dependency Hierarchy: - dropwizard-jetty-1.3.12.jar (Root Library) - :x: **jetty-http-9.4.18.v20190429.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/Baragon-36/commit/3335ef04e9449f11036516e41533318fc21bd8a3">3335ef04e9449f11036516e41533318fc21bd8a3</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> For Eclipse Jetty versions <= 9.4.40, <= 10.0.2, <= 11.0.2, it is possible for requests to the ConcatServlet with a doubly encoded path to access protected resources within the WEB-INF directory. For example a request to `/concat?/%2557EB-INF/web.xml` can retrieve the web.xml file. This can reveal sensitive information regarding the implementation of a web application. <p>Publish Date: 2021-06-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28169>CVE-2021-28169</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/eclipse/jetty.project/security/advisories/GHSA-gwcr-j4wh-j3cq">https://github.com/eclipse/jetty.project/security/advisories/GHSA-gwcr-j4wh-j3cq</a></p> <p>Release Date: 2021-06-09</p> <p>Fix Resolution (org.eclipse.jetty:jetty-http): 9.4.41.v20210516</p> <p>Direct dependency fix Resolution (io.dropwizard:dropwizard-jetty): 2.0.0-rc0+test8</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> CVE-2022-2047</summary> ### Vulnerable Library - <b>jetty-http-9.4.18.v20190429.jar</b></p> <p>The Eclipse Jetty Project</p> <p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p> <p>Path to dependency file: /BaragonData/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar</p> <p> Dependency Hierarchy: - dropwizard-jetty-1.3.12.jar (Root Library) - :x: **jetty-http-9.4.18.v20190429.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/Baragon-36/commit/3335ef04e9449f11036516e41533318fc21bd8a3">3335ef04e9449f11036516e41533318fc21bd8a3</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Eclipse Jetty versions 9.4.0 thru 9.4.46, and 10.0.0 thru 10.0.9, and 11.0.0 thru 11.0.9 versions, the parsing of the authority segment of an http scheme URI, the Jetty HttpURI class improperly detects an invalid input as a hostname. This can lead to failures in a Proxy scenario. <p>Publish Date: 2022-07-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2047>CVE-2022-2047</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>2.7</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/eclipse/jetty.project/security/advisories/GHSA-cj7v-27pg-wf7q">https://github.com/eclipse/jetty.project/security/advisories/GHSA-cj7v-27pg-wf7q</a></p> <p>Release Date: 2022-07-07</p> <p>Fix Resolution (org.eclipse.jetty:jetty-http): 9.4.47.v20220610</p> <p>Direct dependency fix Resolution (io.dropwizard:dropwizard-jetty): 2.0.0-rc0+test8</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details> *** <p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
non_priority
dropwizard jetty jar vulnerabilities highest severity is vulnerable library dropwizard jetty jar path to dependency file baragonservice pom xml path to vulnerable library home wss scanner repository org eclipse jetty jetty http jetty http jar home wss scanner repository org eclipse jetty jetty http jetty http jar home wss scanner repository org eclipse jetty jetty http jetty http jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available medium jetty http jar transitive medium jetty http jar transitive low jetty http jar transitive details cve vulnerable library jetty http jar the eclipse jetty project library home page a href path to dependency file baragondata pom xml path to vulnerable library home wss scanner repository org eclipse jetty jetty http jetty http jar home wss scanner repository org eclipse jetty jetty http jetty http jar home wss scanner repository org eclipse jetty jetty http jetty http jar dependency hierarchy dropwizard jetty jar root library x jetty http jar vulnerable library found in head commit a href found in base branch master vulnerability details in eclipse jetty to inclusive and when jetty handles a request containing multiple accept headers with a large number of “quality” i e q parameters the server may enter a denial of service dos state due to high cpu usage processing those quality values resulting in minutes of cpu time exhausted processing those quality values publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org eclipse jetty jetty http direct dependency fix resolution io dropwizard dropwizard jetty rescue worker helmet automatic remediation is available for this issue cve vulnerable library jetty http jar the eclipse jetty project library home page a href path to dependency file baragondata pom xml path to vulnerable library home wss scanner repository org eclipse jetty jetty http jetty http jar home wss scanner repository org eclipse jetty jetty http jetty http jar home wss scanner repository org eclipse jetty jetty http jetty http jar dependency hierarchy dropwizard jetty jar root library x jetty http jar vulnerable library found in head commit a href found in base branch master vulnerability details for eclipse jetty versions it is possible for requests to the concatservlet with a doubly encoded path to access protected resources within the web inf directory for example a request to concat inf web xml can retrieve the web xml file this can reveal sensitive information regarding the implementation of a web application publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org eclipse jetty jetty http direct dependency fix resolution io dropwizard dropwizard jetty rescue worker helmet automatic remediation is available for this issue cve vulnerable library jetty http jar the eclipse jetty project library home page a href path to dependency file baragondata pom xml path to vulnerable library home wss scanner repository org eclipse jetty jetty http jetty http jar home wss scanner repository org eclipse jetty jetty http jetty http jar home wss scanner repository org eclipse jetty jetty http jetty http jar dependency hierarchy dropwizard jetty jar root library x jetty http jar vulnerable library found in head commit a href found in base branch master vulnerability details in eclipse jetty versions thru and thru and thru versions the parsing of the authority segment of an http scheme uri the jetty httpuri class improperly detects an invalid input as a hostname this can lead to failures in a proxy scenario publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org eclipse jetty jetty http direct dependency fix resolution io dropwizard dropwizard jetty rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue
0
400,332
27,279,239,002
IssuesEvent
2023-02-23 08:48:33
vaadin/flow
https://api.github.com/repos/vaadin/flow
closed
Document vaadin.excludeUrls configuration parameter
enhancement documentation
This parameter was introduced in https://github.com/vaadin/flow/pull/14579, but not yet documented. Can be added into https://vaadin.com/docs/latest/integrations/spring/configuration or into a general chapter about configuration.
1.0
Document vaadin.excludeUrls configuration parameter - This parameter was introduced in https://github.com/vaadin/flow/pull/14579, but not yet documented. Can be added into https://vaadin.com/docs/latest/integrations/spring/configuration or into a general chapter about configuration.
non_priority
document vaadin excludeurls configuration parameter this parameter was introduced in but not yet documented can be added into or into a general chapter about configuration
0
119,529
17,620,304,719
IssuesEvent
2021-08-18 14:36:11
jgeraigery/experian-java
https://api.github.com/repos/jgeraigery/experian-java
closed
CVE-2019-14540 (High) detected in jackson-databind-2.9.2.jar - autoclosed
security vulnerability
## CVE-2019-14540 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.2.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: experian-java/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.2/jackson-databind-2.9.2.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/experian-java/commit/9ade2a959068cca30ecfdbb254939af6f67affb1">9ade2a959068cca30ecfdbb254939af6f67affb1</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind before 2.9.10. It is related to com.zaxxer.hikari.HikariConfig. <p>Publish Date: 2019-09-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14540>CVE-2019-14540</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14540">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14540</a></p> <p>Release Date: 2019-09-15</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.8.11.5,2.9.10,2.10.0.pr3,2.11.0.rc1</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.2","packageFilePaths":["/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.8.11.5,2.9.10,2.10.0.pr3,2.11.0.rc1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-14540","vulnerabilityDetails":"A Polymorphic Typing issue was discovered in FasterXML jackson-databind before 2.9.10. It is related to com.zaxxer.hikari.HikariConfig.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14540","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-14540 (High) detected in jackson-databind-2.9.2.jar - autoclosed - ## CVE-2019-14540 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.2.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: experian-java/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.2/jackson-databind-2.9.2.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/experian-java/commit/9ade2a959068cca30ecfdbb254939af6f67affb1">9ade2a959068cca30ecfdbb254939af6f67affb1</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind before 2.9.10. It is related to com.zaxxer.hikari.HikariConfig. <p>Publish Date: 2019-09-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14540>CVE-2019-14540</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14540">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14540</a></p> <p>Release Date: 2019-09-15</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.8.11.5,2.9.10,2.10.0.pr3,2.11.0.rc1</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.2","packageFilePaths":["/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.8.11.5,2.9.10,2.10.0.pr3,2.11.0.rc1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-14540","vulnerabilityDetails":"A Polymorphic Typing issue was discovered in FasterXML jackson-databind before 2.9.10. It is related to com.zaxxer.hikari.HikariConfig.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14540","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_priority
cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file experian java mavenworkspace bis services lib bis services base pom xml path to vulnerable library canner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind before it is related to com zaxxer hikari hikariconfig publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails a polymorphic typing issue was discovered in fasterxml jackson databind before it is related to com zaxxer hikari hikariconfig vulnerabilityurl
0
260,651
19,680,106,962
IssuesEvent
2022-01-11 15:59:59
vercel/next.js
https://api.github.com/repos/vercel/next.js
closed
CMS preview example demo are missing the secret value
good first issue area: documentation
# Documentation report Would like to try out the demo given in a CMS preview example, e.g. the DatoCMS one but without knowing the preview secret token I can't try out the preview mode on the demo website.
1.0
CMS preview example demo are missing the secret value - # Documentation report Would like to try out the demo given in a CMS preview example, e.g. the DatoCMS one but without knowing the preview secret token I can't try out the preview mode on the demo website.
non_priority
cms preview example demo are missing the secret value documentation report would like to try out the demo given in a cms preview example e g the datocms one but without knowing the preview secret token i can t try out the preview mode on the demo website
0
11,487
4,232,240,720
IssuesEvent
2016-07-04 21:16:38
ReikaKalseki/Reika_Mods_Issues
https://api.github.com/repos/ReikaKalseki/Reika_Mods_Issues
closed
Electricraft OpenComputers function getCurrent() returns voltage
Bug ElectriCraft Mod Interaction Stupid Code
![2016-06-30_20 56 24](https://cloud.githubusercontent.com/assets/2814500/16508859/2d4a4972-3f05-11e6-8e51-8e854c6d30cb.png) Looks like the problem is [here](https://github.com/ReikaKalseki/ElectriCraft/blob/a572d4f1e5446b4fad1d083c5cf634f8dcd9d32e/Auxiliary/Lua/LuaGetCurrent.java) on line 25.
1.0
Electricraft OpenComputers function getCurrent() returns voltage - ![2016-06-30_20 56 24](https://cloud.githubusercontent.com/assets/2814500/16508859/2d4a4972-3f05-11e6-8e51-8e854c6d30cb.png) Looks like the problem is [here](https://github.com/ReikaKalseki/ElectriCraft/blob/a572d4f1e5446b4fad1d083c5cf634f8dcd9d32e/Auxiliary/Lua/LuaGetCurrent.java) on line 25.
non_priority
electricraft opencomputers function getcurrent returns voltage looks like the problem is on line
0
154,132
13,538,867,411
IssuesEvent
2020-09-16 12:43:42
Galusss/OC_project7
https://api.github.com/repos/Galusss/OC_project7
reopened
Models (backend)
documentation
Les models pour 'users' et 'articles' sont mis en place avec Sequelize ORM. Documentation Sequelize v6 : [https://sequelize.org/master/](url)
1.0
Models (backend) - Les models pour 'users' et 'articles' sont mis en place avec Sequelize ORM. Documentation Sequelize v6 : [https://sequelize.org/master/](url)
non_priority
models backend les models pour users et articles sont mis en place avec sequelize orm documentation sequelize url
0
10,210
31,931,633,611
IssuesEvent
2023-09-19 07:50:32
red-hat-storage/ocs-ci
https://api.github.com/repos/red-hat-storage/ocs-ci
opened
test_for_encrypted_pv_ui - make this test work on any ODF version
ui_automation Squad/Black
test is running only against ODF 4.8, remove this marker and adjust test
1.0
test_for_encrypted_pv_ui - make this test work on any ODF version - test is running only against ODF 4.8, remove this marker and adjust test
non_priority
test for encrypted pv ui make this test work on any odf version test is running only against odf remove this marker and adjust test
0
351,871
32,032,971,098
IssuesEvent
2023-09-22 13:33:02
eclipse-openj9/openj9
https://api.github.com/repos/eclipse-openj9/openj9
opened
VirtualThread states changed in jdk22
test failure jdk22
Failure links ------------ The values were changed in https://github.com/ibmruntimes/openj9-openjdk-jdk/commit/ceb174ba8004f6361a307f6d599d786eef9307c7. * https://openj9-jenkins.osuosl.org/job/Test_openjdknext_j9_sanity.functional_s390x_linux_Personal_testList_0/40/consoleText * https://openj9-jenkins.osuosl.org/job/Test_openjdknext_j9_sanity.functional_s390x_linux_Personal_testList_1/40/consoleText Failure output (captured from console output) --------------------------------------------- ``` [2023-09-22T05:48:47.201Z] FAILED: test_verifyJVMTIMacros [2023-09-22T05:48:47.201Z] java.lang.AssertionError: JVMTI_VTHREAD_STATE_YIELDING (7) does not match VirtualThread.YIELDING (10) [2023-09-22T05:48:47.201Z] at org.testng.Assert.fail(Assert.java:96) [2023-09-22T05:48:47.201Z] at org.openj9.test.jep425.VirtualThreadTests.test_verifyJVMTIMacros(VirtualThreadTests.java:321) [2023-09-22T05:48:47.201Z] at java.base/jdk.internal.reflect.DirectMethodHandleAccessor.invoke(DirectMethodHandleAccessor.java:103) [2023-09-22T05:48:47.201Z] at java.base/java.lang.reflect.Method.invoke(Method.java:580) ```
1.0
VirtualThread states changed in jdk22 - Failure links ------------ The values were changed in https://github.com/ibmruntimes/openj9-openjdk-jdk/commit/ceb174ba8004f6361a307f6d599d786eef9307c7. * https://openj9-jenkins.osuosl.org/job/Test_openjdknext_j9_sanity.functional_s390x_linux_Personal_testList_0/40/consoleText * https://openj9-jenkins.osuosl.org/job/Test_openjdknext_j9_sanity.functional_s390x_linux_Personal_testList_1/40/consoleText Failure output (captured from console output) --------------------------------------------- ``` [2023-09-22T05:48:47.201Z] FAILED: test_verifyJVMTIMacros [2023-09-22T05:48:47.201Z] java.lang.AssertionError: JVMTI_VTHREAD_STATE_YIELDING (7) does not match VirtualThread.YIELDING (10) [2023-09-22T05:48:47.201Z] at org.testng.Assert.fail(Assert.java:96) [2023-09-22T05:48:47.201Z] at org.openj9.test.jep425.VirtualThreadTests.test_verifyJVMTIMacros(VirtualThreadTests.java:321) [2023-09-22T05:48:47.201Z] at java.base/jdk.internal.reflect.DirectMethodHandleAccessor.invoke(DirectMethodHandleAccessor.java:103) [2023-09-22T05:48:47.201Z] at java.base/java.lang.reflect.Method.invoke(Method.java:580) ```
non_priority
virtualthread states changed in failure links the values were changed in failure output captured from console output failed test verifyjvmtimacros java lang assertionerror jvmti vthread state yielding does not match virtualthread yielding at org testng assert fail assert java at org test virtualthreadtests test verifyjvmtimacros virtualthreadtests java at java base jdk internal reflect directmethodhandleaccessor invoke directmethodhandleaccessor java at java base java lang reflect method invoke method java
0
278,392
21,076,103,999
IssuesEvent
2022-04-02 06:44:21
AY2122S2-CS2103-F09-3/tp
https://api.github.com/repos/AY2122S2-CS2103-F09-3/tp
closed
[PE-D] Price of insurance can only be integer values
bug documentation
I would expect decimal places to be allowed for inputting prices for adding an insurance command. Doesn't fit well with the product. <!--session: 1648793009038-2837d48d-1dc9-4169-aea4-6e86a2707040--> <!--Version: Web v3.4.2--> ------------- Labels: `type.FeatureFlaw` `severity.Low` original: LapisRaider/ped#19
1.0
[PE-D] Price of insurance can only be integer values - I would expect decimal places to be allowed for inputting prices for adding an insurance command. Doesn't fit well with the product. <!--session: 1648793009038-2837d48d-1dc9-4169-aea4-6e86a2707040--> <!--Version: Web v3.4.2--> ------------- Labels: `type.FeatureFlaw` `severity.Low` original: LapisRaider/ped#19
non_priority
price of insurance can only be integer values i would expect decimal places to be allowed for inputting prices for adding an insurance command doesn t fit well with the product labels type featureflaw severity low original lapisraider ped
0
189,798
15,196,640,359
IssuesEvent
2021-02-16 08:36:04
spring-projects/spring-boot
https://api.github.com/repos/spring-projects/spring-boot
closed
Update the Gradle plugin documentation to recommend the maven-publish plugin over the maven plugin
type: documentation
The documentation doesn't make any recommendation at the moment, other than documenting the maven plugin before the maven-publish plugin. We should swap the order over and also explicitly recommend the maven-publish plugin to align with Gradle's own recommendations and deprecations.
1.0
Update the Gradle plugin documentation to recommend the maven-publish plugin over the maven plugin - The documentation doesn't make any recommendation at the moment, other than documenting the maven plugin before the maven-publish plugin. We should swap the order over and also explicitly recommend the maven-publish plugin to align with Gradle's own recommendations and deprecations.
non_priority
update the gradle plugin documentation to recommend the maven publish plugin over the maven plugin the documentation doesn t make any recommendation at the moment other than documenting the maven plugin before the maven publish plugin we should swap the order over and also explicitly recommend the maven publish plugin to align with gradle s own recommendations and deprecations
0
54,270
13,301,362,685
IssuesEvent
2020-08-25 12:50:56
feelpp/feelpp
https://api.github.com/repos/feelpp/feelpp
opened
Add range-v3 to feelpp contrib
module:build type:feature
range-v3 is the basis for std::range in c++20. in the last release 0.11 range-v3 (which is C++11/14 and 17 compliant) got closer to the C++20 standard. see https://github.com/feelpp/range-v3 for more details see https://en.cppreference.com/w/cpp/header/ranges for more details about std::ranges in the c++20 standard
1.0
Add range-v3 to feelpp contrib - range-v3 is the basis for std::range in c++20. in the last release 0.11 range-v3 (which is C++11/14 and 17 compliant) got closer to the C++20 standard. see https://github.com/feelpp/range-v3 for more details see https://en.cppreference.com/w/cpp/header/ranges for more details about std::ranges in the c++20 standard
non_priority
add range to feelpp contrib range is the basis for std range in c in the last release range which is c and compliant got closer to the c standard see for more details see for more details about std ranges in the c standard
0
128,730
18,070,107,195
IssuesEvent
2021-09-21 01:12:04
dmyers87/matrix-react-sdk
https://api.github.com/repos/dmyers87/matrix-react-sdk
opened
CVE-2021-3803 (Medium) detected in nth-check-1.0.2.tgz
security vulnerability
## CVE-2021-3803 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nth-check-1.0.2.tgz</b></p></summary> <p>performant nth-check parser & compiler</p> <p>Library home page: <a href="https://registry.npmjs.org/nth-check/-/nth-check-1.0.2.tgz">https://registry.npmjs.org/nth-check/-/nth-check-1.0.2.tgz</a></p> <p>Path to dependency file: matrix-react-sdk/node_modules/nth-check/package.json</p> <p>Path to vulnerable library: matrix-react-sdk/node_modules/nth-check/package.json</p> <p> Dependency Hierarchy: - enzyme-3.11.0.tgz (Root Library) - cheerio-1.0.0-rc.3.tgz - css-select-1.2.0.tgz - :x: **nth-check-1.0.2.tgz** (Vulnerable Library) <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> nth-check is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3803>CVE-2021-3803</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/fb55/nth-check/compare/v2.0.0...v2.0.1">https://github.com/fb55/nth-check/compare/v2.0.0...v2.0.1</a></p> <p>Release Date: 2021-09-17</p> <p>Fix Resolution: nth-check - v2.0.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"nth-check","packageVersion":"1.0.2","packageFilePaths":["/node_modules/nth-check/package.json"],"isTransitiveDependency":true,"dependencyTree":"enzyme:3.11.0;cheerio:1.0.0-rc.3;css-select:1.2.0;nth-check:1.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"nth-check - v2.0.1"}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2021-3803","vulnerabilityDetails":"nth-check is vulnerable to Inefficient Regular Expression Complexity","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3803","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-3803 (Medium) detected in nth-check-1.0.2.tgz - ## CVE-2021-3803 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nth-check-1.0.2.tgz</b></p></summary> <p>performant nth-check parser & compiler</p> <p>Library home page: <a href="https://registry.npmjs.org/nth-check/-/nth-check-1.0.2.tgz">https://registry.npmjs.org/nth-check/-/nth-check-1.0.2.tgz</a></p> <p>Path to dependency file: matrix-react-sdk/node_modules/nth-check/package.json</p> <p>Path to vulnerable library: matrix-react-sdk/node_modules/nth-check/package.json</p> <p> Dependency Hierarchy: - enzyme-3.11.0.tgz (Root Library) - cheerio-1.0.0-rc.3.tgz - css-select-1.2.0.tgz - :x: **nth-check-1.0.2.tgz** (Vulnerable Library) <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> nth-check is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3803>CVE-2021-3803</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/fb55/nth-check/compare/v2.0.0...v2.0.1">https://github.com/fb55/nth-check/compare/v2.0.0...v2.0.1</a></p> <p>Release Date: 2021-09-17</p> <p>Fix Resolution: nth-check - v2.0.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"nth-check","packageVersion":"1.0.2","packageFilePaths":["/node_modules/nth-check/package.json"],"isTransitiveDependency":true,"dependencyTree":"enzyme:3.11.0;cheerio:1.0.0-rc.3;css-select:1.2.0;nth-check:1.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"nth-check - v2.0.1"}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2021-3803","vulnerabilityDetails":"nth-check is vulnerable to Inefficient Regular Expression Complexity","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3803","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
non_priority
cve medium detected in nth check tgz cve medium severity vulnerability vulnerable library nth check tgz performant nth check parser compiler library home page a href path to dependency file matrix react sdk node modules nth check package json path to vulnerable library matrix react sdk node modules nth check package json dependency hierarchy enzyme tgz root library cheerio rc tgz css select tgz x nth check tgz vulnerable library found in base branch develop vulnerability details nth check is vulnerable to inefficient regular expression complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution nth check isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree enzyme cheerio rc css select nth check isminimumfixversionavailable true minimumfixversion nth check basebranches vulnerabilityidentifier cve vulnerabilitydetails nth check is vulnerable to inefficient regular expression complexity vulnerabilityurl
0
132,998
18,791,793,970
IssuesEvent
2021-11-08 17:32:58
HTTPArchive/almanac.httparchive.org
https://api.github.com/repos/HTTPArchive/almanac.httparchive.org
closed
Chapter illustrations for new 2021 chapters
design
We have 2 new chapters this year: - Structured Data - WebAssembly We have chapter banners images for each chapter created in 2019, and we managed to create three new ones for the new chapters added in 2020 (Capabilities, Jamstack, and Privacy). So we would need similar ones created for Structured Data and WebAssembly. We have the PSD files to allow you to reuse some of the resources like the characters. Unfortunately the designers that created these in 2019 and 2020 are unavailable this year. @HTTPArchive/designers do any of you have the necessary design skills to do this?
1.0
Chapter illustrations for new 2021 chapters - We have 2 new chapters this year: - Structured Data - WebAssembly We have chapter banners images for each chapter created in 2019, and we managed to create three new ones for the new chapters added in 2020 (Capabilities, Jamstack, and Privacy). So we would need similar ones created for Structured Data and WebAssembly. We have the PSD files to allow you to reuse some of the resources like the characters. Unfortunately the designers that created these in 2019 and 2020 are unavailable this year. @HTTPArchive/designers do any of you have the necessary design skills to do this?
non_priority
chapter illustrations for new chapters we have new chapters this year structured data webassembly we have chapter banners images for each chapter created in and we managed to create three new ones for the new chapters added in capabilities jamstack and privacy so we would need similar ones created for structured data and webassembly we have the psd files to allow you to reuse some of the resources like the characters unfortunately the designers that created these in and are unavailable this year httparchive designers do any of you have the necessary design skills to do this
0
449,295
31,837,617,104
IssuesEvent
2023-09-14 14:23:34
Studio-Yandex-Practicum/RandomCoffeeBotTelegram
https://api.github.com/repos/Studio-Yandex-Practicum/RandomCoffeeBotTelegram
closed
Создать диаграмму состояний бота, оформить всё в drawing
documentation
# Why? Необходимо для упрощения разработки. # How To Do? Реазилуем в диаграмме логику работы бота. НУжно посмотреть в документацию от заказчика и сделать правки в своей диаграмме. Необходимо в readme добавить ссылку # Where? docs
1.0
Создать диаграмму состояний бота, оформить всё в drawing - # Why? Необходимо для упрощения разработки. # How To Do? Реазилуем в диаграмме логику работы бота. НУжно посмотреть в документацию от заказчика и сделать правки в своей диаграмме. Необходимо в readme добавить ссылку # Where? docs
non_priority
создать диаграмму состояний бота оформить всё в drawing why необходимо для упрощения разработки how to do реазилуем в диаграмме логику работы бота нужно посмотреть в документацию от заказчика и сделать правки в своей диаграмме необходимо в readme добавить ссылку where docs
0
68,584
21,717,751,631
IssuesEvent
2022-05-10 19:45:46
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
opened
ZFS Receive of encrypted incremental data stream causes a PANIC
Type: Defect
### System information Type | Version/Name --- | --- Distribution Name | Debian Distribution Version | Bullseye Kernel Version | 5.10.109+truenas Architecture | amd64 OpenZFS Version | zfs-2.1.2-95_g1d2cdd23b zfs-kmod-2.1.2-95_g1d2cdd23b <!-- Command to find OpenZFS version: zfs version Commands to find kernel version: uname -r # Linux freebsd-version -r # FreeBSD --> ### Describe the problem you're observing During an incremental receive, ZFS caused a panic and a system hangup. ### Describe how to reproduce the problem It happens randomly. ### Include any warning/errors/backtraces from the system logs ``` [456678.240841] VERIFY3(0 == dmu_object_set_blocksize(rwa->os, drro->drr_object, drro->drr_blksz, drro->drr_indblkshift, tx)) failed (0 == 95) [456678.243815] PANIC at dmu_recv.c:1776:receive_object() [456678.245141] Showing stack for process 2936808 [456678.246532] CPU: 10 PID: 2936808 Comm: receive_writer Tainted: P OE 5.10.109+truenas #1 [456678.247840] Hardware name: Supermicro X9QR7-TF+/X9QRi-F+/X9QR7-TF+/X9QRi-F+, BIOS 3.0b 05/20/2015 [456678.249138] Call Trace: [456678.250421] dump_stack+0x6b/0x83 [456678.251676] spl_panic+0xd4/0xfc [spl] [456678.253038] ? arc_buf_access+0x14c/0x250 [zfs] [456678.254276] ? dnode_hold_impl+0x4e9/0xef0 [zfs] [456678.255493] ? dnode_set_blksz+0x13b/0x300 [zfs] [456678.256677] ? dnode_rele_and_unlock+0x5c/0xc0 [zfs] [456678.257846] receive_object+0xc2c/0xca0 [zfs] [456678.258984] ? dmu_object_next+0xd6/0x120 [zfs] [456678.260098] ? receive_writer_thread+0xbd/0xad0 [zfs] [456678.261160] ? kfree+0x40c/0x480 [456678.262202] ? _cond_resched+0x16/0x40 [456678.263244] receive_writer_thread+0x1cc/0xad0 [zfs] [456678.264280] ? thread_generic_wrapper+0x62/0x80 [spl] [456678.265252] ? kfree+0x40c/0x480 [456678.266242] ? receive_process_write_record+0x190/0x190 [zfs] [456678.267177] ? thread_generic_wrapper+0x6f/0x80 [spl] [456678.268092] thread_generic_wrapper+0x6f/0x80 [spl] [456678.268988] ? __thread_exit+0x20/0x20 [spl] [456678.269864] kthread+0x11b/0x140 [456678.270706] ? __kthread_bind_mask+0x60/0x60 [456678.271538] ret_from_fork+0x22/0x30 ```
1.0
ZFS Receive of encrypted incremental data stream causes a PANIC - ### System information Type | Version/Name --- | --- Distribution Name | Debian Distribution Version | Bullseye Kernel Version | 5.10.109+truenas Architecture | amd64 OpenZFS Version | zfs-2.1.2-95_g1d2cdd23b zfs-kmod-2.1.2-95_g1d2cdd23b <!-- Command to find OpenZFS version: zfs version Commands to find kernel version: uname -r # Linux freebsd-version -r # FreeBSD --> ### Describe the problem you're observing During an incremental receive, ZFS caused a panic and a system hangup. ### Describe how to reproduce the problem It happens randomly. ### Include any warning/errors/backtraces from the system logs ``` [456678.240841] VERIFY3(0 == dmu_object_set_blocksize(rwa->os, drro->drr_object, drro->drr_blksz, drro->drr_indblkshift, tx)) failed (0 == 95) [456678.243815] PANIC at dmu_recv.c:1776:receive_object() [456678.245141] Showing stack for process 2936808 [456678.246532] CPU: 10 PID: 2936808 Comm: receive_writer Tainted: P OE 5.10.109+truenas #1 [456678.247840] Hardware name: Supermicro X9QR7-TF+/X9QRi-F+/X9QR7-TF+/X9QRi-F+, BIOS 3.0b 05/20/2015 [456678.249138] Call Trace: [456678.250421] dump_stack+0x6b/0x83 [456678.251676] spl_panic+0xd4/0xfc [spl] [456678.253038] ? arc_buf_access+0x14c/0x250 [zfs] [456678.254276] ? dnode_hold_impl+0x4e9/0xef0 [zfs] [456678.255493] ? dnode_set_blksz+0x13b/0x300 [zfs] [456678.256677] ? dnode_rele_and_unlock+0x5c/0xc0 [zfs] [456678.257846] receive_object+0xc2c/0xca0 [zfs] [456678.258984] ? dmu_object_next+0xd6/0x120 [zfs] [456678.260098] ? receive_writer_thread+0xbd/0xad0 [zfs] [456678.261160] ? kfree+0x40c/0x480 [456678.262202] ? _cond_resched+0x16/0x40 [456678.263244] receive_writer_thread+0x1cc/0xad0 [zfs] [456678.264280] ? thread_generic_wrapper+0x62/0x80 [spl] [456678.265252] ? kfree+0x40c/0x480 [456678.266242] ? receive_process_write_record+0x190/0x190 [zfs] [456678.267177] ? thread_generic_wrapper+0x6f/0x80 [spl] [456678.268092] thread_generic_wrapper+0x6f/0x80 [spl] [456678.268988] ? __thread_exit+0x20/0x20 [spl] [456678.269864] kthread+0x11b/0x140 [456678.270706] ? __kthread_bind_mask+0x60/0x60 [456678.271538] ret_from_fork+0x22/0x30 ```
non_priority
zfs receive of encrypted incremental data stream causes a panic system information type version name distribution name debian distribution version bullseye kernel version truenas architecture openzfs version zfs zfs kmod command to find openzfs version zfs version commands to find kernel version uname r linux freebsd version r freebsd describe the problem you re observing during an incremental receive zfs caused a panic and a system hangup describe how to reproduce the problem it happens randomly include any warning errors backtraces from the system logs dmu object set blocksize rwa os drro drr object drro drr blksz drro drr indblkshift tx failed panic at dmu recv c receive object showing stack for process cpu pid comm receive writer tainted p oe truenas hardware name supermicro tf f tf f bios call trace dump stack spl panic arc buf access dnode hold impl dnode set blksz dnode rele and unlock receive object dmu object next receive writer thread kfree cond resched receive writer thread thread generic wrapper kfree receive process write record thread generic wrapper thread generic wrapper thread exit kthread kthread bind mask ret from fork
0
58,699
11,905,066,912
IssuesEvent
2020-03-30 17:55:24
home-assistant/brands
https://api.github.com/repos/home-assistant/brands
opened
August is missing brand images
domain-missing has-codeowner has-config-flow
## The problem The August integration does not have brand images in this repository. We recently started this Brands repository, to create a centralized storage of all brand-related images. These images are used on our website and the Home Assistant frontend. The following images are missing and would ideally be added: - `src/august/icon.png` - `src/august/logo.png` - `src/august/icon@2x.png` - `src/august/logo@2x.png` For image specifications and requirements, please see [README.md](https://github.com/home-assistant/brands/blob/master/README.md). ## Updating the documentation repository Our documentation repository already has a logo for this integration, however, it does not meet the image requirements of this new Brands repository. If adding images to this repository, please open up a PR to the documentation repository as well, removing the `logo: august.png` line from this file: <https://github.com/home-assistant/home-assistant.io/blob/current/source/_integrations/august.markdown> **Note**: The documentation PR needs to be opened against the `current` branch. **Note2**: Please leave the actual logo file in the documentation repository. It will be cleaned up differently. ## Additional information For more information about this repository, read the [README.md](https://github.com/home-assistant/brands/blob/master/README.md) file of this repository. It contains information on how this repository works, and image specification and requirements. ## Codeowner mention Hi there, @bdraco! Mind taking a look at this issue as it is with an integration (august) you are listed as a [codeowner](https://github.com/home-assistant/core/blob/dev/homeassistant/components/august/manifest.json) for? Thanks! Resolving this issue is not limited to codeowners! If you want to help us out, feel free to resolve this issue! Thanks already!
1.0
August is missing brand images - ## The problem The August integration does not have brand images in this repository. We recently started this Brands repository, to create a centralized storage of all brand-related images. These images are used on our website and the Home Assistant frontend. The following images are missing and would ideally be added: - `src/august/icon.png` - `src/august/logo.png` - `src/august/icon@2x.png` - `src/august/logo@2x.png` For image specifications and requirements, please see [README.md](https://github.com/home-assistant/brands/blob/master/README.md). ## Updating the documentation repository Our documentation repository already has a logo for this integration, however, it does not meet the image requirements of this new Brands repository. If adding images to this repository, please open up a PR to the documentation repository as well, removing the `logo: august.png` line from this file: <https://github.com/home-assistant/home-assistant.io/blob/current/source/_integrations/august.markdown> **Note**: The documentation PR needs to be opened against the `current` branch. **Note2**: Please leave the actual logo file in the documentation repository. It will be cleaned up differently. ## Additional information For more information about this repository, read the [README.md](https://github.com/home-assistant/brands/blob/master/README.md) file of this repository. It contains information on how this repository works, and image specification and requirements. ## Codeowner mention Hi there, @bdraco! Mind taking a look at this issue as it is with an integration (august) you are listed as a [codeowner](https://github.com/home-assistant/core/blob/dev/homeassistant/components/august/manifest.json) for? Thanks! Resolving this issue is not limited to codeowners! If you want to help us out, feel free to resolve this issue! Thanks already!
non_priority
august is missing brand images the problem the august integration does not have brand images in this repository we recently started this brands repository to create a centralized storage of all brand related images these images are used on our website and the home assistant frontend the following images are missing and would ideally be added src august icon png src august logo png src august icon png src august logo png for image specifications and requirements please see updating the documentation repository our documentation repository already has a logo for this integration however it does not meet the image requirements of this new brands repository if adding images to this repository please open up a pr to the documentation repository as well removing the logo august png line from this file note the documentation pr needs to be opened against the current branch please leave the actual logo file in the documentation repository it will be cleaned up differently additional information for more information about this repository read the file of this repository it contains information on how this repository works and image specification and requirements codeowner mention hi there bdraco mind taking a look at this issue as it is with an integration august you are listed as a for thanks resolving this issue is not limited to codeowners if you want to help us out feel free to resolve this issue thanks already
0
181,589
14,881,932,373
IssuesEvent
2021-01-20 11:10:02
SAP/code-pal-for-abap
https://api.github.com/repos/SAP/code-pal-for-abap
closed
'Deprecated Key Word' Documentation
documentation
It would be good to have the respective replacement in the check documentation. It can be described in the check documentation or a link to the original system documentation.
1.0
'Deprecated Key Word' Documentation - It would be good to have the respective replacement in the check documentation. It can be described in the check documentation or a link to the original system documentation.
non_priority
deprecated key word documentation it would be good to have the respective replacement in the check documentation it can be described in the check documentation or a link to the original system documentation
0
7,167
16,853,483,406
IssuesEvent
2021-06-21 00:41:09
QutEcoacoustics/workbench-client
https://api.github.com/repos/QutEcoacoustics/workbench-client
opened
Enable prerendering
architecture bug env:production env:staging
Currently prerendering of the SSR files is broken for the website, however making it functional would help SEO performance and should be possible. Currently this is broken because we need to read the configuration file for the instance before building. During SSR we perform this step through the express server, however a different solution is needed for prerending to work. The following guide seems promising: https://www.thirdrocktechkno.com/blog/prerendering-in-angular-9/
1.0
Enable prerendering - Currently prerendering of the SSR files is broken for the website, however making it functional would help SEO performance and should be possible. Currently this is broken because we need to read the configuration file for the instance before building. During SSR we perform this step through the express server, however a different solution is needed for prerending to work. The following guide seems promising: https://www.thirdrocktechkno.com/blog/prerendering-in-angular-9/
non_priority
enable prerendering currently prerendering of the ssr files is broken for the website however making it functional would help seo performance and should be possible currently this is broken because we need to read the configuration file for the instance before building during ssr we perform this step through the express server however a different solution is needed for prerending to work the following guide seems promising
0
55,886
8,033,878,420
IssuesEvent
2018-07-29 12:11:30
PaulHancock/Aegean
https://api.github.com/repos/PaulHancock/Aegean
closed
update wiki to be inline with v 2.0.2
documentation
The wiki is out of date. In particular the help text for `Aegean` and `BANE` needs to be updated, as well as the description of the various command line options.
1.0
update wiki to be inline with v 2.0.2 - The wiki is out of date. In particular the help text for `Aegean` and `BANE` needs to be updated, as well as the description of the various command line options.
non_priority
update wiki to be inline with v the wiki is out of date in particular the help text for aegean and bane needs to be updated as well as the description of the various command line options
0
104,639
22,711,425,479
IssuesEvent
2022-07-05 19:49:25
microsoft/vscode-jupyter
https://api.github.com/repos/microsoft/vscode-jupyter
closed
Remote widget source provider doesn't work if any security required
bug upstream-vscode ipywidgets vscode-web verification-found
1. Remove the CDN sources (make them blank) 2. Connect to a remote server 3. Run something that needs a widget 4. Widget won't be downloaded Root cause is the remote cdn provider doesn't do anything except pass the URI up to the webview for downloading. This doesn't work because the webview doesn't have the auth headers (or token) to make a valid request. I think the solution here is to refactor the CDN downloading code to support the server as a CDN source, but then it would make the appropriate requests with the correct headers, download from the remote jupyter server, and copy the source files to the same spot it does for CDNs. Essentially remove this [source provider ](https://github.com/microsoft/vscode-jupyter/blob/9d9ad7d1f4756789ba9374d1a0351a558279ed8b/src/kernels/ipywidgets-message-coordination/remoteWidgetScriptSourceProvider.ts#L14) and have the CDN code also try kernel's baseURL. Meaning the kernel's base URL is always another CDN source.
2.0
Remote widget source provider doesn't work if any security required - 1. Remove the CDN sources (make them blank) 2. Connect to a remote server 3. Run something that needs a widget 4. Widget won't be downloaded Root cause is the remote cdn provider doesn't do anything except pass the URI up to the webview for downloading. This doesn't work because the webview doesn't have the auth headers (or token) to make a valid request. I think the solution here is to refactor the CDN downloading code to support the server as a CDN source, but then it would make the appropriate requests with the correct headers, download from the remote jupyter server, and copy the source files to the same spot it does for CDNs. Essentially remove this [source provider ](https://github.com/microsoft/vscode-jupyter/blob/9d9ad7d1f4756789ba9374d1a0351a558279ed8b/src/kernels/ipywidgets-message-coordination/remoteWidgetScriptSourceProvider.ts#L14) and have the CDN code also try kernel's baseURL. Meaning the kernel's base URL is always another CDN source.
non_priority
remote widget source provider doesn t work if any security required remove the cdn sources make them blank connect to a remote server run something that needs a widget widget won t be downloaded root cause is the remote cdn provider doesn t do anything except pass the uri up to the webview for downloading this doesn t work because the webview doesn t have the auth headers or token to make a valid request i think the solution here is to refactor the cdn downloading code to support the server as a cdn source but then it would make the appropriate requests with the correct headers download from the remote jupyter server and copy the source files to the same spot it does for cdns essentially remove this and have the cdn code also try kernel s baseurl meaning the kernel s base url is always another cdn source
0
366,923
25,708,751,022
IssuesEvent
2022-12-07 04:02:59
esmero/archipelago-documentation
https://api.github.com/repos/esmero/archipelago-documentation
closed
Add full KeyName Provider Documentation Section
documentation enhancement help wanted
# What? Key Name Providers are fundamental for searching/filtering/faceting. We should have - How to documentation on creating new ones and the different Plugins we provide (including the " Is Date" option) and the Entity Reference One - What/how Archipelago harvests/compiles data using them - How these are exposed to the Search API but also programmatically as properties of a field (for DEVS) - Good practices: e.g if the keys you want to expose are all flat and unique you can use the JSONLD Plugin and type them all into the "extra keys" fields instead of generating too many Key Name Providers, each with a KEY - Some Concrete JMESPATH examples - Some basics on Discovery/Faceting/use of these and troubleshooting using Solr Directly @dmer @alliomeria pretty sure we made a Video for Jack a time ago we can use as starting point
1.0
Add full KeyName Provider Documentation Section - # What? Key Name Providers are fundamental for searching/filtering/faceting. We should have - How to documentation on creating new ones and the different Plugins we provide (including the " Is Date" option) and the Entity Reference One - What/how Archipelago harvests/compiles data using them - How these are exposed to the Search API but also programmatically as properties of a field (for DEVS) - Good practices: e.g if the keys you want to expose are all flat and unique you can use the JSONLD Plugin and type them all into the "extra keys" fields instead of generating too many Key Name Providers, each with a KEY - Some Concrete JMESPATH examples - Some basics on Discovery/Faceting/use of these and troubleshooting using Solr Directly @dmer @alliomeria pretty sure we made a Video for Jack a time ago we can use as starting point
non_priority
add full keyname provider documentation section what key name providers are fundamental for searching filtering faceting we should have how to documentation on creating new ones and the different plugins we provide including the is date option and the entity reference one what how archipelago harvests compiles data using them how these are exposed to the search api but also programmatically as properties of a field for devs good practices e g if the keys you want to expose are all flat and unique you can use the jsonld plugin and type them all into the extra keys fields instead of generating too many key name providers each with a key some concrete jmespath examples some basics on discovery faceting use of these and troubleshooting using solr directly dmer alliomeria pretty sure we made a video for jack a time ago we can use as starting point
0
287,299
31,834,284,494
IssuesEvent
2023-09-14 12:38:40
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Expose "maxSpaces"-like property via Spaces setup and start contracts on both client- and server-side.
Team:Security Feature:Security/Spaces NeededFor:Core NeededFor:DataDiscovery Project:Serverless MVP
## Summary Spaces API consumers also need to determine whether they can show spaces specific UIs. Since the Spaces plugin is enabled even when the UI is hidden, the spaces plugin contract needs to provide a way of communicating whether the UI should be enabled.
True
Expose "maxSpaces"-like property via Spaces setup and start contracts on both client- and server-side. - ## Summary Spaces API consumers also need to determine whether they can show spaces specific UIs. Since the Spaces plugin is enabled even when the UI is hidden, the spaces plugin contract needs to provide a way of communicating whether the UI should be enabled.
non_priority
expose maxspaces like property via spaces setup and start contracts on both client and server side summary spaces api consumers also need to determine whether they can show spaces specific uis since the spaces plugin is enabled even when the ui is hidden the spaces plugin contract needs to provide a way of communicating whether the ui should be enabled
0
162,120
25,482,618,458
IssuesEvent
2022-11-26 01:01:44
CERT-Polska/ursadb
https://api.github.com/repos/CERT-Polska/ursadb
closed
Add `debug parse` command
status:needs more design
We need a way to introspect the running database, or debug some issues without either recompiling the db with debug prints or attaching with gdb. Especially now that I'll be working on query parsing & optimisation, I have a few ideas in mind. Right now I'd like to have something easy - `debug parse` command: ``` debug parse index "hmm" with [gram3]; ``` Will parse the command `index "hmm" with [gram3]` in this case, pretty print the parsed tree and return it to the user. Expected result is for example: ``` index "hmm" index_type_list gram3 ``` (or whatever the real query tree looks like). This can be returned as a structured JSON or as plain text - it's intended for debugging so we don't make any guarantees about the output.
1.0
Add `debug parse` command - We need a way to introspect the running database, or debug some issues without either recompiling the db with debug prints or attaching with gdb. Especially now that I'll be working on query parsing & optimisation, I have a few ideas in mind. Right now I'd like to have something easy - `debug parse` command: ``` debug parse index "hmm" with [gram3]; ``` Will parse the command `index "hmm" with [gram3]` in this case, pretty print the parsed tree and return it to the user. Expected result is for example: ``` index "hmm" index_type_list gram3 ``` (or whatever the real query tree looks like). This can be returned as a structured JSON or as plain text - it's intended for debugging so we don't make any guarantees about the output.
non_priority
add debug parse command we need a way to introspect the running database or debug some issues without either recompiling the db with debug prints or attaching with gdb especially now that i ll be working on query parsing optimisation i have a few ideas in mind right now i d like to have something easy debug parse command debug parse index hmm with will parse the command index hmm with in this case pretty print the parsed tree and return it to the user expected result is for example index hmm index type list or whatever the real query tree looks like this can be returned as a structured json or as plain text it s intended for debugging so we don t make any guarantees about the output
0
221,223
24,599,697,613
IssuesEvent
2022-10-14 11:23:38
arbindo/mimock
https://api.github.com/repos/arbindo/mimock
closed
postgresql-42.3.3.jar: 1 vulnerabilities (highest severity is: 8.0)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postgresql-42.3.3.jar</b></p></summary> <p>PostgreSQL JDBC Driver Postgresql</p> <p>Library home page: <a href="https://jdbc.postgresql.org">https://jdbc.postgresql.org</a></p> <p>Path to dependency file: /mimock-backend/pom.xml</p> <p>Path to vulnerable library: /.m2/repository/org/postgresql/postgresql/42.3.3/postgresql-42.3.3.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/arbindo/mimock/commit/8055f41fefa4c77dfb0ed76c4220dbf3b2d33c6f">8055f41fefa4c77dfb0ed76c4220dbf3b2d33c6f</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2022-31197](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-31197) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.0 | postgresql-42.3.3.jar | Direct | 42.4.1 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-31197</summary> ### Vulnerable Library - <b>postgresql-42.3.3.jar</b></p> <p>PostgreSQL JDBC Driver Postgresql</p> <p>Library home page: <a href="https://jdbc.postgresql.org">https://jdbc.postgresql.org</a></p> <p>Path to dependency file: /mimock-backend/pom.xml</p> <p>Path to vulnerable library: /.m2/repository/org/postgresql/postgresql/42.3.3/postgresql-42.3.3.jar</p> <p> Dependency Hierarchy: - :x: **postgresql-42.3.3.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/arbindo/mimock/commit/8055f41fefa4c77dfb0ed76c4220dbf3b2d33c6f">8055f41fefa4c77dfb0ed76c4220dbf3b2d33c6f</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> PostgreSQL JDBC Driver (PgJDBC for short) allows Java programs to connect to a PostgreSQL database using standard, database independent Java code. The PGJDBC implementation of the `java.sql.ResultRow.refreshRow()` method is not performing escaping of column names so a malicious column name that contains a statement terminator, e.g. `;`, could lead to SQL injection. This could lead to executing additional SQL commands as the application's JDBC user. User applications that do not invoke the `ResultSet.refreshRow()` method are not impacted. User application that do invoke that method are impacted if the underlying database that they are querying via their JDBC application may be under the control of an attacker. The attack requires the attacker to trick the user into executing SQL against a table name who's column names would contain the malicious SQL and subsequently invoke the `refreshRow()` method on the ResultSet. Note that the application's JDBC user and the schema owner need not be the same. A JDBC application that executes as a privileged user querying database schemas owned by potentially malicious less-privileged users would be vulnerable. In that situation it may be possible for the malicious user to craft a schema that causes the application to execute commands as the privileged user. Patched versions will be released as `42.2.26` and `42.4.1`. Users are advised to upgrade. There are no known workarounds for this issue. <p>Publish Date: 2022-08-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-31197>CVE-2022-31197</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>8.0</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/pgjdbc/pgjdbc/security/advisories/GHSA-r38f-c4h4-hqq2">https://github.com/pgjdbc/pgjdbc/security/advisories/GHSA-r38f-c4h4-hqq2</a></p> <p>Release Date: 2022-08-03</p> <p>Fix Resolution: 42.4.1</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
True
postgresql-42.3.3.jar: 1 vulnerabilities (highest severity is: 8.0) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postgresql-42.3.3.jar</b></p></summary> <p>PostgreSQL JDBC Driver Postgresql</p> <p>Library home page: <a href="https://jdbc.postgresql.org">https://jdbc.postgresql.org</a></p> <p>Path to dependency file: /mimock-backend/pom.xml</p> <p>Path to vulnerable library: /.m2/repository/org/postgresql/postgresql/42.3.3/postgresql-42.3.3.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/arbindo/mimock/commit/8055f41fefa4c77dfb0ed76c4220dbf3b2d33c6f">8055f41fefa4c77dfb0ed76c4220dbf3b2d33c6f</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2022-31197](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-31197) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.0 | postgresql-42.3.3.jar | Direct | 42.4.1 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-31197</summary> ### Vulnerable Library - <b>postgresql-42.3.3.jar</b></p> <p>PostgreSQL JDBC Driver Postgresql</p> <p>Library home page: <a href="https://jdbc.postgresql.org">https://jdbc.postgresql.org</a></p> <p>Path to dependency file: /mimock-backend/pom.xml</p> <p>Path to vulnerable library: /.m2/repository/org/postgresql/postgresql/42.3.3/postgresql-42.3.3.jar</p> <p> Dependency Hierarchy: - :x: **postgresql-42.3.3.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/arbindo/mimock/commit/8055f41fefa4c77dfb0ed76c4220dbf3b2d33c6f">8055f41fefa4c77dfb0ed76c4220dbf3b2d33c6f</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> PostgreSQL JDBC Driver (PgJDBC for short) allows Java programs to connect to a PostgreSQL database using standard, database independent Java code. The PGJDBC implementation of the `java.sql.ResultRow.refreshRow()` method is not performing escaping of column names so a malicious column name that contains a statement terminator, e.g. `;`, could lead to SQL injection. This could lead to executing additional SQL commands as the application's JDBC user. User applications that do not invoke the `ResultSet.refreshRow()` method are not impacted. User application that do invoke that method are impacted if the underlying database that they are querying via their JDBC application may be under the control of an attacker. The attack requires the attacker to trick the user into executing SQL against a table name who's column names would contain the malicious SQL and subsequently invoke the `refreshRow()` method on the ResultSet. Note that the application's JDBC user and the schema owner need not be the same. A JDBC application that executes as a privileged user querying database schemas owned by potentially malicious less-privileged users would be vulnerable. In that situation it may be possible for the malicious user to craft a schema that causes the application to execute commands as the privileged user. Patched versions will be released as `42.2.26` and `42.4.1`. Users are advised to upgrade. There are no known workarounds for this issue. <p>Publish Date: 2022-08-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-31197>CVE-2022-31197</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>8.0</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/pgjdbc/pgjdbc/security/advisories/GHSA-r38f-c4h4-hqq2">https://github.com/pgjdbc/pgjdbc/security/advisories/GHSA-r38f-c4h4-hqq2</a></p> <p>Release Date: 2022-08-03</p> <p>Fix Resolution: 42.4.1</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
non_priority
postgresql jar vulnerabilities highest severity is vulnerable library postgresql jar postgresql jdbc driver postgresql library home page a href path to dependency file mimock backend pom xml path to vulnerable library repository org postgresql postgresql postgresql jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high postgresql jar direct details cve vulnerable library postgresql jar postgresql jdbc driver postgresql library home page a href path to dependency file mimock backend pom xml path to vulnerable library repository org postgresql postgresql postgresql jar dependency hierarchy x postgresql jar vulnerable library found in head commit a href found in base branch main vulnerability details postgresql jdbc driver pgjdbc for short allows java programs to connect to a postgresql database using standard database independent java code the pgjdbc implementation of the java sql resultrow refreshrow method is not performing escaping of column names so a malicious column name that contains a statement terminator e g could lead to sql injection this could lead to executing additional sql commands as the application s jdbc user user applications that do not invoke the resultset refreshrow method are not impacted user application that do invoke that method are impacted if the underlying database that they are querying via their jdbc application may be under the control of an attacker the attack requires the attacker to trick the user into executing sql against a table name who s column names would contain the malicious sql and subsequently invoke the refreshrow method on the resultset note that the application s jdbc user and the schema owner need not be the same a jdbc application that executes as a privileged user querying database schemas owned by potentially malicious less privileged users would be vulnerable in that situation it may be possible for the malicious user to craft a schema that causes the application to execute commands as the privileged user patched versions will be released as and users are advised to upgrade there are no known workarounds for this issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
41,079
6,890,715,240
IssuesEvent
2017-11-22 14:53:21
pilosa/pilosa
https://api.github.com/repos/pilosa/pilosa
closed
BSI != range operation documentation
documentation
### Description The new `Range` query `!=` operation need to be documented. This operation works on != null, and != _integer_
1.0
BSI != range operation documentation - ### Description The new `Range` query `!=` operation need to be documented. This operation works on != null, and != _integer_
non_priority
bsi range operation documentation description the new range query operation need to be documented this operation works on null and integer
0
335,698
30,082,047,459
IssuesEvent
2023-06-29 04:57:44
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: failover/system-non-liveness/deadlock/lease=expiration failed
C-test-failure O-robot O-roachtest branch-master release-blocker T-kv
roachtest.failover/system-non-liveness/deadlock/lease=expiration [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestWeeklyBazel/10720111?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestWeeklyBazel/10720111?buildTab=artifacts#/failover/system-non-liveness/deadlock/lease=expiration) on master @ [7fd4c21157221eae9e7d5892d89d2b5a671aba3e](https://github.com/cockroachdb/cockroach/commits/7fd4c21157221eae9e7d5892d89d2b5a671aba3e): ``` (assertions.go:333).Fail: Error Trace: github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/failover.go:1408 github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/failover.go:1059 main/pkg/cmd/roachtest/monitor.go:105 golang.org/x/sync/errgroup/external/org_golang_x_sync/errgroup/errgroup.go:75 GOROOT/src/runtime/asm_amd64.s:1594 Error: Received unexpected error: pq: error getting span statistics - number of spans in request payload (1064) exceeds 'server.span_stats.span_batch_limit' cluster setting limit (500) Test: failover/system-non-liveness/deadlock/lease=expiration (require.go:1360).NoError: FailNow called (monitor.go:137).Wait: monitor failure: monitor task failed: t.Fatal() was called test artifacts and logs in: /artifacts/failover/system-non-liveness/deadlock/lease=expiration/run_1 ``` <p>Parameters: <code>ROACHTEST_arch=amd64</code> , <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=2</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*failover/system-non-liveness/deadlock/lease=expiration.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-29196
2.0
roachtest: failover/system-non-liveness/deadlock/lease=expiration failed - roachtest.failover/system-non-liveness/deadlock/lease=expiration [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestWeeklyBazel/10720111?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestWeeklyBazel/10720111?buildTab=artifacts#/failover/system-non-liveness/deadlock/lease=expiration) on master @ [7fd4c21157221eae9e7d5892d89d2b5a671aba3e](https://github.com/cockroachdb/cockroach/commits/7fd4c21157221eae9e7d5892d89d2b5a671aba3e): ``` (assertions.go:333).Fail: Error Trace: github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/failover.go:1408 github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/failover.go:1059 main/pkg/cmd/roachtest/monitor.go:105 golang.org/x/sync/errgroup/external/org_golang_x_sync/errgroup/errgroup.go:75 GOROOT/src/runtime/asm_amd64.s:1594 Error: Received unexpected error: pq: error getting span statistics - number of spans in request payload (1064) exceeds 'server.span_stats.span_batch_limit' cluster setting limit (500) Test: failover/system-non-liveness/deadlock/lease=expiration (require.go:1360).NoError: FailNow called (monitor.go:137).Wait: monitor failure: monitor task failed: t.Fatal() was called test artifacts and logs in: /artifacts/failover/system-non-liveness/deadlock/lease=expiration/run_1 ``` <p>Parameters: <code>ROACHTEST_arch=amd64</code> , <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=2</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*failover/system-non-liveness/deadlock/lease=expiration.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-29196
non_priority
roachtest failover system non liveness deadlock lease expiration failed roachtest failover system non liveness deadlock lease expiration with on master assertions go fail error trace github com cockroachdb cockroach pkg cmd roachtest tests failover go github com cockroachdb cockroach pkg cmd roachtest tests failover go main pkg cmd roachtest monitor go golang org x sync errgroup external org golang x sync errgroup errgroup go goroot src runtime asm s error received unexpected error pq error getting span statistics number of spans in request payload exceeds server span stats span batch limit cluster setting limit test failover system non liveness deadlock lease expiration require go noerror failnow called monitor go wait monitor failure monitor task failed t fatal was called test artifacts and logs in artifacts failover system non liveness deadlock lease expiration run parameters roachtest arch roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see cc cockroachdb kv triage jira issue crdb
0
26,686
4,239,362,404
IssuesEvent
2016-07-06 09:10:53
e-government-ua/iBP
https://api.github.com/repos/e-government-ua/iBP
closed
Нетішин, Хмельницька обл. - розкрити 7 послуг. Земля
in work ready for release test
Організація ринку, ярмарку 13 Взяття на облік громадян, які потребують поліпшення житлових умов 17 Надання довідок незайнятому населенню про відсутність земельних ділянок для ведення ОСГ 23 Надання інформації, що підтверджує відсутність(наявність) земельної ділянк 657 Видача Акту обстеження земельної ділянки на наявність зелених насаджень 106 Внесення змін до договору оренди землі 1481 Дозвіл на виготовлення технічної документації із землеустрою щодо встановлення меж земельної ділянки 1392
1.0
Нетішин, Хмельницька обл. - розкрити 7 послуг. Земля - Організація ринку, ярмарку 13 Взяття на облік громадян, які потребують поліпшення житлових умов 17 Надання довідок незайнятому населенню про відсутність земельних ділянок для ведення ОСГ 23 Надання інформації, що підтверджує відсутність(наявність) земельної ділянк 657 Видача Акту обстеження земельної ділянки на наявність зелених насаджень 106 Внесення змін до договору оренди землі 1481 Дозвіл на виготовлення технічної документації із землеустрою щодо встановлення меж земельної ділянки 1392
non_priority
нетішин хмельницька обл розкрити послуг земля організація ринку ярмарку взяття на облік громадян які потребують поліпшення житлових умов надання довідок незайнятому населенню про відсутність земельних ділянок для ведення осг надання інформації що підтверджує відсутність наявність земельної ділянк видача акту обстеження земельної ділянки на наявність зелених насаджень внесення змін до договору оренди землі дозвіл на виготовлення технічної документації із землеустрою щодо встановлення меж земельної ділянки
0
32,678
26,900,945,723
IssuesEvent
2023-02-06 15:36:18
evcc-io/evcc
https://api.github.com/repos/evcc-io/evcc
closed
Debian / Ubuntu Service: Separater Nutzer
infrastructure
### Discussed in https://github.com/evcc-io/evcc/discussions/4840 <div type='discussions-op-text'> <sup>Originally posted by **pauxus** October 16, 2022</sup> Gerade nachdem evcc jetzt ja auch state im Dateisystem ablegt: Aktuell ist der Service Nutzer unter Debian "root", was ja aus Sicherheitsaspekten eher unzweckmäßig ist. Sollte hier nicht sinnvollerweise ein eigener Nutzer (evcc) angelegt werden? Damit wäre dann auch die sqlite db z.B. unter "/var/lib/evcc". </div>
1.0
Debian / Ubuntu Service: Separater Nutzer - ### Discussed in https://github.com/evcc-io/evcc/discussions/4840 <div type='discussions-op-text'> <sup>Originally posted by **pauxus** October 16, 2022</sup> Gerade nachdem evcc jetzt ja auch state im Dateisystem ablegt: Aktuell ist der Service Nutzer unter Debian "root", was ja aus Sicherheitsaspekten eher unzweckmäßig ist. Sollte hier nicht sinnvollerweise ein eigener Nutzer (evcc) angelegt werden? Damit wäre dann auch die sqlite db z.B. unter "/var/lib/evcc". </div>
non_priority
debian ubuntu service separater nutzer discussed in originally posted by pauxus october gerade nachdem evcc jetzt ja auch state im dateisystem ablegt aktuell ist der service nutzer unter debian root was ja aus sicherheitsaspekten eher unzweckmäßig ist sollte hier nicht sinnvollerweise ein eigener nutzer evcc angelegt werden damit wäre dann auch die sqlite db z b unter var lib evcc
0
127,526
10,474,349,387
IssuesEvent
2019-09-23 14:21:25
status-im/nim-beacon-chain
https://api.github.com/repos/status-im/nim-beacon-chain
opened
hash_tree_root: pathological case in test_fixture_sanity_blocks
performance test suite
When compiling test_fixture_sanity_blocks as a standalone and running it, I manage to block my machine for a couple minutes on what seems to be faststreams memory manager: ```Nim [Suite] Official - Sanity - Blocks [Preset: mainnet] [OK] [Invalid] Previous slot block transition (prev_slot_block_transition) [OK] [Valid] Same slot block transition (same_slot_block_transition) [OK] [Valid] Empty block transition (empty_block_transition) [OK] [Valid] Skipped Slots (skipped_slots) ^CTraceback (most recent call last) /home/beta/Programming/Status/nim-beacon-chain/tests/official/test_fixture_sanity_blocks.nim(83) test_fixture_sanity_blocks /home/beta/Programming/Status/nim-beacon-chain/tests/official/test_fixture_sanity_blocks.nim(43) testImpl_blck_empty_epoch_transition /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/state_transition.nim(115) state_transition /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/state_transition.nim(63) process_slots /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/state_transition.nim(45) process_slot /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(584) hash_tree_root /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(454) hashTreeRootImpl /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(584) hash_tree_root /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(454) hashTreeRootImpl /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(584) hash_tree_root /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(525) hashTreeRootImpl /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(446) merkelizeSerializedChunks /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(440) merkelizeSerializedChunks /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(426) newSszHashingStream /home/beta/Programming/Status/nim-beacon-chain/vendor/nim-faststreams/faststreams/output_stream.nim(88) initWithSinglePage /home/beta/Programming/Status/nim-beacon-chain/vendor/nim-faststreams/faststreams/output_stream.nim(76) addPage /home/beta/.choosenim/toolchains/nim-0.19.6/lib/system/gc.nim(477) newObj SIGINT: Interrupted by Ctrl-C. Traceback (most recent call last) proxyexe.nim(62) proxyexe proxyexe.nim(49) main osproc.nim(1136) waitForExit SIGINT: Interrupted by Ctrl-C. ```
1.0
hash_tree_root: pathological case in test_fixture_sanity_blocks - When compiling test_fixture_sanity_blocks as a standalone and running it, I manage to block my machine for a couple minutes on what seems to be faststreams memory manager: ```Nim [Suite] Official - Sanity - Blocks [Preset: mainnet] [OK] [Invalid] Previous slot block transition (prev_slot_block_transition) [OK] [Valid] Same slot block transition (same_slot_block_transition) [OK] [Valid] Empty block transition (empty_block_transition) [OK] [Valid] Skipped Slots (skipped_slots) ^CTraceback (most recent call last) /home/beta/Programming/Status/nim-beacon-chain/tests/official/test_fixture_sanity_blocks.nim(83) test_fixture_sanity_blocks /home/beta/Programming/Status/nim-beacon-chain/tests/official/test_fixture_sanity_blocks.nim(43) testImpl_blck_empty_epoch_transition /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/state_transition.nim(115) state_transition /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/state_transition.nim(63) process_slots /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/state_transition.nim(45) process_slot /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(584) hash_tree_root /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(454) hashTreeRootImpl /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(584) hash_tree_root /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(454) hashTreeRootImpl /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(584) hash_tree_root /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(525) hashTreeRootImpl /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(446) merkelizeSerializedChunks /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(440) merkelizeSerializedChunks /home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(426) newSszHashingStream /home/beta/Programming/Status/nim-beacon-chain/vendor/nim-faststreams/faststreams/output_stream.nim(88) initWithSinglePage /home/beta/Programming/Status/nim-beacon-chain/vendor/nim-faststreams/faststreams/output_stream.nim(76) addPage /home/beta/.choosenim/toolchains/nim-0.19.6/lib/system/gc.nim(477) newObj SIGINT: Interrupted by Ctrl-C. Traceback (most recent call last) proxyexe.nim(62) proxyexe proxyexe.nim(49) main osproc.nim(1136) waitForExit SIGINT: Interrupted by Ctrl-C. ```
non_priority
hash tree root pathological case in test fixture sanity blocks when compiling test fixture sanity blocks as a standalone and running it i manage to block my machine for a couple minutes on what seems to be faststreams memory manager nim official sanity blocks previous slot block transition prev slot block transition same slot block transition same slot block transition empty block transition empty block transition skipped slots skipped slots ctraceback most recent call last home beta programming status nim beacon chain tests official test fixture sanity blocks nim test fixture sanity blocks home beta programming status nim beacon chain tests official test fixture sanity blocks nim testimpl blck empty epoch transition home beta programming status nim beacon chain beacon chain state transition nim state transition home beta programming status nim beacon chain beacon chain state transition nim process slots home beta programming status nim beacon chain beacon chain state transition nim process slot home beta programming status nim beacon chain beacon chain ssz nim hash tree root home beta programming status nim beacon chain beacon chain ssz nim hashtreerootimpl home beta programming status nim beacon chain beacon chain ssz nim hash tree root home beta programming status nim beacon chain beacon chain ssz nim hashtreerootimpl home beta programming status nim beacon chain beacon chain ssz nim hash tree root home beta programming status nim beacon chain beacon chain ssz nim hashtreerootimpl home beta programming status nim beacon chain beacon chain ssz nim merkelizeserializedchunks home beta programming status nim beacon chain beacon chain ssz nim merkelizeserializedchunks home beta programming status nim beacon chain beacon chain ssz nim newsszhashingstream home beta programming status nim beacon chain vendor nim faststreams faststreams output stream nim initwithsinglepage home beta programming status nim beacon chain vendor nim faststreams faststreams output stream nim addpage home beta choosenim toolchains nim lib system gc nim newobj sigint interrupted by ctrl c traceback most recent call last proxyexe nim proxyexe proxyexe nim main osproc nim waitforexit sigint interrupted by ctrl c
0
164,479
20,364,488,993
IssuesEvent
2022-02-21 02:54:01
arohablue/skill-india-frontend
https://api.github.com/repos/arohablue/skill-india-frontend
closed
CVE-2018-11698 (High) detected in node-sass-4.11.0.tgz, node-sassv4.12.0 - autoclosed
security vulnerability
## CVE-2018-11698 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.11.0.tgz</b>, <b>node-sassv4.12.0</b></p></summary> <p> <details><summary><b>node-sass-4.11.0.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz</a></p> <p>Path to dependency file: /skill-india-frontend/package.json</p> <p>Path to vulnerable library: skill-india-frontend/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - build-angular-0.13.1.tgz (Root Library) - :x: **node-sass-4.11.0.tgz** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in LibSass through 3.5.4. An out-of-bounds read of a memory region was found in the function Sass::handle_error which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service. <p>Publish Date: 2018-06-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11698>CVE-2018-11698</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sass/libsass/releases/tag/3.5.5">https://github.com/sass/libsass/releases/tag/3.5.5</a></p> <p>Release Date: 2018-06-04</p> <p>Fix Resolution: libsass - 3.5.5;node-sass - 4.14.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-11698 (High) detected in node-sass-4.11.0.tgz, node-sassv4.12.0 - autoclosed - ## CVE-2018-11698 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.11.0.tgz</b>, <b>node-sassv4.12.0</b></p></summary> <p> <details><summary><b>node-sass-4.11.0.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz</a></p> <p>Path to dependency file: /skill-india-frontend/package.json</p> <p>Path to vulnerable library: skill-india-frontend/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - build-angular-0.13.1.tgz (Root Library) - :x: **node-sass-4.11.0.tgz** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in LibSass through 3.5.4. An out-of-bounds read of a memory region was found in the function Sass::handle_error which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service. <p>Publish Date: 2018-06-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11698>CVE-2018-11698</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sass/libsass/releases/tag/3.5.5">https://github.com/sass/libsass/releases/tag/3.5.5</a></p> <p>Release Date: 2018-06-04</p> <p>Fix Resolution: libsass - 3.5.5;node-sass - 4.14.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in node sass tgz node autoclosed cve high severity vulnerability vulnerable libraries node sass tgz node node sass tgz wrapper around libsass library home page a href path to dependency file skill india frontend package json path to vulnerable library skill india frontend node modules node sass package json dependency hierarchy build angular tgz root library x node sass tgz vulnerable library vulnerability details an issue was discovered in libsass through an out of bounds read of a memory region was found in the function sass handle error which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass node sass step up your open source security game with whitesource
0
185,019
21,785,052,467
IssuesEvent
2022-05-14 02:18:17
Yash-Handa/GitHub-Org-Geographics
https://api.github.com/repos/Yash-Handa/GitHub-Org-Geographics
closed
WS-2019-0369 (Medium) detected in handlebars-4.1.1.tgz - autoclosed
security vulnerability
## WS-2019-0369 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.1.1.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.1.1.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.1.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/GitHub-Org-Geographics/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/GitHub-Org-Geographics/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - karma-coverage-istanbul-reporter-2.0.5.tgz (Root Library) - istanbul-api-2.1.1.tgz - istanbul-reports-2.1.1.tgz - :x: **handlebars-4.1.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Yash-Handa/GitHub-Org-Geographics/commit/69e1362228223135794f86d35da5d76214470dda">69e1362228223135794f86d35da5d76214470dda</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype Pollution vulnerability found in handlebars.js before 4.5.3. Attacker may use Remote-Code-Execution exploits. <p>Publish Date: 2020-01-08 <p>URL: <a href=https://github.com/RetireJS/retire.js/commit/f07a7557d3fc1c26b86fe11a5b33cb1b8f3dcf2f>WS-2019-0369</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/wycats/handlebars.js/blob/master/release-notes.md#v453---november-18th-2019">https://github.com/wycats/handlebars.js/blob/master/release-notes.md#v453---november-18th-2019</a></p> <p>Release Date: 2020-01-08</p> <p>Fix Resolution: handlebars - 4.5.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2019-0369 (Medium) detected in handlebars-4.1.1.tgz - autoclosed - ## WS-2019-0369 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.1.1.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.1.1.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.1.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/GitHub-Org-Geographics/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/GitHub-Org-Geographics/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - karma-coverage-istanbul-reporter-2.0.5.tgz (Root Library) - istanbul-api-2.1.1.tgz - istanbul-reports-2.1.1.tgz - :x: **handlebars-4.1.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Yash-Handa/GitHub-Org-Geographics/commit/69e1362228223135794f86d35da5d76214470dda">69e1362228223135794f86d35da5d76214470dda</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype Pollution vulnerability found in handlebars.js before 4.5.3. Attacker may use Remote-Code-Execution exploits. <p>Publish Date: 2020-01-08 <p>URL: <a href=https://github.com/RetireJS/retire.js/commit/f07a7557d3fc1c26b86fe11a5b33cb1b8f3dcf2f>WS-2019-0369</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/wycats/handlebars.js/blob/master/release-notes.md#v453---november-18th-2019">https://github.com/wycats/handlebars.js/blob/master/release-notes.md#v453---november-18th-2019</a></p> <p>Release Date: 2020-01-08</p> <p>Fix Resolution: handlebars - 4.5.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
ws medium detected in handlebars tgz autoclosed ws medium severity vulnerability vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file tmp ws scm github org geographics package json path to vulnerable library tmp ws scm github org geographics node modules handlebars package json dependency hierarchy karma coverage istanbul reporter tgz root library istanbul api tgz istanbul reports tgz x handlebars tgz vulnerable library found in head commit a href vulnerability details prototype pollution vulnerability found in handlebars js before attacker may use remote code execution exploits publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution handlebars step up your open source security game with whitesource
0
43,118
23,122,180,219
IssuesEvent
2022-07-27 23:07:36
open-policy-agent/gatekeeper
https://api.github.com/repos/open-policy-agent/gatekeeper
opened
Do not deserialize admission request object more than once
Performance
With the new Expansion feature in https://github.com/open-policy-agent/gatekeeper/pull/2062, we are deserializing the request object in the validation webhook twice: - once when we expand the object (new code) - again when we actually match the object ([existing code](https://github.com/open-policy-agent/gatekeeper/blob/841e10000acf3e7fc42e35a0ab1be5dfbe0138ae/pkg/target/matcher.go#L67-L87)) We should somehow pass the deserialized object down the chain such that it does not have to be deserialized again when matched.
True
Do not deserialize admission request object more than once - With the new Expansion feature in https://github.com/open-policy-agent/gatekeeper/pull/2062, we are deserializing the request object in the validation webhook twice: - once when we expand the object (new code) - again when we actually match the object ([existing code](https://github.com/open-policy-agent/gatekeeper/blob/841e10000acf3e7fc42e35a0ab1be5dfbe0138ae/pkg/target/matcher.go#L67-L87)) We should somehow pass the deserialized object down the chain such that it does not have to be deserialized again when matched.
non_priority
do not deserialize admission request object more than once with the new expansion feature in we are deserializing the request object in the validation webhook twice once when we expand the object new code again when we actually match the object we should somehow pass the deserialized object down the chain such that it does not have to be deserialized again when matched
0
223,026
17,101,165,779
IssuesEvent
2021-07-09 11:27:27
bitcoin-s/bitcoin-s
https://api.github.com/repos/bitcoin-s/bitcoin-s
closed
Update secp256k1 build docs
bug documentation
**Describe the bug** https://bitcoin-s.org/docs/secp256k1/jni-modify contains "You will need to go to the bitcoin-s/secp256k1" which should now be "bitcoin-s/secp256k1-zkp" May also want to add steps to build on Mac M1 architecture https://github.com/bitcoin-s/secp256k1-zkp/commit/35d09472e1647510f54ee9ebe34fa338a7e6e446
1.0
Update secp256k1 build docs - **Describe the bug** https://bitcoin-s.org/docs/secp256k1/jni-modify contains "You will need to go to the bitcoin-s/secp256k1" which should now be "bitcoin-s/secp256k1-zkp" May also want to add steps to build on Mac M1 architecture https://github.com/bitcoin-s/secp256k1-zkp/commit/35d09472e1647510f54ee9ebe34fa338a7e6e446
non_priority
update build docs describe the bug contains you will need to go to the bitcoin s which should now be bitcoin s zkp may also want to add steps to build on mac architecture
0
208,141
16,104,974,179
IssuesEvent
2021-04-27 13:59:04
GuillaumeFalourd/developers-tips-and-tricks-resources
https://api.github.com/repos/GuillaumeFalourd/developers-tips-and-tricks-resources
opened
Add new cheat-sheets contents
documentation enhancement
## What is expected? Add new cheat-sheets to the [related section](https://github.com/GuillaumeFalourd/developers-tips-and-tricks-resources/tree/master/cheat-sheets) respecting the [repository guidelines](https://github.com/GuillaumeFalourd/developers-tips-and-tricks-resources/blob/master/CONTRIBUTION.md)
1.0
Add new cheat-sheets contents - ## What is expected? Add new cheat-sheets to the [related section](https://github.com/GuillaumeFalourd/developers-tips-and-tricks-resources/tree/master/cheat-sheets) respecting the [repository guidelines](https://github.com/GuillaumeFalourd/developers-tips-and-tricks-resources/blob/master/CONTRIBUTION.md)
non_priority
add new cheat sheets contents what is expected add new cheat sheets to the respecting the
0
31,198
25,408,425,391
IssuesEvent
2022-11-22 16:55:35
SonarSource/sonarlint-visualstudio
https://api.github.com/repos/SonarSource/sonarlint-visualstudio
opened
[Infra] Auto-regenerate the list of Sonar Roslyn rules that can be suppressed on build
Infrastructure
### Description The list of issues C#/VB rules that can be suppress in Connected Mode is generated using a [T4 template](https://github.com/SonarSource/sonarlint-visualstudio/blob/69de0102f4155f7f45aafebbc41b0b79f2958af2/src/Roslyn.Suppressions/Roslyn.Suppressions/SupportedSuppressionBuilder.tt#L1). At present the code is only regenerated manually i.e. we have to remember to do it each time we update the Roslyn analyzers. We should use the T4 build integration to cause the file to be updated automatically on build.
1.0
[Infra] Auto-regenerate the list of Sonar Roslyn rules that can be suppressed on build - ### Description The list of issues C#/VB rules that can be suppress in Connected Mode is generated using a [T4 template](https://github.com/SonarSource/sonarlint-visualstudio/blob/69de0102f4155f7f45aafebbc41b0b79f2958af2/src/Roslyn.Suppressions/Roslyn.Suppressions/SupportedSuppressionBuilder.tt#L1). At present the code is only regenerated manually i.e. we have to remember to do it each time we update the Roslyn analyzers. We should use the T4 build integration to cause the file to be updated automatically on build.
non_priority
auto regenerate the list of sonar roslyn rules that can be suppressed on build description the list of issues c vb rules that can be suppress in connected mode is generated using a at present the code is only regenerated manually i e we have to remember to do it each time we update the roslyn analyzers we should use the build integration to cause the file to be updated automatically on build
0
167,426
20,726,076,551
IssuesEvent
2022-03-14 02:09:35
hitokoto-osc/hitokoto-api
https://api.github.com/repos/hitokoto-osc/hitokoto-api
closed
commitizen-4.1.2.tgz: 4 vulnerabilities (highest severity is: 9.8)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commitizen-4.1.2.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/commitizen/node_modules/lodash/package.json</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2020-28499](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | merge-1.2.1.tgz | Transitive | 4.2.4 | &#10060; | | [CVE-2020-8203](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.4 | lodash-4.17.15.tgz | Transitive | 4.1.3 | &#10060; | | [CVE-2021-23337](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.2 | lodash-4.17.15.tgz | Transitive | 4.1.3 | &#10060; | | [CVE-2020-28500](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | lodash-4.17.15.tgz | Transitive | 4.1.3 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-28499</summary> ### Vulnerable Library - <b>merge-1.2.1.tgz</b></p> <p>Merge multiple objects into one, optionally creating a new cloned object. Similar to the jQuery.extend but more flexible. Works in Node.js and the browser.</p> <p>Library home page: <a href="https://registry.npmjs.org/merge/-/merge-1.2.1.tgz">https://registry.npmjs.org/merge/-/merge-1.2.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/merge/package.json</p> <p> Dependency Hierarchy: - commitizen-4.1.2.tgz (Root Library) - find-node-modules-2.0.0.tgz - :x: **merge-1.2.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> All versions of package merge are vulnerable to Prototype Pollution via _recursiveMerge . <p>Publish Date: 2021-02-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499>CVE-2020-28499</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/yeikos/js.merge/pull/38">https://github.com/yeikos/js.merge/pull/38</a></p> <p>Release Date: 2021-02-18</p> <p>Fix Resolution (merge): 2.1.1</p> <p>Direct dependency fix Resolution (commitizen): 4.2.4</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-8203</summary> ### Vulnerable Library - <b>lodash-4.17.15.tgz</b></p> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/commitizen/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - commitizen-4.1.2.tgz (Root Library) - :x: **lodash-4.17.15.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20. <p>Publish Date: 2020-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203>CVE-2020-8203</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.4</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1523">https://www.npmjs.com/advisories/1523</a></p> <p>Release Date: 2020-07-15</p> <p>Fix Resolution (lodash): 4.17.19</p> <p>Direct dependency fix Resolution (commitizen): 4.1.3</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-23337</summary> ### Vulnerable Library - <b>lodash-4.17.15.tgz</b></p> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/commitizen/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - commitizen-4.1.2.tgz (Root Library) - :x: **lodash-4.17.15.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.2</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution (lodash): 4.17.21</p> <p>Direct dependency fix Resolution (commitizen): 4.1.3</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-28500</summary> ### Vulnerable Library - <b>lodash-4.17.15.tgz</b></p> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/commitizen/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - commitizen-4.1.2.tgz (Root Library) - :x: **lodash-4.17.15.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions. WhiteSource Note: After conducting further research, WhiteSource has determined that CVE-2020-28500 only affects environments with versions 4.0.0 to 4.17.20 of Lodash. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500>CVE-2020-28500</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution (lodash): 4.17.21</p> <p>Direct dependency fix Resolution (commitizen): 4.1.3</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details> <!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"commitizen","packageVersion":"4.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"commitizen:4.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.2.4","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-28499","vulnerabilityDetails":"All versions of package merge are vulnerable to Prototype Pollution via _recursiveMerge .","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"commitizen","packageVersion":"4.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"commitizen:4.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.1.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-8203","vulnerabilityDetails":"Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203","cvss3Severity":"high","cvss3Score":"7.4","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"commitizen","packageVersion":"4.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"commitizen:4.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.1.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23337","vulnerabilityDetails":"Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337","cvss3Severity":"high","cvss3Score":"7.2","cvss3Metrics":{"A":"High","AC":"Low","PR":"High","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"commitizen","packageVersion":"4.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"commitizen:4.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.1.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-28500","vulnerabilityDetails":"Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions.\n WhiteSource Note: After conducting further research, WhiteSource has determined that CVE-2020-28500 only affects environments with versions 4.0.0 to 4.17.20 of Lodash.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}]</REMEDIATE> -->
True
commitizen-4.1.2.tgz: 4 vulnerabilities (highest severity is: 9.8) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commitizen-4.1.2.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/commitizen/node_modules/lodash/package.json</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2020-28499](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | merge-1.2.1.tgz | Transitive | 4.2.4 | &#10060; | | [CVE-2020-8203](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.4 | lodash-4.17.15.tgz | Transitive | 4.1.3 | &#10060; | | [CVE-2021-23337](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.2 | lodash-4.17.15.tgz | Transitive | 4.1.3 | &#10060; | | [CVE-2020-28500](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | lodash-4.17.15.tgz | Transitive | 4.1.3 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-28499</summary> ### Vulnerable Library - <b>merge-1.2.1.tgz</b></p> <p>Merge multiple objects into one, optionally creating a new cloned object. Similar to the jQuery.extend but more flexible. Works in Node.js and the browser.</p> <p>Library home page: <a href="https://registry.npmjs.org/merge/-/merge-1.2.1.tgz">https://registry.npmjs.org/merge/-/merge-1.2.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/merge/package.json</p> <p> Dependency Hierarchy: - commitizen-4.1.2.tgz (Root Library) - find-node-modules-2.0.0.tgz - :x: **merge-1.2.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> All versions of package merge are vulnerable to Prototype Pollution via _recursiveMerge . <p>Publish Date: 2021-02-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499>CVE-2020-28499</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/yeikos/js.merge/pull/38">https://github.com/yeikos/js.merge/pull/38</a></p> <p>Release Date: 2021-02-18</p> <p>Fix Resolution (merge): 2.1.1</p> <p>Direct dependency fix Resolution (commitizen): 4.2.4</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-8203</summary> ### Vulnerable Library - <b>lodash-4.17.15.tgz</b></p> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/commitizen/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - commitizen-4.1.2.tgz (Root Library) - :x: **lodash-4.17.15.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20. <p>Publish Date: 2020-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203>CVE-2020-8203</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.4</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1523">https://www.npmjs.com/advisories/1523</a></p> <p>Release Date: 2020-07-15</p> <p>Fix Resolution (lodash): 4.17.19</p> <p>Direct dependency fix Resolution (commitizen): 4.1.3</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-23337</summary> ### Vulnerable Library - <b>lodash-4.17.15.tgz</b></p> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/commitizen/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - commitizen-4.1.2.tgz (Root Library) - :x: **lodash-4.17.15.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.2</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution (lodash): 4.17.21</p> <p>Direct dependency fix Resolution (commitizen): 4.1.3</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-28500</summary> ### Vulnerable Library - <b>lodash-4.17.15.tgz</b></p> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/commitizen/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - commitizen-4.1.2.tgz (Root Library) - :x: **lodash-4.17.15.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions. WhiteSource Note: After conducting further research, WhiteSource has determined that CVE-2020-28500 only affects environments with versions 4.0.0 to 4.17.20 of Lodash. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500>CVE-2020-28500</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution (lodash): 4.17.21</p> <p>Direct dependency fix Resolution (commitizen): 4.1.3</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details> <!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"commitizen","packageVersion":"4.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"commitizen:4.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.2.4","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-28499","vulnerabilityDetails":"All versions of package merge are vulnerable to Prototype Pollution via _recursiveMerge .","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"commitizen","packageVersion":"4.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"commitizen:4.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.1.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-8203","vulnerabilityDetails":"Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203","cvss3Severity":"high","cvss3Score":"7.4","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"commitizen","packageVersion":"4.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"commitizen:4.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.1.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23337","vulnerabilityDetails":"Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337","cvss3Severity":"high","cvss3Score":"7.2","cvss3Metrics":{"A":"High","AC":"Low","PR":"High","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"commitizen","packageVersion":"4.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"commitizen:4.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.1.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-28500","vulnerabilityDetails":"Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions.\n WhiteSource Note: After conducting further research, WhiteSource has determined that CVE-2020-28500 only affects environments with versions 4.0.0 to 4.17.20 of Lodash.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}]</REMEDIATE> -->
non_priority
commitizen tgz vulnerabilities highest severity is vulnerable library commitizen tgz path to dependency file package json path to vulnerable library node modules commitizen node modules lodash package json vulnerabilities cve severity cvss dependency type fixed in remediation available high merge tgz transitive high lodash tgz transitive high lodash tgz transitive medium lodash tgz transitive details cve vulnerable library merge tgz merge multiple objects into one optionally creating a new cloned object similar to the jquery extend but more flexible works in node js and the browser library home page a href path to dependency file package json path to vulnerable library node modules merge package json dependency hierarchy commitizen tgz root library find node modules tgz x merge tgz vulnerable library found in base branch master vulnerability details all versions of package merge are vulnerable to prototype pollution via recursivemerge publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution merge direct dependency fix resolution commitizen step up your open source security game with whitesource cve vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules commitizen node modules lodash package json dependency hierarchy commitizen tgz root library x lodash tgz vulnerable library found in base branch master vulnerability details prototype pollution attack when using zipobjectdeep in lodash before publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash direct dependency fix resolution commitizen step up your open source security game with whitesource cve vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules commitizen node modules lodash package json dependency hierarchy commitizen tgz root library x lodash tgz vulnerable library found in base branch master vulnerability details lodash versions prior to are vulnerable to command injection via the template function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash direct dependency fix resolution commitizen step up your open source security game with whitesource cve vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules commitizen node modules lodash package json dependency hierarchy commitizen tgz root library x lodash tgz vulnerable library found in base branch master vulnerability details lodash versions prior to are vulnerable to regular expression denial of service redos via the tonumber trim and trimend functions whitesource note after conducting further research whitesource has determined that cve only affects environments with versions to of lodash publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash direct dependency fix resolution commitizen step up your open source security game with whitesource istransitivedependency false dependencytree commitizen isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails all versions of package merge are vulnerable to prototype pollution via recursivemerge vulnerabilityurl istransitivedependency false dependencytree commitizen isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails prototype pollution attack when using zipobjectdeep in lodash before vulnerabilityurl istransitivedependency false dependencytree commitizen isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails lodash versions prior to are vulnerable to command injection via the template function vulnerabilityurl istransitivedependency false dependencytree commitizen isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails lodash versions prior to are vulnerable to regular expression denial of service redos via the tonumber trim and trimend functions n whitesource note after conducting further research whitesource has determined that cve only affects environments with versions to of lodash vulnerabilityurl
0
2,537
4,760,262,236
IssuesEvent
2016-10-25 02:15:16
CactusDev/CactusBot
https://api.github.com/repos/CactusDev/CactusBot
opened
Discord service errors if response length too long
bug service:discord
The maximum length for Discord messages is 2,000 characters. If the response is too long, the bot should split up the message and send individually.
1.0
Discord service errors if response length too long - The maximum length for Discord messages is 2,000 characters. If the response is too long, the bot should split up the message and send individually.
non_priority
discord service errors if response length too long the maximum length for discord messages is characters if the response is too long the bot should split up the message and send individually
0
122,114
17,687,357,294
IssuesEvent
2021-08-24 04:47:11
panasalap/sqlite-3.22.0
https://api.github.com/repos/panasalap/sqlite-3.22.0
opened
CVE-2019-9937 (High) detected in multiple libraries
security vulnerability
## CVE-2019-9937 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>sqliteversion-3.22.0</b>, <b>sqliteversion-3.22.0</b>, <b>sqliteversion-3.22.0</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In SQLite 3.27.2, interleaving reads and writes in a single transaction with an fts5 virtual table will lead to a NULL Pointer Dereference in fts5ChunkIterate in sqlite3.c. This is related to ext/fts5/fts5_hash.c and ext/fts5/fts5_index.c. <p>Publish Date: 2019-03-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-9937>CVE-2019-9937</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.sqlite.org/releaselog/3_28_0.html">https://www.sqlite.org/releaselog/3_28_0.html</a></p> <p>Release Date: 2019-03-22</p> <p>Fix Resolution: 3.28.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-9937 (High) detected in multiple libraries - ## CVE-2019-9937 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>sqliteversion-3.22.0</b>, <b>sqliteversion-3.22.0</b>, <b>sqliteversion-3.22.0</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In SQLite 3.27.2, interleaving reads and writes in a single transaction with an fts5 virtual table will lead to a NULL Pointer Dereference in fts5ChunkIterate in sqlite3.c. This is related to ext/fts5/fts5_hash.c and ext/fts5/fts5_index.c. <p>Publish Date: 2019-03-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-9937>CVE-2019-9937</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.sqlite.org/releaselog/3_28_0.html">https://www.sqlite.org/releaselog/3_28_0.html</a></p> <p>Release Date: 2019-03-22</p> <p>Fix Resolution: 3.28.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries sqliteversion sqliteversion sqliteversion vulnerability details in sqlite interleaving reads and writes in a single transaction with an virtual table will lead to a null pointer dereference in in c this is related to ext hash c and ext index c publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
72,722
19,430,764,592
IssuesEvent
2021-12-21 11:40:12
openego/eGon-data
https://api.github.com/repos/openego/eGon-data
closed
Insert cost assumptions from NEP and technology-data to scenario_parameters
:building_construction: integration
Fuel and CO2 prices will be added to the scenario parameters table according to table 6 in NEP 2035, Version 2021, 1. Entwurf (S.39). Other marginal and capital costs are taken from technology-data repository
1.0
Insert cost assumptions from NEP and technology-data to scenario_parameters - Fuel and CO2 prices will be added to the scenario parameters table according to table 6 in NEP 2035, Version 2021, 1. Entwurf (S.39). Other marginal and capital costs are taken from technology-data repository
non_priority
insert cost assumptions from nep and technology data to scenario parameters fuel and prices will be added to the scenario parameters table according to table in nep version entwurf s other marginal and capital costs are taken from technology data repository
0
97,575
16,236,393,319
IssuesEvent
2021-05-07 01:37:41
michaeldotson/mini-capstone
https://api.github.com/repos/michaeldotson/mini-capstone
opened
CVE-2020-7595 (High) detected in nokogiri-1.10.3.gem
security vulnerability
## CVE-2020-7595 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nokogiri-1.10.3.gem</b></p></summary> <p>Nokogiri (鋸) is an HTML, XML, SAX, and Reader parser. Among Nokogiri's many features is the ability to search documents via XPath or CSS3 selectors.</p> <p>Library home page: <a href="https://rubygems.org/gems/nokogiri-1.10.3.gem">https://rubygems.org/gems/nokogiri-1.10.3.gem</a></p> <p>Path to dependency file: /mini-capstone/Gemfile.lock</p> <p>Path to vulnerable library: /var/lib/gems/2.3.0/cache/nokogiri-1.10.3.gem</p> <p> Dependency Hierarchy: - sass-rails-5.0.7.gem (Root Library) - sprockets-rails-3.2.1.gem - actionpack-5.2.2.gem - rails-html-sanitizer-1.0.4.gem - loofah-2.2.3.gem - :x: **nokogiri-1.10.3.gem** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> xmlStringLenDecodeEntities in parser.c in libxml2 2.9.10 has an infinite loop in a certain end-of-file situation. <p>Publish Date: 2020-01-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7595>CVE-2020-7595</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://security.gentoo.org/glsa/202010-04">https://security.gentoo.org/glsa/202010-04</a></p> <p>Fix Resolution: All libxml2 users should upgrade to the latest version # emerge --sync # emerge --ask --oneshot --verbose >=dev-libs/libxml2-2.9.10 >= </p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7595 (High) detected in nokogiri-1.10.3.gem - ## CVE-2020-7595 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nokogiri-1.10.3.gem</b></p></summary> <p>Nokogiri (鋸) is an HTML, XML, SAX, and Reader parser. Among Nokogiri's many features is the ability to search documents via XPath or CSS3 selectors.</p> <p>Library home page: <a href="https://rubygems.org/gems/nokogiri-1.10.3.gem">https://rubygems.org/gems/nokogiri-1.10.3.gem</a></p> <p>Path to dependency file: /mini-capstone/Gemfile.lock</p> <p>Path to vulnerable library: /var/lib/gems/2.3.0/cache/nokogiri-1.10.3.gem</p> <p> Dependency Hierarchy: - sass-rails-5.0.7.gem (Root Library) - sprockets-rails-3.2.1.gem - actionpack-5.2.2.gem - rails-html-sanitizer-1.0.4.gem - loofah-2.2.3.gem - :x: **nokogiri-1.10.3.gem** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> xmlStringLenDecodeEntities in parser.c in libxml2 2.9.10 has an infinite loop in a certain end-of-file situation. <p>Publish Date: 2020-01-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7595>CVE-2020-7595</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://security.gentoo.org/glsa/202010-04">https://security.gentoo.org/glsa/202010-04</a></p> <p>Fix Resolution: All libxml2 users should upgrade to the latest version # emerge --sync # emerge --ask --oneshot --verbose >=dev-libs/libxml2-2.9.10 >= </p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in nokogiri gem cve high severity vulnerability vulnerable library nokogiri gem nokogiri 鋸 is an html xml sax and reader parser among nokogiri s many features is the ability to search documents via xpath or selectors library home page a href path to dependency file mini capstone gemfile lock path to vulnerable library var lib gems cache nokogiri gem dependency hierarchy sass rails gem root library sprockets rails gem actionpack gem rails html sanitizer gem loofah gem x nokogiri gem vulnerable library vulnerability details xmlstringlendecodeentities in parser c in has an infinite loop in a certain end of file situation publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href fix resolution all users should upgrade to the latest version emerge sync emerge ask oneshot verbose dev libs step up your open source security game with whitesource
0
26,943
4,839,165,210
IssuesEvent
2016-11-09 08:21:20
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
Timeline inside tabView not working
6.0.8 defect
Reported by Pro User; > We are using timeline component and it's working. Now, I need to use a <p:tabView, and when I moved the timeline's code inside this component, then the timeline's event stops working (the call to the method in the backend bean never happens) and when the oncomplete is processed, args are null and the process stops. I need the listener method to be executed before in order to load the specific data and updates the dialog with it. Note: I want the tabView on demand loading. Testing code: xhtml: ``` <script type="text/javascript"> /* <![CDATA[ */ function handleLoadOfferData(xhr, status, args) { alert('args = ' || args); PF('statusDialog').hide(); if (args.notValid || args.validationFailed || !args.dataOk) { return; } PF('detailsTLWdgt').show(); } /* ]]> */ </script> <h:form id="hForm"> <p:tabView id="timelineDetail" dynamic="true" cache="true"> <p:tab title="Tab1"> <h:panelGroup layout="block" styleClass="infoBox" rendered="#{bean.tlModel eq null}"> <span class="ui-icon ui-icon-alert" style="float: left;"></span> <h:outputText value="Model without elements"/> </h:panelGroup> <p:timeline id="timelineNew" value="#{bean.tlModel}" var="element" varGroup="timelineGroup" height="400px" axisOnTop="true" eventStyle="dot" rendered="#{not empty bean.tlModel != null}"> <p:ajax event="select" listener="#{bean.onSelect}" process="@form" onstart="PF('statusDialog').show();" update=":hForm:timelineDetail:detailsTLDlg" oncomplete="handleLoadOfferData(xhr,status,args)"/> <h:panelGrid columns="1"> <h:outputText value="#{element.description}"/> </h:panelGrid> </p:timeline> <!-- Offer Details Modal --> <p:dialog id="detailsTLDlg" widgetVar="detailsTLWdgt" header="Dialog with details" modal="true" resizable="false" draggable="false" width="650" height="400" rendered="#{bean.tlModel != null}"> Data loaded in onSelect method <p:commandButton type="button" value="Close" onclick="PF('detailsTLWdgt').hide()" icon="ui-icon-closethick"/> </p:dialog> </p:tab> <p:tab title="Tab 2"> dfafdsa </p:tab> </p:tabView> </h:form> ``` Backend bean: ``` @ManagedBean @ViewScoped public class Bean implements Serializable { .......... private TimelineModel tlModel; private MpanTimeLineVO detail; .......... public void onSelect(TimelineSelectEvent e) throws Exception { LOG.debug("onSelect"); TimelineEvent timelineEvent = e.getTimelineEvent(); if (timelineEvent.getData() == null) { throw new RuntimeException("No data selected"); } TimelineBasicVO data = (TimelineBasicVO) timelineEvent.getData(); detail = getDataDetail(data.getOfferId()); JSFHelper.getRequestContext().addCallbackParam("dataOk", Boolean.TRUE); } ....... } ```
1.0
Timeline inside tabView not working - Reported by Pro User; > We are using timeline component and it's working. Now, I need to use a <p:tabView, and when I moved the timeline's code inside this component, then the timeline's event stops working (the call to the method in the backend bean never happens) and when the oncomplete is processed, args are null and the process stops. I need the listener method to be executed before in order to load the specific data and updates the dialog with it. Note: I want the tabView on demand loading. Testing code: xhtml: ``` <script type="text/javascript"> /* <![CDATA[ */ function handleLoadOfferData(xhr, status, args) { alert('args = ' || args); PF('statusDialog').hide(); if (args.notValid || args.validationFailed || !args.dataOk) { return; } PF('detailsTLWdgt').show(); } /* ]]> */ </script> <h:form id="hForm"> <p:tabView id="timelineDetail" dynamic="true" cache="true"> <p:tab title="Tab1"> <h:panelGroup layout="block" styleClass="infoBox" rendered="#{bean.tlModel eq null}"> <span class="ui-icon ui-icon-alert" style="float: left;"></span> <h:outputText value="Model without elements"/> </h:panelGroup> <p:timeline id="timelineNew" value="#{bean.tlModel}" var="element" varGroup="timelineGroup" height="400px" axisOnTop="true" eventStyle="dot" rendered="#{not empty bean.tlModel != null}"> <p:ajax event="select" listener="#{bean.onSelect}" process="@form" onstart="PF('statusDialog').show();" update=":hForm:timelineDetail:detailsTLDlg" oncomplete="handleLoadOfferData(xhr,status,args)"/> <h:panelGrid columns="1"> <h:outputText value="#{element.description}"/> </h:panelGrid> </p:timeline> <!-- Offer Details Modal --> <p:dialog id="detailsTLDlg" widgetVar="detailsTLWdgt" header="Dialog with details" modal="true" resizable="false" draggable="false" width="650" height="400" rendered="#{bean.tlModel != null}"> Data loaded in onSelect method <p:commandButton type="button" value="Close" onclick="PF('detailsTLWdgt').hide()" icon="ui-icon-closethick"/> </p:dialog> </p:tab> <p:tab title="Tab 2"> dfafdsa </p:tab> </p:tabView> </h:form> ``` Backend bean: ``` @ManagedBean @ViewScoped public class Bean implements Serializable { .......... private TimelineModel tlModel; private MpanTimeLineVO detail; .......... public void onSelect(TimelineSelectEvent e) throws Exception { LOG.debug("onSelect"); TimelineEvent timelineEvent = e.getTimelineEvent(); if (timelineEvent.getData() == null) { throw new RuntimeException("No data selected"); } TimelineBasicVO data = (TimelineBasicVO) timelineEvent.getData(); detail = getDataDetail(data.getOfferId()); JSFHelper.getRequestContext().addCallbackParam("dataOk", Boolean.TRUE); } ....... } ```
non_priority
timeline inside tabview not working reported by pro user we are using timeline component and it s working now i need to use a p tabview and when i moved the timeline s code inside this component then the timeline s event stops working the call to the method in the backend bean never happens and when the oncomplete is processed args are null and the process stops i need the listener method to be executed before in order to load the specific data and updates the dialog with it note i want the tabview on demand loading testing code xhtml cdata function handleloadofferdata xhr status args alert args args pf statusdialog hide if args notvalid args validationfailed args dataok return pf detailstlwdgt show p timeline id timelinenew value bean tlmodel var element vargroup timelinegroup height axisontop true eventstyle dot rendered not empty bean tlmodel null p ajax event select listener bean onselect process form onstart pf statusdialog show update hform timelinedetail detailstldlg oncomplete handleloadofferdata xhr status args p dialog id detailstldlg widgetvar detailstlwdgt header dialog with details modal true resizable false draggable false width height rendered bean tlmodel null data loaded in onselect method p commandbutton type button value close onclick pf detailstlwdgt hide icon ui icon closethick dfafdsa backend bean managedbean viewscoped public class bean implements serializable private timelinemodel tlmodel private mpantimelinevo detail public void onselect timelineselectevent e throws exception log debug onselect timelineevent timelineevent e gettimelineevent if timelineevent getdata null throw new runtimeexception no data selected timelinebasicvo data timelinebasicvo timelineevent getdata detail getdatadetail data getofferid jsfhelper getrequestcontext addcallbackparam dataok boolean true
0
25,699
7,741,320,151
IssuesEvent
2018-05-29 04:58:46
ShaikASK/Testing
https://api.github.com/repos/ShaikASK/Testing
closed
HR Admin & HR user : Forgot Password : User unable to access the forgot password page
Defect Forgot Password HR Admin Module HR User Module P1 Release#2 Build #1
Steps to Replicate: 1. Launch the url : 2. Navigate to "signin" page 3. Click on "Forgot your Password" link 4. Enter valid email id in the email id field 5. Click on "Submit" button Experienced Behavior : Observed that user is navigated to apache tomcat page Expected Behavior : Ensure that user should be able to update the password using "forgot password page"
1.0
HR Admin & HR user : Forgot Password : User unable to access the forgot password page - Steps to Replicate: 1. Launch the url : 2. Navigate to "signin" page 3. Click on "Forgot your Password" link 4. Enter valid email id in the email id field 5. Click on "Submit" button Experienced Behavior : Observed that user is navigated to apache tomcat page Expected Behavior : Ensure that user should be able to update the password using "forgot password page"
non_priority
hr admin hr user forgot password user unable to access the forgot password page steps to replicate launch the url navigate to signin page click on forgot your password link enter valid email id in the email id field click on submit button experienced behavior observed that user is navigated to apache tomcat page expected behavior ensure that user should be able to update the password using forgot password page
0
70,210
15,057,546,885
IssuesEvent
2021-02-03 21:51:54
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
opened
Security - iterate to improve the management of user access
needs-grooming operations security
## Problem Statement Maintaining credentials is difficult, and we lack the ability to grant fine-grained access to specific platform resources, without granting broad administrative access which is outside of the scope of a developers given role. This is caused by permission redundancy between groups and there being no unified strategy to granting permissions. This results in over-privileged users that have more access than needed to perform their role / job function. This also results in an expectation of needing elevated privileges that are outside of the scope of day-to-day work in order to carry on one's tasks. There is also inconsistency in the existing design of various security-related resources. How might we add easily manage credentials and access to platform resources in a way that grants only the access which is required for a developers given job function? How might we integrate this new credentials management workflow and associated tools into ops-owned applications and services? How might we reduce administrative overhead by automating these processes and workflows for granting access to resources? ## Hypothesis or Bet Our hypothesis is that by implementing a solution for better management of credentials, we can improve platform security, and the administration thereof. ## We will know we're done when... ("Definition of Done") We have migrated permissions from credstash to Amazon's parameter store for all platform resources, ensuring that only required access is granted. There is a clear, documented process in place for managing credentials in a way that is easy to do and requires minimal manual administration. There is plan in place for how applications will use only SSM integration (and can be migrated completely off of cred-stash). When we've considerably reduced policy sprawl (ie clear, straight-forward, easy-to-understand policy in place). When we have a plan for how to automate the securing of non-compliant credentials. Maybe: When we have a plan for how to notify users of non-compliant credentials. ## Known Blockers/Dependencies *List any blockers or dependencies for this work to be completed* ## Projected Launch Date By the end of Q1 2021, we expect to have new permissions management system in place. ## Launch Checklist ### Is this service / tool / feature... ### ... tested? - [ ] Usability test (_TODO: link_) has been performed, to validate that new changes enable users to do what was intended and that these changes don't worsen quality elsewhere. If usability test isn't relevant for this change, document the reason for skipping it. - [ ] ... and issues discovered in usability testing have been addressed. * _Note on skipping: metrics that show the impact of before/after can be a substitute for usability testing._ - [ ] End-to-end [manual QA](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/quality-assurance/README.md) or [UAT](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/research/planning/what-is-uat.md) is complete, to validate there are no high-severity issues before launching - [ ] _(if applicable)_ New functionality has thorough, automated tests running in CI/CD ### ... documented? - [ ] New documentation is written pursuant to our [documentation style guide](https://github.com/department-of-veterans-affairs/va.gov-team/tree/master/platform/documentation/style-guide) - [ ] Product is included in the [List of VSP Products](https://docs.google.com/spreadsheets/d/1Fn2lD419WE3sTZJtN2Ensrjqaz0jH3WvLaBtn812Wjo/edit#gid=0) * _List the existing product that this initiative fits within, or add a new product to this list._ - [ ] Internal-facing: there's a [Product Outline](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/teams/vsp/product-management/product-outline-template.md) checked into [`products/platform/PRODUCT_NAME/`](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/products/platform/) * _Note: the Product Directory Name should match 1:1 with the List of VSP Products_ - [ ] External-facing: a [VFS-facing README](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/teams/vsp/product-management/product-readme-template.md) exists for this product/feature tool - [ ] ... and should be located at `platform/PRODUCT_NAME/README.md` - [ ] External-facing: a [User Guide](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/teams/vsp/product-management/writing-user-guides.md) exists for this product/feature/tool, and is updated for changes from this initiative - [ ] ... and should be linked from the VFS-facing README for your product - [ ] ... and should be located within `platform/PRODUCT_NAME/`, unless you already have another location for it - [ ] _(if applicable)_... and post to [#vsp-content-ia](https://dsva.slack.com/channels/vsp-content-ia) about whether this should be added to the [Documentation homepage](https://department-of-veterans-affairs.github.io/va.gov-team/) - [ ] _(if applicable)_ Post to [#vsp-service-design](https://dsva.slack.com/channels/vsp-service-design) for external communication about this change (e.g. VSP Newsletter, customer-facing meetings) ### ... measurable - [ ] _(if applicable)_ This change has clearly-defined success metrics, with instrumentation of those analytics where possible, or a reason documented for skipping it. * For help, see: [Analytics team](https://github.com/department-of-veterans-affairs/va.gov-team/tree/master/platform/analytics) - [ ] This change has an accompanying [VSP Initiative Release Plan](https://github.com/department-of-veterans-affairs/va.gov-team/issues/new/choose). ## Required Artifacts ### Documentation * **`PRODUCT_NAME`**: _directory name used for your product documentation_ * **Product Outline**: _link to Product Outline_ * **README**: _link to VFS-facing README for your product_ * **User Guide**: _link to User Guide_ ### Testing * **Usability test**: _link to GitHub issue, or provide reason for skipping_ * **Manual QA**: _link to GitHub issue or documented results_ * **Automated tests**: _link to tests, or "N/A"_ ### Measurement * **Success metrics**: - developer onboarding time ~24 hours (match current request SLA) - no empty groups in IAM - only operations team members have administrative privileges (unless service account and/or unless justifiable, documented reason for enhanced privileges for a given user) - full accounting / audit of existing policies * **Release plan**: _link to Release Plan ticket_ ## TODOs - [x] Convert this issue to an epic - [x] Add your team's label to this epic
True
Security - iterate to improve the management of user access - ## Problem Statement Maintaining credentials is difficult, and we lack the ability to grant fine-grained access to specific platform resources, without granting broad administrative access which is outside of the scope of a developers given role. This is caused by permission redundancy between groups and there being no unified strategy to granting permissions. This results in over-privileged users that have more access than needed to perform their role / job function. This also results in an expectation of needing elevated privileges that are outside of the scope of day-to-day work in order to carry on one's tasks. There is also inconsistency in the existing design of various security-related resources. How might we add easily manage credentials and access to platform resources in a way that grants only the access which is required for a developers given job function? How might we integrate this new credentials management workflow and associated tools into ops-owned applications and services? How might we reduce administrative overhead by automating these processes and workflows for granting access to resources? ## Hypothesis or Bet Our hypothesis is that by implementing a solution for better management of credentials, we can improve platform security, and the administration thereof. ## We will know we're done when... ("Definition of Done") We have migrated permissions from credstash to Amazon's parameter store for all platform resources, ensuring that only required access is granted. There is a clear, documented process in place for managing credentials in a way that is easy to do and requires minimal manual administration. There is plan in place for how applications will use only SSM integration (and can be migrated completely off of cred-stash). When we've considerably reduced policy sprawl (ie clear, straight-forward, easy-to-understand policy in place). When we have a plan for how to automate the securing of non-compliant credentials. Maybe: When we have a plan for how to notify users of non-compliant credentials. ## Known Blockers/Dependencies *List any blockers or dependencies for this work to be completed* ## Projected Launch Date By the end of Q1 2021, we expect to have new permissions management system in place. ## Launch Checklist ### Is this service / tool / feature... ### ... tested? - [ ] Usability test (_TODO: link_) has been performed, to validate that new changes enable users to do what was intended and that these changes don't worsen quality elsewhere. If usability test isn't relevant for this change, document the reason for skipping it. - [ ] ... and issues discovered in usability testing have been addressed. * _Note on skipping: metrics that show the impact of before/after can be a substitute for usability testing._ - [ ] End-to-end [manual QA](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/quality-assurance/README.md) or [UAT](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/research/planning/what-is-uat.md) is complete, to validate there are no high-severity issues before launching - [ ] _(if applicable)_ New functionality has thorough, automated tests running in CI/CD ### ... documented? - [ ] New documentation is written pursuant to our [documentation style guide](https://github.com/department-of-veterans-affairs/va.gov-team/tree/master/platform/documentation/style-guide) - [ ] Product is included in the [List of VSP Products](https://docs.google.com/spreadsheets/d/1Fn2lD419WE3sTZJtN2Ensrjqaz0jH3WvLaBtn812Wjo/edit#gid=0) * _List the existing product that this initiative fits within, or add a new product to this list._ - [ ] Internal-facing: there's a [Product Outline](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/teams/vsp/product-management/product-outline-template.md) checked into [`products/platform/PRODUCT_NAME/`](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/products/platform/) * _Note: the Product Directory Name should match 1:1 with the List of VSP Products_ - [ ] External-facing: a [VFS-facing README](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/teams/vsp/product-management/product-readme-template.md) exists for this product/feature tool - [ ] ... and should be located at `platform/PRODUCT_NAME/README.md` - [ ] External-facing: a [User Guide](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/teams/vsp/product-management/writing-user-guides.md) exists for this product/feature/tool, and is updated for changes from this initiative - [ ] ... and should be linked from the VFS-facing README for your product - [ ] ... and should be located within `platform/PRODUCT_NAME/`, unless you already have another location for it - [ ] _(if applicable)_... and post to [#vsp-content-ia](https://dsva.slack.com/channels/vsp-content-ia) about whether this should be added to the [Documentation homepage](https://department-of-veterans-affairs.github.io/va.gov-team/) - [ ] _(if applicable)_ Post to [#vsp-service-design](https://dsva.slack.com/channels/vsp-service-design) for external communication about this change (e.g. VSP Newsletter, customer-facing meetings) ### ... measurable - [ ] _(if applicable)_ This change has clearly-defined success metrics, with instrumentation of those analytics where possible, or a reason documented for skipping it. * For help, see: [Analytics team](https://github.com/department-of-veterans-affairs/va.gov-team/tree/master/platform/analytics) - [ ] This change has an accompanying [VSP Initiative Release Plan](https://github.com/department-of-veterans-affairs/va.gov-team/issues/new/choose). ## Required Artifacts ### Documentation * **`PRODUCT_NAME`**: _directory name used for your product documentation_ * **Product Outline**: _link to Product Outline_ * **README**: _link to VFS-facing README for your product_ * **User Guide**: _link to User Guide_ ### Testing * **Usability test**: _link to GitHub issue, or provide reason for skipping_ * **Manual QA**: _link to GitHub issue or documented results_ * **Automated tests**: _link to tests, or "N/A"_ ### Measurement * **Success metrics**: - developer onboarding time ~24 hours (match current request SLA) - no empty groups in IAM - only operations team members have administrative privileges (unless service account and/or unless justifiable, documented reason for enhanced privileges for a given user) - full accounting / audit of existing policies * **Release plan**: _link to Release Plan ticket_ ## TODOs - [x] Convert this issue to an epic - [x] Add your team's label to this epic
non_priority
security iterate to improve the management of user access problem statement maintaining credentials is difficult and we lack the ability to grant fine grained access to specific platform resources without granting broad administrative access which is outside of the scope of a developers given role this is caused by permission redundancy between groups and there being no unified strategy to granting permissions this results in over privileged users that have more access than needed to perform their role job function this also results in an expectation of needing elevated privileges that are outside of the scope of day to day work in order to carry on one s tasks there is also inconsistency in the existing design of various security related resources how might we add easily manage credentials and access to platform resources in a way that grants only the access which is required for a developers given job function how might we integrate this new credentials management workflow and associated tools into ops owned applications and services how might we reduce administrative overhead by automating these processes and workflows for granting access to resources hypothesis or bet our hypothesis is that by implementing a solution for better management of credentials we can improve platform security and the administration thereof we will know we re done when definition of done we have migrated permissions from credstash to amazon s parameter store for all platform resources ensuring that only required access is granted there is a clear documented process in place for managing credentials in a way that is easy to do and requires minimal manual administration there is plan in place for how applications will use only ssm integration and can be migrated completely off of cred stash when we ve considerably reduced policy sprawl ie clear straight forward easy to understand policy in place when we have a plan for how to automate the securing of non compliant credentials maybe when we have a plan for how to notify users of non compliant credentials known blockers dependencies list any blockers or dependencies for this work to be completed projected launch date by the end of we expect to have new permissions management system in place launch checklist is this service tool feature tested usability test todo link has been performed to validate that new changes enable users to do what was intended and that these changes don t worsen quality elsewhere if usability test isn t relevant for this change document the reason for skipping it and issues discovered in usability testing have been addressed note on skipping metrics that show the impact of before after can be a substitute for usability testing end to end or is complete to validate there are no high severity issues before launching if applicable new functionality has thorough automated tests running in ci cd documented new documentation is written pursuant to our product is included in the list the existing product that this initiative fits within or add a new product to this list internal facing there s a checked into note the product directory name should match with the list of vsp products external facing a exists for this product feature tool and should be located at platform product name readme md external facing a exists for this product feature tool and is updated for changes from this initiative and should be linked from the vfs facing readme for your product and should be located within platform product name unless you already have another location for it if applicable and post to about whether this should be added to the if applicable post to for external communication about this change e g vsp newsletter customer facing meetings measurable if applicable this change has clearly defined success metrics with instrumentation of those analytics where possible or a reason documented for skipping it for help see this change has an accompanying required artifacts documentation product name directory name used for your product documentation product outline link to product outline readme link to vfs facing readme for your product user guide link to user guide testing usability test link to github issue or provide reason for skipping manual qa link to github issue or documented results automated tests link to tests or n a measurement success metrics developer onboarding time hours match current request sla no empty groups in iam only operations team members have administrative privileges unless service account and or unless justifiable documented reason for enhanced privileges for a given user full accounting audit of existing policies release plan link to release plan ticket todos convert this issue to an epic add your team s label to this epic
0
31,262
6,477,744,173
IssuesEvent
2017-08-18 05:06:02
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
ICache.getAndRemove() doesn't invalidate the Near Cache
Team: Core Type: Defect
`ICache.getAndRemove()` doesn't invalidate the Near Cache when called from the Near Cache side. I noticed this while adding more variants of the `remove()` and `delete()` methods to the Near Cache tests. The failures are all on the `onNearCacheAdapter` method and in the `ClientNearCacheBasicTest`. That's a good indication that a call to the `NearCachedClientCacheProxy` is missing, which invalidates the Near Cache locally.
1.0
ICache.getAndRemove() doesn't invalidate the Near Cache - `ICache.getAndRemove()` doesn't invalidate the Near Cache when called from the Near Cache side. I noticed this while adding more variants of the `remove()` and `delete()` methods to the Near Cache tests. The failures are all on the `onNearCacheAdapter` method and in the `ClientNearCacheBasicTest`. That's a good indication that a call to the `NearCachedClientCacheProxy` is missing, which invalidates the Near Cache locally.
non_priority
icache getandremove doesn t invalidate the near cache icache getandremove doesn t invalidate the near cache when called from the near cache side i noticed this while adding more variants of the remove and delete methods to the near cache tests the failures are all on the onnearcacheadapter method and in the clientnearcachebasictest that s a good indication that a call to the nearcachedclientcacheproxy is missing which invalidates the near cache locally
0
86,603
17,027,780,169
IssuesEvent
2021-07-03 22:59:46
TeamAmaze/AmazeFileManager
https://api.github.com/repos/TeamAmaze/AmazeFileManager
opened
Convert all AsyncTasks to Callables called from ReactiveX
Area-CodeHealth Issue-Discussion Issue-Task (low)
`public class ReadFileTask extends AsyncTask<Void, Void, ReturnedValueOnReadFile> {` to `public class ReadFileTask implements Callable<ReturnedValueOnReadFile> {` and used as ```java final ReadFileTask task = new ReadFileTask(getContentResolver(), viewModel.getFile(), getExternalCacheDir(), isRootExplorer()); final Consumer<ReturnedValueOnReadFile> onAsyncTaskFinished = (data) -> { //operations } final Consumer<? super Throwable> onError = error -> { error.printStackTrace(); } Flowable.fromCallable(task) .subscribeOn(Schedulers.io()) .observeOn(Schedulers.single()) .subscribe(onAsyncTaskFinished, onError); ```
1.0
Convert all AsyncTasks to Callables called from ReactiveX - `public class ReadFileTask extends AsyncTask<Void, Void, ReturnedValueOnReadFile> {` to `public class ReadFileTask implements Callable<ReturnedValueOnReadFile> {` and used as ```java final ReadFileTask task = new ReadFileTask(getContentResolver(), viewModel.getFile(), getExternalCacheDir(), isRootExplorer()); final Consumer<ReturnedValueOnReadFile> onAsyncTaskFinished = (data) -> { //operations } final Consumer<? super Throwable> onError = error -> { error.printStackTrace(); } Flowable.fromCallable(task) .subscribeOn(Schedulers.io()) .observeOn(Schedulers.single()) .subscribe(onAsyncTaskFinished, onError); ```
non_priority
convert all asynctasks to callables called from reactivex public class readfiletask extends asynctask to public class readfiletask implements callable and used as java final readfiletask task new readfiletask getcontentresolver viewmodel getfile getexternalcachedir isrootexplorer final consumer onasynctaskfinished data operations final consumer onerror error error printstacktrace flowable fromcallable task subscribeon schedulers io observeon schedulers single subscribe onasynctaskfinished onerror
0
181,268
21,658,345,552
IssuesEvent
2022-05-06 16:17:42
snowdensb/microsoft-teams-apps-company-communicator
https://api.github.com/repos/snowdensb/microsoft-teams-apps-company-communicator
opened
WS-2020-0438 (Medium) detected in i18next-19.8.2.tgz
security vulnerability
## WS-2020-0438 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>i18next-19.8.2.tgz</b></p></summary> <p>i18next internationalization framework</p> <p>Library home page: <a href="https://registry.npmjs.org/i18next/-/i18next-19.8.2.tgz">https://registry.npmjs.org/i18next/-/i18next-19.8.2.tgz</a></p> <p>Path to dependency file: /Source/CompanyCommunicator/ClientApp/package.json</p> <p>Path to vulnerable library: /Source/CompanyCommunicator/ClientApp/node_modules/i18next/package.json</p> <p> Dependency Hierarchy: - :x: **i18next-19.8.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/snowdensb/microsoft-teams-apps-company-communicator/commit/32bf8559b65e6695a5e727d0fdf93e7ae135d32a">32bf8559b65e6695a5e727d0fdf93e7ae135d32a</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In i18next in versions v19.6.0 to v19.8.2 is vulnerable to prototype pollution, it allows to modify the prototype of a base object, which may result in DoS, XSS, RCE, etc. <p>Publish Date: 2020-02-12 <p>URL: <a href=https://github.com/i18next/i18next/commit/ed6169fa21b67d8e73b20d82579aff6f9b808a69>WS-2020-0438</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://hackerone.com/reports/968355">https://hackerone.com/reports/968355</a></p> <p>Release Date: 2020-02-12</p> <p>Fix Resolution: 19.8.3</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"i18next","packageVersion":"19.8.2","packageFilePaths":["/Source/CompanyCommunicator/ClientApp/package.json"],"isTransitiveDependency":false,"dependencyTree":"i18next:19.8.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"19.8.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2020-0438","vulnerabilityDetails":"In i18next in versions v19.6.0 to v19.8.2 is vulnerable to prototype pollution, it allows to modify the prototype of a base object, which may result in DoS, XSS, RCE, etc.","vulnerabilityUrl":"https://github.com/i18next/i18next/commit/ed6169fa21b67d8e73b20d82579aff6f9b808a69","cvss3Severity":"medium","cvss3Score":"4.8","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
WS-2020-0438 (Medium) detected in i18next-19.8.2.tgz - ## WS-2020-0438 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>i18next-19.8.2.tgz</b></p></summary> <p>i18next internationalization framework</p> <p>Library home page: <a href="https://registry.npmjs.org/i18next/-/i18next-19.8.2.tgz">https://registry.npmjs.org/i18next/-/i18next-19.8.2.tgz</a></p> <p>Path to dependency file: /Source/CompanyCommunicator/ClientApp/package.json</p> <p>Path to vulnerable library: /Source/CompanyCommunicator/ClientApp/node_modules/i18next/package.json</p> <p> Dependency Hierarchy: - :x: **i18next-19.8.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/snowdensb/microsoft-teams-apps-company-communicator/commit/32bf8559b65e6695a5e727d0fdf93e7ae135d32a">32bf8559b65e6695a5e727d0fdf93e7ae135d32a</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In i18next in versions v19.6.0 to v19.8.2 is vulnerable to prototype pollution, it allows to modify the prototype of a base object, which may result in DoS, XSS, RCE, etc. <p>Publish Date: 2020-02-12 <p>URL: <a href=https://github.com/i18next/i18next/commit/ed6169fa21b67d8e73b20d82579aff6f9b808a69>WS-2020-0438</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://hackerone.com/reports/968355">https://hackerone.com/reports/968355</a></p> <p>Release Date: 2020-02-12</p> <p>Fix Resolution: 19.8.3</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"i18next","packageVersion":"19.8.2","packageFilePaths":["/Source/CompanyCommunicator/ClientApp/package.json"],"isTransitiveDependency":false,"dependencyTree":"i18next:19.8.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"19.8.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2020-0438","vulnerabilityDetails":"In i18next in versions v19.6.0 to v19.8.2 is vulnerable to prototype pollution, it allows to modify the prototype of a base object, which may result in DoS, XSS, RCE, etc.","vulnerabilityUrl":"https://github.com/i18next/i18next/commit/ed6169fa21b67d8e73b20d82579aff6f9b808a69","cvss3Severity":"medium","cvss3Score":"4.8","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_priority
ws medium detected in tgz ws medium severity vulnerability vulnerable library tgz internationalization framework library home page a href path to dependency file source companycommunicator clientapp package json path to vulnerable library source companycommunicator clientapp node modules package json dependency hierarchy x tgz vulnerable library found in head commit a href found in base branch master vulnerability details in in versions to is vulnerable to prototype pollution it allows to modify the prototype of a base object which may result in dos xss rce etc publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier ws vulnerabilitydetails in in versions to is vulnerable to prototype pollution it allows to modify the prototype of a base object which may result in dos xss rce etc vulnerabilityurl
0
137,214
18,752,659,704
IssuesEvent
2021-11-05 05:45:46
madhans23/linux-4.15
https://api.github.com/repos/madhans23/linux-4.15
opened
CVE-2018-20669 (High) detected in multiple libraries
security vulnerability
## CVE-2018-20669 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxv4.15</b>, <b>linuxv4.15</b>, <b>linuxv4.15</b>, <b>linuxv4.15</b>, <b>linuxv4.15</b>, <b>linuxv4.15</b>, <b>linux-yocto-devv5.4</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue where a provided address with access_ok() is not checked was discovered in i915_gem_execbuffer2_ioctl in drivers/gpu/drm/i915/i915_gem_execbuffer.c in the Linux kernel through 4.19.13. A local attacker can craft a malicious IOCTL function call to overwrite arbitrary kernel memory, resulting in a Denial of Service or privilege escalation. <p>Publish Date: 2019-03-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20669>CVE-2018-20669</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20669">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20669</a></p> <p>Release Date: 2019-03-21</p> <p>Fix Resolution: v5.0-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-20669 (High) detected in multiple libraries - ## CVE-2018-20669 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxv4.15</b>, <b>linuxv4.15</b>, <b>linuxv4.15</b>, <b>linuxv4.15</b>, <b>linuxv4.15</b>, <b>linuxv4.15</b>, <b>linux-yocto-devv5.4</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue where a provided address with access_ok() is not checked was discovered in i915_gem_execbuffer2_ioctl in drivers/gpu/drm/i915/i915_gem_execbuffer.c in the Linux kernel through 4.19.13. A local attacker can craft a malicious IOCTL function call to overwrite arbitrary kernel memory, resulting in a Denial of Service or privilege escalation. <p>Publish Date: 2019-03-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20669>CVE-2018-20669</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20669">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20669</a></p> <p>Release Date: 2019-03-21</p> <p>Fix Resolution: v5.0-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries linux yocto vulnerability details an issue where a provided address with access ok is not checked was discovered in gem ioctl in drivers gpu drm gem execbuffer c in the linux kernel through a local attacker can craft a malicious ioctl function call to overwrite arbitrary kernel memory resulting in a denial of service or privilege escalation publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
255,964
27,539,554,629
IssuesEvent
2023-03-07 07:27:45
istio/istio
https://api.github.com/repos/istio/istio
closed
Github wiki docs: Broken link regarding mTLS troubleshoot
kind/docs area/security
### Bug Description Hey guys! https://github.com/istio/istio/wiki/Mutual-TLS-TroubleShooting This link have multiple broken links that would help us to fix our issues regarding mTLS / HTTPS / Certificates. ### Version ```prose istioctl version client version: 1.17.1 control plane version: 1.17.1 data plane version: 1.17.1 (9 proxies) kubectl version --short Flag --short has been deprecated, and will be removed in the future. The --short output will become the default. Client Version: v1.26.1 Kustomize Version: v4.5.7 Server Version: v1.25.3 helm version --short ``` ### Additional Information not relevant
True
Github wiki docs: Broken link regarding mTLS troubleshoot - ### Bug Description Hey guys! https://github.com/istio/istio/wiki/Mutual-TLS-TroubleShooting This link have multiple broken links that would help us to fix our issues regarding mTLS / HTTPS / Certificates. ### Version ```prose istioctl version client version: 1.17.1 control plane version: 1.17.1 data plane version: 1.17.1 (9 proxies) kubectl version --short Flag --short has been deprecated, and will be removed in the future. The --short output will become the default. Client Version: v1.26.1 Kustomize Version: v4.5.7 Server Version: v1.25.3 helm version --short ``` ### Additional Information not relevant
non_priority
github wiki docs broken link regarding mtls troubleshoot bug description hey guys this link have multiple broken links that would help us to fix our issues regarding mtls https  certificates version prose istioctl version client version control plane version data plane version proxies kubectl version short flag short has been deprecated and will be removed in the future the short output will become the default client version kustomize version server version helm version short additional information not relevant
0
276,282
23,982,831,069
IssuesEvent
2022-09-13 16:21:51
pitt-crc/bank
https://api.github.com/repos/pitt-crc/bank
opened
Make sure SLURM account locking does not bar access to investment hardware
testing
We should check that our locking mechanism (`GrpTresRunMins = 0`) does not interfere with users ability to submit to hardware they've invested in. A possible proxy test for this is to see if a locked account can still submit to `scavenger`, as the billing weights _should_ be set similarly.
1.0
Make sure SLURM account locking does not bar access to investment hardware - We should check that our locking mechanism (`GrpTresRunMins = 0`) does not interfere with users ability to submit to hardware they've invested in. A possible proxy test for this is to see if a locked account can still submit to `scavenger`, as the billing weights _should_ be set similarly.
non_priority
make sure slurm account locking does not bar access to investment hardware we should check that our locking mechanism grptresrunmins does not interfere with users ability to submit to hardware they ve invested in a possible proxy test for this is to see if a locked account can still submit to scavenger as the billing weights should be set similarly
0
44,175
5,765,120,524
IssuesEvent
2017-04-27 00:57:24
red-eight/Tactics
https://api.github.com/repos/red-eight/Tactics
closed
Unit::PerformNextAction() vs ActionController::ProcessNextAction()
Design
A unified naming convention should be used.
1.0
Unit::PerformNextAction() vs ActionController::ProcessNextAction() - A unified naming convention should be used.
non_priority
unit performnextaction vs actioncontroller processnextaction a unified naming convention should be used
0
8,879
4,347,265,059
IssuesEvent
2016-07-29 18:53:33
mapbox/mapbox-gl-native
https://api.github.com/repos/mapbox/mapbox-gl-native
closed
CocoaPods: Unable to satisfy the following requirements...
build iOS
Hi, I try to install Mapbox iOS SDK 3.2 through Cocoapods (0.39.0) but i have this error : [!] Unable to satisfy the following requirements: - `Mapbox-iOS-SDK (~> 3.2.0)` required by `Podfile` Have you any idea of the problem? Do i need to install the pre-version 1.0.0.beta.6 of Cocoapods to use your last version of Mapbox ?
1.0
CocoaPods: Unable to satisfy the following requirements... - Hi, I try to install Mapbox iOS SDK 3.2 through Cocoapods (0.39.0) but i have this error : [!] Unable to satisfy the following requirements: - `Mapbox-iOS-SDK (~> 3.2.0)` required by `Podfile` Have you any idea of the problem? Do i need to install the pre-version 1.0.0.beta.6 of Cocoapods to use your last version of Mapbox ?
non_priority
cocoapods unable to satisfy the following requirements hi i try to install mapbox ios sdk through cocoapods but i have this error unable to satisfy the following requirements mapbox ios sdk required by podfile have you any idea of the problem do i need to install the pre version beta of cocoapods to use your last version of mapbox
0
200,636
15,116,632,971
IssuesEvent
2021-02-09 07:06:47
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
kv/kvserver: TestReplicateRogueRemovedNode failed
C-test-failure O-robot branch-master
[(kv/kvserver).TestReplicateRogueRemovedNode failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2657499&tab=buildLog) on [master@7853fd32de8b6dea869f2a2a92dcd7506f4a8998](https://github.com/cockroachdb/cockroach/commits/7853fd32de8b6dea869f2a2a92dcd7506f4a8998): Fatal error: ``` panic: unexpected error: [NotLeaseHolderError] refusing to acquire lease on follower; r36: replica (n3,s3):3 not lease holder; current lease is repl=(n1,s1):1 seq=0 start=0,0 exp=<nil> ``` Stack: ``` goroutine 341335 [running]: github.com/cockroachdb/cockroach/pkg/kv/kvserver_test.TestReplicateRogueRemovedNode.func2(0xc0027b5470, 0xd, 0x10, 0xc0073edd60, 0xc0073edd70, 0xc00468c000, 0xc00e16ad80, 0x300000003, 0x3, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/kv/kvserver/client_raft_test.go:3338 +0x3a5 created by github.com/cockroachdb/cockroach/pkg/kv/kvserver_test.TestReplicateRogueRemovedNode /go/src/github.com/cockroachdb/cockroach/pkg/kv/kvserver/client_raft_test.go:3322 +0xea5 ``` <details><summary>Log preceding fatal error</summary><p> ``` === RUN TestReplicateRogueRemovedNode test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestReplicateRogueRemovedNode993083353 test_log_scope.go:74: use -show-logs to present logs inline ``` </p></details> <details><summary>More</summary><p> Parameters: - TAGS= - GOFLAGS=-parallel=4 ``` make stressrace TESTS=TestReplicateRogueRemovedNode PKG=./pkg/kv/kvserver TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1 ``` [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestReplicateRogueRemovedNode.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
1.0
kv/kvserver: TestReplicateRogueRemovedNode failed - [(kv/kvserver).TestReplicateRogueRemovedNode failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2657499&tab=buildLog) on [master@7853fd32de8b6dea869f2a2a92dcd7506f4a8998](https://github.com/cockroachdb/cockroach/commits/7853fd32de8b6dea869f2a2a92dcd7506f4a8998): Fatal error: ``` panic: unexpected error: [NotLeaseHolderError] refusing to acquire lease on follower; r36: replica (n3,s3):3 not lease holder; current lease is repl=(n1,s1):1 seq=0 start=0,0 exp=<nil> ``` Stack: ``` goroutine 341335 [running]: github.com/cockroachdb/cockroach/pkg/kv/kvserver_test.TestReplicateRogueRemovedNode.func2(0xc0027b5470, 0xd, 0x10, 0xc0073edd60, 0xc0073edd70, 0xc00468c000, 0xc00e16ad80, 0x300000003, 0x3, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/kv/kvserver/client_raft_test.go:3338 +0x3a5 created by github.com/cockroachdb/cockroach/pkg/kv/kvserver_test.TestReplicateRogueRemovedNode /go/src/github.com/cockroachdb/cockroach/pkg/kv/kvserver/client_raft_test.go:3322 +0xea5 ``` <details><summary>Log preceding fatal error</summary><p> ``` === RUN TestReplicateRogueRemovedNode test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestReplicateRogueRemovedNode993083353 test_log_scope.go:74: use -show-logs to present logs inline ``` </p></details> <details><summary>More</summary><p> Parameters: - TAGS= - GOFLAGS=-parallel=4 ``` make stressrace TESTS=TestReplicateRogueRemovedNode PKG=./pkg/kv/kvserver TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1 ``` [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestReplicateRogueRemovedNode.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
non_priority
kv kvserver testreplicaterogueremovednode failed on fatal error panic unexpected error refusing to acquire lease on follower replica not lease holder current lease is repl seq start exp stack goroutine github com cockroachdb cockroach pkg kv kvserver test testreplicaterogueremovednode go src github com cockroachdb cockroach pkg kv kvserver client raft test go created by github com cockroachdb cockroach pkg kv kvserver test testreplicaterogueremovednode go src github com cockroachdb cockroach pkg kv kvserver client raft test go log preceding fatal error run testreplicaterogueremovednode test log scope go test logs captured to go src github com cockroachdb cockroach artifacts test log scope go use show logs to present logs inline more parameters tags goflags parallel make stressrace tests testreplicaterogueremovednode pkg pkg kv kvserver testtimeout stressflags timeout powered by
0
211,962
16,384,572,381
IssuesEvent
2021-05-17 08:47:20
wazuh/wazuh-qa
https://api.github.com/repos/wazuh/wazuh-qa
opened
Migrate system tests fixes from master to 4.2
bug framework/api system-tests
## Description Hi team! There were many changes to the cluster in the master branch of Wazuh. These changes implied the update of some tests in QA, which were merged in QA master: - https://github.com/wazuh/wazuh-qa/pull/1188 - https://github.com/wazuh/wazuh-qa/pull/1232 - https://github.com/wazuh/wazuh-qa/pull/1237 However, we migrated most of the cluster improvements that were planned for 4.3.0 to 4.2.0 in this PR: https://github.com/wazuh/wazuh/pull/8674 Therefore, all the fixes that were merged in QA master branch need to be moved to 4.2 too. Otherwise, the tests will fail. Regards, Selu.
1.0
Migrate system tests fixes from master to 4.2 - ## Description Hi team! There were many changes to the cluster in the master branch of Wazuh. These changes implied the update of some tests in QA, which were merged in QA master: - https://github.com/wazuh/wazuh-qa/pull/1188 - https://github.com/wazuh/wazuh-qa/pull/1232 - https://github.com/wazuh/wazuh-qa/pull/1237 However, we migrated most of the cluster improvements that were planned for 4.3.0 to 4.2.0 in this PR: https://github.com/wazuh/wazuh/pull/8674 Therefore, all the fixes that were merged in QA master branch need to be moved to 4.2 too. Otherwise, the tests will fail. Regards, Selu.
non_priority
migrate system tests fixes from master to description hi team there were many changes to the cluster in the master branch of wazuh these changes implied the update of some tests in qa which were merged in qa master however we migrated most of the cluster improvements that were planned for to in this pr therefore all the fixes that were merged in qa master branch need to be moved to too otherwise the tests will fail regards selu
0
1,370
2,595,499,384
IssuesEvent
2015-02-20 14:44:11
docker/fig
https://api.github.com/repos/docker/fig
opened
Installation instructions for Compose should point and Docker installation instructions
documentation
It seems odd that we're repeating the installations instructions for Docker in the same documentation: https://github.com/docker/fig/blob/master/docs/install.md
1.0
Installation instructions for Compose should point and Docker installation instructions - It seems odd that we're repeating the installations instructions for Docker in the same documentation: https://github.com/docker/fig/blob/master/docs/install.md
non_priority
installation instructions for compose should point and docker installation instructions it seems odd that we re repeating the installations instructions for docker in the same documentation
0
5,580
27,958,345,405
IssuesEvent
2023-03-24 13:59:38
software-mansion/react-native-reanimated
https://api.github.com/repos/software-mansion/react-native-reanimated
closed
☂️ Deadlock/ANR in performOperations
Platform: Android Platform: iOS Bug Maintainer issue
### Description This is an umbrella issue for ANRs/deadlocks on Android/iOS in NodesManager.performOperations. The bug was introduced in #1215. Android: - #2251 - #3062 iOS: - #3180 - #3862 - #3946 PRs trying to solve this issue: - #3082 - #3194 ### Repro We don't have a repro yet but it needs to use modal or datetime picker as well as animate layout props using Reanimated. ### Reanimated version \>= 2.0.0, >= 3.0.0 ### Platforms Android, iOS
True
☂️ Deadlock/ANR in performOperations - ### Description This is an umbrella issue for ANRs/deadlocks on Android/iOS in NodesManager.performOperations. The bug was introduced in #1215. Android: - #2251 - #3062 iOS: - #3180 - #3862 - #3946 PRs trying to solve this issue: - #3082 - #3194 ### Repro We don't have a repro yet but it needs to use modal or datetime picker as well as animate layout props using Reanimated. ### Reanimated version \>= 2.0.0, >= 3.0.0 ### Platforms Android, iOS
non_priority
☂️ deadlock anr in performoperations description this is an umbrella issue for anrs deadlocks on android ios in nodesmanager performoperations the bug was introduced in android ios prs trying to solve this issue repro we don t have a repro yet but it needs to use modal or datetime picker as well as animate layout props using reanimated reanimated version platforms android ios
0
12,049
9,558,136,101
IssuesEvent
2019-05-03 13:32:17
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
[Logs UI] Add documentation for column customization
:Logs UI :infrastructure [zube]: Ready docs v7.2.0 v8.0.0
The new column customization UI introduced in #34916 should be explained in the documentation. Aspects that are potentially worth mentioning: * Influences the main log view * Configured via the source configuration flyout, 2nd tab * Three column types: * built-in timestamp * built-in message * custom field from underlying document
1.0
[Logs UI] Add documentation for column customization - The new column customization UI introduced in #34916 should be explained in the documentation. Aspects that are potentially worth mentioning: * Influences the main log view * Configured via the source configuration flyout, 2nd tab * Three column types: * built-in timestamp * built-in message * custom field from underlying document
non_priority
add documentation for column customization the new column customization ui introduced in should be explained in the documentation aspects that are potentially worth mentioning influences the main log view configured via the source configuration flyout tab three column types built in timestamp built in message custom field from underlying document
0
8,116
7,242,961,049
IssuesEvent
2018-02-14 09:58:10
stiftungswo/Dime
https://api.github.com/repos/stiftungswo/Dime
opened
Frontend Migration
in progress infrastructure
Ansprechperson: Andi AngularDart v1 ist unmaintained und schlecht dokumentiert. Es soll eine Auffrischung der Technologie für das Dime Frontend evaluiert werden. Die verschiedenen Varianten sollen evaluiert und der zeitliche Aufwand geschätzt werden. Kandidaten: ## AngularDart v4 ### Motivation Die Sprache bleibt dieselbe, aber die Frameworkversion wird auf die aktuelle und weiterentwickelte Version upgegradet. Somit profitiert man von den aktuellen neuen Dokumentationen und neuen Features, die entwickelt werden. Bestehender Code und Projektstruktur können weitgehend wiederverwertet werden. ### Evaluation Die Hammock Library musste angepasst werden, funktioniert aber jetzt soweit. Login, ServiceOverview, ServiceEdit funktionieren soweit. Die Struktur von Code und Templates konnte fast 1:1 übernommen werden. Code auf Branch https://github.com/stiftungswo/Dime/tree/angulardart-4-evaluation Voraussichtlicher Zeitaufwand: 4 Wochen ## Typescript Angular ### Motivation Die Technologie bleibt ähnlich, aber anstatt der Dart Version wird die standard Typescript Version des Frameworks verwendet, das eine grössere Community und somit mehr Support und Zukunftssicherheit bietet. ## React/Inferno ### Motivation Der Vorteil hierbei wäre eine Vereinigung der Frontend Technologien, die in der Stiftung SWO entwickelt werden, da iZivi momentan mit Inferno läuft. Die Einstiegshürde für neue Zivis würde somit verringert.
1.0
Frontend Migration - Ansprechperson: Andi AngularDart v1 ist unmaintained und schlecht dokumentiert. Es soll eine Auffrischung der Technologie für das Dime Frontend evaluiert werden. Die verschiedenen Varianten sollen evaluiert und der zeitliche Aufwand geschätzt werden. Kandidaten: ## AngularDart v4 ### Motivation Die Sprache bleibt dieselbe, aber die Frameworkversion wird auf die aktuelle und weiterentwickelte Version upgegradet. Somit profitiert man von den aktuellen neuen Dokumentationen und neuen Features, die entwickelt werden. Bestehender Code und Projektstruktur können weitgehend wiederverwertet werden. ### Evaluation Die Hammock Library musste angepasst werden, funktioniert aber jetzt soweit. Login, ServiceOverview, ServiceEdit funktionieren soweit. Die Struktur von Code und Templates konnte fast 1:1 übernommen werden. Code auf Branch https://github.com/stiftungswo/Dime/tree/angulardart-4-evaluation Voraussichtlicher Zeitaufwand: 4 Wochen ## Typescript Angular ### Motivation Die Technologie bleibt ähnlich, aber anstatt der Dart Version wird die standard Typescript Version des Frameworks verwendet, das eine grössere Community und somit mehr Support und Zukunftssicherheit bietet. ## React/Inferno ### Motivation Der Vorteil hierbei wäre eine Vereinigung der Frontend Technologien, die in der Stiftung SWO entwickelt werden, da iZivi momentan mit Inferno läuft. Die Einstiegshürde für neue Zivis würde somit verringert.
non_priority
frontend migration ansprechperson andi angulardart ist unmaintained und schlecht dokumentiert es soll eine auffrischung der technologie für das dime frontend evaluiert werden die verschiedenen varianten sollen evaluiert und der zeitliche aufwand geschätzt werden kandidaten angulardart motivation die sprache bleibt dieselbe aber die frameworkversion wird auf die aktuelle und weiterentwickelte version upgegradet somit profitiert man von den aktuellen neuen dokumentationen und neuen features die entwickelt werden bestehender code und projektstruktur können weitgehend wiederverwertet werden evaluation die hammock library musste angepasst werden funktioniert aber jetzt soweit login serviceoverview serviceedit funktionieren soweit die struktur von code und templates konnte fast übernommen werden code auf branch voraussichtlicher zeitaufwand wochen typescript angular motivation die technologie bleibt ähnlich aber anstatt der dart version wird die standard typescript version des frameworks verwendet das eine grössere community und somit mehr support und zukunftssicherheit bietet react inferno motivation der vorteil hierbei wäre eine vereinigung der frontend technologien die in der stiftung swo entwickelt werden da izivi momentan mit inferno läuft die einstiegshürde für neue zivis würde somit verringert
0
326,532
27,998,548,035
IssuesEvent
2023-03-27 10:06:38
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
pkg/ccl/spanconfigccl/spanconfigsqlwatcherccl/spanconfigsqlwatcherccl_test: TestSQLWatcherOnEventError failed
C-test-failure O-robot X-stale T-kv no-test-failure-activity branch-release-22.2
pkg/ccl/spanconfigccl/spanconfigsqlwatcherccl/spanconfigsqlwatcherccl_test.TestSQLWatcherOnEventError [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/8751541?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/8751541?buildTab=artifacts#/) on release-22.2 @ [31eb3ce6ceb18123fd2e7eac464b25e203d7de0f](https://github.com/cockroachdb/cockroach/commits/31eb3ce6ceb18123fd2e7eac464b25e203d7de0f): ``` I230218 08:33:09.958975 890758 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r51/1:/Table/5{1-2}] 507 + 0.619ms 0.619ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:682 [n1,replicate,s1,r51/1:/Table/5{1-2}] add voter - missing voter need=3, have=1, priority=10001.00 I230218 08:33:09.958975 890758 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r51/1:/Table/5{1-2}] 507 + 0.704ms 0.084ms event:kv/kvserver/replicate_queue.go:866 [n1,replicate,s1,r51/1:/Table/5{1-2}] next replica action: add voter I230218 08:33:09.958975 890758 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r51/1:/Table/5{1-2}] 507 + 0.861ms 0.157ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:1037 [n1,replicate,s1,r51/1:/Table/5{1-2}] allocate voter: [] I230218 08:33:09.961667 885282 sql/syntheticprivilegecache/cache.go:206 [n1] 508 warmed privileges for virtual tables in 60.47079ms W230218 08:33:09.965828 811817 kv/kvserver/replica_proposal.go:490 [n1,s1,r4/1:/System{/tsd-tse},raft] 509 lease <empty> expired before being followed by lease repl=(n1,s1):1 seq=1 start=0,0 epo=1 pro=1676709189.964821364,0; foreground traffic may have been impacted I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 error processing replica: 0 of 1 live stores are able to take a new replica for the range (1 already has a voter, 0 already have a non-voter); likely not enough nodes in cluster I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 +trace: I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 + 0.000ms 0.000ms === operation:process replica _unfinished:1 _verbose:1 node:1 replicate: store:1 range:4/1:/System{/tsd-tse} I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 + 0.250ms 0.250ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:682 [n1,replicate,s1,r4/1:/System{/tsd-tse}] add voter - missing voter need=3, have=1, priority=10001.00 I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 + 0.282ms 0.032ms event:kv/kvserver/replicate_queue.go:866 [n1,replicate,s1,r4/1:/System{/tsd-tse}] next replica action: add voter I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 + 0.367ms 0.084ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:1037 [n1,replicate,s1,r4/1:/System{/tsd-tse}] allocate voter: [] I230218 08:33:09.972991 811813 3@pebble/event.go:678 [n1,pebble,s1] 511 [JOB 6] WAL created 000005 I230218 08:33:09.976140 805484 1@server/server_sql.go:1451 [n1] 512 serving sql connections I230218 08:33:09.977398 805484 testutils/testcluster/testcluster.go:1339 [-] 513 WaitForFullReplication I230218 08:33:09.977537 805484 testutils/testcluster/testcluster.go:1343 [-] 514 WaitForFullReplication took: 368ns I230218 08:33:09.977715 805484 2@rpc/context.go:1774 [n1,rnode=1,raddr=127.0.0.1:34123,class=default] 515 dialing I230218 08:33:10.050656 894362 util/log/event_log.go:32 [n1,client=127.0.0.1:37074,user=root] 516 ={"Timestamp":1676709190028595263,"EventType":"set_cluster_setting","Statement":"SET CLUSTER SETTING \"kv.rangefeed.enabled\" = true","Tag":"SET CLUSTER SETTING","User":"root","SettingName":"kv.rangefeed.enabled","Value":"true"} I230218 08:33:10.094869 894362 util/log/event_log.go:32 [n1,client=127.0.0.1:37074,user=root] 517 ={"Timestamp":1676709190058971980,"EventType":"set_cluster_setting","Statement":"SET CLUSTER SETTING \"kv.closed_timestamp.target_duration\" = '100ms'","Tag":"SET CLUSTER SETTING","User":"root","SettingName":"kv.closed_timestamp.target_duration","Value":"00:00:00.1"} W230218 08:33:10.169386 811829 kv/kvserver/replica_proposal.go:490 [n1,s1,r23/1:/Table/2{2-3},raft] 518 lease <empty> expired before being followed by lease repl=(n1,s1):1 seq=1 start=0,0 epo=1 pro=1676709190.167501234,0; foreground traffic may have been impacted I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 error processing replica: 0 of 1 live stores are able to take a new replica for the range (1 already has a voter, 0 already have a non-voter); likely not enough nodes in cluster I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 +trace: I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 + 0.000ms 0.000ms === operation:process replica _unfinished:1 _verbose:1 node:1 replicate: store:1 range:23/1:/Table/2{2-3} I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 + 0.194ms 0.194ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:682 [n1,replicate,s1,r23/1:/Table/2{2-3}] add voter - missing voter need=3, have=1, priority=10001.00 I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 + 0.239ms 0.044ms event:kv/kvserver/replicate_queue.go:866 [n1,replicate,s1,r23/1:/Table/2{2-3}] next replica action: add voter I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 + 0.333ms 0.094ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:1037 [n1,replicate,s1,r23/1:/Table/2{2-3}] allocate voter: [] I230218 08:33:10.220490 894362 5@util/log/event_log.go:32 [n1,client=127.0.0.1:37074,user=root] 520 ={"Timestamp":1676709190117699231,"EventType":"create_table","Statement":"CREATE TABLE defaultdb.public.t ()","Tag":"CREATE TABLE","User":"root","DescriptorID":104,"TableName":"defaultdb.public.t"} I230218 08:33:10.222227 805484 spanconfig/spanconfigsqlwatcher/sqlwatcher.go:244 [-] 521 established range feed over system.descriptors starting at time 1676709190.095893637,0 I230218 08:33:10.222410 805484 spanconfig/spanconfigsqlwatcher/sqlwatcher.go:303 [-] 522 established range feed over system.zones starting at time 1676709190.095893637,0 I230218 08:33:10.224092 805484 spanconfig/spanconfigsqlwatcher/sqlwatcher.go:399 [-] 523 established range feed over system.protected_ts_records starting at time 1676709190.095893637,0 sqlwatcher_test.go:460: handler should never run I230218 08:33:10.234792 805484 testutils/testcluster/testcluster.go:128 [-] 524 TestCluster quiescing nodes W230218 08:33:10.235023 811565 kv/kvserver/replica_proposal.go:490 [n1,s1,r33/1:/Table/3{2-3},raft] 525 lease <empty> expired before being followed by lease repl=(n1,s1):1 seq=1 start=0,0 epo=1 pro=1676709190.233308221,0; foreground traffic may have been impacted I230218 08:33:10.236189 893316 2@rpc/context.go:1393 [n1] 526 closing [{n1: 127.0.0.1:34123 (default)} {n0: 127.0.0.1:34123 (default)}] W230218 08:33:10.245611 816472 sql/sqlliveness/slinstance/slinstance.go:242 [n1] 527 exiting heartbeat loop W230218 08:33:10.246151 866759 kv/kvserver/intentresolver/intent_resolver.go:799 [-] 528 failed to gc transaction record: could not GC completed transaction anchored at /Table/30/1/100/0/"public"/4/1: node unavailable; try another peer W230218 08:33:10.246541 811565 kv/kvserver/store.go:2585 [n1,s1,r33/1:/Table/3{2-3},raft] 529 unable to gossip on capacity change: node unavailable; try another peer W230218 08:33:10.246541 811565 kv/kvserver/store.go:2585 [n1,s1,r33/1:/Table/3{2-3},raft] 529 +(1) node unavailable; try another peer W230218 08:33:10.246541 811565 kv/kvserver/store.go:2585 [n1,s1,r33/1:/Table/3{2-3},raft] 529 +Error types: (1) *roachpb.NodeUnavailableError W230218 08:33:10.247802 817426 kv/txn.go:705 [n1] 530 failure aborting transaction: node unavailable; try another peer; abort caused by: expire-sessions: failed to read query result: context canceled E230218 08:33:10.250288 817426 jobs/registry.go:744 [n1] 531 error expiring job sessions: expire-sessions: failed to read query result: context canceled W230218 08:33:10.250467 817426 kv/txn.go:705 [n1] 532 failure aborting transaction: node unavailable; try another peer; abort caused by: txn exec: context canceled E230218 08:33:10.250544 817426 jobs/registry.go:752 [n1] 533 failed to serve pause and cancel requests: txn exec: context canceled W230218 08:33:10.250604 817426 jobs/registry.go:821 [n1] 534 canceling all adopted jobs due to stopper quiescing W230218 08:33:10.253153 903355 kv/txn.go:705 [n1,intExec=select-running/get-claimed-jobs] 535 failure aborting transaction: node unavailable; try another peer; abort caused by: connExecutor closing E230218 08:33:10.253580 817877 jobs/registry.go:805 [n1] 536 error processing claimed jobs: could not query for claimed jobs: select-running/get-claimed-jobs: failed to read query result: context canceled E230218 08:33:10.265287 816760 sql/stats/automatic_stats.go:580 [n1] 537 failed to get tables for automatic stats: get-tables: node unavailable; try another peer I230218 08:33:10.265586 816760 sql/stats/automatic_stats.go:509 [n1] 538 quiescing auto stats refresher W230218 08:33:10.266359 904143 kv/txn.go:705 [n1,client=127.0.0.1:37074,user=root] 539 failure aborting transaction: node unavailable; try another peer; abort caused by: result is ambiguous: server shutdown W230218 08:33:10.266541 904143 1@sql/event_log.go:641 [n1,client=127.0.0.1:37074,user=root] 540 unable to save 1 entries to system.eventlog: result is ambiguous: server shutdown --- FAIL: TestSQLWatcherOnEventError (1.43s) ``` <p>Parameters: <code>TAGS=bazel,gss,deadlock</code> </p> <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #82031 pkg/ccl/spanconfigccl/spanconfigsqlwatcherccl/spanconfigsqlwatcherccl_test: TestSQLWatcherOnEventError failed [C-test-failure O-robot branch-release-22.1 sync-me sync-me-5] - #76765 pkg/ccl/spanconfigccl/spanconfigsqlwatcherccl/spanconfigsqlwatcherccl_test: TestSQLWatcherOnEventError failed [C-test-failure O-robot branch-master sync-me-8] </p> </details> /cc @cockroachdb/kv <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestSQLWatcherOnEventError.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-24633
2.0
pkg/ccl/spanconfigccl/spanconfigsqlwatcherccl/spanconfigsqlwatcherccl_test: TestSQLWatcherOnEventError failed - pkg/ccl/spanconfigccl/spanconfigsqlwatcherccl/spanconfigsqlwatcherccl_test.TestSQLWatcherOnEventError [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/8751541?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/8751541?buildTab=artifacts#/) on release-22.2 @ [31eb3ce6ceb18123fd2e7eac464b25e203d7de0f](https://github.com/cockroachdb/cockroach/commits/31eb3ce6ceb18123fd2e7eac464b25e203d7de0f): ``` I230218 08:33:09.958975 890758 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r51/1:/Table/5{1-2}] 507 + 0.619ms 0.619ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:682 [n1,replicate,s1,r51/1:/Table/5{1-2}] add voter - missing voter need=3, have=1, priority=10001.00 I230218 08:33:09.958975 890758 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r51/1:/Table/5{1-2}] 507 + 0.704ms 0.084ms event:kv/kvserver/replicate_queue.go:866 [n1,replicate,s1,r51/1:/Table/5{1-2}] next replica action: add voter I230218 08:33:09.958975 890758 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r51/1:/Table/5{1-2}] 507 + 0.861ms 0.157ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:1037 [n1,replicate,s1,r51/1:/Table/5{1-2}] allocate voter: [] I230218 08:33:09.961667 885282 sql/syntheticprivilegecache/cache.go:206 [n1] 508 warmed privileges for virtual tables in 60.47079ms W230218 08:33:09.965828 811817 kv/kvserver/replica_proposal.go:490 [n1,s1,r4/1:/System{/tsd-tse},raft] 509 lease <empty> expired before being followed by lease repl=(n1,s1):1 seq=1 start=0,0 epo=1 pro=1676709189.964821364,0; foreground traffic may have been impacted I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 error processing replica: 0 of 1 live stores are able to take a new replica for the range (1 already has a voter, 0 already have a non-voter); likely not enough nodes in cluster I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 +trace: I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 + 0.000ms 0.000ms === operation:process replica _unfinished:1 _verbose:1 node:1 replicate: store:1 range:4/1:/System{/tsd-tse} I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 + 0.250ms 0.250ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:682 [n1,replicate,s1,r4/1:/System{/tsd-tse}] add voter - missing voter need=3, have=1, priority=10001.00 I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 + 0.282ms 0.032ms event:kv/kvserver/replicate_queue.go:866 [n1,replicate,s1,r4/1:/System{/tsd-tse}] next replica action: add voter I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 + 0.367ms 0.084ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:1037 [n1,replicate,s1,r4/1:/System{/tsd-tse}] allocate voter: [] I230218 08:33:09.972991 811813 3@pebble/event.go:678 [n1,pebble,s1] 511 [JOB 6] WAL created 000005 I230218 08:33:09.976140 805484 1@server/server_sql.go:1451 [n1] 512 serving sql connections I230218 08:33:09.977398 805484 testutils/testcluster/testcluster.go:1339 [-] 513 WaitForFullReplication I230218 08:33:09.977537 805484 testutils/testcluster/testcluster.go:1343 [-] 514 WaitForFullReplication took: 368ns I230218 08:33:09.977715 805484 2@rpc/context.go:1774 [n1,rnode=1,raddr=127.0.0.1:34123,class=default] 515 dialing I230218 08:33:10.050656 894362 util/log/event_log.go:32 [n1,client=127.0.0.1:37074,user=root] 516 ={"Timestamp":1676709190028595263,"EventType":"set_cluster_setting","Statement":"SET CLUSTER SETTING \"kv.rangefeed.enabled\" = true","Tag":"SET CLUSTER SETTING","User":"root","SettingName":"kv.rangefeed.enabled","Value":"true"} I230218 08:33:10.094869 894362 util/log/event_log.go:32 [n1,client=127.0.0.1:37074,user=root] 517 ={"Timestamp":1676709190058971980,"EventType":"set_cluster_setting","Statement":"SET CLUSTER SETTING \"kv.closed_timestamp.target_duration\" = '100ms'","Tag":"SET CLUSTER SETTING","User":"root","SettingName":"kv.closed_timestamp.target_duration","Value":"00:00:00.1"} W230218 08:33:10.169386 811829 kv/kvserver/replica_proposal.go:490 [n1,s1,r23/1:/Table/2{2-3},raft] 518 lease <empty> expired before being followed by lease repl=(n1,s1):1 seq=1 start=0,0 epo=1 pro=1676709190.167501234,0; foreground traffic may have been impacted I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 error processing replica: 0 of 1 live stores are able to take a new replica for the range (1 already has a voter, 0 already have a non-voter); likely not enough nodes in cluster I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 +trace: I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 + 0.000ms 0.000ms === operation:process replica _unfinished:1 _verbose:1 node:1 replicate: store:1 range:23/1:/Table/2{2-3} I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 + 0.194ms 0.194ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:682 [n1,replicate,s1,r23/1:/Table/2{2-3}] add voter - missing voter need=3, have=1, priority=10001.00 I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 + 0.239ms 0.044ms event:kv/kvserver/replicate_queue.go:866 [n1,replicate,s1,r23/1:/Table/2{2-3}] next replica action: add voter I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 + 0.333ms 0.094ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:1037 [n1,replicate,s1,r23/1:/Table/2{2-3}] allocate voter: [] I230218 08:33:10.220490 894362 5@util/log/event_log.go:32 [n1,client=127.0.0.1:37074,user=root] 520 ={"Timestamp":1676709190117699231,"EventType":"create_table","Statement":"CREATE TABLE defaultdb.public.t ()","Tag":"CREATE TABLE","User":"root","DescriptorID":104,"TableName":"defaultdb.public.t"} I230218 08:33:10.222227 805484 spanconfig/spanconfigsqlwatcher/sqlwatcher.go:244 [-] 521 established range feed over system.descriptors starting at time 1676709190.095893637,0 I230218 08:33:10.222410 805484 spanconfig/spanconfigsqlwatcher/sqlwatcher.go:303 [-] 522 established range feed over system.zones starting at time 1676709190.095893637,0 I230218 08:33:10.224092 805484 spanconfig/spanconfigsqlwatcher/sqlwatcher.go:399 [-] 523 established range feed over system.protected_ts_records starting at time 1676709190.095893637,0 sqlwatcher_test.go:460: handler should never run I230218 08:33:10.234792 805484 testutils/testcluster/testcluster.go:128 [-] 524 TestCluster quiescing nodes W230218 08:33:10.235023 811565 kv/kvserver/replica_proposal.go:490 [n1,s1,r33/1:/Table/3{2-3},raft] 525 lease <empty> expired before being followed by lease repl=(n1,s1):1 seq=1 start=0,0 epo=1 pro=1676709190.233308221,0; foreground traffic may have been impacted I230218 08:33:10.236189 893316 2@rpc/context.go:1393 [n1] 526 closing [{n1: 127.0.0.1:34123 (default)} {n0: 127.0.0.1:34123 (default)}] W230218 08:33:10.245611 816472 sql/sqlliveness/slinstance/slinstance.go:242 [n1] 527 exiting heartbeat loop W230218 08:33:10.246151 866759 kv/kvserver/intentresolver/intent_resolver.go:799 [-] 528 failed to gc transaction record: could not GC completed transaction anchored at /Table/30/1/100/0/"public"/4/1: node unavailable; try another peer W230218 08:33:10.246541 811565 kv/kvserver/store.go:2585 [n1,s1,r33/1:/Table/3{2-3},raft] 529 unable to gossip on capacity change: node unavailable; try another peer W230218 08:33:10.246541 811565 kv/kvserver/store.go:2585 [n1,s1,r33/1:/Table/3{2-3},raft] 529 +(1) node unavailable; try another peer W230218 08:33:10.246541 811565 kv/kvserver/store.go:2585 [n1,s1,r33/1:/Table/3{2-3},raft] 529 +Error types: (1) *roachpb.NodeUnavailableError W230218 08:33:10.247802 817426 kv/txn.go:705 [n1] 530 failure aborting transaction: node unavailable; try another peer; abort caused by: expire-sessions: failed to read query result: context canceled E230218 08:33:10.250288 817426 jobs/registry.go:744 [n1] 531 error expiring job sessions: expire-sessions: failed to read query result: context canceled W230218 08:33:10.250467 817426 kv/txn.go:705 [n1] 532 failure aborting transaction: node unavailable; try another peer; abort caused by: txn exec: context canceled E230218 08:33:10.250544 817426 jobs/registry.go:752 [n1] 533 failed to serve pause and cancel requests: txn exec: context canceled W230218 08:33:10.250604 817426 jobs/registry.go:821 [n1] 534 canceling all adopted jobs due to stopper quiescing W230218 08:33:10.253153 903355 kv/txn.go:705 [n1,intExec=select-running/get-claimed-jobs] 535 failure aborting transaction: node unavailable; try another peer; abort caused by: connExecutor closing E230218 08:33:10.253580 817877 jobs/registry.go:805 [n1] 536 error processing claimed jobs: could not query for claimed jobs: select-running/get-claimed-jobs: failed to read query result: context canceled E230218 08:33:10.265287 816760 sql/stats/automatic_stats.go:580 [n1] 537 failed to get tables for automatic stats: get-tables: node unavailable; try another peer I230218 08:33:10.265586 816760 sql/stats/automatic_stats.go:509 [n1] 538 quiescing auto stats refresher W230218 08:33:10.266359 904143 kv/txn.go:705 [n1,client=127.0.0.1:37074,user=root] 539 failure aborting transaction: node unavailable; try another peer; abort caused by: result is ambiguous: server shutdown W230218 08:33:10.266541 904143 1@sql/event_log.go:641 [n1,client=127.0.0.1:37074,user=root] 540 unable to save 1 entries to system.eventlog: result is ambiguous: server shutdown --- FAIL: TestSQLWatcherOnEventError (1.43s) ``` <p>Parameters: <code>TAGS=bazel,gss,deadlock</code> </p> <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #82031 pkg/ccl/spanconfigccl/spanconfigsqlwatcherccl/spanconfigsqlwatcherccl_test: TestSQLWatcherOnEventError failed [C-test-failure O-robot branch-release-22.1 sync-me sync-me-5] - #76765 pkg/ccl/spanconfigccl/spanconfigsqlwatcherccl/spanconfigsqlwatcherccl_test: TestSQLWatcherOnEventError failed [C-test-failure O-robot branch-master sync-me-8] </p> </details> /cc @cockroachdb/kv <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestSQLWatcherOnEventError.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-24633
non_priority
pkg ccl spanconfigccl spanconfigsqlwatcherccl spanconfigsqlwatcherccl test testsqlwatcheroneventerror failed pkg ccl spanconfigccl spanconfigsqlwatcherccl spanconfigsqlwatcherccl test testsqlwatcheroneventerror with on release kv kvserver replicate queue go event kv kvserver allocator allocatorimpl allocator go add voter missing voter need have priority kv kvserver replicate queue go event kv kvserver replicate queue go next replica action add voter kv kvserver replicate queue go event kv kvserver allocator allocatorimpl allocator go allocate voter sql syntheticprivilegecache cache go warmed privileges for virtual tables in kv kvserver replica proposal go lease expired before being followed by lease repl seq start epo pro foreground traffic may have been impacted kv kvserver replicate queue go error processing replica of live stores are able to take a new replica for the range already has a voter already have a non voter likely not enough nodes in cluster kv kvserver replicate queue go trace kv kvserver replicate queue go operation process replica unfinished verbose node replicate store range system tsd tse kv kvserver replicate queue go event kv kvserver allocator allocatorimpl allocator go add voter missing voter need have priority kv kvserver replicate queue go event kv kvserver replicate queue go next replica action add voter kv kvserver replicate queue go event kv kvserver allocator allocatorimpl allocator go allocate voter pebble event go wal created server server sql go serving sql connections testutils testcluster testcluster go waitforfullreplication testutils testcluster testcluster go waitforfullreplication took rpc context go dialing util log event log go timestamp eventtype set cluster setting statement set cluster setting kv rangefeed enabled true tag set cluster setting user root settingname kv rangefeed enabled value true util log event log go timestamp eventtype set cluster setting statement set cluster setting kv closed timestamp target duration tag set cluster setting user root settingname kv closed timestamp target duration value kv kvserver replica proposal go lease expired before being followed by lease repl seq start epo pro foreground traffic may have been impacted kv kvserver replicate queue go error processing replica of live stores are able to take a new replica for the range already has a voter already have a non voter likely not enough nodes in cluster kv kvserver replicate queue go trace kv kvserver replicate queue go operation process replica unfinished verbose node replicate store range table kv kvserver replicate queue go event kv kvserver allocator allocatorimpl allocator go add voter missing voter need have priority kv kvserver replicate queue go event kv kvserver replicate queue go next replica action add voter kv kvserver replicate queue go event kv kvserver allocator allocatorimpl allocator go allocate voter util log event log go timestamp eventtype create table statement create table defaultdb public t tag create table user root descriptorid tablename defaultdb public t spanconfig spanconfigsqlwatcher sqlwatcher go established range feed over system descriptors starting at time spanconfig spanconfigsqlwatcher sqlwatcher go established range feed over system zones starting at time spanconfig spanconfigsqlwatcher sqlwatcher go established range feed over system protected ts records starting at time sqlwatcher test go handler should never run testutils testcluster testcluster go testcluster quiescing nodes kv kvserver replica proposal go lease expired before being followed by lease repl seq start epo pro foreground traffic may have been impacted rpc context go closing sql sqlliveness slinstance slinstance go exiting heartbeat loop kv kvserver intentresolver intent resolver go failed to gc transaction record could not gc completed transaction anchored at table public node unavailable try another peer kv kvserver store go unable to gossip on capacity change node unavailable try another peer kv kvserver store go node unavailable try another peer kv kvserver store go error types roachpb nodeunavailableerror kv txn go failure aborting transaction node unavailable try another peer abort caused by expire sessions failed to read query result context canceled jobs registry go error expiring job sessions expire sessions failed to read query result context canceled kv txn go failure aborting transaction node unavailable try another peer abort caused by txn exec context canceled jobs registry go failed to serve pause and cancel requests txn exec context canceled jobs registry go canceling all adopted jobs due to stopper quiescing kv txn go failure aborting transaction node unavailable try another peer abort caused by connexecutor closing jobs registry go error processing claimed jobs could not query for claimed jobs select running get claimed jobs failed to read query result context canceled sql stats automatic stats go failed to get tables for automatic stats get tables node unavailable try another peer sql stats automatic stats go quiescing auto stats refresher kv txn go failure aborting transaction node unavailable try another peer abort caused by result is ambiguous server shutdown sql event log go unable to save entries to system eventlog result is ambiguous server shutdown fail testsqlwatcheroneventerror parameters tags bazel gss deadlock help see also same failure on other branches pkg ccl spanconfigccl spanconfigsqlwatcherccl spanconfigsqlwatcherccl test testsqlwatcheroneventerror failed pkg ccl spanconfigccl spanconfigsqlwatcherccl spanconfigsqlwatcherccl test testsqlwatcheroneventerror failed cc cockroachdb kv jira issue crdb
0