Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 957 | labels stringlengths 4 795 | body stringlengths 1 259k | index stringclasses 12
values | text_combine stringlengths 96 259k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
852 | 2,648,610,319 | IssuesEvent | 2015-03-14 02:37:20 | WPN-XM/WPN-XM | https://api.github.com/repos/WPN-XM/WPN-XM | closed | try to compile the innosetup scripts on ci server | build tools installation wizard | **Try to compile the InnoSetup installation wizards on a CI server/unix**
- [x] travis-ci or
- [ ] jenkins-ci
- [x] use wine
- [x] install innosetup into wine
- we already have innosetup in /bin/innosetup
- it's no problem to execute the compiler `wineconsole issc`or `wine issc`
- [x] run the build file
- fix possible windows path issues
- maybe clone the build file for unix?
- run build file with phing and ant at the same time
- ant would be used for the parts relying on git. we dont need to install git into wine.
**Build Artifact Handling**
- [ ] upload build artifacts
- to a nightly folder ?
- http://docs.travis-ci.com/user/deployment/releases/
- `after_success: php upload.php` ?
- $GITHUB_TOKEN
- $TRAVIS_COMMIT
- file to upload
- release filename
- $TRAVIS_BRANCH
- [ ] add "do-release" task, to push a new release to github
- [ ] add "add-artifact" task, to push a new artifact to a release tag on github
- [ ] add "add-artifacts" task, to push all artifacts (executables) via add-artifact
**Automatic Deployment**
- [ ] official wpnxm server via sftp/scp
- [ ] github releases (curl)
- https://developer.github.com/v3/repos/releases/#upload-a-release-asset
- https://developer.github.com/changes/2013-09-25-releases-api/
- **CURL POST** `$ curl -H "Authorization: token TOKEN" \
-H "Accept: application/vnd.github.manifold-preview" \
-H "Content-Type: application/zip" \
--data-binary @build/mac/package.zip \
"https://uploads.github.com/repos/hubot/singularity/releases/123/assets?name=1.0.0-mac.zip"`
- [ ] sourceforge (sftp/scp/curl)
- https://sourceforge.net/p/forge/community-docs/Using%20the%20Release%20API/
- https://sourceforge.net/p/forge/documentation/Release%20Files%20for%20Download/#scp | 1.0 | try to compile the innosetup scripts on ci server - **Try to compile the InnoSetup installation wizards on a CI server/unix**
- [x] travis-ci or
- [ ] jenkins-ci
- [x] use wine
- [x] install innosetup into wine
- we already have innosetup in /bin/innosetup
- it's no problem to execute the compiler `wineconsole issc`or `wine issc`
- [x] run the build file
- fix possible windows path issues
- maybe clone the build file for unix?
- run build file with phing and ant at the same time
- ant would be used for the parts relying on git. we dont need to install git into wine.
**Build Artifact Handling**
- [ ] upload build artifacts
- to a nightly folder ?
- http://docs.travis-ci.com/user/deployment/releases/
- `after_success: php upload.php` ?
- $GITHUB_TOKEN
- $TRAVIS_COMMIT
- file to upload
- release filename
- $TRAVIS_BRANCH
- [ ] add "do-release" task, to push a new release to github
- [ ] add "add-artifact" task, to push a new artifact to a release tag on github
- [ ] add "add-artifacts" task, to push all artifacts (executables) via add-artifact
**Automatic Deployment**
- [ ] official wpnxm server via sftp/scp
- [ ] github releases (curl)
- https://developer.github.com/v3/repos/releases/#upload-a-release-asset
- https://developer.github.com/changes/2013-09-25-releases-api/
- **CURL POST** `$ curl -H "Authorization: token TOKEN" \
-H "Accept: application/vnd.github.manifold-preview" \
-H "Content-Type: application/zip" \
--data-binary @build/mac/package.zip \
"https://uploads.github.com/repos/hubot/singularity/releases/123/assets?name=1.0.0-mac.zip"`
- [ ] sourceforge (sftp/scp/curl)
- https://sourceforge.net/p/forge/community-docs/Using%20the%20Release%20API/
- https://sourceforge.net/p/forge/documentation/Release%20Files%20for%20Download/#scp | non_priority | try to compile the innosetup scripts on ci server try to compile the innosetup installation wizards on a ci server unix travis ci or jenkins ci use wine install innosetup into wine we already have innosetup in bin innosetup it s no problem to execute the compiler wineconsole issc or wine issc run the build file fix possible windows path issues maybe clone the build file for unix run build file with phing and ant at the same time ant would be used for the parts relying on git we dont need to install git into wine build artifact handling upload build artifacts to a nightly folder after success php upload php github token travis commit file to upload release filename travis branch add do release task to push a new release to github add add artifact task to push a new artifact to a release tag on github add add artifacts task to push all artifacts executables via add artifact automatic deployment official wpnxm server via sftp scp github releases curl curl post curl h authorization token token h accept application vnd github manifold preview h content type application zip data binary build mac package zip sourceforge sftp scp curl | 0 |
166,827 | 12,973,078,912 | IssuesEvent | 2020-07-21 13:31:55 | LiskHQ/lisk-sdk | https://api.github.com/repos/LiskHQ/lisk-sdk | opened | Refactor synchronization unit tests | type: refactoring type: test | ### Description
Currently synchronization unit tests (fast chain switching and block syncrhonization) are using complex setup of stubs.
If any of the implementation detail changes, it is very hard to update the tests.
It should be cleaned up to be more modifiable and readable with less mock setups. | 1.0 | Refactor synchronization unit tests - ### Description
Currently synchronization unit tests (fast chain switching and block syncrhonization) are using complex setup of stubs.
If any of the implementation detail changes, it is very hard to update the tests.
It should be cleaned up to be more modifiable and readable with less mock setups. | non_priority | refactor synchronization unit tests description currently synchronization unit tests fast chain switching and block syncrhonization are using complex setup of stubs if any of the implementation detail changes it is very hard to update the tests it should be cleaned up to be more modifiable and readable with less mock setups | 0 |
22,046 | 11,462,903,789 | IssuesEvent | 2020-02-07 15:01:10 | motss/app-datepicker | https://api.github.com/repos/motss/app-datepicker | closed | Recommended performance improvement | enhancement help-wanted help-welcome in-progress lit-element needs investigation performance | ### Description
The Google Chrome browser warns (console) about a performance hit with non-passive event listeners. It points to tracker.js.
A recommendation for improvement is documented here: [passive-event-listeners](https://developers.google.com/web/tools/lighthouse/audits/passive-event-listeners)
### Modifications tested
My apologies, I didn't have much time to look at this but to quickly test, and serve as an example I made the following updates to tracker.js:
1. in class Tracker, add the passive option to the event listeners. In practice the recommendation is to check if the passive option is support. In my case it wasn't necessary, hence:
```
_element.addEventListener('mousedown', this._down, {passive:true});
_element.addEventListener('touchstart', this._down, {passive:true});
_element.addEventListener('touchmove', this._move, {passive:true});
_element.addEventListener('touchend', this._up, {passive:true});
```
2. in _onDown(down), remove:
`ev.preventDefault()`
### Result
The Google warning went away and the element appeared to work OK.
Hopefully this is useful to you. | True | Recommended performance improvement - ### Description
The Google Chrome browser warns (console) about a performance hit with non-passive event listeners. It points to tracker.js.
A recommendation for improvement is documented here: [passive-event-listeners](https://developers.google.com/web/tools/lighthouse/audits/passive-event-listeners)
### Modifications tested
My apologies, I didn't have much time to look at this but to quickly test, and serve as an example I made the following updates to tracker.js:
1. in class Tracker, add the passive option to the event listeners. In practice the recommendation is to check if the passive option is support. In my case it wasn't necessary, hence:
```
_element.addEventListener('mousedown', this._down, {passive:true});
_element.addEventListener('touchstart', this._down, {passive:true});
_element.addEventListener('touchmove', this._move, {passive:true});
_element.addEventListener('touchend', this._up, {passive:true});
```
2. in _onDown(down), remove:
`ev.preventDefault()`
### Result
The Google warning went away and the element appeared to work OK.
Hopefully this is useful to you. | non_priority | recommended performance improvement description the google chrome browser warns console about a performance hit with non passive event listeners it points to tracker js a recommendation for improvement is documented here modifications tested my apologies i didn t have much time to look at this but to quickly test and serve as an example i made the following updates to tracker js in class tracker add the passive option to the event listeners in practice the recommendation is to check if the passive option is support in my case it wasn t necessary hence element addeventlistener mousedown this down passive true element addeventlistener touchstart this down passive true element addeventlistener touchmove this move passive true element addeventlistener touchend this up passive true in ondown down remove ev preventdefault result the google warning went away and the element appeared to work ok hopefully this is useful to you | 0 |
45,455 | 11,672,031,528 | IssuesEvent | 2020-03-04 05:19:31 | spring-projects/spring-security | https://api.github.com/repos/spring-projects/spring-security | closed | Update to spring-build-conventions:0.0.31.RELEASE | in: build type: dependency-upgrade | <!--
For Security Vulnerabilities, please use https://pivotal.io/security#reporting
-->
### Summary
Update to spring-build-conventions:0.0.31.RELEASE
This will fix gh-8025 | 1.0 | Update to spring-build-conventions:0.0.31.RELEASE - <!--
For Security Vulnerabilities, please use https://pivotal.io/security#reporting
-->
### Summary
Update to spring-build-conventions:0.0.31.RELEASE
This will fix gh-8025 | non_priority | update to spring build conventions release for security vulnerabilities please use summary update to spring build conventions release this will fix gh | 0 |
349,874 | 24,959,666,910 | IssuesEvent | 2022-11-01 14:36:15 | AY2223S1-CS2103T-T14-3/tp | https://api.github.com/repos/AY2223S1-CS2103T-T14-3/tp | closed | [PE-D][Tester E] UG: Misleading birthdate format for 'add' command | documentation Must Fix | In the UG, for 'add' command, the tip given for birthdate is as follows:
"Tip: Birthdate must be given in DD-MM-YYYY HHmm format!"
This is not accurate, as the correct format should have been DD-MM-YYYY instead! Be careful!
<!--session: 1666945098506-7b9a3a6b-9771-404c-be16-3420d7464ef5-->
<!--Version: Web v3.4.4-->
-------------
Labels: `type.DocumentationBug` `severity.High`
original: eugenelmj/ped#9 | 1.0 | [PE-D][Tester E] UG: Misleading birthdate format for 'add' command - In the UG, for 'add' command, the tip given for birthdate is as follows:
"Tip: Birthdate must be given in DD-MM-YYYY HHmm format!"
This is not accurate, as the correct format should have been DD-MM-YYYY instead! Be careful!
<!--session: 1666945098506-7b9a3a6b-9771-404c-be16-3420d7464ef5-->
<!--Version: Web v3.4.4-->
-------------
Labels: `type.DocumentationBug` `severity.High`
original: eugenelmj/ped#9 | non_priority | ug misleading birthdate format for add command in the ug for add command the tip given for birthdate is as follows tip birthdate must be given in dd mm yyyy hhmm format this is not accurate as the correct format should have been dd mm yyyy instead be careful labels type documentationbug severity high original eugenelmj ped | 0 |
38,357 | 8,461,339,707 | IssuesEvent | 2018-10-22 21:30:37 | MicrosoftDocs/live-share | https://api.github.com/repos/MicrosoftDocs/live-share | closed | [VS Code] The user code is invalid or expired. Try signing in again. | area: identity and sign-in duplicate os: linux vscode | <!--
For Visual Studio problems/feedback, please use the "Report a Problem..." feature built into the tool. See https://aka.ms/vsls-vsproblem.
For VS Code issues, attach verbose logs as follows:
1. Press F1 (or Ctrl-Shift-P), type "export logs" and run the "Live Share: Export Logs" command.
2. Drag and drop the zip to the issue on this screen and wait for it to upload before creating the issue.
For feature requests, please include enough of this same info so we know if the request is tool or language/platform specific.
-->
## Error:
The user code is invalid or expired. Try signing in again.
## Steps to Reproduce:
1.
2.
||Version Data|
|-:|:-|
|**extensionName**|VSLS|
|**extensionVersion**|0.3.788|
|**protocolVersion**|2.2|
|**applicationName**|VSCode|
|**applicationVersion**|1.28.0|
|**platformName**|Linux|
|**platformVersion**|4.15.0-36-generic| | 1.0 | [VS Code] The user code is invalid or expired. Try signing in again. - <!--
For Visual Studio problems/feedback, please use the "Report a Problem..." feature built into the tool. See https://aka.ms/vsls-vsproblem.
For VS Code issues, attach verbose logs as follows:
1. Press F1 (or Ctrl-Shift-P), type "export logs" and run the "Live Share: Export Logs" command.
2. Drag and drop the zip to the issue on this screen and wait for it to upload before creating the issue.
For feature requests, please include enough of this same info so we know if the request is tool or language/platform specific.
-->
## Error:
The user code is invalid or expired. Try signing in again.
## Steps to Reproduce:
1.
2.
||Version Data|
|-:|:-|
|**extensionName**|VSLS|
|**extensionVersion**|0.3.788|
|**protocolVersion**|2.2|
|**applicationName**|VSCode|
|**applicationVersion**|1.28.0|
|**platformName**|Linux|
|**platformVersion**|4.15.0-36-generic| | non_priority | the user code is invalid or expired try signing in again for visual studio problems feedback please use the report a problem feature built into the tool see for vs code issues attach verbose logs as follows press or ctrl shift p type export logs and run the live share export logs command drag and drop the zip to the issue on this screen and wait for it to upload before creating the issue for feature requests please include enough of this same info so we know if the request is tool or language platform specific error the user code is invalid or expired try signing in again steps to reproduce version data extensionname vsls extensionversion protocolversion applicationname vscode applicationversion platformname linux platformversion generic | 0 |
301,796 | 22,775,237,020 | IssuesEvent | 2022-07-08 13:53:26 | brpocock/grizzards | https://api.github.com/repos/brpocock/grizzards | closed | Final map graphic for manual | documentation enhancement | At least to include the Demo area and hints of Anchor Village, Lost Mine, Spiral Woods, Port Lion, and Shadow Valley. | 1.0 | Final map graphic for manual - At least to include the Demo area and hints of Anchor Village, Lost Mine, Spiral Woods, Port Lion, and Shadow Valley. | non_priority | final map graphic for manual at least to include the demo area and hints of anchor village lost mine spiral woods port lion and shadow valley | 0 |
22,877 | 20,405,588,530 | IssuesEvent | 2022-02-23 04:44:25 | geneontology/go-site | https://api.github.com/repos/geneontology/go-site | closed | Get validate.py to run all GORULES again | bug (B: affects usability) | @pgaudet Observed in 2021-05-22 `snapshot` that most GORULES were not running. This appears due to https://github.com/biolink/ontobio/issues/564. In the short term, to bypass the wonky default, we can just explicitly add `--rule-set all` to this `validate.py` cmd:
https://github.com/geneontology/go-site/blob/0f6c248b5aa2f253cfe3a9a736cb424f7851e91c/pipeline/Makefile#L89
Tagging @kltm | True | Get validate.py to run all GORULES again - @pgaudet Observed in 2021-05-22 `snapshot` that most GORULES were not running. This appears due to https://github.com/biolink/ontobio/issues/564. In the short term, to bypass the wonky default, we can just explicitly add `--rule-set all` to this `validate.py` cmd:
https://github.com/geneontology/go-site/blob/0f6c248b5aa2f253cfe3a9a736cb424f7851e91c/pipeline/Makefile#L89
Tagging @kltm | non_priority | get validate py to run all gorules again pgaudet observed in snapshot that most gorules were not running this appears due to in the short term to bypass the wonky default we can just explicitly add rule set all to this validate py cmd tagging kltm | 0 |
6,441 | 8,737,904,727 | IssuesEvent | 2018-12-12 00:30:36 | scylladb/scylla | https://api.github.com/repos/scylladb/scylla | closed | Cassandra's parameter for compression chunk configuration not accepted by Scylla | CQL cassandra 2.2 compatibility n00b | Cassandra configures the chunk size of compression with the parameter `chunk_length_in_kb`. Our parameter is a bit different, and it was probably an overlook on our side: `chunk_length_kb`
Because we now would have Scylla deployments using our parameter, best would be to accept both.
| True | Cassandra's parameter for compression chunk configuration not accepted by Scylla - Cassandra configures the chunk size of compression with the parameter `chunk_length_in_kb`. Our parameter is a bit different, and it was probably an overlook on our side: `chunk_length_kb`
Because we now would have Scylla deployments using our parameter, best would be to accept both.
| non_priority | cassandra s parameter for compression chunk configuration not accepted by scylla cassandra configures the chunk size of compression with the parameter chunk length in kb our parameter is a bit different and it was probably an overlook on our side chunk length kb because we now would have scylla deployments using our parameter best would be to accept both | 0 |
50,297 | 13,187,429,124 | IssuesEvent | 2020-08-13 03:23:08 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | closed | I3ParticleVector pybindings missing bases (Trac #479) | Migrated from Trac dataclasses defect | I3ParticleVector pybindings are missing the I3FrameObject base, and maybe other things.
We also have a similar pybinding by the name of I3VectorI3Particle that does work. Let's resolve this naming duplication.
Also examine other I3Vector classes for similar problems. I3RecoPulseSeries was noted.
<details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/479
, reported by david.schultz and owned by olivas_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-02-12T06:52:34",
"description": "I3ParticleVector pybindings are missing the I3FrameObject base, and maybe other things.\n\nWe also have a similar pybinding by the name of I3VectorI3Particle that does work. Let's resolve this naming duplication.\n\nAlso examine other I3Vector classes for similar problems. I3RecoPulseSeries was noted.",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1423723954189338",
"component": "dataclasses",
"summary": "I3ParticleVector pybindings missing bases",
"priority": "normal",
"keywords": "",
"time": "2014-01-22T03:15:57",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
| 1.0 | I3ParticleVector pybindings missing bases (Trac #479) - I3ParticleVector pybindings are missing the I3FrameObject base, and maybe other things.
We also have a similar pybinding by the name of I3VectorI3Particle that does work. Let's resolve this naming duplication.
Also examine other I3Vector classes for similar problems. I3RecoPulseSeries was noted.
<details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/479
, reported by david.schultz and owned by olivas_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-02-12T06:52:34",
"description": "I3ParticleVector pybindings are missing the I3FrameObject base, and maybe other things.\n\nWe also have a similar pybinding by the name of I3VectorI3Particle that does work. Let's resolve this naming duplication.\n\nAlso examine other I3Vector classes for similar problems. I3RecoPulseSeries was noted.",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1423723954189338",
"component": "dataclasses",
"summary": "I3ParticleVector pybindings missing bases",
"priority": "normal",
"keywords": "",
"time": "2014-01-22T03:15:57",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
| non_priority | pybindings missing bases trac pybindings are missing the base and maybe other things we also have a similar pybinding by the name of that does work let s resolve this naming duplication also examine other classes for similar problems was noted migrated from reported by david schultz and owned by olivas json status closed changetime description pybindings are missing the base and maybe other things n nwe also have a similar pybinding by the name of that does work let s resolve this naming duplication n nalso examine other classes for similar problems was noted reporter david schultz cc resolution fixed ts component dataclasses summary pybindings missing bases priority normal keywords time milestone owner olivas type defect | 0 |
226,504 | 17,353,857,320 | IssuesEvent | 2021-07-29 12:12:38 | STooDs-tools/RSTooDs | https://api.github.com/repos/STooDs-tools/RSTooDs | opened | Improvements to vignettes | documentation | * All vignettes: set knit directory to working directory (probably: knitr::opts_knit$set(root.dir = getwd()) in setup chunk).
* Vignette "covariate": plot the pmf of the estimated Bernoulli by consistency with the first "getting started" vignette.
* Create new vignette "Bayesian Basics" reproducing existing teaching material with RSTooDs. | 1.0 | Improvements to vignettes - * All vignettes: set knit directory to working directory (probably: knitr::opts_knit$set(root.dir = getwd()) in setup chunk).
* Vignette "covariate": plot the pmf of the estimated Bernoulli by consistency with the first "getting started" vignette.
* Create new vignette "Bayesian Basics" reproducing existing teaching material with RSTooDs. | non_priority | improvements to vignettes all vignettes set knit directory to working directory probably knitr opts knit set root dir getwd in setup chunk vignette covariate plot the pmf of the estimated bernoulli by consistency with the first getting started vignette create new vignette bayesian basics reproducing existing teaching material with rstoods | 0 |
337,347 | 24,536,223,210 | IssuesEvent | 2022-10-11 21:01:26 | trimble-oss/dba-dash | https://api.github.com/repos/trimble-oss/dba-dash | closed | Performance Summary riddles and puzzels | documentation | Hi
Today I was reviewing the "PerfMon Summary" information (Version used 2.22.8) and I was wondering about following points below. It is worth to mention, all the data presented in all the issues are real world data and they are collected from productive systems refelecting real scenarios.
Starting from top to bottom (in the image):
- We got a column called "MB/Sec". I couldn't **clearly** decipher what this values is. Maybe a better column header would help. There are also no Tool Tips which explain the columns. This even could be implemnted as an additional feature: Tool Tip which explains the column headers in all the table grids, application wide.
- Column "Read Latency" shows a painted cell as warning. I was seraching on the server itself for a quite while and I couldn't find anything which would help me to identify and localize the issue reported **on the spot**. I then searched even more and eventually I discovered a way to get some values and responsibles for the latency. This was hard and to some extent cryptical but maybe there are better ways. Anyhow, this was my approach:
1. Switch to single server.
2. Click on "Performance" Tab
3. On the "IO Perfromance" chart click "View Table Summary" (Which is yet another Modal Window I discovered :D) and then review the reported values which by the way they aren't painted at this stage so if one isn't paying attention he/she will then miss the point and leave.
The above just tried to demonstrate how cumbersome is to identify the reported problem in detail. I hope you would come up with a easier / better way to solve this case.
- Columns "Read Latecy" and "Write Latecy" are being painted as green. I really tried to understand **why**. I couldn't come up with any logic.
- There are some cells in "Critical Wait" column which are gray colored. **Why**? Value 0.00 can't be gray AND green at the same time!
- In "Lock Wait" and "Total Wait" columns, I find values way high in compare to the other similar values but these cells aren't painted at all. **Why**? Aren't we using any color indicators in theses columns? Shouldn't we do that?
- There are some cells in the "Read Latecy" with no values at all. **Why?** Isn't there any values existing or are they so small that could be ignored? If so, then why can't we show 0,0?
I am sure each one of the above questions can be answered one way or another but the MAIN point here is:
The end user needs to **work** with these data and try not to decipher them. One way or another everyone (with enough
persistence) can get along but that is also not the point of having the values collected. Application should do the most job for the user (because it is the tool, it is hard to keep track of these values and using visual effect one can transfer messages which helps the process.) and the user mostly needs to focus on identifying, anlasying of the collected data. This of course implies, collected data need to be precise and relyable on one hand and the application should try to avoid ambiguity on the other hand.

| 1.0 | Performance Summary riddles and puzzels - Hi
Today I was reviewing the "PerfMon Summary" information (Version used 2.22.8) and I was wondering about following points below. It is worth to mention, all the data presented in all the issues are real world data and they are collected from productive systems refelecting real scenarios.
Starting from top to bottom (in the image):
- We got a column called "MB/Sec". I couldn't **clearly** decipher what this values is. Maybe a better column header would help. There are also no Tool Tips which explain the columns. This even could be implemnted as an additional feature: Tool Tip which explains the column headers in all the table grids, application wide.
- Column "Read Latency" shows a painted cell as warning. I was seraching on the server itself for a quite while and I couldn't find anything which would help me to identify and localize the issue reported **on the spot**. I then searched even more and eventually I discovered a way to get some values and responsibles for the latency. This was hard and to some extent cryptical but maybe there are better ways. Anyhow, this was my approach:
1. Switch to single server.
2. Click on "Performance" Tab
3. On the "IO Perfromance" chart click "View Table Summary" (Which is yet another Modal Window I discovered :D) and then review the reported values which by the way they aren't painted at this stage so if one isn't paying attention he/she will then miss the point and leave.
The above just tried to demonstrate how cumbersome is to identify the reported problem in detail. I hope you would come up with a easier / better way to solve this case.
- Columns "Read Latecy" and "Write Latecy" are being painted as green. I really tried to understand **why**. I couldn't come up with any logic.
- There are some cells in "Critical Wait" column which are gray colored. **Why**? Value 0.00 can't be gray AND green at the same time!
- In "Lock Wait" and "Total Wait" columns, I find values way high in compare to the other similar values but these cells aren't painted at all. **Why**? Aren't we using any color indicators in theses columns? Shouldn't we do that?
- There are some cells in the "Read Latecy" with no values at all. **Why?** Isn't there any values existing or are they so small that could be ignored? If so, then why can't we show 0,0?
I am sure each one of the above questions can be answered one way or another but the MAIN point here is:
The end user needs to **work** with these data and try not to decipher them. One way or another everyone (with enough
persistence) can get along but that is also not the point of having the values collected. Application should do the most job for the user (because it is the tool, it is hard to keep track of these values and using visual effect one can transfer messages which helps the process.) and the user mostly needs to focus on identifying, anlasying of the collected data. This of course implies, collected data need to be precise and relyable on one hand and the application should try to avoid ambiguity on the other hand.

| non_priority | performance summary riddles and puzzels hi today i was reviewing the perfmon summary information version used and i was wondering about following points below it is worth to mention all the data presented in all the issues are real world data and they are collected from productive systems refelecting real scenarios starting from top to bottom in the image we got a column called mb sec i couldn t clearly decipher what this values is maybe a better column header would help there are also no tool tips which explain the columns this even could be implemnted as an additional feature tool tip which explains the column headers in all the table grids application wide column read latency shows a painted cell as warning i was seraching on the server itself for a quite while and i couldn t find anything which would help me to identify and localize the issue reported on the spot i then searched even more and eventually i discovered a way to get some values and responsibles for the latency this was hard and to some extent cryptical but maybe there are better ways anyhow this was my approach switch to single server click on performance tab on the io perfromance chart click view table summary which is yet another modal window i discovered d and then review the reported values which by the way they aren t painted at this stage so if one isn t paying attention he she will then miss the point and leave the above just tried to demonstrate how cumbersome is to identify the reported problem in detail i hope you would come up with a easier better way to solve this case columns read latecy and write latecy are being painted as green i really tried to understand why i couldn t come up with any logic there are some cells in critical wait column which are gray colored why value can t be gray and green at the same time in lock wait and total wait columns i find values way high in compare to the other similar values but these cells aren t painted at all why aren t we using any color indicators in theses columns shouldn t we do that there are some cells in the read latecy with no values at all why isn t there any values existing or are they so small that could be ignored if so then why can t we show i am sure each one of the above questions can be answered one way or another but the main point here is the end user needs to work with these data and try not to decipher them one way or another everyone with enough persistence can get along but that is also not the point of having the values collected application should do the most job for the user because it is the tool it is hard to keep track of these values and using visual effect one can transfer messages which helps the process and the user mostly needs to focus on identifying anlasying of the collected data this of course implies collected data need to be precise and relyable on one hand and the application should try to avoid ambiguity on the other hand | 0 |
84,915 | 10,421,114,040 | IssuesEvent | 2019-09-16 04:32:57 | YugaByte/yugabyte-db | https://api.github.com/repos/YugaByte/yugabyte-db | closed | [docs] Fix copy button copying for prompts | area/documentation kind/improve-ux | With the recent change of default username from `postgres` to `yugabyte` in commit 713233a, some docs pages have started to migrate towards using `yugabyte` in their prompts, and that is the intended plan as detailed in issue #2319. Unfortunately, this breaks the smart copying functionality that previously worked when `postgres` was used in the prompts. Previously, the copy button would only copy commands and not the prompt `postgres#= `. Now, the copy button simply copies it all. Restore the previous functionality. | 1.0 | [docs] Fix copy button copying for prompts - With the recent change of default username from `postgres` to `yugabyte` in commit 713233a, some docs pages have started to migrate towards using `yugabyte` in their prompts, and that is the intended plan as detailed in issue #2319. Unfortunately, this breaks the smart copying functionality that previously worked when `postgres` was used in the prompts. Previously, the copy button would only copy commands and not the prompt `postgres#= `. Now, the copy button simply copies it all. Restore the previous functionality. | non_priority | fix copy button copying for prompts with the recent change of default username from postgres to yugabyte in commit some docs pages have started to migrate towards using yugabyte in their prompts and that is the intended plan as detailed in issue unfortunately this breaks the smart copying functionality that previously worked when postgres was used in the prompts previously the copy button would only copy commands and not the prompt postgres now the copy button simply copies it all restore the previous functionality | 0 |
76,280 | 9,932,663,810 | IssuesEvent | 2019-07-02 10:21:00 | microsoft/vcpkg | https://api.github.com/repos/microsoft/vcpkg | closed | CmakeSettings example uses absolute path ? | documentation | I'm pretty new to C++ and Windows dev ... from what I've seen CMakeSettings.json should be in git.
The [example here](https://github.com/Microsoft/vcpkg/blob/master/docs/examples/installing-and-using-packages.md#cmake-toolchain-file) adds a path in ```D:\\src\\vcpkg\\....\\vcpkg.cmake```.
Is there any way this could be somewhere under the user profile ?
I can't expect other users of my code to have the same drive mappings as I do, even if I do move it from ```D:\```.
Or should I add a copy vcpkg.cmake somewhere in my own source repo ? | 1.0 | CmakeSettings example uses absolute path ? - I'm pretty new to C++ and Windows dev ... from what I've seen CMakeSettings.json should be in git.
The [example here](https://github.com/Microsoft/vcpkg/blob/master/docs/examples/installing-and-using-packages.md#cmake-toolchain-file) adds a path in ```D:\\src\\vcpkg\\....\\vcpkg.cmake```.
Is there any way this could be somewhere under the user profile ?
I can't expect other users of my code to have the same drive mappings as I do, even if I do move it from ```D:\```.
Or should I add a copy vcpkg.cmake somewhere in my own source repo ? | non_priority | cmakesettings example uses absolute path i m pretty new to c and windows dev from what i ve seen cmakesettings json should be in git the adds a path in d src vcpkg vcpkg cmake is there any way this could be somewhere under the user profile i can t expect other users of my code to have the same drive mappings as i do even if i do move it from d or should i add a copy vcpkg cmake somewhere in my own source repo | 0 |
134,001 | 18,403,273,437 | IssuesEvent | 2021-10-12 18:50:05 | susanstwss/WebGoat | https://api.github.com/repos/susanstwss/WebGoat | opened | CVE-2020-26259 (Medium) detected in xstream-1.4.5.jar | security vulnerability | ## CVE-2020-26259 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.5.jar</b></p></summary>
<p>XStream is a serialization library from Java objects to XML and back.</p>
<p>Path to dependency file: WebGoat/webgoat-server/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar,m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar</p>
<p>
Dependency Hierarchy:
- :x: **xstream-1.4.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/susanstwss/WebGoat/commit/2b2a58a6bee86bd0e98aa367052cda5f99597abb">2b2a58a6bee86bd0e98aa367052cda5f99597abb</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.15, is vulnerable to an Arbitrary File Deletion on the local host when unmarshalling. The vulnerability may allow a remote attacker to delete arbitrary know files on the host as log as the executing process has sufficient rights only by manipulating the processed input stream. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.15. The reported vulnerability does not exist running Java 15 or higher. No user is affected, who followed the recommendation to setup XStream's Security Framework with a whitelist! Anyone relying on XStream's default blacklist can immediately switch to a whilelist for the allowed types to avoid the vulnerability. Users of XStream 1.4.14 or below who still want to use XStream default blacklist can use a workaround described in more detailed in the referenced advisories.
<p>Publish Date: 2020-12-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26259>CVE-2020-26259</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-jfvx-7wrx-43fh">https://github.com/x-stream/xstream/security/advisories/GHSA-jfvx-7wrx-43fh</a></p>
<p>Release Date: 2020-12-16</p>
<p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.15</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.thoughtworks.xstream","packageName":"xstream","packageVersion":"1.4.5","packageFilePaths":["/webgoat-server/pom.xml","/webgoat-lessons/vulnerable-components/pom.xml","/webgoat-integration-tests/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.thoughtworks.xstream:xstream:1.4.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.thoughtworks.xstream:xstream:1.4.15"}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2020-26259","vulnerabilityDetails":"XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.15, is vulnerable to an Arbitrary File Deletion on the local host when unmarshalling. The vulnerability may allow a remote attacker to delete arbitrary know files on the host as log as the executing process has sufficient rights only by manipulating the processed input stream. If you rely on XStream\u0027s default blacklist of the Security Framework, you will have to use at least version 1.4.15. The reported vulnerability does not exist running Java 15 or higher. No user is affected, who followed the recommendation to setup XStream\u0027s Security Framework with a whitelist! Anyone relying on XStream\u0027s default blacklist can immediately switch to a whilelist for the allowed types to avoid the vulnerability. Users of XStream 1.4.14 or below who still want to use XStream default blacklist can use a workaround described in more detailed in the referenced advisories.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26259","cvss3Severity":"medium","cvss3Score":"6.8","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Changed","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-26259 (Medium) detected in xstream-1.4.5.jar - ## CVE-2020-26259 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.5.jar</b></p></summary>
<p>XStream is a serialization library from Java objects to XML and back.</p>
<p>Path to dependency file: WebGoat/webgoat-server/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar,m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar</p>
<p>
Dependency Hierarchy:
- :x: **xstream-1.4.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/susanstwss/WebGoat/commit/2b2a58a6bee86bd0e98aa367052cda5f99597abb">2b2a58a6bee86bd0e98aa367052cda5f99597abb</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.15, is vulnerable to an Arbitrary File Deletion on the local host when unmarshalling. The vulnerability may allow a remote attacker to delete arbitrary know files on the host as log as the executing process has sufficient rights only by manipulating the processed input stream. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.15. The reported vulnerability does not exist running Java 15 or higher. No user is affected, who followed the recommendation to setup XStream's Security Framework with a whitelist! Anyone relying on XStream's default blacklist can immediately switch to a whilelist for the allowed types to avoid the vulnerability. Users of XStream 1.4.14 or below who still want to use XStream default blacklist can use a workaround described in more detailed in the referenced advisories.
<p>Publish Date: 2020-12-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26259>CVE-2020-26259</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-jfvx-7wrx-43fh">https://github.com/x-stream/xstream/security/advisories/GHSA-jfvx-7wrx-43fh</a></p>
<p>Release Date: 2020-12-16</p>
<p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.15</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.thoughtworks.xstream","packageName":"xstream","packageVersion":"1.4.5","packageFilePaths":["/webgoat-server/pom.xml","/webgoat-lessons/vulnerable-components/pom.xml","/webgoat-integration-tests/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.thoughtworks.xstream:xstream:1.4.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.thoughtworks.xstream:xstream:1.4.15"}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2020-26259","vulnerabilityDetails":"XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.15, is vulnerable to an Arbitrary File Deletion on the local host when unmarshalling. The vulnerability may allow a remote attacker to delete arbitrary know files on the host as log as the executing process has sufficient rights only by manipulating the processed input stream. If you rely on XStream\u0027s default blacklist of the Security Framework, you will have to use at least version 1.4.15. The reported vulnerability does not exist running Java 15 or higher. No user is affected, who followed the recommendation to setup XStream\u0027s Security Framework with a whitelist! Anyone relying on XStream\u0027s default blacklist can immediately switch to a whilelist for the allowed types to avoid the vulnerability. Users of XStream 1.4.14 or below who still want to use XStream default blacklist can use a workaround described in more detailed in the referenced advisories.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26259","cvss3Severity":"medium","cvss3Score":"6.8","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Changed","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_priority | cve medium detected in xstream jar cve medium severity vulnerability vulnerable library xstream jar xstream is a serialization library from java objects to xml and back path to dependency file webgoat webgoat server pom xml path to vulnerable library home wss scanner repository com thoughtworks xstream xstream xstream jar repository com thoughtworks xstream xstream xstream jar home wss scanner repository com thoughtworks xstream xstream xstream jar dependency hierarchy x xstream jar vulnerable library found in head commit a href found in base branch develop vulnerability details xstream is a java library to serialize objects to xml and back again in xstream before version is vulnerable to an arbitrary file deletion on the local host when unmarshalling the vulnerability may allow a remote attacker to delete arbitrary know files on the host as log as the executing process has sufficient rights only by manipulating the processed input stream if you rely on xstream s default blacklist of the security framework you will have to use at least version the reported vulnerability does not exist running java or higher no user is affected who followed the recommendation to setup xstream s security framework with a whitelist anyone relying on xstream s default blacklist can immediately switch to a whilelist for the allowed types to avoid the vulnerability users of xstream or below who still want to use xstream default blacklist can use a workaround described in more detailed in the referenced advisories publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope changed impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com thoughtworks xstream xstream check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com thoughtworks xstream xstream isminimumfixversionavailable true minimumfixversion com thoughtworks xstream xstream basebranches vulnerabilityidentifier cve vulnerabilitydetails xstream is a java library to serialize objects to xml and back again in xstream before version is vulnerable to an arbitrary file deletion on the local host when unmarshalling the vulnerability may allow a remote attacker to delete arbitrary know files on the host as log as the executing process has sufficient rights only by manipulating the processed input stream if you rely on xstream default blacklist of the security framework you will have to use at least version the reported vulnerability does not exist running java or higher no user is affected who followed the recommendation to setup xstream security framework with a whitelist anyone relying on xstream default blacklist can immediately switch to a whilelist for the allowed types to avoid the vulnerability users of xstream or below who still want to use xstream default blacklist can use a workaround described in more detailed in the referenced advisories vulnerabilityurl | 0 |
145,644 | 19,339,776,022 | IssuesEvent | 2021-12-15 02:09:53 | madhans23/linux-4.15 | https://api.github.com/repos/madhans23/linux-4.15 | opened | CVE-2021-43975 (Medium) detected in linux-stagingv5.15 | security vulnerability | ## CVE-2021-43975 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stagingv5.15</b></p></summary>
<p>
<p>hwmon staging tree</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/groeck/linux-staging.git>https://git.kernel.org/pub/scm/linux/kernel/git/groeck/linux-staging.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_utils.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In the Linux kernel through 5.15.2, hw_atl_utils_fw_rpc_wait in drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_utils.c allows an attacker (who can introduce a crafted device) to trigger an out-of-bounds write via a crafted length value.
<p>Publish Date: 2021-11-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43975>CVE-2021-43975</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-43975">https://www.linuxkernelcves.com/cves/CVE-2021-43975</a></p>
<p>Release Date: 2021-11-17</p>
<p>Fix Resolution: v5.16-rc2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-43975 (Medium) detected in linux-stagingv5.15 - ## CVE-2021-43975 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stagingv5.15</b></p></summary>
<p>
<p>hwmon staging tree</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/groeck/linux-staging.git>https://git.kernel.org/pub/scm/linux/kernel/git/groeck/linux-staging.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_utils.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In the Linux kernel through 5.15.2, hw_atl_utils_fw_rpc_wait in drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_utils.c allows an attacker (who can introduce a crafted device) to trigger an out-of-bounds write via a crafted length value.
<p>Publish Date: 2021-11-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43975>CVE-2021-43975</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-43975">https://www.linuxkernelcves.com/cves/CVE-2021-43975</a></p>
<p>Release Date: 2021-11-17</p>
<p>Fix Resolution: v5.16-rc2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in linux cve medium severity vulnerability vulnerable library linux hwmon staging tree library home page a href found in base branch master vulnerable source files drivers net ethernet aquantia atlantic hw atl hw atl utils c vulnerability details in the linux kernel through hw atl utils fw rpc wait in drivers net ethernet aquantia atlantic hw atl hw atl utils c allows an attacker who can introduce a crafted device to trigger an out of bounds write via a crafted length value publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
13,237 | 8,380,412,972 | IssuesEvent | 2018-10-07 14:31:59 | godotengine/godot | https://api.github.com/repos/godotengine/godot | closed | New property inspector: Scripts don't show filename | enhancement topic:editor usability | **Godot version:**
ba134d4
**Issue description:**
This is more of a personal nitpick. While I find the script thumb preview *marginally* useful, seeing a filename at a glance would be much more useful, or in a tooltip if must be. I'm assuming this is simply a feature in testing (for supporting property field previews which weren't feasible before?), but nevertheless it seems to be an issue.
| True | New property inspector: Scripts don't show filename - **Godot version:**
ba134d4
**Issue description:**
This is more of a personal nitpick. While I find the script thumb preview *marginally* useful, seeing a filename at a glance would be much more useful, or in a tooltip if must be. I'm assuming this is simply a feature in testing (for supporting property field previews which weren't feasible before?), but nevertheless it seems to be an issue.
| non_priority | new property inspector scripts don t show filename godot version issue description this is more of a personal nitpick while i find the script thumb preview marginally useful seeing a filename at a glance would be much more useful or in a tooltip if must be i m assuming this is simply a feature in testing for supporting property field previews which weren t feasible before but nevertheless it seems to be an issue | 0 |
70,401 | 30,660,454,586 | IssuesEvent | 2023-07-25 14:41:59 | hashicorp/terraform-provider-azurerm | https://api.github.com/repos/hashicorp/terraform-provider-azurerm | closed | azurerm_application_gateway Health Prode - Body should not be required field | bug waiting-response service/application-gateway v/3.x | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Community Note
<!--- Please keep this note for the community --->
* Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform Version
1.3.7
### AzureRM Provider Version
3.42.0
### Affected Resource(s)/Data Source(s)
azurerm_application_gateway
### Terraform Configuration Files
```hcl
resource "azurerm_application_gateway" "network" {
name = "example-appgateway"
resource_group_name = azurerm_resource_group.example.name
location = azurerm_resource_group.example.location
sku {
name = "Standard_Small"
tier = "Standard"
capacity = 2
}
gateway_ip_configuration {
name = "my-gateway-ip-configuration"
subnet_id = azurerm_subnet.frontend.id
}
frontend_port {
name = local.frontend_port_name
port = 80
}
frontend_ip_configuration {
name = local.frontend_ip_configuration_name
public_ip_address_id = azurerm_public_ip.example.id
}
backend_address_pool {
name = local.backend_address_pool_name
}
backend_http_settings {
name = local.http_setting_name
cookie_based_affinity = "Disabled"
path = "/path1/"
port = 80
protocol = "Http"
request_timeout = 60
}
http_listener {
name = local.listener_name
frontend_ip_configuration_name = local.frontend_ip_configuration_name
frontend_port_name = local.frontend_port_name
protocol = "Http"
}
request_routing_rule {
name = local.request_routing_rule_name
rule_type = "Basic"
http_listener_name = local.listener_name
backend_address_pool_name = local.backend_address_pool_name
backend_http_settings_name = local.http_setting_name
}
probe {
name = "health-probe-default-https"
pick_host_name_from_backend_http_settings = true
path = "/"
protocol = "Https"
port = 443
interval = 30
timeout = 15
unhealthy_threshold = 3
match {
body = ""
status_code = ["200-399"]
}
}
```
### Debug Output/Panic Output
```shell
Error: Missing required argument
│
│ on app-gw-prod\main.tf line 203, in resource "azurerm_application_gateway" "external-prod-centralus":
│ 203: match {
│
│ The argument "body" is required, but no definition was found.
```
### Expected Behaviour
Body should not be a required field. Only status_code should be required. Can workaround by adding _body = ""_
### Actual Behaviour
Failed TF build
### Steps to Reproduce
terraform apply
### Important Factoids
_No response_
### References
_No response_ | 1.0 | azurerm_application_gateway Health Prode - Body should not be required field - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Community Note
<!--- Please keep this note for the community --->
* Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform Version
1.3.7
### AzureRM Provider Version
3.42.0
### Affected Resource(s)/Data Source(s)
azurerm_application_gateway
### Terraform Configuration Files
```hcl
resource "azurerm_application_gateway" "network" {
name = "example-appgateway"
resource_group_name = azurerm_resource_group.example.name
location = azurerm_resource_group.example.location
sku {
name = "Standard_Small"
tier = "Standard"
capacity = 2
}
gateway_ip_configuration {
name = "my-gateway-ip-configuration"
subnet_id = azurerm_subnet.frontend.id
}
frontend_port {
name = local.frontend_port_name
port = 80
}
frontend_ip_configuration {
name = local.frontend_ip_configuration_name
public_ip_address_id = azurerm_public_ip.example.id
}
backend_address_pool {
name = local.backend_address_pool_name
}
backend_http_settings {
name = local.http_setting_name
cookie_based_affinity = "Disabled"
path = "/path1/"
port = 80
protocol = "Http"
request_timeout = 60
}
http_listener {
name = local.listener_name
frontend_ip_configuration_name = local.frontend_ip_configuration_name
frontend_port_name = local.frontend_port_name
protocol = "Http"
}
request_routing_rule {
name = local.request_routing_rule_name
rule_type = "Basic"
http_listener_name = local.listener_name
backend_address_pool_name = local.backend_address_pool_name
backend_http_settings_name = local.http_setting_name
}
probe {
name = "health-probe-default-https"
pick_host_name_from_backend_http_settings = true
path = "/"
protocol = "Https"
port = 443
interval = 30
timeout = 15
unhealthy_threshold = 3
match {
body = ""
status_code = ["200-399"]
}
}
```
### Debug Output/Panic Output
```shell
Error: Missing required argument
│
│ on app-gw-prod\main.tf line 203, in resource "azurerm_application_gateway" "external-prod-centralus":
│ 203: match {
│
│ The argument "body" is required, but no definition was found.
```
### Expected Behaviour
Body should not be a required field. Only status_code should be required. Can workaround by adding _body = ""_
### Actual Behaviour
Failed TF build
### Steps to Reproduce
terraform apply
### Important Factoids
_No response_
### References
_No response_ | non_priority | azurerm application gateway health prode body should not be required field is there an existing issue for this i have searched the existing issues community note please vote on this issue by adding a thumbsup to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform version azurerm provider version affected resource s data source s azurerm application gateway terraform configuration files hcl resource azurerm application gateway network name example appgateway resource group name azurerm resource group example name location azurerm resource group example location sku name standard small tier standard capacity gateway ip configuration name my gateway ip configuration subnet id azurerm subnet frontend id frontend port name local frontend port name port frontend ip configuration name local frontend ip configuration name public ip address id azurerm public ip example id backend address pool name local backend address pool name backend http settings name local http setting name cookie based affinity disabled path port protocol http request timeout http listener name local listener name frontend ip configuration name local frontend ip configuration name frontend port name local frontend port name protocol http request routing rule name local request routing rule name rule type basic http listener name local listener name backend address pool name local backend address pool name backend http settings name local http setting name probe name health probe default https pick host name from backend http settings true path protocol https port interval timeout unhealthy threshold match body status code debug output panic output shell error missing required argument │ │ on app gw prod main tf line in resource azurerm application gateway external prod centralus │ match │ │ the argument body is required but no definition was found expected behaviour body should not be a required field only status code should be required can workaround by adding body actual behaviour failed tf build steps to reproduce terraform apply important factoids no response references no response | 0 |
267,835 | 23,323,111,440 | IssuesEvent | 2022-08-08 18:20:35 | mozilla-mobile/mobile-test-eng | https://api.github.com/repos/mozilla-mobile/mobile-test-eng | opened | Create experimental Fenix cron for FTL improved emulator beta | waiting android infra:mobile infra:ui-test | This depends on if we're accepted into their beta (waiting label).
Context: FTL has a beta sign-up for new emulators with improved speeds and they're looking for early adopters.
Scope: In the same fashion of as `legacy-x86` cron, we can create a similar cron on schedule with a new Flank configuration targeting the new emulator name. It can run the full-suite. | 1.0 | Create experimental Fenix cron for FTL improved emulator beta - This depends on if we're accepted into their beta (waiting label).
Context: FTL has a beta sign-up for new emulators with improved speeds and they're looking for early adopters.
Scope: In the same fashion of as `legacy-x86` cron, we can create a similar cron on schedule with a new Flank configuration targeting the new emulator name. It can run the full-suite. | non_priority | create experimental fenix cron for ftl improved emulator beta this depends on if we re accepted into their beta waiting label context ftl has a beta sign up for new emulators with improved speeds and they re looking for early adopters scope in the same fashion of as legacy cron we can create a similar cron on schedule with a new flank configuration targeting the new emulator name it can run the full suite | 0 |
96,737 | 16,164,449,356 | IssuesEvent | 2021-05-01 07:50:34 | AlexRogalskiy/wotd | https://api.github.com/repos/AlexRogalskiy/wotd | opened | CVE-2020-11022 (Medium) detected in jquery-1.9.1.js, jquery-1.8.1.min.js | security vulnerability | ## CVE-2020-11022 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.9.1.js</b>, <b>jquery-1.8.1.min.js</b></p></summary>
<p>
<details><summary><b>jquery-1.9.1.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.js</a></p>
<p>Path to dependency file: wotd/node_modules/tinygradient/bower_components/tinycolor/index.html</p>
<p>Path to vulnerable library: wotd/node_modules/tinygradient/bower_components/tinycolor/demo/jquery-1.9.1.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.9.1.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.8.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p>
<p>Path to dependency file: wotd/node_modules/redeyed/examples/browser/index.html</p>
<p>Path to vulnerable library: wotd/node_modules/redeyed/examples/browser/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.8.1.min.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/wotd/commit/f6150f55f68ef1a1f68c2ccaa5f94d7cc6a7c5fb">f6150f55f68ef1a1f68c2ccaa5f94d7cc6a7c5fb</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-11022 (Medium) detected in jquery-1.9.1.js, jquery-1.8.1.min.js - ## CVE-2020-11022 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.9.1.js</b>, <b>jquery-1.8.1.min.js</b></p></summary>
<p>
<details><summary><b>jquery-1.9.1.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.js</a></p>
<p>Path to dependency file: wotd/node_modules/tinygradient/bower_components/tinycolor/index.html</p>
<p>Path to vulnerable library: wotd/node_modules/tinygradient/bower_components/tinycolor/demo/jquery-1.9.1.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.9.1.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.8.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p>
<p>Path to dependency file: wotd/node_modules/redeyed/examples/browser/index.html</p>
<p>Path to vulnerable library: wotd/node_modules/redeyed/examples/browser/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.8.1.min.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/wotd/commit/f6150f55f68ef1a1f68c2ccaa5f94d7cc6a7c5fb">f6150f55f68ef1a1f68c2ccaa5f94d7cc6a7c5fb</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in jquery js jquery min js cve medium severity vulnerability vulnerable libraries jquery js jquery min js jquery js javascript library for dom operations library home page a href path to dependency file wotd node modules tinygradient bower components tinycolor index html path to vulnerable library wotd node modules tinygradient bower components tinycolor demo jquery js dependency hierarchy x jquery js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file wotd node modules redeyed examples browser index html path to vulnerable library wotd node modules redeyed examples browser index html dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource | 0 |
126,856 | 26,929,163,850 | IssuesEvent | 2023-02-07 15:41:42 | joomla/joomla-cms | https://api.github.com/repos/joomla/joomla-cms | closed | PHPMail is rejected by the mail server with Error when using PHP 8.X | No Code Attached Yet Information Required | ### PHPMail is rejected by the mail server with an "overlength date field" error when using PHP 8.X. Using PHP 7.4 does not produce the error
Steps to reproduce the issue
Set mail as PHPMail in Global Configuration and set PHP to 8.X
goto "Forgot Logon" and try to get a new password.
Email will be rejected
### Expected result
### Actual result
### System information (as much as possible)
### Additional comments
| 1.0 | PHPMail is rejected by the mail server with Error when using PHP 8.X - ### PHPMail is rejected by the mail server with an "overlength date field" error when using PHP 8.X. Using PHP 7.4 does not produce the error
Steps to reproduce the issue
Set mail as PHPMail in Global Configuration and set PHP to 8.X
goto "Forgot Logon" and try to get a new password.
Email will be rejected
### Expected result
### Actual result
### System information (as much as possible)
### Additional comments
| non_priority | phpmail is rejected by the mail server with error when using php x phpmail is rejected by the mail server with an overlength date field error when using php x using php does not produce the error steps to reproduce the issue set mail as phpmail in global configuration and set php to x goto forgot logon and try to get a new password email will be rejected expected result actual result system information as much as possible additional comments | 0 |
129,011 | 10,559,756,337 | IssuesEvent | 2019-10-04 12:24:42 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | PartitionMigrationListenerTest.testMigrationStats_afterPartitionsLost_when_NO_MIGRATION | Module: Partitioning Source: Internal Team: Core Type: Test-Failure | http://jenkins.hazelcast.com/job/Hazelcast-pr-builder/3601/
```
Expected: a value equal to or greater than <4>
but: <3> was less than <4>
Stacktrace
java.lang.AssertionError:
Expected: a value equal to or greater than <4>
but: <3> was less than <4>
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
at org.junit.Assert.assertThat(Assert.java:956)
at org.junit.Assert.assertThat(Assert.java:923)
at com.hazelcast.partition.PartitionMigrationListenerTest$EventCollectingMigrationListener.lambda$awaitEventPacksComplete$0(PartitionMigrationListenerTest.java:445)
at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:1308)
at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:1410)
at com.hazelcast.partition.PartitionMigrationListenerTest$EventCollectingMigrationListener.awaitEventPacksComplete(PartitionMigrationListenerTest.java:444)
at com.hazelcast.partition.PartitionMigrationListenerTest$EventCollectingMigrationListener.ensureAndGetEventPacks(PartitionMigrationListenerTest.java:435)
at com.hazelcast.partition.PartitionMigrationListenerTest.testMigrationStats_afterPartitionsLost_when_NO_MIGRATION(PartitionMigrationListenerTest.java:178)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:114)
at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:106)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.lang.Thread.run(Thread.java:748)
``` | 1.0 | PartitionMigrationListenerTest.testMigrationStats_afterPartitionsLost_when_NO_MIGRATION - http://jenkins.hazelcast.com/job/Hazelcast-pr-builder/3601/
```
Expected: a value equal to or greater than <4>
but: <3> was less than <4>
Stacktrace
java.lang.AssertionError:
Expected: a value equal to or greater than <4>
but: <3> was less than <4>
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
at org.junit.Assert.assertThat(Assert.java:956)
at org.junit.Assert.assertThat(Assert.java:923)
at com.hazelcast.partition.PartitionMigrationListenerTest$EventCollectingMigrationListener.lambda$awaitEventPacksComplete$0(PartitionMigrationListenerTest.java:445)
at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:1308)
at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:1410)
at com.hazelcast.partition.PartitionMigrationListenerTest$EventCollectingMigrationListener.awaitEventPacksComplete(PartitionMigrationListenerTest.java:444)
at com.hazelcast.partition.PartitionMigrationListenerTest$EventCollectingMigrationListener.ensureAndGetEventPacks(PartitionMigrationListenerTest.java:435)
at com.hazelcast.partition.PartitionMigrationListenerTest.testMigrationStats_afterPartitionsLost_when_NO_MIGRATION(PartitionMigrationListenerTest.java:178)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:114)
at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:106)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.lang.Thread.run(Thread.java:748)
``` | non_priority | partitionmigrationlistenertest testmigrationstats afterpartitionslost when no migration expected a value equal to or greater than but was less than stacktrace java lang assertionerror expected a value equal to or greater than but was less than at org hamcrest matcherassert assertthat matcherassert java at org junit assert assertthat assert java at org junit assert assertthat assert java at com hazelcast partition partitionmigrationlistenertest eventcollectingmigrationlistener lambda awaiteventpackscomplete partitionmigrationlistenertest java at com hazelcast test hazelcasttestsupport asserttrueeventually hazelcasttestsupport java at com hazelcast test hazelcasttestsupport asserttrueeventually hazelcasttestsupport java at com hazelcast partition partitionmigrationlistenertest eventcollectingmigrationlistener awaiteventpackscomplete partitionmigrationlistenertest java at com hazelcast partition partitionmigrationlistenertest eventcollectingmigrationlistener ensureandgeteventpacks partitionmigrationlistenertest java at com hazelcast partition partitionmigrationlistenertest testmigrationstats afterpartitionslost when no migration partitionmigrationlistenertest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at java util concurrent futuretask run futuretask java at java lang thread run thread java | 0 |
92,452 | 18,869,526,185 | IssuesEvent | 2021-11-13 00:37:03 | nopSolutions/nopCommerce | https://api.github.com/repos/nopSolutions/nopCommerce | closed | Add javascript events | refactoring / source code | We started working on Google Analytics Enhanced ecommerce plugin for nopCommerce.
Unfortunately nopCommerce is lacking any javascript events for any user actions. For example there is no way to know when a product has been added to the cart.
It would be nice if you raise javascript events like "productAddedToCart" etc. for the actions that happen via Ajax.
For example when you add a product via Ajax it would be nice to have something like this:
$.event.trigger({ type: "nopAjaxCartProductAddedToCartEvent", productId: productId, quantity: addToCartProductQuantity });
This way we can subscribe to such events and submit the data to Google Anaytics for example.
Unfortunately the Remove from Cart does not use Ajax, so there is no way this to be done for now.
Source: http://www.nopcommerce.com/boards/t/38391/suggestions-for-the-upcoming-version-370.aspx?p=2#156114
also see http://www.nopcommerce.com/boards/t/43297/improvements-before-380-release.aspx#171728
also see http://www.nopcommerce.com/boards/t/39325/add-javascript-events-in-the-payment-flow.aspx
P.S. We already have such event defined in \Views\Product_ProductAttributes.cshtml
| 1.0 | Add javascript events - We started working on Google Analytics Enhanced ecommerce plugin for nopCommerce.
Unfortunately nopCommerce is lacking any javascript events for any user actions. For example there is no way to know when a product has been added to the cart.
It would be nice if you raise javascript events like "productAddedToCart" etc. for the actions that happen via Ajax.
For example when you add a product via Ajax it would be nice to have something like this:
$.event.trigger({ type: "nopAjaxCartProductAddedToCartEvent", productId: productId, quantity: addToCartProductQuantity });
This way we can subscribe to such events and submit the data to Google Anaytics for example.
Unfortunately the Remove from Cart does not use Ajax, so there is no way this to be done for now.
Source: http://www.nopcommerce.com/boards/t/38391/suggestions-for-the-upcoming-version-370.aspx?p=2#156114
also see http://www.nopcommerce.com/boards/t/43297/improvements-before-380-release.aspx#171728
also see http://www.nopcommerce.com/boards/t/39325/add-javascript-events-in-the-payment-flow.aspx
P.S. We already have such event defined in \Views\Product_ProductAttributes.cshtml
| non_priority | add javascript events we started working on google analytics enhanced ecommerce plugin for nopcommerce unfortunately nopcommerce is lacking any javascript events for any user actions for example there is no way to know when a product has been added to the cart it would be nice if you raise javascript events like productaddedtocart etc for the actions that happen via ajax for example when you add a product via ajax it would be nice to have something like this event trigger type nopajaxcartproductaddedtocartevent productid productid quantity addtocartproductquantity this way we can subscribe to such events and submit the data to google anaytics for example unfortunately the remove from cart does not use ajax so there is no way this to be done for now source also see also see p s we already have such event defined in views product productattributes cshtml | 0 |
230,194 | 17,605,605,552 | IssuesEvent | 2021-08-17 16:39:37 | microsoft/vscode-docker | https://api.github.com/repos/microsoft/vscode-docker | closed | SSH tunnel setup doesn't work in PowerShell | documentation | In ["Connect to remote Docker over SSH/Set up SSH Tunneling"](https://code.visualstudio.com/docs/containers/ssh#_set-up-ssh-tunneling) it is advised to run `sc config ssh-agent start=auto`, but it only works in old `cmd` terminal. PowerShell will complain:
```
> sc config ssh-agent start=auto
Set-Content : A positional parameter cannot be found that accepts argument 'start=auto'.
At line:1 char:1
```
Please update the documentation with correct steps:
```
Set-Service ssh-agent -StartupType "Automatic"
Start-Service ssh-agent
``` | 1.0 | SSH tunnel setup doesn't work in PowerShell - In ["Connect to remote Docker over SSH/Set up SSH Tunneling"](https://code.visualstudio.com/docs/containers/ssh#_set-up-ssh-tunneling) it is advised to run `sc config ssh-agent start=auto`, but it only works in old `cmd` terminal. PowerShell will complain:
```
> sc config ssh-agent start=auto
Set-Content : A positional parameter cannot be found that accepts argument 'start=auto'.
At line:1 char:1
```
Please update the documentation with correct steps:
```
Set-Service ssh-agent -StartupType "Automatic"
Start-Service ssh-agent
``` | non_priority | ssh tunnel setup doesn t work in powershell in it is advised to run sc config ssh agent start auto but it only works in old cmd terminal powershell will complain sc config ssh agent start auto set content a positional parameter cannot be found that accepts argument start auto at line char please update the documentation with correct steps set service ssh agent startuptype automatic start service ssh agent | 0 |
53,529 | 13,261,827,449 | IssuesEvent | 2020-08-20 20:36:16 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | CVMFS - doc builds are choking on the lack of `napoleon` (Trac #1608) | Migrated from Trac defect infrastructure | david- could you add napoleon, or update sphinx?
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1608">https://code.icecube.wisc.edu/projects/icecube/ticket/1608</a>, reported by negaand owned by david.schultz</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-03-29T19:54:11",
"_ts": "1459281251636355",
"description": "david- could you add napoleon, or update sphinx?",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"time": "2016-03-28T21:20:33",
"component": "infrastructure",
"summary": "CVMFS - doc builds are choking on the lack of `napoleon`",
"priority": "normal",
"keywords": "cvmfs sphinx documentation",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
</p>
</details>
| 1.0 | CVMFS - doc builds are choking on the lack of `napoleon` (Trac #1608) - david- could you add napoleon, or update sphinx?
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1608">https://code.icecube.wisc.edu/projects/icecube/ticket/1608</a>, reported by negaand owned by david.schultz</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-03-29T19:54:11",
"_ts": "1459281251636355",
"description": "david- could you add napoleon, or update sphinx?",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"time": "2016-03-28T21:20:33",
"component": "infrastructure",
"summary": "CVMFS - doc builds are choking on the lack of `napoleon`",
"priority": "normal",
"keywords": "cvmfs sphinx documentation",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
</p>
</details>
| non_priority | cvmfs doc builds are choking on the lack of napoleon trac david could you add napoleon or update sphinx migrated from json status closed changetime ts description david could you add napoleon or update sphinx reporter nega cc resolution fixed time component infrastructure summary cvmfs doc builds are choking on the lack of napoleon priority normal keywords cvmfs sphinx documentation milestone owner david schultz type defect | 0 |
10,472 | 4,074,264,334 | IssuesEvent | 2016-05-28 09:51:53 | SleepyTrousers/EnderIO | https://api.github.com/repos/SleepyTrousers/EnderIO | closed | TODO: Rename changed GUI textures before release | Code Complete Will Implement | otherwise we'll get many reports like this:
 | 1.0 | TODO: Rename changed GUI textures before release - otherwise we'll get many reports like this:
 | non_priority | todo rename changed gui textures before release otherwise we ll get many reports like this | 0 |
2,142 | 4,460,926,581 | IssuesEvent | 2016-08-24 02:11:11 | diamondinrough/app | https://api.github.com/repos/diamondinrough/app | closed | Create basic hybrid app package | dev requirement | 1. create the basic code base for client side app, upload it to github
2. connect to the server side to make sure the client side and server side integration works.
Please document and share out the related staffs. | 1.0 | Create basic hybrid app package - 1. create the basic code base for client side app, upload it to github
2. connect to the server side to make sure the client side and server side integration works.
Please document and share out the related staffs. | non_priority | create basic hybrid app package create the basic code base for client side app upload it to github connect to the server side to make sure the client side and server side integration works please document and share out the related staffs | 0 |
101,748 | 16,528,033,422 | IssuesEvent | 2021-05-26 23:32:10 | alpersonalwebsite/cards | https://api.github.com/repos/alpersonalwebsite/cards | opened | CVE-2020-28500 (Medium) detected in lodash-4.17.11.tgz | security vulnerability | ## CVE-2020-28500 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.11.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p>
<p>Path to dependency file: cards/package.json</p>
<p>Path to vulnerable library: cards/node_modules/lodash</p>
<p>
Dependency Hierarchy:
- react-native-0.62.0.tgz (Root Library)
- cli-4.14.0.tgz
- inquirer-3.3.0.tgz
- :x: **lodash-4.17.11.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/alpersonalwebsite/cards/commit/959053f33fb963a3ec04cdb4be3c6f705f0312c6">959053f33fb963a3ec04cdb4be3c6f705f0312c6</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500>CVE-2020-28500</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution: lodash-4.17.21</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-28500 (Medium) detected in lodash-4.17.11.tgz - ## CVE-2020-28500 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.11.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p>
<p>Path to dependency file: cards/package.json</p>
<p>Path to vulnerable library: cards/node_modules/lodash</p>
<p>
Dependency Hierarchy:
- react-native-0.62.0.tgz (Root Library)
- cli-4.14.0.tgz
- inquirer-3.3.0.tgz
- :x: **lodash-4.17.11.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/alpersonalwebsite/cards/commit/959053f33fb963a3ec04cdb4be3c6f705f0312c6">959053f33fb963a3ec04cdb4be3c6f705f0312c6</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500>CVE-2020-28500</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution: lodash-4.17.21</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in lodash tgz cve medium severity vulnerability vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file cards package json path to vulnerable library cards node modules lodash dependency hierarchy react native tgz root library cli tgz inquirer tgz x lodash tgz vulnerable library found in head commit a href vulnerability details lodash versions prior to are vulnerable to regular expression denial of service redos via the tonumber trim and trimend functions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash step up your open source security game with whitesource | 0 |
32,688 | 4,786,530,536 | IssuesEvent | 2016-10-29 13:42:13 | infiniteautomation/ma-core-public | https://api.github.com/repos/infiniteautomation/ma-core-public | closed | Watchlist - Chart legend only displays names | Bug Ready for Testing | It was observed [in the forum](http://infiniteautomation.com/forum/topic/2441/watchlist-chart-only-shows-names-in-legend-without-devicenames) that somewhere in 2.5 or after the watchlist chart's legend ceased displaying the device name of the point in the legend, and only displays the name. | 1.0 | Watchlist - Chart legend only displays names - It was observed [in the forum](http://infiniteautomation.com/forum/topic/2441/watchlist-chart-only-shows-names-in-legend-without-devicenames) that somewhere in 2.5 or after the watchlist chart's legend ceased displaying the device name of the point in the legend, and only displays the name. | non_priority | watchlist chart legend only displays names it was observed that somewhere in or after the watchlist chart s legend ceased displaying the device name of the point in the legend and only displays the name | 0 |
324,673 | 27,814,729,344 | IssuesEvent | 2023-03-18 14:48:51 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | kv/kvserver: TestReplicateQueueExpirationLeasesOnly failed | C-test-failure O-robot branch-master T-kv-replication | kv/kvserver.TestReplicateQueueExpirationLeasesOnly [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/9122901?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/9122901?buildTab=artifacts#/) on master @ [0a72a49eb6e4ce5823b53d02125441c3da2e94f5](https://github.com/cockroachdb/cockroach/commits/0a72a49eb6e4ce5823b53d02125441c3da2e94f5):
```
=== RUN TestReplicateQueueExpirationLeasesOnly
test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/33e1d369c27b9c01b2b6009c561815a3/logTestReplicateQueueExpirationLeasesOnly2320733262
test_log_scope.go:79: use -show-logs to present logs inline
replicate_queue_test.go:2406: initial: epochLeases=316 expLeases=2
replicate_queue_test.go:2411:
Error Trace: github.com/cockroachdb/cockroach/pkg/kv/kvserver_test/pkg/kv/kvserver/replicate_queue_test.go:2411
Error: Condition never satisfied
Test: TestReplicateQueueExpirationLeasesOnly
panic.go:522: -- test log scope end --
--- FAIL: TestReplicateQueueExpirationLeasesOnly (319.95s)
```
<p>Parameters: <code>TAGS=bazel,gss,deadlock</code>
</p>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/kv
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestReplicateQueueExpirationLeasesOnly.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-25608 | 1.0 | kv/kvserver: TestReplicateQueueExpirationLeasesOnly failed - kv/kvserver.TestReplicateQueueExpirationLeasesOnly [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/9122901?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/9122901?buildTab=artifacts#/) on master @ [0a72a49eb6e4ce5823b53d02125441c3da2e94f5](https://github.com/cockroachdb/cockroach/commits/0a72a49eb6e4ce5823b53d02125441c3da2e94f5):
```
=== RUN TestReplicateQueueExpirationLeasesOnly
test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/33e1d369c27b9c01b2b6009c561815a3/logTestReplicateQueueExpirationLeasesOnly2320733262
test_log_scope.go:79: use -show-logs to present logs inline
replicate_queue_test.go:2406: initial: epochLeases=316 expLeases=2
replicate_queue_test.go:2411:
Error Trace: github.com/cockroachdb/cockroach/pkg/kv/kvserver_test/pkg/kv/kvserver/replicate_queue_test.go:2411
Error: Condition never satisfied
Test: TestReplicateQueueExpirationLeasesOnly
panic.go:522: -- test log scope end --
--- FAIL: TestReplicateQueueExpirationLeasesOnly (319.95s)
```
<p>Parameters: <code>TAGS=bazel,gss,deadlock</code>
</p>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/kv
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestReplicateQueueExpirationLeasesOnly.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-25608 | non_priority | kv kvserver testreplicatequeueexpirationleasesonly failed kv kvserver testreplicatequeueexpirationleasesonly with on master run testreplicatequeueexpirationleasesonly test log scope go test logs captured to artifacts tmp tmp test log scope go use show logs to present logs inline replicate queue test go initial epochleases expleases replicate queue test go error trace github com cockroachdb cockroach pkg kv kvserver test pkg kv kvserver replicate queue test go error condition never satisfied test testreplicatequeueexpirationleasesonly panic go test log scope end fail testreplicatequeueexpirationleasesonly parameters tags bazel gss deadlock help see also cc cockroachdb kv jira issue crdb | 0 |
62,301 | 7,569,918,375 | IssuesEvent | 2018-04-23 07:12:14 | akvo/akvo-web | https://api.github.com/repos/akvo/akvo-web | closed | All RSR Project Updates feed - images not optimised | 1 - Content 1 - Design 2 - Maintenance | I see the RSR update feed continues to render small thumbnail images at full resolution they were downloaded at (e.g. 1.5 MB pic for one tiny pic I saw just now). Is there any likelihood of a fix for this to serve images at much smaller file sizes? Aware that this page must perform very slowly in some of the countries in which Akvo is active (and where it is marketing / selling its tools).
http://akvodev.org/seeithappen/all-rsr-project-updates
| 1.0 | All RSR Project Updates feed - images not optimised - I see the RSR update feed continues to render small thumbnail images at full resolution they were downloaded at (e.g. 1.5 MB pic for one tiny pic I saw just now). Is there any likelihood of a fix for this to serve images at much smaller file sizes? Aware that this page must perform very slowly in some of the countries in which Akvo is active (and where it is marketing / selling its tools).
http://akvodev.org/seeithappen/all-rsr-project-updates
| non_priority | all rsr project updates feed images not optimised i see the rsr update feed continues to render small thumbnail images at full resolution they were downloaded at e g mb pic for one tiny pic i saw just now is there any likelihood of a fix for this to serve images at much smaller file sizes aware that this page must perform very slowly in some of the countries in which akvo is active and where it is marketing selling its tools | 0 |
19,062 | 26,490,364,383 | IssuesEvent | 2023-01-17 22:00:43 | storybookjs/storybook | https://api.github.com/repos/storybookjs/storybook | closed | Storybook fails to load story when generated code from @openapitools/openapi-generator-cli is used | question / support help wanted compatibility with other tools | **Describe the bug**
Storybook fails to load a story as soon as generated code from @openapitools/openapi-generator-cli is used.
The context is a NX monorepo but there is no issue when self-written code is used. We could not find what exactly breaks storybook.
**To Reproduce**
Steps to reproduce the behavior:
I created a minimal example: https://github.com/mgansler/openapi-storybook
1. Clone Repo
2. Install dependencies (`yarn install`)
3. Start Storybook: `nx run react-app:storybook`
4. http://localhost:4400/?path=/story/app--my-app-story should show a story
5. Remove usage of generated code: remove line 9 from `app.tsx` https://github.com/mgansler/openapi-storybook/blob/38517464b39eb781b825bc6c0fc10bf923a4589f/apps/react-app/src/app/app.tsx#L9
6. Story shows up
**Expected behavior**
Story should show up when generated code is used.
**Code snippets**
https://github.com/mgansler/openapi-storybook
**System**
Environment Info:
System:
OS: macOS 10.15.7
CPU: (8) x64 Intel(R) Core(TM) i7-7820HQ CPU @ 2.90GHz
Binaries:
Node: 12.19.0 - ~/.nvm/versions/node/v12.19.0/bin/node
Yarn: 1.22.10 - ~/.nvm/versions/node/v12.19.0/bin/yarn
npm: 7.5.2 - ~/.nvm/versions/node/v12.19.0/bin/npm
Browsers:
Chrome: 89.0.4389.90
Firefox: 87.0
Safari: 14.0.3
**Additional context**
I did not try the code "in production" in this example but in our project (which I cannot share) the App itself works without any issues. Only storybook fails to show the story. | True | Storybook fails to load story when generated code from @openapitools/openapi-generator-cli is used - **Describe the bug**
Storybook fails to load a story as soon as generated code from @openapitools/openapi-generator-cli is used.
The context is a NX monorepo but there is no issue when self-written code is used. We could not find what exactly breaks storybook.
**To Reproduce**
Steps to reproduce the behavior:
I created a minimal example: https://github.com/mgansler/openapi-storybook
1. Clone Repo
2. Install dependencies (`yarn install`)
3. Start Storybook: `nx run react-app:storybook`
4. http://localhost:4400/?path=/story/app--my-app-story should show a story
5. Remove usage of generated code: remove line 9 from `app.tsx` https://github.com/mgansler/openapi-storybook/blob/38517464b39eb781b825bc6c0fc10bf923a4589f/apps/react-app/src/app/app.tsx#L9
6. Story shows up
**Expected behavior**
Story should show up when generated code is used.
**Code snippets**
https://github.com/mgansler/openapi-storybook
**System**
Environment Info:
System:
OS: macOS 10.15.7
CPU: (8) x64 Intel(R) Core(TM) i7-7820HQ CPU @ 2.90GHz
Binaries:
Node: 12.19.0 - ~/.nvm/versions/node/v12.19.0/bin/node
Yarn: 1.22.10 - ~/.nvm/versions/node/v12.19.0/bin/yarn
npm: 7.5.2 - ~/.nvm/versions/node/v12.19.0/bin/npm
Browsers:
Chrome: 89.0.4389.90
Firefox: 87.0
Safari: 14.0.3
**Additional context**
I did not try the code "in production" in this example but in our project (which I cannot share) the App itself works without any issues. Only storybook fails to show the story. | non_priority | storybook fails to load story when generated code from openapitools openapi generator cli is used describe the bug storybook fails to load a story as soon as generated code from openapitools openapi generator cli is used the context is a nx monorepo but there is no issue when self written code is used we could not find what exactly breaks storybook to reproduce steps to reproduce the behavior i created a minimal example clone repo install dependencies yarn install start storybook nx run react app storybook should show a story remove usage of generated code remove line from app tsx story shows up expected behavior story should show up when generated code is used code snippets system environment info system os macos cpu intel r core tm cpu binaries node nvm versions node bin node yarn nvm versions node bin yarn npm nvm versions node bin npm browsers chrome firefox safari additional context i did not try the code in production in this example but in our project which i cannot share the app itself works without any issues only storybook fails to show the story | 0 |
20,974 | 11,033,665,391 | IssuesEvent | 2019-12-06 23:55:15 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | opt: common 5-way virtual table join takes 45ms to plan | C-performance | The following simplified query is used frequently by ActiveRecord. It currently takes 30-45ms to plan, which feels excessive, especially since the query is run so frequently. I've simplified the query to remove a bunch of extra cruft that makes it harder to read.
```
root@:26257/activerecord_unittest> explain(opt) SELECT
attr.attname,
nsp.nspname,
def.adbin
FROM
pg_class AS t
JOIN pg_attribute AS attr ON t.oid = attrelid
JOIN pg_attrdef AS def ON adrelid = attrelid AND adnum = attnum
JOIN pg_constraint AS cons ON conrelid = adrelid AND adnum = conkey[1]
JOIN pg_namespace AS nsp ON t.relnamespace = nsp.oid
WHERE
t.oid = 3;
text
+-----------------------------------------------------------------------------------+
project
└── inner-join (hash)
├── virtual-scan pg_namespace
├── inner-join (hash)
│ ├── project
│ │ ├── virtual-scan pg_constraint
│ │ └── projections
│ │ └── conkey[1]
│ ├── inner-join (hash)
│ │ ├── select
│ │ │ ├── virtual-scan pg_class@pg_class_oid_idx
│ │ │ │ └── constraint: /1: [/3 - /3]
│ │ │ └── filters
│ │ │ └── pg_class.oid = 3
│ │ ├── inner-join (hash)
│ │ │ ├── virtual-scan pg_attrdef
│ │ │ ├── select
│ │ │ │ ├── virtual-scan pg_attribute@pg_attribute_attrelid_idx
│ │ │ │ │ └── constraint: /29: [/3 - /3]
│ │ │ │ └── filters
│ │ │ │ └── attrelid = 3
│ │ │ └── filters
│ │ │ ├── adrelid = attrelid
│ │ │ └── adnum = attnum
│ │ └── filters
│ │ └── pg_class.oid = attrelid
│ └── filters
│ ├── conrelid = adrelid
│ └── adnum = column81
└── filters
└── relnamespace = pg_namespace.oid
(31 rows)
Time: 50.133ms
```
Is there anything to be done about this? For comparison, Postgres plans the same query in 1-3ms. | True | opt: common 5-way virtual table join takes 45ms to plan - The following simplified query is used frequently by ActiveRecord. It currently takes 30-45ms to plan, which feels excessive, especially since the query is run so frequently. I've simplified the query to remove a bunch of extra cruft that makes it harder to read.
```
root@:26257/activerecord_unittest> explain(opt) SELECT
attr.attname,
nsp.nspname,
def.adbin
FROM
pg_class AS t
JOIN pg_attribute AS attr ON t.oid = attrelid
JOIN pg_attrdef AS def ON adrelid = attrelid AND adnum = attnum
JOIN pg_constraint AS cons ON conrelid = adrelid AND adnum = conkey[1]
JOIN pg_namespace AS nsp ON t.relnamespace = nsp.oid
WHERE
t.oid = 3;
text
+-----------------------------------------------------------------------------------+
project
└── inner-join (hash)
├── virtual-scan pg_namespace
├── inner-join (hash)
│ ├── project
│ │ ├── virtual-scan pg_constraint
│ │ └── projections
│ │ └── conkey[1]
│ ├── inner-join (hash)
│ │ ├── select
│ │ │ ├── virtual-scan pg_class@pg_class_oid_idx
│ │ │ │ └── constraint: /1: [/3 - /3]
│ │ │ └── filters
│ │ │ └── pg_class.oid = 3
│ │ ├── inner-join (hash)
│ │ │ ├── virtual-scan pg_attrdef
│ │ │ ├── select
│ │ │ │ ├── virtual-scan pg_attribute@pg_attribute_attrelid_idx
│ │ │ │ │ └── constraint: /29: [/3 - /3]
│ │ │ │ └── filters
│ │ │ │ └── attrelid = 3
│ │ │ └── filters
│ │ │ ├── adrelid = attrelid
│ │ │ └── adnum = attnum
│ │ └── filters
│ │ └── pg_class.oid = attrelid
│ └── filters
│ ├── conrelid = adrelid
│ └── adnum = column81
└── filters
└── relnamespace = pg_namespace.oid
(31 rows)
Time: 50.133ms
```
Is there anything to be done about this? For comparison, Postgres plans the same query in 1-3ms. | non_priority | opt common way virtual table join takes to plan the following simplified query is used frequently by activerecord it currently takes to plan which feels excessive especially since the query is run so frequently i ve simplified the query to remove a bunch of extra cruft that makes it harder to read root activerecord unittest explain opt select attr attname nsp nspname def adbin from pg class as t join pg attribute as attr on t oid attrelid join pg attrdef as def on adrelid attrelid and adnum attnum join pg constraint as cons on conrelid adrelid and adnum conkey join pg namespace as nsp on t relnamespace nsp oid where t oid text project └── inner join hash ├── virtual scan pg namespace ├── inner join hash │ ├── project │ │ ├── virtual scan pg constraint │ │ └── projections │ │ └── conkey │ ├── inner join hash │ │ ├── select │ │ │ ├── virtual scan pg class pg class oid idx │ │ │ │ └── constraint │ │ │ └── filters │ │ │ └── pg class oid │ │ ├── inner join hash │ │ │ ├── virtual scan pg attrdef │ │ │ ├── select │ │ │ │ ├── virtual scan pg attribute pg attribute attrelid idx │ │ │ │ │ └── constraint │ │ │ │ └── filters │ │ │ │ └── attrelid │ │ │ └── filters │ │ │ ├── adrelid attrelid │ │ │ └── adnum attnum │ │ └── filters │ │ └── pg class oid attrelid │ └── filters │ ├── conrelid adrelid │ └── adnum └── filters └── relnamespace pg namespace oid rows time is there anything to be done about this for comparison postgres plans the same query in | 0 |
36,624 | 5,074,641,476 | IssuesEvent | 2016-12-27 15:25:32 | MajkiIT/polish-ads-filter | https://api.github.com/repos/MajkiIT/polish-ads-filter | closed | 300polityka.pl | reguły gotowe/testowanie reklama | Boxy po reklamach:

identyfikowane jako:
```
###text-2
```

```
###text-3
``` | 1.0 | 300polityka.pl - Boxy po reklamach:

identyfikowane jako:
```
###text-2
```

```
###text-3
``` | non_priority | pl boxy po reklamach identyfikowane jako text text | 0 |
201,409 | 15,802,273,604 | IssuesEvent | 2021-04-03 08:56:01 | chenling1022/ped | https://api.github.com/repos/chenling1022/ped | opened | Brackets in documentation | severity.VeryLow type.DocumentationBug | No details provided.

In command description, brackets are used, however, brackets are not used in the real command examples.
i.e.
view t/E [a/ACTIVITY_NAME] [date/DD-MM-YYYY]
<!--session: 1617437413881-36cf2e4d-b336-4d21-91df-8378b3b68f93--> | 1.0 | Brackets in documentation - No details provided.

In command description, brackets are used, however, brackets are not used in the real command examples.
i.e.
view t/E [a/ACTIVITY_NAME] [date/DD-MM-YYYY]
<!--session: 1617437413881-36cf2e4d-b336-4d21-91df-8378b3b68f93--> | non_priority | brackets in documentation no details provided in command description brackets are used however brackets are not used in the real command examples i e view t e | 0 |
34,096 | 6,289,965,642 | IssuesEvent | 2017-07-19 20:26:34 | Naoghuman/lib-logger | https://api.github.com/repos/Naoghuman/lib-logger | closed | [JavaDoc] Use {@code xy} instead the tag <code> in JavaDoc. | documentation refactoring | [JavaDoc] Use {@code xy} instead the tag <code> in JavaDoc. | 1.0 | [JavaDoc] Use {@code xy} instead the tag <code> in JavaDoc. - [JavaDoc] Use {@code xy} instead the tag <code> in JavaDoc. | non_priority | use code xy instead the tag in javadoc use code xy instead the tag lt code gt in javadoc | 0 |
73,802 | 7,359,079,826 | IssuesEvent | 2018-03-10 01:52:23 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | sql/logictest: TestParallel failed under stress | Robot test-failure | SHA: https://github.com/cockroachdb/cockroach/commits/ea5f8b22ba951799b76c84502aaab4456fe00a46
Parameters:
```
TAGS=
GOFLAGS=-race
```
Stress build found a failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=552372&tab=buildLog
```
I180308 09:27:49.787154 4678 storage/replica_raftstorage.go:732 [n3,s3,r9/?:/Table/1{2-3}] applied preemptive snapshot in 8ms [clear=0ms batch=0ms entries=7ms commit=0ms]
I180308 09:27:49.902388 835 testutils/testcluster/testcluster.go:534 [n1,s1] has 1 underreplicated ranges
W180308 09:27:49.965287 1035 storage/store.go:3866 [n1,s1] handle raft ready: 1.0s [processed=0]
I180308 09:27:49.992899 4679 storage/replica.go:3209 [n1,s1,r3/1:/System/NodeLiveness{-Max}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n4,s4):2 (n2,s2):3] next=4
I180308 09:27:49.998836 4659 storage/replica_command.go:1760 [replicate,n2,s2,r9/2:/Table/1{2-3}] change replicas (ADD_REPLICA (n3,s3):4): read existing descriptor r9:/Table/1{2-3} [(n1,s1):1, (n2,s2):2, (n5,s5):3, next=4]
W180308 09:27:50.033177 1806 storage/node_liveness.go:501 [n3,hb] slow heartbeat took 1.1s
I180308 09:27:50.226728 835 testutils/testcluster/testcluster.go:534 [n1,s1] has 1 underreplicated ranges
I180308 09:27:50.327811 4695 storage/replica_raftstorage.go:520 [replicate,n1,s1,r10/1:/Table/1{3-4}] generated preemptive snapshot ae1bb9ed at index 153
I180308 09:27:50.408223 4615 storage/store.go:2350 [replicaGC,n1,s1,r16/1:/Table/{19-20}] removing replica
I180308 09:27:50.482741 4695 storage/store.go:3745 [replicate,n1,s1,r10/1:/Table/1{3-4}] streamed snapshot to (n5,s5):?: kv pairs: 407, log entries: 51, rate-limit: 2.0 MiB/sec, 154ms
I180308 09:27:50.524357 4615 storage/replica.go:836 [replicaGC,n1,s1,r16/1:/Table/{19-20}] removed 11 (0+11) keys in 40ms [clear=0ms commit=40ms]
I180308 09:27:50.738648 4697 storage/replica_raftstorage.go:726 [n5,s5,r10/?:{-}] applying preemptive snapshot at index 153 (id=ae1bb9ed, encoded size=81088, 1 rocksdb batches, 51 log entries)
I180308 09:27:50.786926 4653 storage/replica.go:3209 [n4,s4,r18/2:/Table/2{1-2}] proposing ADD_REPLICA((n3,s3):4): updated=[(n1,s1):1 (n4,s4):2 (n5,s5):3 (n3,s3):4] next=5
I180308 09:27:50.817425 1748 gossip/gossip.go:1306 [n3] node has connected to cluster via gossip
I180308 09:27:50.818926 1748 storage/stores.go:350 [n3] wrote 4 node addresses to persistent storage
I180308 09:27:50.857137 4684 storage/replica.go:3209 [n3,s3,r13/2:/Table/1{6-7}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n4,s4):4 (n3,s3):2 (n5,s5):3] next=5
I180308 09:27:51.086023 4697 storage/replica_raftstorage.go:732 [n5,s5,r10/?:/Table/1{3-4}] applied preemptive snapshot in 347ms [clear=0ms batch=0ms entries=187ms commit=1ms]
I180308 09:27:51.243460 4695 storage/replica_command.go:1760 [replicate,n1,s1,r10/1:/Table/1{3-4}] change replicas (ADD_REPLICA (n5,s5):4): read existing descriptor r10:/Table/1{3-4} [(n1,s1):1, (n3,s3):2, (n2,s2):3, next=4]
I180308 09:27:51.381230 4761 storage/replica_command.go:1760 [replicate,n4,s4,r18/2:/Table/2{1-2}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r18:/Table/2{1-2} [(n1,s1):1, (n4,s4):2, (n5,s5):3, (n3,s3):4, next=5]
I180308 09:27:51.509691 4735 storage/replica_raftstorage.go:520 [replicate,n3,s3,r6/2:/{System/tse-Table/System…}] generated preemptive snapshot 27aebd8b at index 30
I180308 09:27:51.563486 835 sql/logictest/parallel_test.go:213 Setting range split size: 32768
I180308 09:27:51.601539 4735 storage/store.go:3745 [replicate,n3,s3,r6/2:/{System/tse-Table/System…}] streamed snapshot to (n2,s2):?: kv pairs: 14, log entries: 20, rate-limit: 2.0 MiB/sec, 91ms
I180308 09:27:51.619862 1901 storage/store.go:3479 [n1,s1,r13/1:/Table/1{6-7}] added to replica GC queue (peer suggestion)
I180308 09:27:51.630538 4789 storage/replica_raftstorage.go:726 [n2,s2,r6/?:{-}] applying preemptive snapshot at index 30 (id=27aebd8b, encoded size=8130, 1 rocksdb batches, 20 log entries)
I180308 09:27:51.728873 4789 storage/replica_raftstorage.go:732 [n2,s2,r6/?:/{System/tse-Table/System…}] applied preemptive snapshot in 98ms [clear=0ms batch=0ms entries=94ms commit=3ms]
I180308 09:27:51.807803 4735 storage/replica_command.go:1760 [replicate,n3,s3,r6/2:/{System/tse-Table/System…}] change replicas (ADD_REPLICA (n2,s2):4): read existing descriptor r6:/{System/tse-Table/SystemConfigSpan/Start} [(n1,s1):1, (n3,s3):2, (n4,s4):3, next=4]
I180308 09:27:51.828942 4767 storage/replica_raftstorage.go:520 [raftsnapshot,n4,s4,r18/2:/Table/2{1-2}] generated Raft snapshot df1ffd91 at index 31
I180308 09:27:51.834992 4767 storage/store.go:3745 [raftsnapshot,n4,s4,r18/2:/Table/2{1-2}] streamed snapshot to (n3,s3):4: kv pairs: 18, log entries: 21, rate-limit: 8.0 MiB/sec, 5ms
I180308 09:27:51.837164 4834 storage/replica_raftstorage.go:726 [n3,s3,r18/4:{-}] applying Raft snapshot at index 31 (id=df1ffd91, encoded size=7716, 1 rocksdb batches, 21 log entries)
I180308 09:27:51.842351 4834 storage/replica_raftstorage.go:732 [n3,s3,r18/4:/Table/2{1-2}] applied Raft snapshot in 5ms [clear=0ms batch=0ms entries=3ms commit=0ms]
I180308 09:27:52.277701 4836 storage/store.go:2350 [replicaGC,n5,s5,r10/?:/Table/1{3-4}] removing replica
I180308 09:27:52.302073 4671 storage/store.go:2350 [replicaGC,n1,s1,r13/1:/Table/1{6-7}] removing replica
I180308 09:27:52.431403 4836 storage/replica.go:836 [replicaGC,n5,s5,r10/?:/Table/1{3-4}] removed 389 (379+10) keys in 153ms [clear=0ms commit=153ms]
I180308 09:27:52.549535 4671 storage/replica.go:836 [replicaGC,n1,s1,r13/1:/Table/1{6-7}] removed 12 (0+12) keys in 219ms [clear=0ms commit=219ms]
I180308 09:27:52.628009 4816 storage/replica.go:3209 [n2,s2,r9/2:/Table/1{2-3}] proposing ADD_REPLICA((n3,s3):4): updated=[(n1,s1):1 (n2,s2):2 (n5,s5):3 (n3,s3):4] next=5
I180308 09:27:52.699806 1055 server/status/runtime.go:219 [n1] runtime stats: 1.1 GiB RSS, 1117 goroutines, 76 MiB/16 MiB/119 MiB GO alloc/idle/total, 154 MiB/186 MiB CGO alloc/total, 862.81cgo/sec, 0.95/0.07 %(u/s)time, 0.05 %gc (5x)
I180308 09:27:52.842796 4890 storage/replica_command.go:1760 [replicate,n2,s2,r9/2:/Table/1{2-3}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r9:/Table/1{2-3} [(n1,s1):1, (n2,s2):2, (n5,s5):3, (n3,s3):4, next=5]
I180308 09:27:53.053227 4909 storage/replica.go:3209 [n3,s3,r6/2:/{System/tse-Table/System…}] proposing ADD_REPLICA((n2,s2):4): updated=[(n1,s1):1 (n3,s3):2 (n4,s4):3 (n2,s2):4] next=5
I180308 09:27:53.204274 4783 storage/replica_command.go:1760 [replicate,n3,s3,r6/2:/{System/tse-Table/System…}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r6:/{System/tse-Table/SystemConfigSpan/Start} [(n1,s1):1, (n3,s3):2, (n4,s4):3, (n2,s2):4, next=5]
I180308 09:27:53.385647 4933 storage/replica.go:3209 [n4,s4,r18/2:/Table/2{1-2}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n3,s3):4 (n4,s4):2 (n5,s5):3] next=5
I180308 09:27:53.654566 835 sql/logictest/parallel_test.go:227 Creating database
I180308 09:27:53.714521 1518 gossip/gossip.go:1306 [n2] node has connected to cluster via gossip
I180308 09:27:53.812099 1518 storage/stores.go:350 [n2] wrote 4 node addresses to persistent storage
I180308 09:27:53.949703 2370 storage/store.go:3479 [n1,s1,r18/1:/Table/2{1-2}] added to replica GC queue (peer suggestion)
I180308 09:27:53.950132 2370 storage/store.go:3479 [n1,s1,r18/1:/Table/2{1-2}] added to replica GC queue (peer suggestion)
I180308 09:27:54.187036 4918 storage/replica_command.go:838 [split,n1,s1,r5/1:/System/ts{d-e}] initiating a split of this range at key /System/tsd/cr.node.gossip.infos.sent/4/10s/2018-03-08T09:00:00Z [r21]
I180308 09:27:54.490089 4921 storage/replica_command.go:838 [split,n1,s1,r4/1:/System/{NodeLive…-tsd}] initiating a split of this range at key /System/StatusNode/2 [r22]
I180308 09:27:54.641807 4982 storage/replica_command.go:838 [split,n1,s1,r10/1:/Table/1{3-4}] initiating a split of this range at key /Table/13/1/2018-03-08T09:27:29.777287Z/329091808619495425 [r23]
I180308 09:27:54.767000 4695 storage/replica.go:3209 [n1,s1,r10/1:/Table/1{3-4}] proposing ADD_REPLICA((n5,s5):4): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3 (n5,s5):4] next=5
I180308 09:27:54.924274 4936 storage/store.go:2350 [replicaGC,n1,s1,r18/1:/Table/2{1-2}] removing replica
I180308 09:27:55.227444 4936 storage/replica.go:836 [replicaGC,n1,s1,r18/1:/Table/2{1-2}] removed 11 (0+11) keys in 51ms [clear=0ms commit=50ms]
W180308 09:27:55.283305 4861 storage/replica.go:3016 [n1,s1,r10/1:/Table/1{3-4}] context canceled before proposing: 1 HeartbeatTxn
I180308 09:27:55.373122 5042 storage/replica_command.go:838 [split,n1,s1,r10/1:/Table/1{3-4}] initiating a split of this range at key /Table/13/1/2018-03-08T09:27:29.777287Z/329091808619495425 [r24]
I180308 09:27:55.591922 5003 storage/replica.go:3209 [n2,s2,r9/2:/Table/1{2-3}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n3,s3):4 (n2,s2):2 (n5,s5):3] next=5
I180308 09:27:55.659550 1544 server/status/runtime.go:219 [n2] runtime stats: 1.1 GiB RSS, 1126 goroutines, 57 MiB/35 MiB/120 MiB GO alloc/idle/total, 155 MiB/187 MiB CGO alloc/total, 798.50cgo/sec, 0.95/0.07 %(u/s)time, 0.03 %gc (5x)
I180308 09:27:56.074863 2121 storage/store.go:3479 [n1,s1,r9/1:/Table/1{2-3}] added to replica GC queue (peer suggestion)
I180308 09:27:56.304086 4874 sql/event_log.go:121 [n1,client=127.0.0.1:43858,user=root] Event: "create_database", target: 50, info: {DatabaseName:test Statement:CREATE DATABASE test User:root}
I180308 09:27:56.307831 2121 storage/store.go:3479 [n1,s1,r9/1:/Table/1{2-3}] added to replica GC queue (peer suggestion)
I180308 09:27:56.399786 1635 storage/replica.go:3277 [n2,s2,r2/3:/System/{-NodeLive…}] not quiescing: 1 pending commands
I180308 09:27:56.412847 2121 storage/store.go:3479 [n1,s1,r9/1:/Table/1{2-3}] added to replica GC queue (peer suggestion)
I180308 09:27:56.413411 2121 storage/store.go:3479 [n1,s1,r9/1:/Table/1{2-3}] added to replica GC queue (peer suggestion)
I180308 09:27:56.429913 5110 storage/store.go:2350 [replicaGC,n1,s1,r9/1:/Table/1{2-3}] removing replica
I180308 09:27:56.448755 1797 server/status/runtime.go:219 [n3] runtime stats: 1.1 GiB RSS, 1135 goroutines, 66 MiB/26 MiB/120 MiB GO alloc/idle/total, 156 MiB/187 MiB CGO alloc/total, 773.55cgo/sec, 0.94/0.07 %(u/s)time, 0.03 %gc (5x)
I180308 09:27:56.688465 5110 storage/replica.go:836 [replicaGC,n1,s1,r9/1:/Table/1{2-3}] removed 71 (60+11) keys in 257ms [clear=0ms commit=257ms]
I180308 09:27:57.234316 5009 storage/replica_command.go:838 [split,n4,s4,r20/2:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r41]
I180308 09:27:57.715312 5007 storage/replica_raftstorage.go:520 [raftsnapshot,n1,s1,r10/1:/Table/1{3-4}] generated Raft snapshot 729bfb72 at index 173
W180308 09:27:57.734210 982 storage/store.go:3866 [n1,s1] handle raft ready: 0.6s [processed=0]
I180308 09:27:57.771678 2044 server/status/runtime.go:219 [n4] runtime stats: 1.2 GiB RSS, 1138 goroutines, 51 MiB/40 MiB/120 MiB GO alloc/idle/total, 156 MiB/187 MiB CGO alloc/total, 717.70cgo/sec, 0.95/0.07 %(u/s)time, 0.03 %gc (5x)
I180308 09:27:57.853130 5007 storage/store.go:3745 [raftsnapshot,n1,s1,r10/1:/Table/1{3-4}] streamed snapshot to (n5,s5):4: kv pairs: 456, log entries: 71, rate-limit: 8.0 MiB/sec, 104ms
I180308 09:27:57.904274 5196 storage/replica_raftstorage.go:726 [n5,s5,r10/4:{-}] applying Raft snapshot at index 173 (id=729bfb72, encoded size=103834, 1 rocksdb batches, 71 log entries)
W180308 09:27:58.307548 1553 storage/node_liveness.go:501 [n2,hb] slow heartbeat took 1.2s
I180308 09:27:58.890124 5216 storage/replica.go:3209 [n3,s3,r6/2:/{System/tse-Table/System…}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n2,s2):4 (n3,s3):2 (n4,s4):3] next=5
I180308 09:27:59.234979 2484 server/status/runtime.go:219 [n5] runtime stats: 1.2 GiB RSS, 1139 goroutines, 78 MiB/15 MiB/120 MiB GO alloc/idle/total, 156 MiB/188 MiB CGO alloc/total, 720.50cgo/sec, 0.94/0.06 %(u/s)time, 0.03 %gc (5x)
I180308 09:27:59.534800 1901 storage/store.go:3479 [n1,s1,r6/1:/{System/tse-Table/System…}] added to replica GC queue (peer suggestion)
I180308 09:27:59.617241 1901 storage/store.go:3479 [n1,s1,r6/1:/{System/tse-Table/System…}] added to replica GC queue (peer suggestion)
W180308 09:27:59.658775 1806 storage/node_liveness.go:501 [n3,hb] slow heartbeat took 1.7s
I180308 09:27:59.883988 835 sql/logictest/parallel_test.go:236 Test setup done
I180308 09:27:59.884248 835 sql/logictest/parallel_test.go:159 testdata/parallel_test/subquery_retry_multinode/test.yaml: run list 0: 0:setup
W180308 09:28:00.194433 1096 storage/node_liveness.go:501 [n1,hb] slow heartbeat took 1.4s
I180308 09:28:00.297625 5240 storage/store.go:2350 [replicaGC,n1,s1,r6/1:/{System/tse-Table/System…}] removing replica
I180308 09:28:00.329573 1028 storage/replica_proposal.go:201 [n1,s1,r4/1:/System/{NodeLive…-tsd}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0
I180308 09:28:00.379562 5248 storage/replica_command.go:838 [split,n1,s1,r22/1:/System/{StatusNo…-tsd}] initiating a split of this range at key /System/StatusNode/3 [r25]
I180308 09:28:00.448234 5240 storage/replica.go:836 [replicaGC,n1,s1,r6/1:/{System/tse-Table/System…}] removed 12 (0+12) keys in 91ms [clear=0ms commit=91ms]
W180308 09:28:00.707751 2147 storage/node_liveness.go:501 [n4,hb] slow heartbeat took 1.4s
W180308 09:28:00.731795 5352 storage/replica_backpressure.go:135 [n1,s1,r5/1:/System/ts{d-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.security.certificate.expiration.ca/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.security.certificate.expiration.node/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.connections.incoming/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.connections.refused/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.bytes.received/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.bytes.sent/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.infos.received/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.infos.sent/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.connections.outgoing/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.batches/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.batches.partial/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.rpc.sent/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.rpc.sent.local/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.rpc.sent.nextreplicaerror/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.errors.notleaseholder/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.requests.slow.distsender/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.aborts/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/1/10s/2018-03-08T09:00:00Z,/Min), ... 405 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/1/10s/2018-03-08T09:00:00Z,/Min)
I180308 09:28:01.257194 983 storage/replica_proposal.go:201 [n1,s1,r10/1:/Table/1{3-4}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.325864565,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.325864565,0
I180308 09:28:01.263525 5297 storage/replica_command.go:838 [split,n1,s1,r24/1:/Table/1{3/1/2018…-4}] initiating a split of this range at key /Table/13/1/2018-03-08T09:27:40.666166Z/329091846156877826 [r26]
I180308 09:28:01.312145 5196 storage/replica_raftstorage.go:732 [n5,s5,r10/4:/Table/1{3-4}] applied Raft snapshot in 3408ms [clear=0ms batch=0ms entries=3374ms commit=21ms]
I180308 09:28:01.454527 2220 storage/replica_proposal.go:201 [n4,s4,r20/2:/{Table/23-Max}] new range lease repl=(n4,s4):2 seq=3 start=1520501251.624682287,1 epo=1 pro=1520501255.402825609,0 following repl=(n4,s4):2 seq=3 start=1520501251.624682287,1 epo=1 pro=1520501255.402825609,0
I180308 09:28:01.482106 1005 storage/replica_proposal.go:201 [n1,s1,r5/1:/System/ts{d-e}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0
W180308 09:28:01.575951 5351 storage/replica.go:3016 [n1,s1,r10/1:/Table/13{-/1/2018-…}] context canceled before proposing: 1 HeartbeatTxn
I180308 09:28:01.658992 5386 storage/replica_command.go:838 [split,n1,s1,r21/1:/System/ts{d/cr.no…-e}] initiating a split of this range at key /System/tsd/cr.node.sql.bytesin/3/10s/2018-03-08T09:00:00Z [r27]
I180308 09:28:01.678252 5296 storage/replica_command.go:1760 [replicate,n1,s1,r10/1:/Table/13{-/1/2018-…}] change replicas (REMOVE_REPLICA (n3,s3):2): read existing descriptor r10:/Table/13{-/1/2018-03-08T09:27:29.777287Z/329091808619495425} [(n1,s1):1, (n3,s3):2, (n2,s2):3, (n5,s5):4, next=5]
I180308 09:28:01.932018 5423 storage/replica_raftstorage.go:520 [replicate,n4,s4,r41/2:/{Table/50-Max}] generated preemptive snapshot e7efcdd4 at index 10
I180308 09:28:01.954883 5491 storage/replica_raftstorage.go:520 [replicate,n3,s3,r17/2:/Table/2{0-1}] generated preemptive snapshot b7e6a194 at index 36
E180308 09:28:02.175447 5423 storage/queue.go:778 [replicate,n4,s4,r41/2:/{Table/50-Max}] raft status not initialized
I180308 09:28:02.188640 4874 sql/event_log.go:121 [n1,client=127.0.0.1:43858,user=root] Event: "create_table", target: 51, info: {TableName:test.public.t Statement:CREATE TABLE t (k INT, filler STRING) User:root}
--- progress: testdata/parallel_test/subquery_retry_multinode/setup: 1 statements/queries
I180308 09:28:02.231950 5276 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/setup: 1 statements/queries
I180308 09:28:02.309261 5491 storage/store.go:3745 [replicate,n3,s3,r17/2:/Table/2{0-1}] streamed snapshot to (n1,s1):?: kv pairs: 16, log entries: 26, rate-limit: 2.0 MiB/sec, 97ms
I180308 09:28:02.727426 5515 storage/replica_command.go:838 [split,n4,s4,r41/2:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r42]
I180308 09:28:02.767793 1979 gossip/gossip.go:1306 [n4] node has connected to cluster via gossip
I180308 09:28:02.791521 1055 server/status/runtime.go:219 [n1] runtime stats: 1.2 GiB RSS, 1166 goroutines, 65 MiB/26 MiB/120 MiB GO alloc/idle/total, 157 MiB/187 MiB CGO alloc/total, 1065.63cgo/sec, 0.94/0.07 %(u/s)time, 0.02 %gc (5x)
I180308 09:28:03.002229 5486 storage/replica_raftstorage.go:726 [n1,s1,r17/?:{-}] applying preemptive snapshot at index 36 (id=b7e6a194, encoded size=8224, 1 rocksdb batches, 26 log entries)
I180308 09:28:03.005301 1979 storage/stores.go:350 [n4] wrote 4 node addresses to persistent storage
I180308 09:28:03.154898 5486 storage/replica_raftstorage.go:732 [n1,s1,r17/?:/Table/2{0-1}] applied preemptive snapshot in 152ms [clear=0ms batch=0ms entries=127ms commit=12ms]
W180308 09:28:03.190672 5455 storage/replica_backpressure.go:135 [n2,s2,r21/2:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.security.certificate.expiration.ca/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.security.certificate.expiration.node/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.requests.slow.distsender/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.aborts/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), ... 379 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/2/10s/2018-03-08T09:00:00Z,/Min)
I180308 09:28:03.513617 5491 storage/replica_command.go:1760 [replicate,n3,s3,r17/2:/Table/2{0-1}] change replicas (ADD_REPLICA (n1,s1):5): read existing descriptor r17:/Table/2{0-1} [(n5,s5):4, (n3,s3):2, (n2,s2):3, next=5]
W180308 09:28:04.034472 5607 storage/replica_backpressure.go:135 [n1,s1,r21/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.security.certificate.expiration.ca/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.security.certificate.expiration.node/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.infos.sent/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.requests.slow.distsender/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.aborts/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/4/10s/2018-03-08T09:00:00Z,/Min), ... 380 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/4/10s/2018-03-08T09:00:00Z,/Min)
W180308 09:28:04.459926 5676 storage/replica_backpressure.go:135 [n1,s1,r21/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.security.certificate.expiration.ca/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.security.certificate.expiration.node/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.infos.sent/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.requests.slow.distsender/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.aborts/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/5/10s/2018-03-08T09:00:00Z,/Min), ... 380 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/5/10s/2018-03-08T09:00:00Z,/Min)
--- progress: testdata/parallel_test/subquery_retry_multinode/setup: 2 statements/queries
I180308 09:28:04.488626 5276 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/setup: 2 statements/queries
--- done: testdata/parallel_test/subquery_retry_multinode/setup with config : 2 tests, 0 failures
I180308 09:28:04.488940 5276 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/setup with config : 2 tests, 0 failures
I180308 09:28:04.494592 835 sql/logictest/parallel_test.go:159 testdata/parallel_test/subquery_retry_multinode/test.yaml: run list 1: 0:txn, 1:txn, 2:txn, 3:txn, 4:txn
I180308 09:28:04.692538 5642 storage/replica.go:3209 [n1,s1,r10/1:/Table/13{-/1/2018-…}] proposing REMOVE_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n5,s5):4 (n2,s2):3] next=5
I180308 09:28:05.177727 1003 storage/replica_proposal.go:201 [n1,s1,r22/1:/System/{StatusNo…-tsd}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0
I180308 09:28:05.184170 1972 storage/store.go:3479 [n3,s3,r10/2:/Table/13{-/1/2018-…}] added to replica GC queue (peer suggestion)
I180308 09:28:05.275031 5772 storage/replica_command.go:838 [split,n1,s1,r25/1:/System/{StatusNo…-tsd}] initiating a split of this range at key /System/StatusNode/4 [r28]
I180308 09:28:05.629470 1544 server/status/runtime.go:219 [n2] runtime stats: 1.2 GiB RSS, 1171 goroutines, 87 MiB/11 MiB/125 MiB GO alloc/idle/total, 157 MiB/189 MiB CGO alloc/total, 1173.83cgo/sec, 0.94/0.07 %(u/s)time, 0.02 %gc (4x)
I180308 09:28:05.634410 1827 storage/replica_proposal.go:201 [n3,s3,r4/2:/System/{NodeLive…-StatusNo…}] new range lease repl=(n3,s3):2 seq=4 start=1520501285.016867674,0 epo=1 pro=1520501285.016882538,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0
W180308 09:28:06.031522 5844 storage/replica.go:3016 [n1,s1,r22/1:/System/StatusNode/{2-3}] context canceled before proposing: 1 HeartbeatTxn
--- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries
I180308 09:28:06.562121 5731 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries
--- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures
I180308 09:28:06.562430 5731 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures
I180308 09:28:06.630349 1797 server/status/runtime.go:219 [n3] runtime stats: 1.2 GiB RSS, 1178 goroutines, 97 MiB/5.9 MiB/129 MiB GO alloc/idle/total, 158 MiB/189 MiB CGO alloc/total, 1136.76cgo/sec, 0.94/0.07 %(u/s)time, 0.03 %gc (5x)
W180308 09:28:06.862124 2493 storage/node_liveness.go:501 [n5,hb] slow heartbeat took 1.7s
W180308 09:28:07.341217 1553 storage/node_liveness.go:501 [n2,hb] slow heartbeat took 1.2s
I180308 09:28:07.775014 2044 server/status/runtime.go:219 [n4] runtime stats: 1.2 GiB RSS, 1174 goroutines, 75 MiB/24 MiB/129 MiB GO alloc/idle/total, 157 MiB/189 MiB CGO alloc/total, 1173.62cgo/sec, 0.94/0.07 %(u/s)time, 0.03 %gc (4x)
I180308 09:28:07.899393 993 storage/replica_proposal.go:201 [n1,s1,r21/1:/System/ts{d/cr.no…-e}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0
I180308 09:28:08.092040 6071 storage/replica_command.go:838 [split,n1,s1,r27/1:/System/ts{d/cr.no…-e}] initiating a split of this range at key /System/tsd/cr.node.sql.distsql.service.latency-p99.999/1/10s/2018-03-08T09:00:00Z [r29]
W180308 09:28:09.182946 1850 storage/store.go:3866 [n3,s3] handle raft ready: 0.7s [processed=0]
I180308 09:28:09.228630 2484 server/status/runtime.go:219 [n5] runtime stats: 1.3 GiB RSS, 1216 goroutines, 96 MiB/9.2 MiB/132 MiB GO alloc/idle/total, 160 MiB/191 MiB CGO alloc/total, 1213.47cgo/sec, 0.94/0.07 %(u/s)time, 0.03 %gc (5x)
W180308 09:28:09.431040 987 storage/store.go:3866 [n1,s1] handle raft ready: 0.6s [processed=0]
W180308 09:28:09.702153 1029 storage/store.go:3866 [n1,s1] handle raft ready: 0.5s [processed=0]
I180308 09:28:09.884861 6241 storage/replica_raftstorage.go:520 [replicate,n1,s1,r5/1:/System/tsd{-/cr.nod…}] generated preemptive snapshot c879f910 at index 68
W180308 09:28:09.920664 1096 storage/node_liveness.go:501 [n1,hb] slow heartbeat took 2.2s
I180308 09:28:09.926471 6051 storage/store.go:2350 [replicaGC,n1,s1,r17/?:/Table/2{0-1}] removing replica
I180308 09:28:10.164342 6051 storage/replica.go:836 [replicaGC,n1,s1,r17/?:/Table/2{0-1}] removed 11 (0+11) keys in 177ms [clear=0ms commit=177ms]
I180308 09:28:10.208745 6080 storage/replica_raftstorage.go:520 [replicate,n2,s2,r2/3:/System/{-NodeLive…}] generated preemptive snapshot c8cdd480 at index 34
I180308 09:28:10.482965 6080 storage/store.go:3745 [replicate,n2,s2,r2/3:/System/{-NodeLive…}] streamed snapshot to (n5,s5):?: kv pairs: 13, log entries: 24, rate-limit: 2.0 MiB/sec, 273ms
I180308 09:28:10.627902 6241 storage/store.go:3745 [replicate,n1,s1,r5/1:/System/tsd{-/cr.nod…}] streamed snapshot to (n4,s4):?: kv pairs: 143, log entries: 13, rate-limit: 2.0 MiB/sec, 677ms
I180308 09:28:10.693165 6287 storage/replica_raftstorage.go:726 [n5,s5,r2/?:{-}] applying preemptive snapshot at index 34 (id=c8cdd480, encoded size=56977, 1 rocksdb batches, 24 log entries)
I180308 09:28:10.790504 6345 storage/replica_raftstorage.go:726 [n4,s4,r5/?:{-}] applying preemptive snapshot at index 68 (id=c879f910, encoded size=36756, 1 rocksdb batches, 13 log entries)
W180308 09:28:10.795289 6380 storage/replica_backpressure.go:135 [n1,s1,r27/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.txn.aborts/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p90/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p75/3/10s/2018-03-08T09:00:00Z,/Min), ... 355 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/3/10s/2018-03-08T09:00:00Z,/Min)
W180308 09:28:10.836796 1826 storage/store.go:3866 [n3,s3] handle raft ready: 0.5s [processed=0]
W180308 09:28:10.963543 2147 storage/node_liveness.go:501 [n4,hb] slow heartbeat took 2.7s
W180308 09:28:11.405336 6410 storage/replica_backpressure.go:135 [n1,s1,r27/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.txn.aborts/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p90/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p75/2/10s/2018-03-08T09:00:00Z,/Min), ... 354 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/2/10s/2018-03-08T09:00:00Z,/Min)
I180308 09:28:11.779388 987 storage/replica_proposal.go:201 [n1,s1,r24/1:/Table/1{3/1/2018…-4}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.325864565,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.325864565,0
W180308 09:28:11.899345 2493 storage/node_liveness.go:501 [n5,hb] slow heartbeat took 2.2s
W180308 09:28:12.551481 1553 storage/node_liveness.go:501 [n2,hb] slow heartbeat took 1.9s
I180308 09:28:12.571934 6345 storage/replica_raftstorage.go:732 [n4,s4,r5/?:/System/tsd{-/cr.nod…}] applied preemptive snapshot in 1781ms [clear=0ms batch=0ms entries=1729ms commit=33ms]
I180308 09:28:12.630942 6241 storage/replica_command.go:1760 [replicate,n1,s1,r5/1:/System/tsd{-/cr.nod…}] change replicas (ADD_REPLICA (n4,s4):4): read existing descriptor r5:/System/tsd{-/cr.node.gossip.infos.sent/4/10s/2018-03-08T09:00:00Z} [(n1,s1):1, (n2,s2):2, (n5,s5):3, next=4]
W180308 09:28:12.654684 1806 storage/node_liveness.go:501 [n3,hb] slow heartbeat took 1.2s
I180308 09:28:12.667081 1055 server/status/runtime.go:219 [n1] runtime stats: 1.3 GiB RSS, 1182 goroutines, 78 MiB/23 MiB/132 MiB GO alloc/idle/total, 160 MiB/194 MiB CGO alloc/total, 934.67cgo/sec, 0.92/0.07 %(u/s)time, 0.02 %gc (4x)
I180308 09:28:13.552826 6287 storage/replica_raftstorage.go:732 [n5,s5,r2/?:/System/{-NodeLive…}] applied preemptive snapshot in 2833ms [clear=0ms batch=0ms entries=2609ms commit=164ms]
I180308 09:28:13.769924 6080 storage/replica_command.go:1760 [replicate,n2,s2,r2/3:/System/{-NodeLive…}] change replicas (ADD_REPLICA (n5,s5):4): read existing descriptor r2:/System/{-NodeLiveness} [(n1,s1):1, (n3,s3):2, (n2,s2):3, next=4]
W180308 09:28:13.950338 2147 storage/node_liveness.go:501 [n4,hb] slow heartbeat took 1.2s
I180308 09:28:14.047030 6643 storage/replica.go:3209 [n3,s3,r17/2:/Table/2{0-1}] proposing ADD_REPLICA((n1,s1):5): updated=[(n5,s5):4 (n3,s3):2 (n2,s2):3 (n1,s1):5] next=6
--- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries
I180308 09:28:14.991532 5735 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries
--- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures
I180308 09:28:14.991818 5735 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures
W180308 09:28:15.210739 2493 storage/node_liveness.go:501 [n5,hb] slow heartbeat took 1.1s
I180308 09:28:15.459366 2539 storage/replica_proposal.go:201 [n5,s5,r17/4:/Table/2{0-1}] new range lease repl=(n5,s5):4 seq=4 start=1520501294.765769961,0 epo=1 pro=1520501294.765803960,0 following repl=(n3,s3):2 seq=3 start=1520501251.624682287,1 epo=1 pro=1520501257.825367317,0
I180308 09:28:15.554730 6653 storage/replica_raftstorage.go:520 [raftsnapshot,n3,s3,r17/2:/Table/2{0-1}] generated Raft snapshot 901a1124 at index 49
I180308 09:28:15.697368 1544 server/status/runtime.go:219 [n2] runtime stats: 1.3 GiB RSS, 1181 goroutines, 73 MiB/35 MiB/138 MiB GO alloc/idle/total, 160 MiB/195 MiB CGO alloc/total, 840.48cgo/sec, 0.90/0.07 %(u/s)time, 0.02 %gc (4x)
I180308 09:28:15.764467 6653 storage/store.go:3745 [raftsnapshot,n3,s3,r17/2:/Table/2{0-1}] streamed snapshot to (n1,s1):5: kv pairs: 18, log entries: 39, rate-limit: 8.0 MiB/sec, 61ms
I180308 09:28:15.778517 6602 storage/replica_raftstorage.go:726 [n1,s1,r17/5:{-}] applying Raft snapshot at index 49 (id=901a1124, encoded size=12107, 1 rocksdb batches, 39 log entries)
I180308 09:28:16.177052 6521 storage/store.go:2350 [replicaGC,n4,s4,r5/?:/System/tsd{-/cr.nod…}] removing replica
I180308 09:28:16.294600 6521 storage/replica.go:836 [replicaGC,n4,s4,r5/?:/System/tsd{-/cr.nod…}] removed 139 (128+11) keys in 117ms [clear=0ms commit=116ms]
W180308 09:28:16.347616 1553 storage/node_liveness.go:501 [n2,hb] slow heartbeat took 1.3s
I180308 09:28:16.372076 994 storage/replica_proposal.go:201 [n1,s1,r25/1:/System/{StatusNo…-tsd}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0
I180308 09:28:16.421976 5631 storage/store.go:2350 [replicaGC,n3,s3,r10/2:/Table/13{-/1/2018-…}] removing replica
I180308 09:28:16.665522 1797 server/status/runtime.go:219 [n3] runtime stats: 1.3 GiB RSS, 1171 goroutines, 88 MiB/20 MiB/138 MiB GO alloc/idle/total, 162 MiB/196 MiB CGO alloc/total, 884.78cgo/sec, 0.89/0.07 %(u/s)time, 0.02 %gc (3x)
I180308 09:28:16.681548 6668 storage/replica_command.go:838 [split,n1,s1,r28/1:/System/{StatusNo…-tsd}] initiating a split of this range at key /System/StatusNode/5 [r30]
I180308 09:28:16.801652 6602 storage/replica_raftstorage.go:732 [n1,s1,r17/5:/Table/2{0-1}] applied Raft snapshot in 1003ms [clear=2ms batch=0ms entries=991ms commit=8ms]
I180308 09:28:16.806835 5631 storage/replica.go:836 [replicaGC,n3,s3,r10/2:/Table/13{-/1/2018-…}] removed 181 (169+12) keys in 384ms [clear=0ms commit=384ms]
W180308 09:28:17.412329 1806 storage/node_liveness.go:501 [n3,hb] slow heartbeat took 1.5s
I180308 09:28:17.506774 6629 storage/replica_consistency.go:129 [replica consistency checker,n1,s1,r21/1:/System/tsd/cr.node.{go…-sq…}] triggering stats recomputation to resolve delta of {ContainsEstimates:true LastUpdateNanos:1520501288989219219 IntentAge:0 GCBytesAge:0 LiveBytes:4092 LiveCount:0 KeyBytes:0 KeyCount:0 ValBytes:4092 ValCount:0 IntentBytes:0 IntentCount:0 SysBytes:0 SysCount:0}
I180308 09:28:17.510115 2214 storage/replica_proposal.go:201 [n4,s4,r41/2:/{Table/50-Max}] new range lease repl=(n4,s4):2 seq=3 start=1520501251.624682287,1 epo=1 pro=1520501255.402825609,0 following repl=(n4,s4):2 seq=3 start=1520501251.624682287,1 epo=1 pro=1520501255.402825609,0
I180308 09:28:17.785167 2044 server/status/runtime.go:219 [n4] runtime stats: 1.3 GiB RSS, 1172 goroutines, 62 MiB/44 MiB/138 MiB GO alloc/idle/total, 162 MiB/195 MiB CGO alloc/total, 981.39cgo/sec, 0.90/0.07 %(u/s)time, 0.02 %gc (4x)
I180308 09:28:17.789384 6779 storage/replica.go:3209 [n2,s2,r2/3:/System/{-NodeLive…}] proposing ADD_REPLICA((n5,s5):4): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3 (n5,s5):4] next=5
W180308 09:28:17.792842 6603 storage/replica.go:3016 [n4,s4,r41/2:/Table/5{0-1}] context canceled before proposing: 1 HeartbeatTxn
W180308 09:28:18.718662 1096 storage/node_liveness.go:501 [n1,hb] slow heartbeat took 2.0s
I180308 09:28:18.788824 6931 storage/replica_command.go:1760 [replicate,n5,s5,r17/4:/Table/2{0-1}] change replicas (REMOVE_REPLICA (n1,s1):5): read existing descriptor r17:/Table/2{0-1} [(n5,s5):4, (n3,s3):2, (n2,s2):3, (n1,s1):5, next=6]
I180308 09:28:19.053142 1010 storage/replica_proposal.go:201 [n1,s1,r27/1:/System/ts{d/cr.no…-e}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0
E180308 09:28:19.095691 5732 sql/logictest/logic.go:2124
testdata/parallel_test/subquery_retry_multinode/txn:9:
expected success, but found
pq: TransactionStatusError: does not exist
--- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 0 tests, 1 failures
I180308 09:28:19.153442 5732 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 0 tests, 1 failures
E180308 09:28:19.153715 5732 sql/logictest/parallel_test.go:84 error processing testdata/parallel_test/subquery_retry_multinode/txn: pq: TransactionStatusError: does not exist
I180308 09:28:19.146805 2484 server/status/runtime.go:219 [n5] runtime stats: 1.3 GiB RSS, 1162 goroutines, 95 MiB/13 MiB/138 MiB GO alloc/idle/total, 161 MiB/196 MiB CGO alloc/total, 1029.88cgo/sec, 0.90/0.08 %(u/s)time, 0.02 %gc (3x)
W180308 09:28:19.277944 2147 storage/node_liveness.go:501 [n4,hb] slow heartbeat took 2.0s
I180308 09:28:19.513328 7000 storage/replica_command.go:838 [split,n1,s1,r29/1:/System/ts{d/cr.no…-e}] initiating a split of this range at key /System/tsd/cr.node.sql.mem.admin.session.max-p99.9/1/10s/2018-03-08T09:00:00Z [r31]
I180308 09:28:19.527756 6832 storage/replica.go:3209 [n1,s1,r5/1:/System/tsd{-/cr.nod…}] proposing ADD_REPLICA((n4,s4):4): updated=[(n1,s1):1 (n2,s2):2 (n5,s5):3 (n4,s4):4] next=5
I180308 09:28:20.040589 7039 storage/replica_command.go:1760 [replicate,n1,s1,r24/1:/Table/13/1/2018-03-08T09:27:…] change replicas (REMOVE_REPLICA (n2,s2):3): read existing descriptor r24:/Table/13/1/2018-03-08T09:27:{29.777287Z/329091808619495425-40.666166Z/329091846156877826} [(n1,s1):1, (n3,s3):2, (n2,s2):3, (n5,s5):4, next=5]
W180308 09:28:20.196189 7047 storage/replica_backpressure.go:135 [n1,s1,r29/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.txn.aborts/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p90/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p75/1/10s/2018-03-08T09:00:00Z,/Min), ... 330 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/1/10s/2018-03-08T09:00:00Z,/Min)
I180308 09:28:20.209629 7061 storage/replica_raftstorage.go:520 [raftsnapshot,n1,s1,r5/1:/System/tsd{-/cr.nod…}] generated Raft snapshot a58aa6bc at index 74
I180308 09:28:20.318346 7061 storage/store.go:3745 [raftsnapshot,n1,s1,r5/1:/System/tsd{-/cr.nod…}] streamed snapshot to (n4,s4):4: kv pairs: 145, log entries: 19, rate-limit: 8.0 MiB/sec, 103ms
I180308 09:28:20.321961 7122 storage/replica_raftstorage.go:726 [n4,s4,r5/4:{-}] applying Raft snapshot at index 74 (id=a58aa6bc, encoded size=39489, 1 rocksdb batches, 19 log entries)
--- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries
I180308 09:28:20.675686 5734 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries
--- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures
I180308 09:28:20.676049 5734 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures
W180308 09:28:20.841923 7144 storage/replica_backpressure.go:135 [n1,s1,r29/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.txn.aborts/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p90/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p75/3/10s/2018-03-08T09:00:00Z,/Min), ... 330 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/3/10s/2018-03-08T09:00:00Z,/Min)
W180308 09:28:21.168305 2493 storage/node_liveness.go:501 [n5,hb] slow heartbeat took 2.5s
W180308 09:28:21.335806 1028 storage/store.go:3866 [n1,s1] handle raft ready: 0.6s [processed=0]
W180308 09:28:21.372717 7073 storage/replica_backpressure.go:135 [n5,s5,r29/3:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.txn.aborts/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p90/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p75/2/10s/2018-03-08T09:00:00Z,/Min), ... 330 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/2/10s/2018-03-08T09:00:00Z,/Min)
W180308 09:28:21.566361 1553 storage/node_liveness.go:501 [n2,hb] slow heartbeat took 2.0s
I180308 09:28:21.934342 7122 storage/replica_raftstorage.go:732 [n4,s4,r5/4:/System/tsd{-/cr.nod…}] applied Raft snapshot in 1612ms [clear=0ms batch=0ms entries=1609ms commit=2ms]
--- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries
I180308 09:28:22.172925 5733 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries
--- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures
I180308 09:28:22.173248 5733 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures
I180308 09:28:22.173462 835 sql/logictest/parallel_test.go:159 testdata/parallel_test/subquery_retry_multinode/test.yaml: run list 2: 0:final
W180308 09:28:22.273947 1096 storage/node_liveness.go:501 [n1,hb] slow heartbeat took 1.1s
W180308 09:28:22.317366 1806 storage/node_liveness.go:501 [n3,hb] slow heartbeat took 1.9s
--- done: testdata/parallel_test/subquery_retry_multinode/final with config : 2 tests, 0 failures
I180308 09:28:22.631831 7227 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/final with config : 2 tests, 0 failures
I180308 09:28:22.633272 7200 util/stop/stopper.go:473 quiescing; tasks left:
1 ts.poller: poll
I180308 09:28:22.636235 7199 util/stop/stopper.go:473 quiescing; tasks left:
1 ts.poller: poll
W180308 09:28:22.636697 3779 storage/raft_transport.go:465 [n3] raft transport stream to node 4 failed: rpc error: code = Unavailable desc = transport is closing
I180308 09:28:22.637631 7198 util/stop/stopper.go:473 quiescing; tasks left:
6 node.Node: batch
3 [async] kv.TxnCoordSender: heartbeat loop
2 [async] storage.split: processing replica
1 ts.poller: poll
1 [async] transport racer
1 [async] storage.replicate: processing replica
W180308 09:28:22.638338 2119 storage/raft_transport.go:465 [n1] raft transport stream to node 2 failed: EOF
W180308 09:28:22.645863 7000 internal/client/txn.go:531 [split,n1,s1,r29/1:/System/ts{d/cr.no…-e}] failure aborting transaction: node unavailable; try another peer; abort caused by: node unavailable; try another peer
W180308 09:28:22.658964 1047 storage/store.go:1480 [n1,s1,r7/1:/Table/{SystemCon…-11}] could not gossip system config: node unavailable; try another peer
W180308 09:28:22.659676 2336 storage/raft_transport.go:465 [n1] raft transport stream to node 4 failed: rpc error: code = Unavailable desc = transport is closing
I180308 09:28:22.663545 7198 util/stop/stopper.go:473 quiescing; tasks left:
3 node.Node: batch
3 [async] kv.TxnCoordSender: heartbeat loop
1 ts.poller: poll
1 [async] transport racer
1 [async] storage.split: processing replica
1 [async] storage.replicate: processing replica
W180308 09:28:22.666411 6668 internal/client/txn.go:531 [split,n1,s1,r28/1:/System/{StatusNo…-tsd}] failure aborting transaction: node unavailable; try another peer; abort caused by: result is ambiguous (server shutdown)
I180308 09:28:22.669226 7198 util/stop/stopper.go:473 quiescing; tasks left:
3 [async] kv.TxnCoordSender: heartbeat loop
1 ts.poller: poll
1 node.Node: batch
1 [async] transport racer
1 [async] storage.split: processing replica
1 [async] storage.replicate: processing replica
W180308 09:28:22.670449 1545 ts/db.go:154 [n2,ts-poll] error writing time series data: the batch experienced mixed success and failure
W180308 09:28:22.679032 1545 ts/db.go:157 [n2,ts-poll] node unavailable; try another peer
I180308 09:28:22.679527 7198 util/stop/stopper.go:473 quiescing; tasks left:
3 [async] kv.TxnCoordSender: heartbeat loop
1 ts.poller: poll
1 node.Node: batch
1 [async] transport racer
1 [async] storage.replicate: processing replica
I180308 09:28:22.705284 1055 server/status/runtime.go:219 [n1] runtime stats: 1.3 GiB RSS, 1118 goroutines, 67 MiB/42 MiB/140 MiB GO alloc/idle/total, 158 MiB/197 MiB CGO alloc/total, 881.83cgo/sec, 0.89/0.09 %(u/s)time, 0.03 %gc (4x)
W180308 09:28:22.706902 1048 storage/store.go:1480 [n1,s1,r3/1:/System/NodeLiveness{-Max}] could not gossip node liveness: node unavailable; try another peer
I180308 09:28:22.707664 7201 util/stop/stopper.go:473 quiescing; tasks left:
1 ts.poller: poll
I180308 09:28:22.710501 7330 util/stop/stopper.go:473 quiescing; tasks left:
1 ts.poller: poll
1 [async] storage.replicate: processing replica
1 [async] kv.TxnCoordSender: heartbeat loop
W180308 09:28:22.711733 1047 storage/store.go:1480 [n1,s1,r7/1:/Table/{SystemCon…-11}] could not gossip system config: node unavailable; try another peer
W180308 09:28:22.716304 1046 gossip/infostore.go:303 [n1] node unavailable; try another peer
W180308 09:28:22.716688 1046 storage/store.go:1480 [n1,s1,r1/1:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer
I180308 09:28:22.717708 1053 gossip/gossip.go:487 [n1] gossip status (ok, 5 nodes)
gossip client (0/3 cur/max conns)
gossip server (3/3 cur/max conns, infos 2303/150 sent/received, bytes 774296B/46012B sent/received)
3: 127.0.0.1:46629 (56s)
4: 127.0.0.1:38315 (55s)
2: 127.0.0.1:37987 (57s)
W180308 09:28:22.720545 1053 gossip/infostore.go:303 [n1] node unavailable; try another peer
W180308 09:28:22.722669 2373 storage/raft_transport.go:465 [n4] raft transport stream to node 1 failed: rpc error: code = Unavailable desc = transport is closing
W180308 09:28:22.727955 2485 ts/db.go:154 [n5,ts-poll] error writing time series data: the batch experienced mixed success and failure
W180308 09:28:22.728252 2485 ts/db.go:157 [n5,ts-poll] node unavailable; try another peer
I180308 09:28:22.728512 7330 util/stop/stopper.go:473 quiescing; tasks left:
1 [async] storage.replicate: processing replica
1 [async] kv.TxnCoordSender: heartbeat loop
W180308 09:28:22.732195 2770 storage/raft_transport.go:465 [n1] raft transport stream to node 5 failed: rpc error: code = Unavailable desc = transport is closing
W180308 09:28:22.737226 4045 storage/raft_transport.go:465 [n2] raft transport stream to node 3 failed: EOF
W180308 09:28:22.742157 1899 storage/raft_transport.go:465 [n1] raft transport stream to node 3 failed: rpc error: code = Unavailable desc = transport is closing
W180308 09:28:22.747260 4042 storage/raft_transport.go:465 [n3] raft transport stream to node 2 failed: rpc error: code = FailedPrecondition desc = grpc: the client connection is closing
W180308 09:28:22.760150 1970 storage/raft_transport.go:465 [n3] raft transport stream to node 1 failed: rpc error: code = Unavailable desc = transport is closing
W180308 09:28:22.763583 2045 ts/db.go:154 [n4,ts-poll] error writing time series data: the batch experienced mixed success and failure
W180308 09:28:22.770300 2045 ts/db.go:157 [n4,ts-poll] node unavailable; try another peer
I180308 09:28:22.763970 750 kv/transport_race.go:67 transport race promotion: ran 45 iterations on up to 2482 requests
W180308 09:28:22.764917 3208 storage/raft_transport.go:465 [n4] raft transport stream to node 5 failed: rpc error: code = FailedPrecondition desc = grpc: the client connection is closing
I180308 09:28:22.772524 7198 util/stop/stopper.go:473 quiescing; tasks left:
3 [async] kv.TxnCoordSender: heartbeat loop
1 ts.poller: poll
1 node.Node: batch
1 [async] storage.replicate: processing replica
W180308 09:28:22.775756 1518 gossip/gossip.go:1292 [n2] no incoming or outgoing connections
W180308 09:28:22.776260 6938 kv/txn_coord_sender.go:1036 [split,s1,r28/1:/System/{StatusNo…-tsd},n1,txn=fcc1998d] node unavailable; try another peer
W180308 09:28:22.777253 3848 storage/raft_transport.go:465 [n4] raft transport stream to node 2 failed: rpc error: code = Unavailable desc = transport is closing
W180308 09:28:22.792605 2178 storage/raft_transport.go:465 [n2] raft transport stream to node 1 failed: rpc error: code = Unavailable desc = transport is closing
W180308 09:28:22.793349 3922 storage/raft_transport.go:465 [n2] raft transport stream to node 4 failed: rpc error: code = Unavailable desc = transport is closing
W180308 09:28:22.803019 1056 ts/db.go:154 [n1,ts-poll] error writing time series data: the batch experienced mixed success and failure
W180308 09:28:22.812175 1056 ts/db.go:157 [n1,ts-poll] node unavailable; try another peer
W180308 09:28:22.803505 1046 gossip/infostore.go:303 [n1] node unavailable; try another peer
W180308 09:28:22.812587 1046 storage/store.go:1480 [n1,s1,r1/1:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer
W180308 09:28:22.807111 1048 storage/store.go:1480 [n1,s1,r3/1:/System/NodeLiveness{-Max}] could not gossip node liveness: node unavailable; try another peer
W180308 09:28:22.807396 1057 server/node.go:776 [n1,summaries] error recording status summaries: node unavailable; try another peer
W180308 09:28:22.808727 7039 internal/client/txn.go:531 [replicate,n1,s1,r24/1:/Table/13/1/2018-03-08T09:27:…] failure aborting transaction: node unavailable; try another peer; abort caused by: node unavailable; try another peer
E180308 09:28:22.813785 7039 storage/queue.go:778 [replicate,n1,s1,r24/1:/Table/13/1/2018-03-08T09:27:…] change replicas of r24 failed: node unavailable; try another peer
W180308 09:28:22.809209 3173 storage/raft_transport.go:465 [n2] raft transport stream to node 5 failed: rpc error: code = Unavailable desc = transport is closing
W180308 09:28:22.815172 7191 kv/txn_coord_sender.go:1036 [split,s1,r29/1:/System/ts{d/cr.no…-e},n1,txn=db46a2ed] node unavailable; try another peer
W180308 09:28:22.818081 6931 internal/client/txn.go:531 [replicate,n5,s5,r17/4:/Table/2{0-1}] failure aborting transaction: node unavailable; try another peer; abort caused by: node unavailable; try another peer
W180308 09:28:22.820510 2743 storage/raft_transport.go:465 [n5] raft transport stream to node 1 failed: EOF
W180308 09:28:22.824887 2238 gossip/gossip.go:1292 [n5] no incoming or outgoing connections
W180308 09:28:22.825931 1047 storage/store.go:1480 [n1,s1,r7/1:/Table/{SystemCon…-11}] could not gossip system config: node unavailable; try another peer
W180308 09:28:22.835973 3199 storage/raft_transport.go:465 [n5] raft transport stream to node 4 failed: EOF
W180308 09:28:22.837525 3335 storage/raft_transport.go:465 [n5] raft transport stream to node 3 failed: EOF
W180308 09:28:22.842099 1748 gossip/gossip.go:1292 [n3] no incoming or outgoing connections
W180308 09:28:22.842893 3245 storage/raft_transport.go:465 [n3] raft transport stream to node 5 failed: rpc error: code = FailedPrecondition desc = grpc: the client connection is closing
W180308 09:28:22.845802 3832 storage/raft_transport.go:465 [n4] raft transport stream to node 3 failed: rpc error: code = Unavailable desc = transport is closing
W180308 09:28:22.853934 1798 ts/db.go:154 [n3,ts-poll] error writing time series data: the batch experienced mixed success and failure
W180308 09:28:22.855078 1979 gossip/gossip.go:1292 [n4] no incoming or outgoing connections
W180308 09:28:22.856870 3102 storage/raft_transport.go:465 [n5] raft transport stream to node 2 failed: EOF
W180308 09:28:22.890039 1798 ts/db.go:157 [n3,ts-poll] node unavailable; try another peer
I180308 09:28:23.130857 6 sql/logictest/parallel_test.go:266 1 out of 2 parallel tests failed
``` | 1.0 | sql/logictest: TestParallel failed under stress - SHA: https://github.com/cockroachdb/cockroach/commits/ea5f8b22ba951799b76c84502aaab4456fe00a46
Parameters:
```
TAGS=
GOFLAGS=-race
```
Stress build found a failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=552372&tab=buildLog
```
I180308 09:27:49.787154 4678 storage/replica_raftstorage.go:732 [n3,s3,r9/?:/Table/1{2-3}] applied preemptive snapshot in 8ms [clear=0ms batch=0ms entries=7ms commit=0ms]
I180308 09:27:49.902388 835 testutils/testcluster/testcluster.go:534 [n1,s1] has 1 underreplicated ranges
W180308 09:27:49.965287 1035 storage/store.go:3866 [n1,s1] handle raft ready: 1.0s [processed=0]
I180308 09:27:49.992899 4679 storage/replica.go:3209 [n1,s1,r3/1:/System/NodeLiveness{-Max}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n4,s4):2 (n2,s2):3] next=4
I180308 09:27:49.998836 4659 storage/replica_command.go:1760 [replicate,n2,s2,r9/2:/Table/1{2-3}] change replicas (ADD_REPLICA (n3,s3):4): read existing descriptor r9:/Table/1{2-3} [(n1,s1):1, (n2,s2):2, (n5,s5):3, next=4]
W180308 09:27:50.033177 1806 storage/node_liveness.go:501 [n3,hb] slow heartbeat took 1.1s
I180308 09:27:50.226728 835 testutils/testcluster/testcluster.go:534 [n1,s1] has 1 underreplicated ranges
I180308 09:27:50.327811 4695 storage/replica_raftstorage.go:520 [replicate,n1,s1,r10/1:/Table/1{3-4}] generated preemptive snapshot ae1bb9ed at index 153
I180308 09:27:50.408223 4615 storage/store.go:2350 [replicaGC,n1,s1,r16/1:/Table/{19-20}] removing replica
I180308 09:27:50.482741 4695 storage/store.go:3745 [replicate,n1,s1,r10/1:/Table/1{3-4}] streamed snapshot to (n5,s5):?: kv pairs: 407, log entries: 51, rate-limit: 2.0 MiB/sec, 154ms
I180308 09:27:50.524357 4615 storage/replica.go:836 [replicaGC,n1,s1,r16/1:/Table/{19-20}] removed 11 (0+11) keys in 40ms [clear=0ms commit=40ms]
I180308 09:27:50.738648 4697 storage/replica_raftstorage.go:726 [n5,s5,r10/?:{-}] applying preemptive snapshot at index 153 (id=ae1bb9ed, encoded size=81088, 1 rocksdb batches, 51 log entries)
I180308 09:27:50.786926 4653 storage/replica.go:3209 [n4,s4,r18/2:/Table/2{1-2}] proposing ADD_REPLICA((n3,s3):4): updated=[(n1,s1):1 (n4,s4):2 (n5,s5):3 (n3,s3):4] next=5
I180308 09:27:50.817425 1748 gossip/gossip.go:1306 [n3] node has connected to cluster via gossip
I180308 09:27:50.818926 1748 storage/stores.go:350 [n3] wrote 4 node addresses to persistent storage
I180308 09:27:50.857137 4684 storage/replica.go:3209 [n3,s3,r13/2:/Table/1{6-7}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n4,s4):4 (n3,s3):2 (n5,s5):3] next=5
I180308 09:27:51.086023 4697 storage/replica_raftstorage.go:732 [n5,s5,r10/?:/Table/1{3-4}] applied preemptive snapshot in 347ms [clear=0ms batch=0ms entries=187ms commit=1ms]
I180308 09:27:51.243460 4695 storage/replica_command.go:1760 [replicate,n1,s1,r10/1:/Table/1{3-4}] change replicas (ADD_REPLICA (n5,s5):4): read existing descriptor r10:/Table/1{3-4} [(n1,s1):1, (n3,s3):2, (n2,s2):3, next=4]
I180308 09:27:51.381230 4761 storage/replica_command.go:1760 [replicate,n4,s4,r18/2:/Table/2{1-2}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r18:/Table/2{1-2} [(n1,s1):1, (n4,s4):2, (n5,s5):3, (n3,s3):4, next=5]
I180308 09:27:51.509691 4735 storage/replica_raftstorage.go:520 [replicate,n3,s3,r6/2:/{System/tse-Table/System…}] generated preemptive snapshot 27aebd8b at index 30
I180308 09:27:51.563486 835 sql/logictest/parallel_test.go:213 Setting range split size: 32768
I180308 09:27:51.601539 4735 storage/store.go:3745 [replicate,n3,s3,r6/2:/{System/tse-Table/System…}] streamed snapshot to (n2,s2):?: kv pairs: 14, log entries: 20, rate-limit: 2.0 MiB/sec, 91ms
I180308 09:27:51.619862 1901 storage/store.go:3479 [n1,s1,r13/1:/Table/1{6-7}] added to replica GC queue (peer suggestion)
I180308 09:27:51.630538 4789 storage/replica_raftstorage.go:726 [n2,s2,r6/?:{-}] applying preemptive snapshot at index 30 (id=27aebd8b, encoded size=8130, 1 rocksdb batches, 20 log entries)
I180308 09:27:51.728873 4789 storage/replica_raftstorage.go:732 [n2,s2,r6/?:/{System/tse-Table/System…}] applied preemptive snapshot in 98ms [clear=0ms batch=0ms entries=94ms commit=3ms]
I180308 09:27:51.807803 4735 storage/replica_command.go:1760 [replicate,n3,s3,r6/2:/{System/tse-Table/System…}] change replicas (ADD_REPLICA (n2,s2):4): read existing descriptor r6:/{System/tse-Table/SystemConfigSpan/Start} [(n1,s1):1, (n3,s3):2, (n4,s4):3, next=4]
I180308 09:27:51.828942 4767 storage/replica_raftstorage.go:520 [raftsnapshot,n4,s4,r18/2:/Table/2{1-2}] generated Raft snapshot df1ffd91 at index 31
I180308 09:27:51.834992 4767 storage/store.go:3745 [raftsnapshot,n4,s4,r18/2:/Table/2{1-2}] streamed snapshot to (n3,s3):4: kv pairs: 18, log entries: 21, rate-limit: 8.0 MiB/sec, 5ms
I180308 09:27:51.837164 4834 storage/replica_raftstorage.go:726 [n3,s3,r18/4:{-}] applying Raft snapshot at index 31 (id=df1ffd91, encoded size=7716, 1 rocksdb batches, 21 log entries)
I180308 09:27:51.842351 4834 storage/replica_raftstorage.go:732 [n3,s3,r18/4:/Table/2{1-2}] applied Raft snapshot in 5ms [clear=0ms batch=0ms entries=3ms commit=0ms]
I180308 09:27:52.277701 4836 storage/store.go:2350 [replicaGC,n5,s5,r10/?:/Table/1{3-4}] removing replica
I180308 09:27:52.302073 4671 storage/store.go:2350 [replicaGC,n1,s1,r13/1:/Table/1{6-7}] removing replica
I180308 09:27:52.431403 4836 storage/replica.go:836 [replicaGC,n5,s5,r10/?:/Table/1{3-4}] removed 389 (379+10) keys in 153ms [clear=0ms commit=153ms]
I180308 09:27:52.549535 4671 storage/replica.go:836 [replicaGC,n1,s1,r13/1:/Table/1{6-7}] removed 12 (0+12) keys in 219ms [clear=0ms commit=219ms]
I180308 09:27:52.628009 4816 storage/replica.go:3209 [n2,s2,r9/2:/Table/1{2-3}] proposing ADD_REPLICA((n3,s3):4): updated=[(n1,s1):1 (n2,s2):2 (n5,s5):3 (n3,s3):4] next=5
I180308 09:27:52.699806 1055 server/status/runtime.go:219 [n1] runtime stats: 1.1 GiB RSS, 1117 goroutines, 76 MiB/16 MiB/119 MiB GO alloc/idle/total, 154 MiB/186 MiB CGO alloc/total, 862.81cgo/sec, 0.95/0.07 %(u/s)time, 0.05 %gc (5x)
I180308 09:27:52.842796 4890 storage/replica_command.go:1760 [replicate,n2,s2,r9/2:/Table/1{2-3}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r9:/Table/1{2-3} [(n1,s1):1, (n2,s2):2, (n5,s5):3, (n3,s3):4, next=5]
I180308 09:27:53.053227 4909 storage/replica.go:3209 [n3,s3,r6/2:/{System/tse-Table/System…}] proposing ADD_REPLICA((n2,s2):4): updated=[(n1,s1):1 (n3,s3):2 (n4,s4):3 (n2,s2):4] next=5
I180308 09:27:53.204274 4783 storage/replica_command.go:1760 [replicate,n3,s3,r6/2:/{System/tse-Table/System…}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r6:/{System/tse-Table/SystemConfigSpan/Start} [(n1,s1):1, (n3,s3):2, (n4,s4):3, (n2,s2):4, next=5]
I180308 09:27:53.385647 4933 storage/replica.go:3209 [n4,s4,r18/2:/Table/2{1-2}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n3,s3):4 (n4,s4):2 (n5,s5):3] next=5
I180308 09:27:53.654566 835 sql/logictest/parallel_test.go:227 Creating database
I180308 09:27:53.714521 1518 gossip/gossip.go:1306 [n2] node has connected to cluster via gossip
I180308 09:27:53.812099 1518 storage/stores.go:350 [n2] wrote 4 node addresses to persistent storage
I180308 09:27:53.949703 2370 storage/store.go:3479 [n1,s1,r18/1:/Table/2{1-2}] added to replica GC queue (peer suggestion)
I180308 09:27:53.950132 2370 storage/store.go:3479 [n1,s1,r18/1:/Table/2{1-2}] added to replica GC queue (peer suggestion)
I180308 09:27:54.187036 4918 storage/replica_command.go:838 [split,n1,s1,r5/1:/System/ts{d-e}] initiating a split of this range at key /System/tsd/cr.node.gossip.infos.sent/4/10s/2018-03-08T09:00:00Z [r21]
I180308 09:27:54.490089 4921 storage/replica_command.go:838 [split,n1,s1,r4/1:/System/{NodeLive…-tsd}] initiating a split of this range at key /System/StatusNode/2 [r22]
I180308 09:27:54.641807 4982 storage/replica_command.go:838 [split,n1,s1,r10/1:/Table/1{3-4}] initiating a split of this range at key /Table/13/1/2018-03-08T09:27:29.777287Z/329091808619495425 [r23]
I180308 09:27:54.767000 4695 storage/replica.go:3209 [n1,s1,r10/1:/Table/1{3-4}] proposing ADD_REPLICA((n5,s5):4): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3 (n5,s5):4] next=5
I180308 09:27:54.924274 4936 storage/store.go:2350 [replicaGC,n1,s1,r18/1:/Table/2{1-2}] removing replica
I180308 09:27:55.227444 4936 storage/replica.go:836 [replicaGC,n1,s1,r18/1:/Table/2{1-2}] removed 11 (0+11) keys in 51ms [clear=0ms commit=50ms]
W180308 09:27:55.283305 4861 storage/replica.go:3016 [n1,s1,r10/1:/Table/1{3-4}] context canceled before proposing: 1 HeartbeatTxn
I180308 09:27:55.373122 5042 storage/replica_command.go:838 [split,n1,s1,r10/1:/Table/1{3-4}] initiating a split of this range at key /Table/13/1/2018-03-08T09:27:29.777287Z/329091808619495425 [r24]
I180308 09:27:55.591922 5003 storage/replica.go:3209 [n2,s2,r9/2:/Table/1{2-3}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n3,s3):4 (n2,s2):2 (n5,s5):3] next=5
I180308 09:27:55.659550 1544 server/status/runtime.go:219 [n2] runtime stats: 1.1 GiB RSS, 1126 goroutines, 57 MiB/35 MiB/120 MiB GO alloc/idle/total, 155 MiB/187 MiB CGO alloc/total, 798.50cgo/sec, 0.95/0.07 %(u/s)time, 0.03 %gc (5x)
I180308 09:27:56.074863 2121 storage/store.go:3479 [n1,s1,r9/1:/Table/1{2-3}] added to replica GC queue (peer suggestion)
I180308 09:27:56.304086 4874 sql/event_log.go:121 [n1,client=127.0.0.1:43858,user=root] Event: "create_database", target: 50, info: {DatabaseName:test Statement:CREATE DATABASE test User:root}
I180308 09:27:56.307831 2121 storage/store.go:3479 [n1,s1,r9/1:/Table/1{2-3}] added to replica GC queue (peer suggestion)
I180308 09:27:56.399786 1635 storage/replica.go:3277 [n2,s2,r2/3:/System/{-NodeLive…}] not quiescing: 1 pending commands
I180308 09:27:56.412847 2121 storage/store.go:3479 [n1,s1,r9/1:/Table/1{2-3}] added to replica GC queue (peer suggestion)
I180308 09:27:56.413411 2121 storage/store.go:3479 [n1,s1,r9/1:/Table/1{2-3}] added to replica GC queue (peer suggestion)
I180308 09:27:56.429913 5110 storage/store.go:2350 [replicaGC,n1,s1,r9/1:/Table/1{2-3}] removing replica
I180308 09:27:56.448755 1797 server/status/runtime.go:219 [n3] runtime stats: 1.1 GiB RSS, 1135 goroutines, 66 MiB/26 MiB/120 MiB GO alloc/idle/total, 156 MiB/187 MiB CGO alloc/total, 773.55cgo/sec, 0.94/0.07 %(u/s)time, 0.03 %gc (5x)
I180308 09:27:56.688465 5110 storage/replica.go:836 [replicaGC,n1,s1,r9/1:/Table/1{2-3}] removed 71 (60+11) keys in 257ms [clear=0ms commit=257ms]
I180308 09:27:57.234316 5009 storage/replica_command.go:838 [split,n4,s4,r20/2:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r41]
I180308 09:27:57.715312 5007 storage/replica_raftstorage.go:520 [raftsnapshot,n1,s1,r10/1:/Table/1{3-4}] generated Raft snapshot 729bfb72 at index 173
W180308 09:27:57.734210 982 storage/store.go:3866 [n1,s1] handle raft ready: 0.6s [processed=0]
I180308 09:27:57.771678 2044 server/status/runtime.go:219 [n4] runtime stats: 1.2 GiB RSS, 1138 goroutines, 51 MiB/40 MiB/120 MiB GO alloc/idle/total, 156 MiB/187 MiB CGO alloc/total, 717.70cgo/sec, 0.95/0.07 %(u/s)time, 0.03 %gc (5x)
I180308 09:27:57.853130 5007 storage/store.go:3745 [raftsnapshot,n1,s1,r10/1:/Table/1{3-4}] streamed snapshot to (n5,s5):4: kv pairs: 456, log entries: 71, rate-limit: 8.0 MiB/sec, 104ms
I180308 09:27:57.904274 5196 storage/replica_raftstorage.go:726 [n5,s5,r10/4:{-}] applying Raft snapshot at index 173 (id=729bfb72, encoded size=103834, 1 rocksdb batches, 71 log entries)
W180308 09:27:58.307548 1553 storage/node_liveness.go:501 [n2,hb] slow heartbeat took 1.2s
I180308 09:27:58.890124 5216 storage/replica.go:3209 [n3,s3,r6/2:/{System/tse-Table/System…}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n2,s2):4 (n3,s3):2 (n4,s4):3] next=5
I180308 09:27:59.234979 2484 server/status/runtime.go:219 [n5] runtime stats: 1.2 GiB RSS, 1139 goroutines, 78 MiB/15 MiB/120 MiB GO alloc/idle/total, 156 MiB/188 MiB CGO alloc/total, 720.50cgo/sec, 0.94/0.06 %(u/s)time, 0.03 %gc (5x)
I180308 09:27:59.534800 1901 storage/store.go:3479 [n1,s1,r6/1:/{System/tse-Table/System…}] added to replica GC queue (peer suggestion)
I180308 09:27:59.617241 1901 storage/store.go:3479 [n1,s1,r6/1:/{System/tse-Table/System…}] added to replica GC queue (peer suggestion)
W180308 09:27:59.658775 1806 storage/node_liveness.go:501 [n3,hb] slow heartbeat took 1.7s
I180308 09:27:59.883988 835 sql/logictest/parallel_test.go:236 Test setup done
I180308 09:27:59.884248 835 sql/logictest/parallel_test.go:159 testdata/parallel_test/subquery_retry_multinode/test.yaml: run list 0: 0:setup
W180308 09:28:00.194433 1096 storage/node_liveness.go:501 [n1,hb] slow heartbeat took 1.4s
I180308 09:28:00.297625 5240 storage/store.go:2350 [replicaGC,n1,s1,r6/1:/{System/tse-Table/System…}] removing replica
I180308 09:28:00.329573 1028 storage/replica_proposal.go:201 [n1,s1,r4/1:/System/{NodeLive…-tsd}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0
I180308 09:28:00.379562 5248 storage/replica_command.go:838 [split,n1,s1,r22/1:/System/{StatusNo…-tsd}] initiating a split of this range at key /System/StatusNode/3 [r25]
I180308 09:28:00.448234 5240 storage/replica.go:836 [replicaGC,n1,s1,r6/1:/{System/tse-Table/System…}] removed 12 (0+12) keys in 91ms [clear=0ms commit=91ms]
W180308 09:28:00.707751 2147 storage/node_liveness.go:501 [n4,hb] slow heartbeat took 1.4s
W180308 09:28:00.731795 5352 storage/replica_backpressure.go:135 [n1,s1,r5/1:/System/ts{d-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.security.certificate.expiration.ca/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.security.certificate.expiration.node/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.connections.incoming/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.connections.refused/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.bytes.received/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.bytes.sent/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.infos.received/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.infos.sent/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.connections.outgoing/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.batches/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.batches.partial/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.rpc.sent/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.rpc.sent.local/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.rpc.sent.nextreplicaerror/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.distsender.errors.notleaseholder/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.requests.slow.distsender/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.aborts/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/1/10s/2018-03-08T09:00:00Z,/Min), ... 405 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/1/10s/2018-03-08T09:00:00Z,/Min)
I180308 09:28:01.257194 983 storage/replica_proposal.go:201 [n1,s1,r10/1:/Table/1{3-4}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.325864565,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.325864565,0
I180308 09:28:01.263525 5297 storage/replica_command.go:838 [split,n1,s1,r24/1:/Table/1{3/1/2018…-4}] initiating a split of this range at key /Table/13/1/2018-03-08T09:27:40.666166Z/329091846156877826 [r26]
I180308 09:28:01.312145 5196 storage/replica_raftstorage.go:732 [n5,s5,r10/4:/Table/1{3-4}] applied Raft snapshot in 3408ms [clear=0ms batch=0ms entries=3374ms commit=21ms]
I180308 09:28:01.454527 2220 storage/replica_proposal.go:201 [n4,s4,r20/2:/{Table/23-Max}] new range lease repl=(n4,s4):2 seq=3 start=1520501251.624682287,1 epo=1 pro=1520501255.402825609,0 following repl=(n4,s4):2 seq=3 start=1520501251.624682287,1 epo=1 pro=1520501255.402825609,0
I180308 09:28:01.482106 1005 storage/replica_proposal.go:201 [n1,s1,r5/1:/System/ts{d-e}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0
W180308 09:28:01.575951 5351 storage/replica.go:3016 [n1,s1,r10/1:/Table/13{-/1/2018-…}] context canceled before proposing: 1 HeartbeatTxn
I180308 09:28:01.658992 5386 storage/replica_command.go:838 [split,n1,s1,r21/1:/System/ts{d/cr.no…-e}] initiating a split of this range at key /System/tsd/cr.node.sql.bytesin/3/10s/2018-03-08T09:00:00Z [r27]
I180308 09:28:01.678252 5296 storage/replica_command.go:1760 [replicate,n1,s1,r10/1:/Table/13{-/1/2018-…}] change replicas (REMOVE_REPLICA (n3,s3):2): read existing descriptor r10:/Table/13{-/1/2018-03-08T09:27:29.777287Z/329091808619495425} [(n1,s1):1, (n3,s3):2, (n2,s2):3, (n5,s5):4, next=5]
I180308 09:28:01.932018 5423 storage/replica_raftstorage.go:520 [replicate,n4,s4,r41/2:/{Table/50-Max}] generated preemptive snapshot e7efcdd4 at index 10
I180308 09:28:01.954883 5491 storage/replica_raftstorage.go:520 [replicate,n3,s3,r17/2:/Table/2{0-1}] generated preemptive snapshot b7e6a194 at index 36
E180308 09:28:02.175447 5423 storage/queue.go:778 [replicate,n4,s4,r41/2:/{Table/50-Max}] raft status not initialized
I180308 09:28:02.188640 4874 sql/event_log.go:121 [n1,client=127.0.0.1:43858,user=root] Event: "create_table", target: 51, info: {TableName:test.public.t Statement:CREATE TABLE t (k INT, filler STRING) User:root}
--- progress: testdata/parallel_test/subquery_retry_multinode/setup: 1 statements/queries
I180308 09:28:02.231950 5276 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/setup: 1 statements/queries
I180308 09:28:02.309261 5491 storage/store.go:3745 [replicate,n3,s3,r17/2:/Table/2{0-1}] streamed snapshot to (n1,s1):?: kv pairs: 16, log entries: 26, rate-limit: 2.0 MiB/sec, 97ms
I180308 09:28:02.727426 5515 storage/replica_command.go:838 [split,n4,s4,r41/2:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r42]
I180308 09:28:02.767793 1979 gossip/gossip.go:1306 [n4] node has connected to cluster via gossip
I180308 09:28:02.791521 1055 server/status/runtime.go:219 [n1] runtime stats: 1.2 GiB RSS, 1166 goroutines, 65 MiB/26 MiB/120 MiB GO alloc/idle/total, 157 MiB/187 MiB CGO alloc/total, 1065.63cgo/sec, 0.94/0.07 %(u/s)time, 0.02 %gc (5x)
I180308 09:28:03.002229 5486 storage/replica_raftstorage.go:726 [n1,s1,r17/?:{-}] applying preemptive snapshot at index 36 (id=b7e6a194, encoded size=8224, 1 rocksdb batches, 26 log entries)
I180308 09:28:03.005301 1979 storage/stores.go:350 [n4] wrote 4 node addresses to persistent storage
I180308 09:28:03.154898 5486 storage/replica_raftstorage.go:732 [n1,s1,r17/?:/Table/2{0-1}] applied preemptive snapshot in 152ms [clear=0ms batch=0ms entries=127ms commit=12ms]
W180308 09:28:03.190672 5455 storage/replica_backpressure.go:135 [n2,s2,r21/2:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.security.certificate.expiration.ca/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.security.certificate.expiration.node/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.requests.slow.distsender/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.aborts/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), ... 379 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/2/10s/2018-03-08T09:00:00Z,/Min)
I180308 09:28:03.513617 5491 storage/replica_command.go:1760 [replicate,n3,s3,r17/2:/Table/2{0-1}] change replicas (ADD_REPLICA (n1,s1):5): read existing descriptor r17:/Table/2{0-1} [(n5,s5):4, (n3,s3):2, (n2,s2):3, next=5]
W180308 09:28:04.034472 5607 storage/replica_backpressure.go:135 [n1,s1,r21/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.security.certificate.expiration.ca/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.security.certificate.expiration.node/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.infos.sent/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.requests.slow.distsender/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.aborts/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/4/10s/2018-03-08T09:00:00Z,/Min), ... 380 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/4/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/4/10s/2018-03-08T09:00:00Z,/Min)
W180308 09:28:04.459926 5676 storage/replica_backpressure.go:135 [n1,s1,r21/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.security.certificate.expiration.ca/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.security.certificate.expiration.node/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.gossip.infos.sent/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.requests.slow.distsender/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.aborts/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/5/10s/2018-03-08T09:00:00Z,/Min), ... 380 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/5/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/5/10s/2018-03-08T09:00:00Z,/Min)
--- progress: testdata/parallel_test/subquery_retry_multinode/setup: 2 statements/queries
I180308 09:28:04.488626 5276 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/setup: 2 statements/queries
--- done: testdata/parallel_test/subquery_retry_multinode/setup with config : 2 tests, 0 failures
I180308 09:28:04.488940 5276 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/setup with config : 2 tests, 0 failures
I180308 09:28:04.494592 835 sql/logictest/parallel_test.go:159 testdata/parallel_test/subquery_retry_multinode/test.yaml: run list 1: 0:txn, 1:txn, 2:txn, 3:txn, 4:txn
I180308 09:28:04.692538 5642 storage/replica.go:3209 [n1,s1,r10/1:/Table/13{-/1/2018-…}] proposing REMOVE_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n5,s5):4 (n2,s2):3] next=5
I180308 09:28:05.177727 1003 storage/replica_proposal.go:201 [n1,s1,r22/1:/System/{StatusNo…-tsd}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0
I180308 09:28:05.184170 1972 storage/store.go:3479 [n3,s3,r10/2:/Table/13{-/1/2018-…}] added to replica GC queue (peer suggestion)
I180308 09:28:05.275031 5772 storage/replica_command.go:838 [split,n1,s1,r25/1:/System/{StatusNo…-tsd}] initiating a split of this range at key /System/StatusNode/4 [r28]
I180308 09:28:05.629470 1544 server/status/runtime.go:219 [n2] runtime stats: 1.2 GiB RSS, 1171 goroutines, 87 MiB/11 MiB/125 MiB GO alloc/idle/total, 157 MiB/189 MiB CGO alloc/total, 1173.83cgo/sec, 0.94/0.07 %(u/s)time, 0.02 %gc (4x)
I180308 09:28:05.634410 1827 storage/replica_proposal.go:201 [n3,s3,r4/2:/System/{NodeLive…-StatusNo…}] new range lease repl=(n3,s3):2 seq=4 start=1520501285.016867674,0 epo=1 pro=1520501285.016882538,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0
W180308 09:28:06.031522 5844 storage/replica.go:3016 [n1,s1,r22/1:/System/StatusNode/{2-3}] context canceled before proposing: 1 HeartbeatTxn
--- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries
I180308 09:28:06.562121 5731 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries
--- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures
I180308 09:28:06.562430 5731 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures
I180308 09:28:06.630349 1797 server/status/runtime.go:219 [n3] runtime stats: 1.2 GiB RSS, 1178 goroutines, 97 MiB/5.9 MiB/129 MiB GO alloc/idle/total, 158 MiB/189 MiB CGO alloc/total, 1136.76cgo/sec, 0.94/0.07 %(u/s)time, 0.03 %gc (5x)
W180308 09:28:06.862124 2493 storage/node_liveness.go:501 [n5,hb] slow heartbeat took 1.7s
W180308 09:28:07.341217 1553 storage/node_liveness.go:501 [n2,hb] slow heartbeat took 1.2s
I180308 09:28:07.775014 2044 server/status/runtime.go:219 [n4] runtime stats: 1.2 GiB RSS, 1174 goroutines, 75 MiB/24 MiB/129 MiB GO alloc/idle/total, 157 MiB/189 MiB CGO alloc/total, 1173.62cgo/sec, 0.94/0.07 %(u/s)time, 0.03 %gc (4x)
I180308 09:28:07.899393 993 storage/replica_proposal.go:201 [n1,s1,r21/1:/System/ts{d/cr.no…-e}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0
I180308 09:28:08.092040 6071 storage/replica_command.go:838 [split,n1,s1,r27/1:/System/ts{d/cr.no…-e}] initiating a split of this range at key /System/tsd/cr.node.sql.distsql.service.latency-p99.999/1/10s/2018-03-08T09:00:00Z [r29]
W180308 09:28:09.182946 1850 storage/store.go:3866 [n3,s3] handle raft ready: 0.7s [processed=0]
I180308 09:28:09.228630 2484 server/status/runtime.go:219 [n5] runtime stats: 1.3 GiB RSS, 1216 goroutines, 96 MiB/9.2 MiB/132 MiB GO alloc/idle/total, 160 MiB/191 MiB CGO alloc/total, 1213.47cgo/sec, 0.94/0.07 %(u/s)time, 0.03 %gc (5x)
W180308 09:28:09.431040 987 storage/store.go:3866 [n1,s1] handle raft ready: 0.6s [processed=0]
W180308 09:28:09.702153 1029 storage/store.go:3866 [n1,s1] handle raft ready: 0.5s [processed=0]
I180308 09:28:09.884861 6241 storage/replica_raftstorage.go:520 [replicate,n1,s1,r5/1:/System/tsd{-/cr.nod…}] generated preemptive snapshot c879f910 at index 68
W180308 09:28:09.920664 1096 storage/node_liveness.go:501 [n1,hb] slow heartbeat took 2.2s
I180308 09:28:09.926471 6051 storage/store.go:2350 [replicaGC,n1,s1,r17/?:/Table/2{0-1}] removing replica
I180308 09:28:10.164342 6051 storage/replica.go:836 [replicaGC,n1,s1,r17/?:/Table/2{0-1}] removed 11 (0+11) keys in 177ms [clear=0ms commit=177ms]
I180308 09:28:10.208745 6080 storage/replica_raftstorage.go:520 [replicate,n2,s2,r2/3:/System/{-NodeLive…}] generated preemptive snapshot c8cdd480 at index 34
I180308 09:28:10.482965 6080 storage/store.go:3745 [replicate,n2,s2,r2/3:/System/{-NodeLive…}] streamed snapshot to (n5,s5):?: kv pairs: 13, log entries: 24, rate-limit: 2.0 MiB/sec, 273ms
I180308 09:28:10.627902 6241 storage/store.go:3745 [replicate,n1,s1,r5/1:/System/tsd{-/cr.nod…}] streamed snapshot to (n4,s4):?: kv pairs: 143, log entries: 13, rate-limit: 2.0 MiB/sec, 677ms
I180308 09:28:10.693165 6287 storage/replica_raftstorage.go:726 [n5,s5,r2/?:{-}] applying preemptive snapshot at index 34 (id=c8cdd480, encoded size=56977, 1 rocksdb batches, 24 log entries)
I180308 09:28:10.790504 6345 storage/replica_raftstorage.go:726 [n4,s4,r5/?:{-}] applying preemptive snapshot at index 68 (id=c879f910, encoded size=36756, 1 rocksdb batches, 13 log entries)
W180308 09:28:10.795289 6380 storage/replica_backpressure.go:135 [n1,s1,r27/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.txn.aborts/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p90/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p75/3/10s/2018-03-08T09:00:00Z,/Min), ... 355 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/3/10s/2018-03-08T09:00:00Z,/Min)
W180308 09:28:10.836796 1826 storage/store.go:3866 [n3,s3] handle raft ready: 0.5s [processed=0]
W180308 09:28:10.963543 2147 storage/node_liveness.go:501 [n4,hb] slow heartbeat took 2.7s
W180308 09:28:11.405336 6410 storage/replica_backpressure.go:135 [n1,s1,r27/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.txn.aborts/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p90/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p75/2/10s/2018-03-08T09:00:00Z,/Min), ... 354 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/2/10s/2018-03-08T09:00:00Z,/Min)
I180308 09:28:11.779388 987 storage/replica_proposal.go:201 [n1,s1,r24/1:/Table/1{3/1/2018…-4}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.325864565,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.325864565,0
W180308 09:28:11.899345 2493 storage/node_liveness.go:501 [n5,hb] slow heartbeat took 2.2s
W180308 09:28:12.551481 1553 storage/node_liveness.go:501 [n2,hb] slow heartbeat took 1.9s
I180308 09:28:12.571934 6345 storage/replica_raftstorage.go:732 [n4,s4,r5/?:/System/tsd{-/cr.nod…}] applied preemptive snapshot in 1781ms [clear=0ms batch=0ms entries=1729ms commit=33ms]
I180308 09:28:12.630942 6241 storage/replica_command.go:1760 [replicate,n1,s1,r5/1:/System/tsd{-/cr.nod…}] change replicas (ADD_REPLICA (n4,s4):4): read existing descriptor r5:/System/tsd{-/cr.node.gossip.infos.sent/4/10s/2018-03-08T09:00:00Z} [(n1,s1):1, (n2,s2):2, (n5,s5):3, next=4]
W180308 09:28:12.654684 1806 storage/node_liveness.go:501 [n3,hb] slow heartbeat took 1.2s
I180308 09:28:12.667081 1055 server/status/runtime.go:219 [n1] runtime stats: 1.3 GiB RSS, 1182 goroutines, 78 MiB/23 MiB/132 MiB GO alloc/idle/total, 160 MiB/194 MiB CGO alloc/total, 934.67cgo/sec, 0.92/0.07 %(u/s)time, 0.02 %gc (4x)
I180308 09:28:13.552826 6287 storage/replica_raftstorage.go:732 [n5,s5,r2/?:/System/{-NodeLive…}] applied preemptive snapshot in 2833ms [clear=0ms batch=0ms entries=2609ms commit=164ms]
I180308 09:28:13.769924 6080 storage/replica_command.go:1760 [replicate,n2,s2,r2/3:/System/{-NodeLive…}] change replicas (ADD_REPLICA (n5,s5):4): read existing descriptor r2:/System/{-NodeLiveness} [(n1,s1):1, (n3,s3):2, (n2,s2):3, next=4]
W180308 09:28:13.950338 2147 storage/node_liveness.go:501 [n4,hb] slow heartbeat took 1.2s
I180308 09:28:14.047030 6643 storage/replica.go:3209 [n3,s3,r17/2:/Table/2{0-1}] proposing ADD_REPLICA((n1,s1):5): updated=[(n5,s5):4 (n3,s3):2 (n2,s2):3 (n1,s1):5] next=6
--- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries
I180308 09:28:14.991532 5735 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries
--- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures
I180308 09:28:14.991818 5735 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures
W180308 09:28:15.210739 2493 storage/node_liveness.go:501 [n5,hb] slow heartbeat took 1.1s
I180308 09:28:15.459366 2539 storage/replica_proposal.go:201 [n5,s5,r17/4:/Table/2{0-1}] new range lease repl=(n5,s5):4 seq=4 start=1520501294.765769961,0 epo=1 pro=1520501294.765803960,0 following repl=(n3,s3):2 seq=3 start=1520501251.624682287,1 epo=1 pro=1520501257.825367317,0
I180308 09:28:15.554730 6653 storage/replica_raftstorage.go:520 [raftsnapshot,n3,s3,r17/2:/Table/2{0-1}] generated Raft snapshot 901a1124 at index 49
I180308 09:28:15.697368 1544 server/status/runtime.go:219 [n2] runtime stats: 1.3 GiB RSS, 1181 goroutines, 73 MiB/35 MiB/138 MiB GO alloc/idle/total, 160 MiB/195 MiB CGO alloc/total, 840.48cgo/sec, 0.90/0.07 %(u/s)time, 0.02 %gc (4x)
I180308 09:28:15.764467 6653 storage/store.go:3745 [raftsnapshot,n3,s3,r17/2:/Table/2{0-1}] streamed snapshot to (n1,s1):5: kv pairs: 18, log entries: 39, rate-limit: 8.0 MiB/sec, 61ms
I180308 09:28:15.778517 6602 storage/replica_raftstorage.go:726 [n1,s1,r17/5:{-}] applying Raft snapshot at index 49 (id=901a1124, encoded size=12107, 1 rocksdb batches, 39 log entries)
I180308 09:28:16.177052 6521 storage/store.go:2350 [replicaGC,n4,s4,r5/?:/System/tsd{-/cr.nod…}] removing replica
I180308 09:28:16.294600 6521 storage/replica.go:836 [replicaGC,n4,s4,r5/?:/System/tsd{-/cr.nod…}] removed 139 (128+11) keys in 117ms [clear=0ms commit=116ms]
W180308 09:28:16.347616 1553 storage/node_liveness.go:501 [n2,hb] slow heartbeat took 1.3s
I180308 09:28:16.372076 994 storage/replica_proposal.go:201 [n1,s1,r25/1:/System/{StatusNo…-tsd}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501254.079920771,0
I180308 09:28:16.421976 5631 storage/store.go:2350 [replicaGC,n3,s3,r10/2:/Table/13{-/1/2018-…}] removing replica
I180308 09:28:16.665522 1797 server/status/runtime.go:219 [n3] runtime stats: 1.3 GiB RSS, 1171 goroutines, 88 MiB/20 MiB/138 MiB GO alloc/idle/total, 162 MiB/196 MiB CGO alloc/total, 884.78cgo/sec, 0.89/0.07 %(u/s)time, 0.02 %gc (3x)
I180308 09:28:16.681548 6668 storage/replica_command.go:838 [split,n1,s1,r28/1:/System/{StatusNo…-tsd}] initiating a split of this range at key /System/StatusNode/5 [r30]
I180308 09:28:16.801652 6602 storage/replica_raftstorage.go:732 [n1,s1,r17/5:/Table/2{0-1}] applied Raft snapshot in 1003ms [clear=2ms batch=0ms entries=991ms commit=8ms]
I180308 09:28:16.806835 5631 storage/replica.go:836 [replicaGC,n3,s3,r10/2:/Table/13{-/1/2018-…}] removed 181 (169+12) keys in 384ms [clear=0ms commit=384ms]
W180308 09:28:17.412329 1806 storage/node_liveness.go:501 [n3,hb] slow heartbeat took 1.5s
I180308 09:28:17.506774 6629 storage/replica_consistency.go:129 [replica consistency checker,n1,s1,r21/1:/System/tsd/cr.node.{go…-sq…}] triggering stats recomputation to resolve delta of {ContainsEstimates:true LastUpdateNanos:1520501288989219219 IntentAge:0 GCBytesAge:0 LiveBytes:4092 LiveCount:0 KeyBytes:0 KeyCount:0 ValBytes:4092 ValCount:0 IntentBytes:0 IntentCount:0 SysBytes:0 SysCount:0}
I180308 09:28:17.510115 2214 storage/replica_proposal.go:201 [n4,s4,r41/2:/{Table/50-Max}] new range lease repl=(n4,s4):2 seq=3 start=1520501251.624682287,1 epo=1 pro=1520501255.402825609,0 following repl=(n4,s4):2 seq=3 start=1520501251.624682287,1 epo=1 pro=1520501255.402825609,0
I180308 09:28:17.785167 2044 server/status/runtime.go:219 [n4] runtime stats: 1.3 GiB RSS, 1172 goroutines, 62 MiB/44 MiB/138 MiB GO alloc/idle/total, 162 MiB/195 MiB CGO alloc/total, 981.39cgo/sec, 0.90/0.07 %(u/s)time, 0.02 %gc (4x)
I180308 09:28:17.789384 6779 storage/replica.go:3209 [n2,s2,r2/3:/System/{-NodeLive…}] proposing ADD_REPLICA((n5,s5):4): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3 (n5,s5):4] next=5
W180308 09:28:17.792842 6603 storage/replica.go:3016 [n4,s4,r41/2:/Table/5{0-1}] context canceled before proposing: 1 HeartbeatTxn
W180308 09:28:18.718662 1096 storage/node_liveness.go:501 [n1,hb] slow heartbeat took 2.0s
I180308 09:28:18.788824 6931 storage/replica_command.go:1760 [replicate,n5,s5,r17/4:/Table/2{0-1}] change replicas (REMOVE_REPLICA (n1,s1):5): read existing descriptor r17:/Table/2{0-1} [(n5,s5):4, (n3,s3):2, (n2,s2):3, (n1,s1):5, next=6]
I180308 09:28:19.053142 1010 storage/replica_proposal.go:201 [n1,s1,r27/1:/System/ts{d/cr.no…-e}] new range lease repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0 following repl=(n1,s1):1 seq=3 start=1520501242.569793796,0 epo=1 pro=1520501251.298975011,0
E180308 09:28:19.095691 5732 sql/logictest/logic.go:2124
testdata/parallel_test/subquery_retry_multinode/txn:9:
expected success, but found
pq: TransactionStatusError: does not exist
--- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 0 tests, 1 failures
I180308 09:28:19.153442 5732 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 0 tests, 1 failures
E180308 09:28:19.153715 5732 sql/logictest/parallel_test.go:84 error processing testdata/parallel_test/subquery_retry_multinode/txn: pq: TransactionStatusError: does not exist
I180308 09:28:19.146805 2484 server/status/runtime.go:219 [n5] runtime stats: 1.3 GiB RSS, 1162 goroutines, 95 MiB/13 MiB/138 MiB GO alloc/idle/total, 161 MiB/196 MiB CGO alloc/total, 1029.88cgo/sec, 0.90/0.08 %(u/s)time, 0.02 %gc (3x)
W180308 09:28:19.277944 2147 storage/node_liveness.go:501 [n4,hb] slow heartbeat took 2.0s
I180308 09:28:19.513328 7000 storage/replica_command.go:838 [split,n1,s1,r29/1:/System/ts{d/cr.no…-e}] initiating a split of this range at key /System/tsd/cr.node.sql.mem.admin.session.max-p99.9/1/10s/2018-03-08T09:00:00Z [r31]
I180308 09:28:19.527756 6832 storage/replica.go:3209 [n1,s1,r5/1:/System/tsd{-/cr.nod…}] proposing ADD_REPLICA((n4,s4):4): updated=[(n1,s1):1 (n2,s2):2 (n5,s5):3 (n4,s4):4] next=5
I180308 09:28:20.040589 7039 storage/replica_command.go:1760 [replicate,n1,s1,r24/1:/Table/13/1/2018-03-08T09:27:…] change replicas (REMOVE_REPLICA (n2,s2):3): read existing descriptor r24:/Table/13/1/2018-03-08T09:27:{29.777287Z/329091808619495425-40.666166Z/329091846156877826} [(n1,s1):1, (n3,s3):2, (n2,s2):3, (n5,s5):4, next=5]
W180308 09:28:20.196189 7047 storage/replica_backpressure.go:135 [n1,s1,r29/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.txn.aborts/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p90/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p75/1/10s/2018-03-08T09:00:00Z,/Min), ... 330 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/1/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/1/10s/2018-03-08T09:00:00Z,/Min)
I180308 09:28:20.209629 7061 storage/replica_raftstorage.go:520 [raftsnapshot,n1,s1,r5/1:/System/tsd{-/cr.nod…}] generated Raft snapshot a58aa6bc at index 74
I180308 09:28:20.318346 7061 storage/store.go:3745 [raftsnapshot,n1,s1,r5/1:/System/tsd{-/cr.nod…}] streamed snapshot to (n4,s4):4: kv pairs: 145, log entries: 19, rate-limit: 8.0 MiB/sec, 103ms
I180308 09:28:20.321961 7122 storage/replica_raftstorage.go:726 [n4,s4,r5/4:{-}] applying Raft snapshot at index 74 (id=a58aa6bc, encoded size=39489, 1 rocksdb batches, 19 log entries)
--- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries
I180308 09:28:20.675686 5734 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries
--- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures
I180308 09:28:20.676049 5734 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures
W180308 09:28:20.841923 7144 storage/replica_backpressure.go:135 [n1,s1,r29/1:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.txn.aborts/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p90/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p75/3/10s/2018-03-08T09:00:00Z,/Min), ... 330 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/3/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/3/10s/2018-03-08T09:00:00Z,/Min)
W180308 09:28:21.168305 2493 storage/node_liveness.go:501 [n5,hb] slow heartbeat took 2.5s
W180308 09:28:21.335806 1028 storage/store.go:3866 [n1,s1] handle raft ready: 0.6s [processed=0]
W180308 09:28:21.372717 7073 storage/replica_backpressure.go:135 [n5,s5,r29/3:/System/ts{d/cr.no…-e}] applying backpressure to limit range growth on batch Merge [/System/tsd/cr.node.txn.aborts/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.commits1PC/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.autoretries/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.abandons/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p90/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p75/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.durations-p50/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-max/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.999/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99.9/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p99/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p90/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.node.txn.restarts-p75/2/10s/2018-03-08T09:00:00Z,/Min), ... 330 skipped ..., Merge [/System/tsd/cr.store.queue.replicate.removereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.removedeadreplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.rebalancereplica/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicate.transferlease/2/10s/2018-03-08T09:00:00Z,/Min), Merge [/System/tsd/cr.store.queue.replicagc.removereplica/2/10s/2018-03-08T09:00:00Z,/Min)
W180308 09:28:21.566361 1553 storage/node_liveness.go:501 [n2,hb] slow heartbeat took 2.0s
I180308 09:28:21.934342 7122 storage/replica_raftstorage.go:732 [n4,s4,r5/4:/System/tsd{-/cr.nod…}] applied Raft snapshot in 1612ms [clear=0ms batch=0ms entries=1609ms commit=2ms]
--- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries
I180308 09:28:22.172925 5733 sql/logictest/logic.go:796 --- progress: testdata/parallel_test/subquery_retry_multinode/txn: 1 statements/queries
--- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures
I180308 09:28:22.173248 5733 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/txn with config : 1 tests, 0 failures
I180308 09:28:22.173462 835 sql/logictest/parallel_test.go:159 testdata/parallel_test/subquery_retry_multinode/test.yaml: run list 2: 0:final
W180308 09:28:22.273947 1096 storage/node_liveness.go:501 [n1,hb] slow heartbeat took 1.1s
W180308 09:28:22.317366 1806 storage/node_liveness.go:501 [n3,hb] slow heartbeat took 1.9s
--- done: testdata/parallel_test/subquery_retry_multinode/final with config : 2 tests, 0 failures
I180308 09:28:22.631831 7227 sql/logictest/logic.go:796 --- done: testdata/parallel_test/subquery_retry_multinode/final with config : 2 tests, 0 failures
I180308 09:28:22.633272 7200 util/stop/stopper.go:473 quiescing; tasks left:
1 ts.poller: poll
I180308 09:28:22.636235 7199 util/stop/stopper.go:473 quiescing; tasks left:
1 ts.poller: poll
W180308 09:28:22.636697 3779 storage/raft_transport.go:465 [n3] raft transport stream to node 4 failed: rpc error: code = Unavailable desc = transport is closing
I180308 09:28:22.637631 7198 util/stop/stopper.go:473 quiescing; tasks left:
6 node.Node: batch
3 [async] kv.TxnCoordSender: heartbeat loop
2 [async] storage.split: processing replica
1 ts.poller: poll
1 [async] transport racer
1 [async] storage.replicate: processing replica
W180308 09:28:22.638338 2119 storage/raft_transport.go:465 [n1] raft transport stream to node 2 failed: EOF
W180308 09:28:22.645863 7000 internal/client/txn.go:531 [split,n1,s1,r29/1:/System/ts{d/cr.no…-e}] failure aborting transaction: node unavailable; try another peer; abort caused by: node unavailable; try another peer
W180308 09:28:22.658964 1047 storage/store.go:1480 [n1,s1,r7/1:/Table/{SystemCon…-11}] could not gossip system config: node unavailable; try another peer
W180308 09:28:22.659676 2336 storage/raft_transport.go:465 [n1] raft transport stream to node 4 failed: rpc error: code = Unavailable desc = transport is closing
I180308 09:28:22.663545 7198 util/stop/stopper.go:473 quiescing; tasks left:
3 node.Node: batch
3 [async] kv.TxnCoordSender: heartbeat loop
1 ts.poller: poll
1 [async] transport racer
1 [async] storage.split: processing replica
1 [async] storage.replicate: processing replica
W180308 09:28:22.666411 6668 internal/client/txn.go:531 [split,n1,s1,r28/1:/System/{StatusNo…-tsd}] failure aborting transaction: node unavailable; try another peer; abort caused by: result is ambiguous (server shutdown)
I180308 09:28:22.669226 7198 util/stop/stopper.go:473 quiescing; tasks left:
3 [async] kv.TxnCoordSender: heartbeat loop
1 ts.poller: poll
1 node.Node: batch
1 [async] transport racer
1 [async] storage.split: processing replica
1 [async] storage.replicate: processing replica
W180308 09:28:22.670449 1545 ts/db.go:154 [n2,ts-poll] error writing time series data: the batch experienced mixed success and failure
W180308 09:28:22.679032 1545 ts/db.go:157 [n2,ts-poll] node unavailable; try another peer
I180308 09:28:22.679527 7198 util/stop/stopper.go:473 quiescing; tasks left:
3 [async] kv.TxnCoordSender: heartbeat loop
1 ts.poller: poll
1 node.Node: batch
1 [async] transport racer
1 [async] storage.replicate: processing replica
I180308 09:28:22.705284 1055 server/status/runtime.go:219 [n1] runtime stats: 1.3 GiB RSS, 1118 goroutines, 67 MiB/42 MiB/140 MiB GO alloc/idle/total, 158 MiB/197 MiB CGO alloc/total, 881.83cgo/sec, 0.89/0.09 %(u/s)time, 0.03 %gc (4x)
W180308 09:28:22.706902 1048 storage/store.go:1480 [n1,s1,r3/1:/System/NodeLiveness{-Max}] could not gossip node liveness: node unavailable; try another peer
I180308 09:28:22.707664 7201 util/stop/stopper.go:473 quiescing; tasks left:
1 ts.poller: poll
I180308 09:28:22.710501 7330 util/stop/stopper.go:473 quiescing; tasks left:
1 ts.poller: poll
1 [async] storage.replicate: processing replica
1 [async] kv.TxnCoordSender: heartbeat loop
W180308 09:28:22.711733 1047 storage/store.go:1480 [n1,s1,r7/1:/Table/{SystemCon…-11}] could not gossip system config: node unavailable; try another peer
W180308 09:28:22.716304 1046 gossip/infostore.go:303 [n1] node unavailable; try another peer
W180308 09:28:22.716688 1046 storage/store.go:1480 [n1,s1,r1/1:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer
I180308 09:28:22.717708 1053 gossip/gossip.go:487 [n1] gossip status (ok, 5 nodes)
gossip client (0/3 cur/max conns)
gossip server (3/3 cur/max conns, infos 2303/150 sent/received, bytes 774296B/46012B sent/received)
3: 127.0.0.1:46629 (56s)
4: 127.0.0.1:38315 (55s)
2: 127.0.0.1:37987 (57s)
W180308 09:28:22.720545 1053 gossip/infostore.go:303 [n1] node unavailable; try another peer
W180308 09:28:22.722669 2373 storage/raft_transport.go:465 [n4] raft transport stream to node 1 failed: rpc error: code = Unavailable desc = transport is closing
W180308 09:28:22.727955 2485 ts/db.go:154 [n5,ts-poll] error writing time series data: the batch experienced mixed success and failure
W180308 09:28:22.728252 2485 ts/db.go:157 [n5,ts-poll] node unavailable; try another peer
I180308 09:28:22.728512 7330 util/stop/stopper.go:473 quiescing; tasks left:
1 [async] storage.replicate: processing replica
1 [async] kv.TxnCoordSender: heartbeat loop
W180308 09:28:22.732195 2770 storage/raft_transport.go:465 [n1] raft transport stream to node 5 failed: rpc error: code = Unavailable desc = transport is closing
W180308 09:28:22.737226 4045 storage/raft_transport.go:465 [n2] raft transport stream to node 3 failed: EOF
W180308 09:28:22.742157 1899 storage/raft_transport.go:465 [n1] raft transport stream to node 3 failed: rpc error: code = Unavailable desc = transport is closing
W180308 09:28:22.747260 4042 storage/raft_transport.go:465 [n3] raft transport stream to node 2 failed: rpc error: code = FailedPrecondition desc = grpc: the client connection is closing
W180308 09:28:22.760150 1970 storage/raft_transport.go:465 [n3] raft transport stream to node 1 failed: rpc error: code = Unavailable desc = transport is closing
W180308 09:28:22.763583 2045 ts/db.go:154 [n4,ts-poll] error writing time series data: the batch experienced mixed success and failure
W180308 09:28:22.770300 2045 ts/db.go:157 [n4,ts-poll] node unavailable; try another peer
I180308 09:28:22.763970 750 kv/transport_race.go:67 transport race promotion: ran 45 iterations on up to 2482 requests
W180308 09:28:22.764917 3208 storage/raft_transport.go:465 [n4] raft transport stream to node 5 failed: rpc error: code = FailedPrecondition desc = grpc: the client connection is closing
I180308 09:28:22.772524 7198 util/stop/stopper.go:473 quiescing; tasks left:
3 [async] kv.TxnCoordSender: heartbeat loop
1 ts.poller: poll
1 node.Node: batch
1 [async] storage.replicate: processing replica
W180308 09:28:22.775756 1518 gossip/gossip.go:1292 [n2] no incoming or outgoing connections
W180308 09:28:22.776260 6938 kv/txn_coord_sender.go:1036 [split,s1,r28/1:/System/{StatusNo…-tsd},n1,txn=fcc1998d] node unavailable; try another peer
W180308 09:28:22.777253 3848 storage/raft_transport.go:465 [n4] raft transport stream to node 2 failed: rpc error: code = Unavailable desc = transport is closing
W180308 09:28:22.792605 2178 storage/raft_transport.go:465 [n2] raft transport stream to node 1 failed: rpc error: code = Unavailable desc = transport is closing
W180308 09:28:22.793349 3922 storage/raft_transport.go:465 [n2] raft transport stream to node 4 failed: rpc error: code = Unavailable desc = transport is closing
W180308 09:28:22.803019 1056 ts/db.go:154 [n1,ts-poll] error writing time series data: the batch experienced mixed success and failure
W180308 09:28:22.812175 1056 ts/db.go:157 [n1,ts-poll] node unavailable; try another peer
W180308 09:28:22.803505 1046 gossip/infostore.go:303 [n1] node unavailable; try another peer
W180308 09:28:22.812587 1046 storage/store.go:1480 [n1,s1,r1/1:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer
W180308 09:28:22.807111 1048 storage/store.go:1480 [n1,s1,r3/1:/System/NodeLiveness{-Max}] could not gossip node liveness: node unavailable; try another peer
W180308 09:28:22.807396 1057 server/node.go:776 [n1,summaries] error recording status summaries: node unavailable; try another peer
W180308 09:28:22.808727 7039 internal/client/txn.go:531 [replicate,n1,s1,r24/1:/Table/13/1/2018-03-08T09:27:…] failure aborting transaction: node unavailable; try another peer; abort caused by: node unavailable; try another peer
E180308 09:28:22.813785 7039 storage/queue.go:778 [replicate,n1,s1,r24/1:/Table/13/1/2018-03-08T09:27:…] change replicas of r24 failed: node unavailable; try another peer
W180308 09:28:22.809209 3173 storage/raft_transport.go:465 [n2] raft transport stream to node 5 failed: rpc error: code = Unavailable desc = transport is closing
W180308 09:28:22.815172 7191 kv/txn_coord_sender.go:1036 [split,s1,r29/1:/System/ts{d/cr.no…-e},n1,txn=db46a2ed] node unavailable; try another peer
W180308 09:28:22.818081 6931 internal/client/txn.go:531 [replicate,n5,s5,r17/4:/Table/2{0-1}] failure aborting transaction: node unavailable; try another peer; abort caused by: node unavailable; try another peer
W180308 09:28:22.820510 2743 storage/raft_transport.go:465 [n5] raft transport stream to node 1 failed: EOF
W180308 09:28:22.824887 2238 gossip/gossip.go:1292 [n5] no incoming or outgoing connections
W180308 09:28:22.825931 1047 storage/store.go:1480 [n1,s1,r7/1:/Table/{SystemCon…-11}] could not gossip system config: node unavailable; try another peer
W180308 09:28:22.835973 3199 storage/raft_transport.go:465 [n5] raft transport stream to node 4 failed: EOF
W180308 09:28:22.837525 3335 storage/raft_transport.go:465 [n5] raft transport stream to node 3 failed: EOF
W180308 09:28:22.842099 1748 gossip/gossip.go:1292 [n3] no incoming or outgoing connections
W180308 09:28:22.842893 3245 storage/raft_transport.go:465 [n3] raft transport stream to node 5 failed: rpc error: code = FailedPrecondition desc = grpc: the client connection is closing
W180308 09:28:22.845802 3832 storage/raft_transport.go:465 [n4] raft transport stream to node 3 failed: rpc error: code = Unavailable desc = transport is closing
W180308 09:28:22.853934 1798 ts/db.go:154 [n3,ts-poll] error writing time series data: the batch experienced mixed success and failure
W180308 09:28:22.855078 1979 gossip/gossip.go:1292 [n4] no incoming or outgoing connections
W180308 09:28:22.856870 3102 storage/raft_transport.go:465 [n5] raft transport stream to node 2 failed: EOF
W180308 09:28:22.890039 1798 ts/db.go:157 [n3,ts-poll] node unavailable; try another peer
I180308 09:28:23.130857 6 sql/logictest/parallel_test.go:266 1 out of 2 parallel tests failed
``` | non_priority | sql logictest testparallel failed under stress sha parameters tags goflags race stress build found a failed test storage replica raftstorage go applied preemptive snapshot in testutils testcluster testcluster go has underreplicated ranges storage store go handle raft ready storage replica go proposing add replica updated next storage replica command go change replicas add replica read existing descriptor table storage node liveness go slow heartbeat took testutils testcluster testcluster go has underreplicated ranges storage replica raftstorage go generated preemptive snapshot at index storage store go removing replica storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica go removed keys in storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica go proposing add replica updated next gossip gossip go node has connected to cluster via gossip storage stores go wrote node addresses to persistent storage storage replica go proposing remove replica updated next storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica command go change replicas remove replica read existing descriptor table storage replica raftstorage go generated preemptive snapshot at index sql logictest parallel test go setting range split size storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage store go added to replica gc queue peer suggestion storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor system tse table systemconfigspan start storage replica raftstorage go generated raft snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying raft snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied raft snapshot in storage store go removing replica storage store go removing replica storage replica go removed keys in storage replica go removed keys in storage replica go proposing add replica updated next server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage replica command go change replicas remove replica read existing descriptor table storage replica go proposing add replica updated next storage replica command go change replicas remove replica read existing descriptor system tse table systemconfigspan start storage replica go proposing remove replica updated next sql logictest parallel test go creating database gossip gossip go node has connected to cluster via gossip storage stores go wrote node addresses to persistent storage storage store go added to replica gc queue peer suggestion storage store go added to replica gc queue peer suggestion storage replica command go initiating a split of this range at key system tsd cr node gossip infos sent storage replica command go initiating a split of this range at key system statusnode storage replica command go initiating a split of this range at key table storage replica go proposing add replica updated next storage store go removing replica storage replica go removed keys in storage replica go context canceled before proposing heartbeattxn storage replica command go initiating a split of this range at key table storage replica go proposing remove replica updated next server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage store go added to replica gc queue peer suggestion sql event log go event create database target info databasename test statement create database test user root storage store go added to replica gc queue peer suggestion storage replica go not quiescing pending commands storage store go added to replica gc queue peer suggestion storage store go added to replica gc queue peer suggestion storage store go removing replica server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage replica go removed keys in storage replica command go initiating a split of this range at key table storage replica raftstorage go generated raft snapshot at index storage store go handle raft ready server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying raft snapshot at index id encoded size rocksdb batches log entries storage node liveness go slow heartbeat took storage replica go proposing remove replica updated next server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage store go added to replica gc queue peer suggestion storage store go added to replica gc queue peer suggestion storage node liveness go slow heartbeat took sql logictest parallel test go test setup done sql logictest parallel test go testdata parallel test subquery retry multinode test yaml run list setup storage node liveness go slow heartbeat took storage store go removing replica storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica command go initiating a split of this range at key system statusnode storage replica go removed keys in storage node liveness go slow heartbeat took storage replica backpressure go applying backpressure to limit range growth on batch merge system tsd cr node security certificate expiration ca min merge system tsd cr node security certificate expiration node min merge system tsd cr node gossip connections incoming min merge system tsd cr node gossip connections refused min merge system tsd cr node gossip bytes received min merge system tsd cr node gossip bytes sent min merge system tsd cr node gossip infos received min merge system tsd cr node gossip infos sent min merge system tsd cr node gossip connections outgoing min merge system tsd cr node distsender batches min merge system tsd cr node distsender batches partial min merge system tsd cr node distsender rpc sent min merge system tsd cr node distsender rpc sent local min merge system tsd cr node distsender rpc sent nextreplicaerror min merge system tsd cr node distsender errors notleaseholder min merge system tsd cr node requests slow distsender min merge system tsd cr node txn aborts min merge system tsd cr node txn commits min merge system tsd cr node txn min merge system tsd cr node txn autoretries min skipped merge system tsd cr store queue replicate removereplica min merge system tsd cr store queue replicate removedeadreplica min merge system tsd cr store queue replicate rebalancereplica min merge system tsd cr store queue replicate transferlease min merge system tsd cr store queue replicagc removereplica min storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica command go initiating a split of this range at key table storage replica raftstorage go applied raft snapshot in storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica go context canceled before proposing heartbeattxn storage replica command go initiating a split of this range at key system tsd cr node sql bytesin storage replica command go change replicas remove replica read existing descriptor table storage replica raftstorage go generated preemptive snapshot at index storage replica raftstorage go generated preemptive snapshot at index storage queue go raft status not initialized sql event log go event create table target info tablename test public t statement create table t k int filler string user root progress testdata parallel test subquery retry multinode setup statements queries sql logictest logic go progress testdata parallel test subquery retry multinode setup statements queries storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica command go initiating a split of this range at key table gossip gossip go node has connected to cluster via gossip server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage stores go wrote node addresses to persistent storage storage replica raftstorage go applied preemptive snapshot in storage replica backpressure go applying backpressure to limit range growth on batch merge system tsd cr node security certificate expiration ca min merge system tsd cr node security certificate expiration node min merge system tsd cr node requests slow distsender min merge system tsd cr node txn aborts min merge system tsd cr node txn commits min merge system tsd cr node txn min merge system tsd cr node txn autoretries min merge system tsd cr node txn abandons min merge system tsd cr node txn durations max min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn restarts max min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min skipped merge system tsd cr store queue replicate removereplica min merge system tsd cr store queue replicate removedeadreplica min merge system tsd cr store queue replicate rebalancereplica min merge system tsd cr store queue replicate transferlease min merge system tsd cr store queue replicagc removereplica min storage replica command go change replicas add replica read existing descriptor table storage replica backpressure go applying backpressure to limit range growth on batch merge system tsd cr node security certificate expiration ca min merge system tsd cr node security certificate expiration node min merge system tsd cr node gossip infos sent min merge system tsd cr node requests slow distsender min merge system tsd cr node txn aborts min merge system tsd cr node txn commits min merge system tsd cr node txn min merge system tsd cr node txn autoretries min merge system tsd cr node txn abandons min merge system tsd cr node txn durations max min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn restarts max min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min skipped merge system tsd cr store queue replicate removereplica min merge system tsd cr store queue replicate removedeadreplica min merge system tsd cr store queue replicate rebalancereplica min merge system tsd cr store queue replicate transferlease min merge system tsd cr store queue replicagc removereplica min storage replica backpressure go applying backpressure to limit range growth on batch merge system tsd cr node security certificate expiration ca min merge system tsd cr node security certificate expiration node min merge system tsd cr node gossip infos sent min merge system tsd cr node requests slow distsender min merge system tsd cr node txn aborts min merge system tsd cr node txn commits min merge system tsd cr node txn min merge system tsd cr node txn autoretries min merge system tsd cr node txn abandons min merge system tsd cr node txn durations max min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn restarts max min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min skipped merge system tsd cr store queue replicate removereplica min merge system tsd cr store queue replicate removedeadreplica min merge system tsd cr store queue replicate rebalancereplica min merge system tsd cr store queue replicate transferlease min merge system tsd cr store queue replicagc removereplica min progress testdata parallel test subquery retry multinode setup statements queries sql logictest logic go progress testdata parallel test subquery retry multinode setup statements queries done testdata parallel test subquery retry multinode setup with config tests failures sql logictest logic go done testdata parallel test subquery retry multinode setup with config tests failures sql logictest parallel test go testdata parallel test subquery retry multinode test yaml run list txn txn txn txn txn storage replica go proposing remove replica updated next storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage store go added to replica gc queue peer suggestion storage replica command go initiating a split of this range at key system statusnode server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica go context canceled before proposing heartbeattxn progress testdata parallel test subquery retry multinode txn statements queries sql logictest logic go progress testdata parallel test subquery retry multinode txn statements queries done testdata parallel test subquery retry multinode txn with config tests failures sql logictest logic go done testdata parallel test subquery retry multinode txn with config tests failures server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage node liveness go slow heartbeat took storage node liveness go slow heartbeat took server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica command go initiating a split of this range at key system tsd cr node sql distsql service latency storage store go handle raft ready server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage store go handle raft ready storage store go handle raft ready storage replica raftstorage go generated preemptive snapshot at index storage node liveness go slow heartbeat took storage store go removing replica storage replica go removed keys in storage replica raftstorage go generated preemptive snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica backpressure go applying backpressure to limit range growth on batch merge system tsd cr node txn aborts min merge system tsd cr node txn commits min merge system tsd cr node txn min merge system tsd cr node txn autoretries min merge system tsd cr node txn abandons min merge system tsd cr node txn durations max min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn restarts max min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min skipped merge system tsd cr store queue replicate removereplica min merge system tsd cr store queue replicate removedeadreplica min merge system tsd cr store queue replicate rebalancereplica min merge system tsd cr store queue replicate transferlease min merge system tsd cr store queue replicagc removereplica min storage store go handle raft ready storage node liveness go slow heartbeat took storage replica backpressure go applying backpressure to limit range growth on batch merge system tsd cr node txn aborts min merge system tsd cr node txn commits min merge system tsd cr node txn min merge system tsd cr node txn autoretries min merge system tsd cr node txn abandons min merge system tsd cr node txn durations max min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn restarts max min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min skipped merge system tsd cr store queue replicate removereplica min merge system tsd cr store queue replicate removedeadreplica min merge system tsd cr store queue replicate rebalancereplica min merge system tsd cr store queue replicate transferlease min merge system tsd cr store queue replicagc removereplica min storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage node liveness go slow heartbeat took storage node liveness go slow heartbeat took storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor system tsd cr node gossip infos sent storage node liveness go slow heartbeat took server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor system nodeliveness storage node liveness go slow heartbeat took storage replica go proposing add replica updated next progress testdata parallel test subquery retry multinode txn statements queries sql logictest logic go progress testdata parallel test subquery retry multinode txn statements queries done testdata parallel test subquery retry multinode txn with config tests failures sql logictest logic go done testdata parallel test subquery retry multinode txn with config tests failures storage node liveness go slow heartbeat took storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica raftstorage go generated raft snapshot at index server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying raft snapshot at index id encoded size rocksdb batches log entries storage store go removing replica storage replica go removed keys in storage node liveness go slow heartbeat took storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage store go removing replica server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage replica command go initiating a split of this range at key system statusnode storage replica raftstorage go applied raft snapshot in storage replica go removed keys in storage node liveness go slow heartbeat took storage replica consistency go triggering stats recomputation to resolve delta of containsestimates true lastupdatenanos intentage gcbytesage livebytes livecount keybytes keycount valbytes valcount intentbytes intentcount sysbytes syscount storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage replica go proposing add replica updated next storage replica go context canceled before proposing heartbeattxn storage node liveness go slow heartbeat took storage replica command go change replicas remove replica read existing descriptor table storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro sql logictest logic go testdata parallel test subquery retry multinode txn expected success but found pq transactionstatuserror does not exist done testdata parallel test subquery retry multinode txn with config tests failures sql logictest logic go done testdata parallel test subquery retry multinode txn with config tests failures sql logictest parallel test go error processing testdata parallel test subquery retry multinode txn pq transactionstatuserror does not exist server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage node liveness go slow heartbeat took storage replica command go initiating a split of this range at key system tsd cr node sql mem admin session max storage replica go proposing add replica updated next storage replica command go change replicas remove replica read existing descriptor table storage replica backpressure go applying backpressure to limit range growth on batch merge system tsd cr node txn aborts min merge system tsd cr node txn commits min merge system tsd cr node txn min merge system tsd cr node txn autoretries min merge system tsd cr node txn abandons min merge system tsd cr node txn durations max min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn restarts max min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min skipped merge system tsd cr store queue replicate removereplica min merge system tsd cr store queue replicate removedeadreplica min merge system tsd cr store queue replicate rebalancereplica min merge system tsd cr store queue replicate transferlease min merge system tsd cr store queue replicagc removereplica min storage replica raftstorage go generated raft snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying raft snapshot at index id encoded size rocksdb batches log entries progress testdata parallel test subquery retry multinode txn statements queries sql logictest logic go progress testdata parallel test subquery retry multinode txn statements queries done testdata parallel test subquery retry multinode txn with config tests failures sql logictest logic go done testdata parallel test subquery retry multinode txn with config tests failures storage replica backpressure go applying backpressure to limit range growth on batch merge system tsd cr node txn aborts min merge system tsd cr node txn commits min merge system tsd cr node txn min merge system tsd cr node txn autoretries min merge system tsd cr node txn abandons min merge system tsd cr node txn durations max min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn restarts max min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min skipped merge system tsd cr store queue replicate removereplica min merge system tsd cr store queue replicate removedeadreplica min merge system tsd cr store queue replicate rebalancereplica min merge system tsd cr store queue replicate transferlease min merge system tsd cr store queue replicagc removereplica min storage node liveness go slow heartbeat took storage store go handle raft ready storage replica backpressure go applying backpressure to limit range growth on batch merge system tsd cr node txn aborts min merge system tsd cr node txn commits min merge system tsd cr node txn min merge system tsd cr node txn autoretries min merge system tsd cr node txn abandons min merge system tsd cr node txn durations max min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn durations min merge system tsd cr node txn restarts max min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min merge system tsd cr node txn restarts min skipped merge system tsd cr store queue replicate removereplica min merge system tsd cr store queue replicate removedeadreplica min merge system tsd cr store queue replicate rebalancereplica min merge system tsd cr store queue replicate transferlease min merge system tsd cr store queue replicagc removereplica min storage node liveness go slow heartbeat took storage replica raftstorage go applied raft snapshot in progress testdata parallel test subquery retry multinode txn statements queries sql logictest logic go progress testdata parallel test subquery retry multinode txn statements queries done testdata parallel test subquery retry multinode txn with config tests failures sql logictest logic go done testdata parallel test subquery retry multinode txn with config tests failures sql logictest parallel test go testdata parallel test subquery retry multinode test yaml run list final storage node liveness go slow heartbeat took storage node liveness go slow heartbeat took done testdata parallel test subquery retry multinode final with config tests failures sql logictest logic go done testdata parallel test subquery retry multinode final with config tests failures util stop stopper go quiescing tasks left ts poller poll util stop stopper go quiescing tasks left ts poller poll storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing util stop stopper go quiescing tasks left node node batch kv txncoordsender heartbeat loop storage split processing replica ts poller poll transport racer storage replicate processing replica storage raft transport go raft transport stream to node failed eof internal client txn go failure aborting transaction node unavailable try another peer abort caused by node unavailable try another peer storage store go could not gossip system config node unavailable try another peer storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing util stop stopper go quiescing tasks left node node batch kv txncoordsender heartbeat loop ts poller poll transport racer storage split processing replica storage replicate processing replica internal client txn go failure aborting transaction node unavailable try another peer abort caused by result is ambiguous server shutdown util stop stopper go quiescing tasks left kv txncoordsender heartbeat loop ts poller poll node node batch transport racer storage split processing replica storage replicate processing replica ts db go error writing time series data the batch experienced mixed success and failure ts db go node unavailable try another peer util stop stopper go quiescing tasks left kv txncoordsender heartbeat loop ts poller poll node node batch transport racer storage replicate processing replica server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage store go could not gossip node liveness node unavailable try another peer util stop stopper go quiescing tasks left ts poller poll util stop stopper go quiescing tasks left ts poller poll storage replicate processing replica kv txncoordsender heartbeat loop storage store go could not gossip system config node unavailable try another peer gossip infostore go node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer gossip gossip go gossip status ok nodes gossip client cur max conns gossip server cur max conns infos sent received bytes sent received gossip infostore go node unavailable try another peer storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing ts db go error writing time series data the batch experienced mixed success and failure ts db go node unavailable try another peer util stop stopper go quiescing tasks left storage replicate processing replica kv txncoordsender heartbeat loop storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing storage raft transport go raft transport stream to node failed eof storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing storage raft transport go raft transport stream to node failed rpc error code failedprecondition desc grpc the client connection is closing storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing ts db go error writing time series data the batch experienced mixed success and failure ts db go node unavailable try another peer kv transport race go transport race promotion ran iterations on up to requests storage raft transport go raft transport stream to node failed rpc error code failedprecondition desc grpc the client connection is closing util stop stopper go quiescing tasks left kv txncoordsender heartbeat loop ts poller poll node node batch storage replicate processing replica gossip gossip go no incoming or outgoing connections kv txn coord sender go node unavailable try another peer storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing ts db go error writing time series data the batch experienced mixed success and failure ts db go node unavailable try another peer gossip infostore go node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip node liveness node unavailable try another peer server node go error recording status summaries node unavailable try another peer internal client txn go failure aborting transaction node unavailable try another peer abort caused by node unavailable try another peer storage queue go change replicas of failed node unavailable try another peer storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing kv txn coord sender go node unavailable try another peer internal client txn go failure aborting transaction node unavailable try another peer abort caused by node unavailable try another peer storage raft transport go raft transport stream to node failed eof gossip gossip go no incoming or outgoing connections storage store go could not gossip system config node unavailable try another peer storage raft transport go raft transport stream to node failed eof storage raft transport go raft transport stream to node failed eof gossip gossip go no incoming or outgoing connections storage raft transport go raft transport stream to node failed rpc error code failedprecondition desc grpc the client connection is closing storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing ts db go error writing time series data the batch experienced mixed success and failure gossip gossip go no incoming or outgoing connections storage raft transport go raft transport stream to node failed eof ts db go node unavailable try another peer sql logictest parallel test go out of parallel tests failed | 0 |
296,575 | 25,560,204,902 | IssuesEvent | 2022-11-30 10:11:30 | saleor/saleor-dashboard | https://api.github.com/repos/saleor/saleor-dashboard | closed | Enable "noImplicitThis" in tsconfig | maitenance static tests | Enable rule `noImplicitThis` in tsconfig. If the rule needs massive changes in the code, please create separate smaller issues to solve it gradually. | 1.0 | Enable "noImplicitThis" in tsconfig - Enable rule `noImplicitThis` in tsconfig. If the rule needs massive changes in the code, please create separate smaller issues to solve it gradually. | non_priority | enable noimplicitthis in tsconfig enable rule noimplicitthis in tsconfig if the rule needs massive changes in the code please create separate smaller issues to solve it gradually | 0 |
84,908 | 10,420,980,834 | IssuesEvent | 2019-09-16 03:48:18 | tarantool/tarantool-java | https://api.github.com/repos/tarantool/tarantool-java | closed | List and briefly describe client and cluster client configuration options | documentation | I mean TarantoolClientConfig and TarantoolClusterClientConfig fields. It is okay to just write it in README. | 1.0 | List and briefly describe client and cluster client configuration options - I mean TarantoolClientConfig and TarantoolClusterClientConfig fields. It is okay to just write it in README. | non_priority | list and briefly describe client and cluster client configuration options i mean tarantoolclientconfig and tarantoolclusterclientconfig fields it is okay to just write it in readme | 0 |
10,927 | 9,194,954,949 | IssuesEvent | 2019-03-07 00:21:24 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | Role driven permissions structure | assigned-to-author azure-analysis-services/svc in-progress product-question triaged | Hi,
We have a tabular model on Azure SSAS and we are generating PowerBI reports based on that model. We want to build a dynamic steering of the access of cube data based on which user logs in. By that I mean that depending on which report the user runs that a different slice of the cube is performed(with the help of dax filters). I have created different roles in the cube but I noticed that because it is the effective name that is used to pull the data from the cube, the rights of the two different roles are combined when creating the power bi report. I decided then to create two different Azure AD groups(email security) and use the groups to differentiate the data pull from the cube. The group emails were added to both power bi app and the cube roles. The result however was again the same. When one powerbi report is created, mapped to only one of the Azure AD groups and respective cube roles, data from both roles where the user belongs to are still defining the cube slice. Is there any way for one user to get different slices of a cube by being assigned two different roles(with different dax filters) in the cube itself?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 3f0e1195-b6a2-d789-a950-4454798f265e
* Version Independent ID: d8ea7415-3b97-5985-44e7-1b89bf1196b6
* Content: [Manage database roles and users in Azure Analysis Services](https://docs.microsoft.com/en-us/azure/analysis-services/analysis-services-database-users#feedback)
* Content Source: [articles/analysis-services/analysis-services-database-users.md](https://github.com/Microsoft/azure-docs/blob/master/articles/analysis-services/analysis-services-database-users.md)
* Service: **azure-analysis-services**
* GitHub Login: @Minewiskan
* Microsoft Alias: **owend** | 1.0 | Role driven permissions structure - Hi,
We have a tabular model on Azure SSAS and we are generating PowerBI reports based on that model. We want to build a dynamic steering of the access of cube data based on which user logs in. By that I mean that depending on which report the user runs that a different slice of the cube is performed(with the help of dax filters). I have created different roles in the cube but I noticed that because it is the effective name that is used to pull the data from the cube, the rights of the two different roles are combined when creating the power bi report. I decided then to create two different Azure AD groups(email security) and use the groups to differentiate the data pull from the cube. The group emails were added to both power bi app and the cube roles. The result however was again the same. When one powerbi report is created, mapped to only one of the Azure AD groups and respective cube roles, data from both roles where the user belongs to are still defining the cube slice. Is there any way for one user to get different slices of a cube by being assigned two different roles(with different dax filters) in the cube itself?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 3f0e1195-b6a2-d789-a950-4454798f265e
* Version Independent ID: d8ea7415-3b97-5985-44e7-1b89bf1196b6
* Content: [Manage database roles and users in Azure Analysis Services](https://docs.microsoft.com/en-us/azure/analysis-services/analysis-services-database-users#feedback)
* Content Source: [articles/analysis-services/analysis-services-database-users.md](https://github.com/Microsoft/azure-docs/blob/master/articles/analysis-services/analysis-services-database-users.md)
* Service: **azure-analysis-services**
* GitHub Login: @Minewiskan
* Microsoft Alias: **owend** | non_priority | role driven permissions structure hi we have a tabular model on azure ssas and we are generating powerbi reports based on that model we want to build a dynamic steering of the access of cube data based on which user logs in by that i mean that depending on which report the user runs that a different slice of the cube is performed with the help of dax filters i have created different roles in the cube but i noticed that because it is the effective name that is used to pull the data from the cube the rights of the two different roles are combined when creating the power bi report i decided then to create two different azure ad groups email security and use the groups to differentiate the data pull from the cube the group emails were added to both power bi app and the cube roles the result however was again the same when one powerbi report is created mapped to only one of the azure ad groups and respective cube roles data from both roles where the user belongs to are still defining the cube slice is there any way for one user to get different slices of a cube by being assigned two different roles with different dax filters in the cube itself document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service azure analysis services github login minewiskan microsoft alias owend | 0 |
49,753 | 13,187,264,423 | IssuesEvent | 2020-08-13 02:51:56 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | [stochastics] remove minuit dependency (Trac #2007) | Incomplete Migration Migrated from Trac combo reconstruction defect | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2007">https://code.icecube.wisc.edu/ticket/2007</a>, reported by kjmeagher and owned by jgonzalez</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:14:55",
"description": "either replace it with the implementation of minuit in lilliput or use the gulliver framework",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1550067295757382",
"component": "combo reconstruction",
"summary": "[stochastics] remove minuit dependency",
"priority": "normal",
"keywords": "",
"time": "2017-05-09T17:07:35",
"milestone": "",
"owner": "jgonzalez",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [stochastics] remove minuit dependency (Trac #2007) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2007">https://code.icecube.wisc.edu/ticket/2007</a>, reported by kjmeagher and owned by jgonzalez</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:14:55",
"description": "either replace it with the implementation of minuit in lilliput or use the gulliver framework",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1550067295757382",
"component": "combo reconstruction",
"summary": "[stochastics] remove minuit dependency",
"priority": "normal",
"keywords": "",
"time": "2017-05-09T17:07:35",
"milestone": "",
"owner": "jgonzalez",
"type": "defect"
}
```
</p>
</details>
| non_priority | remove minuit dependency trac migrated from json status closed changetime description either replace it with the implementation of minuit in lilliput or use the gulliver framework reporter kjmeagher cc resolution fixed ts component combo reconstruction summary remove minuit dependency priority normal keywords time milestone owner jgonzalez type defect | 0 |
105,394 | 13,183,295,712 | IssuesEvent | 2020-08-12 17:15:19 | tektoncd/experimental | https://api.github.com/repos/tektoncd/experimental | closed | Add 'Repository' and 'Branch' filters to PipelineRuns | design-help-wanted webhooks | This issue is part of `experimental` epic #37 but concerns development changes in `dashboard`. The webhooks extension needs to be able to filter pipeline runs by repository and branch: it makes sense to surface this capability to dashboard users.
Under Dashboard > PipelineRuns, users should be able to filter the pipeline runs shown by 'git repository' and 'branch'. These can be passed as query params to the /pipelineruns REST API. We should apply labels to pipeline runs so as to be able to filter by these labels when obtaining pipeline runs via the Go client. | 1.0 | Add 'Repository' and 'Branch' filters to PipelineRuns - This issue is part of `experimental` epic #37 but concerns development changes in `dashboard`. The webhooks extension needs to be able to filter pipeline runs by repository and branch: it makes sense to surface this capability to dashboard users.
Under Dashboard > PipelineRuns, users should be able to filter the pipeline runs shown by 'git repository' and 'branch'. These can be passed as query params to the /pipelineruns REST API. We should apply labels to pipeline runs so as to be able to filter by these labels when obtaining pipeline runs via the Go client. | non_priority | add repository and branch filters to pipelineruns this issue is part of experimental epic but concerns development changes in dashboard the webhooks extension needs to be able to filter pipeline runs by repository and branch it makes sense to surface this capability to dashboard users under dashboard pipelineruns users should be able to filter the pipeline runs shown by git repository and branch these can be passed as query params to the pipelineruns rest api we should apply labels to pipeline runs so as to be able to filter by these labels when obtaining pipeline runs via the go client | 0 |
207,221 | 23,434,919,563 | IssuesEvent | 2022-08-15 08:44:44 | Gal-Doron/Baragon-36 | https://api.github.com/repos/Gal-Doron/Baragon-36 | opened | dropwizard-jetty-1.3.12.jar: 3 vulnerabilities (highest severity is: 5.3) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>dropwizard-jetty-1.3.12.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /BaragonService/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/Baragon-36/commit/3335ef04e9449f11036516e41533318fc21bd8a3">3335ef04e9449f11036516e41533318fc21bd8a3</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2020-27223](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27223) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | jetty-http-9.4.18.v20190429.jar | Transitive | 2.0.0-rc0+test8 | ✅ |
| [CVE-2021-28169](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28169) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | jetty-http-9.4.18.v20190429.jar | Transitive | 2.0.0-rc0+test8 | ✅ |
| [CVE-2022-2047](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2047) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 2.7 | jetty-http-9.4.18.v20190429.jar | Transitive | 2.0.0-rc0+test8 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-27223</summary>
### Vulnerable Library - <b>jetty-http-9.4.18.v20190429.jar</b></p>
<p>The Eclipse Jetty Project</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to dependency file: /BaragonData/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar</p>
<p>
Dependency Hierarchy:
- dropwizard-jetty-1.3.12.jar (Root Library)
- :x: **jetty-http-9.4.18.v20190429.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/Baragon-36/commit/3335ef04e9449f11036516e41533318fc21bd8a3">3335ef04e9449f11036516e41533318fc21bd8a3</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Eclipse Jetty 9.4.6.v20170531 to 9.4.36.v20210114 (inclusive), 10.0.0, and 11.0.0 when Jetty handles a request containing multiple Accept headers with a large number of “quality” (i.e. q) parameters, the server may enter a denial of service (DoS) state due to high CPU usage processing those quality values, resulting in minutes of CPU time exhausted processing those quality values.
<p>Publish Date: 2021-02-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27223>CVE-2020-27223</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/eclipse/jetty.project/security/advisories/GHSA-m394-8rww-3jr7">https://github.com/eclipse/jetty.project/security/advisories/GHSA-m394-8rww-3jr7</a></p>
<p>Release Date: 2021-02-26</p>
<p>Fix Resolution (org.eclipse.jetty:jetty-http): 9.4.37.v20210219</p>
<p>Direct dependency fix Resolution (io.dropwizard:dropwizard-jetty): 2.0.0-rc0+test8</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-28169</summary>
### Vulnerable Library - <b>jetty-http-9.4.18.v20190429.jar</b></p>
<p>The Eclipse Jetty Project</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to dependency file: /BaragonData/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar</p>
<p>
Dependency Hierarchy:
- dropwizard-jetty-1.3.12.jar (Root Library)
- :x: **jetty-http-9.4.18.v20190429.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/Baragon-36/commit/3335ef04e9449f11036516e41533318fc21bd8a3">3335ef04e9449f11036516e41533318fc21bd8a3</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
For Eclipse Jetty versions <= 9.4.40, <= 10.0.2, <= 11.0.2, it is possible for requests to the ConcatServlet with a doubly encoded path to access protected resources within the WEB-INF directory. For example a request to `/concat?/%2557EB-INF/web.xml` can retrieve the web.xml file. This can reveal sensitive information regarding the implementation of a web application.
<p>Publish Date: 2021-06-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28169>CVE-2021-28169</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/eclipse/jetty.project/security/advisories/GHSA-gwcr-j4wh-j3cq">https://github.com/eclipse/jetty.project/security/advisories/GHSA-gwcr-j4wh-j3cq</a></p>
<p>Release Date: 2021-06-09</p>
<p>Fix Resolution (org.eclipse.jetty:jetty-http): 9.4.41.v20210516</p>
<p>Direct dependency fix Resolution (io.dropwizard:dropwizard-jetty): 2.0.0-rc0+test8</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> CVE-2022-2047</summary>
### Vulnerable Library - <b>jetty-http-9.4.18.v20190429.jar</b></p>
<p>The Eclipse Jetty Project</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to dependency file: /BaragonData/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar</p>
<p>
Dependency Hierarchy:
- dropwizard-jetty-1.3.12.jar (Root Library)
- :x: **jetty-http-9.4.18.v20190429.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/Baragon-36/commit/3335ef04e9449f11036516e41533318fc21bd8a3">3335ef04e9449f11036516e41533318fc21bd8a3</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Eclipse Jetty versions 9.4.0 thru 9.4.46, and 10.0.0 thru 10.0.9, and 11.0.0 thru 11.0.9 versions, the parsing of the authority segment of an http scheme URI, the Jetty HttpURI class improperly detects an invalid input as a hostname. This can lead to failures in a Proxy scenario.
<p>Publish Date: 2022-07-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2047>CVE-2022-2047</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>2.7</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/eclipse/jetty.project/security/advisories/GHSA-cj7v-27pg-wf7q">https://github.com/eclipse/jetty.project/security/advisories/GHSA-cj7v-27pg-wf7q</a></p>
<p>Release Date: 2022-07-07</p>
<p>Fix Resolution (org.eclipse.jetty:jetty-http): 9.4.47.v20220610</p>
<p>Direct dependency fix Resolution (io.dropwizard:dropwizard-jetty): 2.0.0-rc0+test8</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> | True | dropwizard-jetty-1.3.12.jar: 3 vulnerabilities (highest severity is: 5.3) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>dropwizard-jetty-1.3.12.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /BaragonService/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/Baragon-36/commit/3335ef04e9449f11036516e41533318fc21bd8a3">3335ef04e9449f11036516e41533318fc21bd8a3</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2020-27223](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27223) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | jetty-http-9.4.18.v20190429.jar | Transitive | 2.0.0-rc0+test8 | ✅ |
| [CVE-2021-28169](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28169) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | jetty-http-9.4.18.v20190429.jar | Transitive | 2.0.0-rc0+test8 | ✅ |
| [CVE-2022-2047](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2047) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 2.7 | jetty-http-9.4.18.v20190429.jar | Transitive | 2.0.0-rc0+test8 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-27223</summary>
### Vulnerable Library - <b>jetty-http-9.4.18.v20190429.jar</b></p>
<p>The Eclipse Jetty Project</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to dependency file: /BaragonData/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar</p>
<p>
Dependency Hierarchy:
- dropwizard-jetty-1.3.12.jar (Root Library)
- :x: **jetty-http-9.4.18.v20190429.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/Baragon-36/commit/3335ef04e9449f11036516e41533318fc21bd8a3">3335ef04e9449f11036516e41533318fc21bd8a3</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Eclipse Jetty 9.4.6.v20170531 to 9.4.36.v20210114 (inclusive), 10.0.0, and 11.0.0 when Jetty handles a request containing multiple Accept headers with a large number of “quality” (i.e. q) parameters, the server may enter a denial of service (DoS) state due to high CPU usage processing those quality values, resulting in minutes of CPU time exhausted processing those quality values.
<p>Publish Date: 2021-02-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27223>CVE-2020-27223</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/eclipse/jetty.project/security/advisories/GHSA-m394-8rww-3jr7">https://github.com/eclipse/jetty.project/security/advisories/GHSA-m394-8rww-3jr7</a></p>
<p>Release Date: 2021-02-26</p>
<p>Fix Resolution (org.eclipse.jetty:jetty-http): 9.4.37.v20210219</p>
<p>Direct dependency fix Resolution (io.dropwizard:dropwizard-jetty): 2.0.0-rc0+test8</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-28169</summary>
### Vulnerable Library - <b>jetty-http-9.4.18.v20190429.jar</b></p>
<p>The Eclipse Jetty Project</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to dependency file: /BaragonData/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar</p>
<p>
Dependency Hierarchy:
- dropwizard-jetty-1.3.12.jar (Root Library)
- :x: **jetty-http-9.4.18.v20190429.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/Baragon-36/commit/3335ef04e9449f11036516e41533318fc21bd8a3">3335ef04e9449f11036516e41533318fc21bd8a3</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
For Eclipse Jetty versions <= 9.4.40, <= 10.0.2, <= 11.0.2, it is possible for requests to the ConcatServlet with a doubly encoded path to access protected resources within the WEB-INF directory. For example a request to `/concat?/%2557EB-INF/web.xml` can retrieve the web.xml file. This can reveal sensitive information regarding the implementation of a web application.
<p>Publish Date: 2021-06-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28169>CVE-2021-28169</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/eclipse/jetty.project/security/advisories/GHSA-gwcr-j4wh-j3cq">https://github.com/eclipse/jetty.project/security/advisories/GHSA-gwcr-j4wh-j3cq</a></p>
<p>Release Date: 2021-06-09</p>
<p>Fix Resolution (org.eclipse.jetty:jetty-http): 9.4.41.v20210516</p>
<p>Direct dependency fix Resolution (io.dropwizard:dropwizard-jetty): 2.0.0-rc0+test8</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> CVE-2022-2047</summary>
### Vulnerable Library - <b>jetty-http-9.4.18.v20190429.jar</b></p>
<p>The Eclipse Jetty Project</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to dependency file: /BaragonData/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar,/home/wss-scanner/.m2/repository/org/eclipse/jetty/jetty-http/9.4.18.v20190429/jetty-http-9.4.18.v20190429.jar</p>
<p>
Dependency Hierarchy:
- dropwizard-jetty-1.3.12.jar (Root Library)
- :x: **jetty-http-9.4.18.v20190429.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/Baragon-36/commit/3335ef04e9449f11036516e41533318fc21bd8a3">3335ef04e9449f11036516e41533318fc21bd8a3</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Eclipse Jetty versions 9.4.0 thru 9.4.46, and 10.0.0 thru 10.0.9, and 11.0.0 thru 11.0.9 versions, the parsing of the authority segment of an http scheme URI, the Jetty HttpURI class improperly detects an invalid input as a hostname. This can lead to failures in a Proxy scenario.
<p>Publish Date: 2022-07-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2047>CVE-2022-2047</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>2.7</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/eclipse/jetty.project/security/advisories/GHSA-cj7v-27pg-wf7q">https://github.com/eclipse/jetty.project/security/advisories/GHSA-cj7v-27pg-wf7q</a></p>
<p>Release Date: 2022-07-07</p>
<p>Fix Resolution (org.eclipse.jetty:jetty-http): 9.4.47.v20220610</p>
<p>Direct dependency fix Resolution (io.dropwizard:dropwizard-jetty): 2.0.0-rc0+test8</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> | non_priority | dropwizard jetty jar vulnerabilities highest severity is vulnerable library dropwizard jetty jar path to dependency file baragonservice pom xml path to vulnerable library home wss scanner repository org eclipse jetty jetty http jetty http jar home wss scanner repository org eclipse jetty jetty http jetty http jar home wss scanner repository org eclipse jetty jetty http jetty http jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available medium jetty http jar transitive medium jetty http jar transitive low jetty http jar transitive details cve vulnerable library jetty http jar the eclipse jetty project library home page a href path to dependency file baragondata pom xml path to vulnerable library home wss scanner repository org eclipse jetty jetty http jetty http jar home wss scanner repository org eclipse jetty jetty http jetty http jar home wss scanner repository org eclipse jetty jetty http jetty http jar dependency hierarchy dropwizard jetty jar root library x jetty http jar vulnerable library found in head commit a href found in base branch master vulnerability details in eclipse jetty to inclusive and when jetty handles a request containing multiple accept headers with a large number of “quality” i e q parameters the server may enter a denial of service dos state due to high cpu usage processing those quality values resulting in minutes of cpu time exhausted processing those quality values publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org eclipse jetty jetty http direct dependency fix resolution io dropwizard dropwizard jetty rescue worker helmet automatic remediation is available for this issue cve vulnerable library jetty http jar the eclipse jetty project library home page a href path to dependency file baragondata pom xml path to vulnerable library home wss scanner repository org eclipse jetty jetty http jetty http jar home wss scanner repository org eclipse jetty jetty http jetty http jar home wss scanner repository org eclipse jetty jetty http jetty http jar dependency hierarchy dropwizard jetty jar root library x jetty http jar vulnerable library found in head commit a href found in base branch master vulnerability details for eclipse jetty versions it is possible for requests to the concatservlet with a doubly encoded path to access protected resources within the web inf directory for example a request to concat inf web xml can retrieve the web xml file this can reveal sensitive information regarding the implementation of a web application publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org eclipse jetty jetty http direct dependency fix resolution io dropwizard dropwizard jetty rescue worker helmet automatic remediation is available for this issue cve vulnerable library jetty http jar the eclipse jetty project library home page a href path to dependency file baragondata pom xml path to vulnerable library home wss scanner repository org eclipse jetty jetty http jetty http jar home wss scanner repository org eclipse jetty jetty http jetty http jar home wss scanner repository org eclipse jetty jetty http jetty http jar dependency hierarchy dropwizard jetty jar root library x jetty http jar vulnerable library found in head commit a href found in base branch master vulnerability details in eclipse jetty versions thru and thru and thru versions the parsing of the authority segment of an http scheme uri the jetty httpuri class improperly detects an invalid input as a hostname this can lead to failures in a proxy scenario publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org eclipse jetty jetty http direct dependency fix resolution io dropwizard dropwizard jetty rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue | 0 |
400,332 | 27,279,239,002 | IssuesEvent | 2023-02-23 08:48:33 | vaadin/flow | https://api.github.com/repos/vaadin/flow | closed | Document vaadin.excludeUrls configuration parameter | enhancement documentation | This parameter was introduced in https://github.com/vaadin/flow/pull/14579, but not yet documented.
Can be added into https://vaadin.com/docs/latest/integrations/spring/configuration or into a general chapter about configuration. | 1.0 | Document vaadin.excludeUrls configuration parameter - This parameter was introduced in https://github.com/vaadin/flow/pull/14579, but not yet documented.
Can be added into https://vaadin.com/docs/latest/integrations/spring/configuration or into a general chapter about configuration. | non_priority | document vaadin excludeurls configuration parameter this parameter was introduced in but not yet documented can be added into or into a general chapter about configuration | 0 |
119,529 | 17,620,304,719 | IssuesEvent | 2021-08-18 14:36:11 | jgeraigery/experian-java | https://api.github.com/repos/jgeraigery/experian-java | closed | CVE-2019-14540 (High) detected in jackson-databind-2.9.2.jar - autoclosed | security vulnerability | ## CVE-2019-14540 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: experian-java/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.2/jackson-databind-2.9.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/experian-java/commit/9ade2a959068cca30ecfdbb254939af6f67affb1">9ade2a959068cca30ecfdbb254939af6f67affb1</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Polymorphic Typing issue was discovered in FasterXML jackson-databind before 2.9.10. It is related to com.zaxxer.hikari.HikariConfig.
<p>Publish Date: 2019-09-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14540>CVE-2019-14540</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14540">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14540</a></p>
<p>Release Date: 2019-09-15</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.8.11.5,2.9.10,2.10.0.pr3,2.11.0.rc1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.2","packageFilePaths":["/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.8.11.5,2.9.10,2.10.0.pr3,2.11.0.rc1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-14540","vulnerabilityDetails":"A Polymorphic Typing issue was discovered in FasterXML jackson-databind before 2.9.10. It is related to com.zaxxer.hikari.HikariConfig.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14540","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2019-14540 (High) detected in jackson-databind-2.9.2.jar - autoclosed - ## CVE-2019-14540 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: experian-java/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.2/jackson-databind-2.9.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/experian-java/commit/9ade2a959068cca30ecfdbb254939af6f67affb1">9ade2a959068cca30ecfdbb254939af6f67affb1</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Polymorphic Typing issue was discovered in FasterXML jackson-databind before 2.9.10. It is related to com.zaxxer.hikari.HikariConfig.
<p>Publish Date: 2019-09-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14540>CVE-2019-14540</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14540">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14540</a></p>
<p>Release Date: 2019-09-15</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.8.11.5,2.9.10,2.10.0.pr3,2.11.0.rc1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.2","packageFilePaths":["/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.8.11.5,2.9.10,2.10.0.pr3,2.11.0.rc1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-14540","vulnerabilityDetails":"A Polymorphic Typing issue was discovered in FasterXML jackson-databind before 2.9.10. It is related to com.zaxxer.hikari.HikariConfig.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14540","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file experian java mavenworkspace bis services lib bis services base pom xml path to vulnerable library canner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind before it is related to com zaxxer hikari hikariconfig publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails a polymorphic typing issue was discovered in fasterxml jackson databind before it is related to com zaxxer hikari hikariconfig vulnerabilityurl | 0 |
260,651 | 19,680,106,962 | IssuesEvent | 2022-01-11 15:59:59 | vercel/next.js | https://api.github.com/repos/vercel/next.js | closed | CMS preview example demo are missing the secret value | good first issue area: documentation | # Documentation report
Would like to try out the demo given in a CMS preview example, e.g. the DatoCMS one but without knowing the preview secret token I can't try out the preview mode on the demo website.
| 1.0 | CMS preview example demo are missing the secret value - # Documentation report
Would like to try out the demo given in a CMS preview example, e.g. the DatoCMS one but without knowing the preview secret token I can't try out the preview mode on the demo website.
| non_priority | cms preview example demo are missing the secret value documentation report would like to try out the demo given in a cms preview example e g the datocms one but without knowing the preview secret token i can t try out the preview mode on the demo website | 0 |
11,487 | 4,232,240,720 | IssuesEvent | 2016-07-04 21:16:38 | ReikaKalseki/Reika_Mods_Issues | https://api.github.com/repos/ReikaKalseki/Reika_Mods_Issues | closed | Electricraft OpenComputers function getCurrent() returns voltage | Bug ElectriCraft Mod Interaction Stupid Code | 
Looks like the problem is [here](https://github.com/ReikaKalseki/ElectriCraft/blob/a572d4f1e5446b4fad1d083c5cf634f8dcd9d32e/Auxiliary/Lua/LuaGetCurrent.java) on line 25. | 1.0 | Electricraft OpenComputers function getCurrent() returns voltage - 
Looks like the problem is [here](https://github.com/ReikaKalseki/ElectriCraft/blob/a572d4f1e5446b4fad1d083c5cf634f8dcd9d32e/Auxiliary/Lua/LuaGetCurrent.java) on line 25. | non_priority | electricraft opencomputers function getcurrent returns voltage looks like the problem is on line | 0 |
154,132 | 13,538,867,411 | IssuesEvent | 2020-09-16 12:43:42 | Galusss/OC_project7 | https://api.github.com/repos/Galusss/OC_project7 | reopened | Models (backend) | documentation | Les models pour 'users' et 'articles' sont mis en place avec Sequelize ORM.
Documentation Sequelize v6 : [https://sequelize.org/master/](url) | 1.0 | Models (backend) - Les models pour 'users' et 'articles' sont mis en place avec Sequelize ORM.
Documentation Sequelize v6 : [https://sequelize.org/master/](url) | non_priority | models backend les models pour users et articles sont mis en place avec sequelize orm documentation sequelize url | 0 |
10,210 | 31,931,633,611 | IssuesEvent | 2023-09-19 07:50:32 | red-hat-storage/ocs-ci | https://api.github.com/repos/red-hat-storage/ocs-ci | opened | test_for_encrypted_pv_ui - make this test work on any ODF version | ui_automation Squad/Black | test is running only against ODF 4.8, remove this marker and adjust test | 1.0 | test_for_encrypted_pv_ui - make this test work on any ODF version - test is running only against ODF 4.8, remove this marker and adjust test | non_priority | test for encrypted pv ui make this test work on any odf version test is running only against odf remove this marker and adjust test | 0 |
351,871 | 32,032,971,098 | IssuesEvent | 2023-09-22 13:33:02 | eclipse-openj9/openj9 | https://api.github.com/repos/eclipse-openj9/openj9 | opened | VirtualThread states changed in jdk22 | test failure jdk22 | Failure links
------------
The values were changed in https://github.com/ibmruntimes/openj9-openjdk-jdk/commit/ceb174ba8004f6361a307f6d599d786eef9307c7.
* https://openj9-jenkins.osuosl.org/job/Test_openjdknext_j9_sanity.functional_s390x_linux_Personal_testList_0/40/consoleText
* https://openj9-jenkins.osuosl.org/job/Test_openjdknext_j9_sanity.functional_s390x_linux_Personal_testList_1/40/consoleText
Failure output (captured from console output)
---------------------------------------------
```
[2023-09-22T05:48:47.201Z] FAILED: test_verifyJVMTIMacros
[2023-09-22T05:48:47.201Z] java.lang.AssertionError: JVMTI_VTHREAD_STATE_YIELDING (7) does not match VirtualThread.YIELDING (10)
[2023-09-22T05:48:47.201Z] at org.testng.Assert.fail(Assert.java:96)
[2023-09-22T05:48:47.201Z] at org.openj9.test.jep425.VirtualThreadTests.test_verifyJVMTIMacros(VirtualThreadTests.java:321)
[2023-09-22T05:48:47.201Z] at java.base/jdk.internal.reflect.DirectMethodHandleAccessor.invoke(DirectMethodHandleAccessor.java:103)
[2023-09-22T05:48:47.201Z] at java.base/java.lang.reflect.Method.invoke(Method.java:580)
``` | 1.0 | VirtualThread states changed in jdk22 - Failure links
------------
The values were changed in https://github.com/ibmruntimes/openj9-openjdk-jdk/commit/ceb174ba8004f6361a307f6d599d786eef9307c7.
* https://openj9-jenkins.osuosl.org/job/Test_openjdknext_j9_sanity.functional_s390x_linux_Personal_testList_0/40/consoleText
* https://openj9-jenkins.osuosl.org/job/Test_openjdknext_j9_sanity.functional_s390x_linux_Personal_testList_1/40/consoleText
Failure output (captured from console output)
---------------------------------------------
```
[2023-09-22T05:48:47.201Z] FAILED: test_verifyJVMTIMacros
[2023-09-22T05:48:47.201Z] java.lang.AssertionError: JVMTI_VTHREAD_STATE_YIELDING (7) does not match VirtualThread.YIELDING (10)
[2023-09-22T05:48:47.201Z] at org.testng.Assert.fail(Assert.java:96)
[2023-09-22T05:48:47.201Z] at org.openj9.test.jep425.VirtualThreadTests.test_verifyJVMTIMacros(VirtualThreadTests.java:321)
[2023-09-22T05:48:47.201Z] at java.base/jdk.internal.reflect.DirectMethodHandleAccessor.invoke(DirectMethodHandleAccessor.java:103)
[2023-09-22T05:48:47.201Z] at java.base/java.lang.reflect.Method.invoke(Method.java:580)
``` | non_priority | virtualthread states changed in failure links the values were changed in failure output captured from console output failed test verifyjvmtimacros java lang assertionerror jvmti vthread state yielding does not match virtualthread yielding at org testng assert fail assert java at org test virtualthreadtests test verifyjvmtimacros virtualthreadtests java at java base jdk internal reflect directmethodhandleaccessor invoke directmethodhandleaccessor java at java base java lang reflect method invoke method java | 0 |
278,392 | 21,076,103,999 | IssuesEvent | 2022-04-02 06:44:21 | AY2122S2-CS2103-F09-3/tp | https://api.github.com/repos/AY2122S2-CS2103-F09-3/tp | closed | [PE-D] Price of insurance can only be integer values | bug documentation | I would expect decimal places to be allowed for inputting prices for adding an insurance command. Doesn't fit well with the product.
<!--session: 1648793009038-2837d48d-1dc9-4169-aea4-6e86a2707040-->
<!--Version: Web v3.4.2-->
-------------
Labels: `type.FeatureFlaw` `severity.Low`
original: LapisRaider/ped#19 | 1.0 | [PE-D] Price of insurance can only be integer values - I would expect decimal places to be allowed for inputting prices for adding an insurance command. Doesn't fit well with the product.
<!--session: 1648793009038-2837d48d-1dc9-4169-aea4-6e86a2707040-->
<!--Version: Web v3.4.2-->
-------------
Labels: `type.FeatureFlaw` `severity.Low`
original: LapisRaider/ped#19 | non_priority | price of insurance can only be integer values i would expect decimal places to be allowed for inputting prices for adding an insurance command doesn t fit well with the product labels type featureflaw severity low original lapisraider ped | 0 |
189,798 | 15,196,640,359 | IssuesEvent | 2021-02-16 08:36:04 | spring-projects/spring-boot | https://api.github.com/repos/spring-projects/spring-boot | closed | Update the Gradle plugin documentation to recommend the maven-publish plugin over the maven plugin | type: documentation | The documentation doesn't make any recommendation at the moment, other than documenting the maven plugin before the maven-publish plugin. We should swap the order over and also explicitly recommend the maven-publish plugin to align with Gradle's own recommendations and deprecations. | 1.0 | Update the Gradle plugin documentation to recommend the maven-publish plugin over the maven plugin - The documentation doesn't make any recommendation at the moment, other than documenting the maven plugin before the maven-publish plugin. We should swap the order over and also explicitly recommend the maven-publish plugin to align with Gradle's own recommendations and deprecations. | non_priority | update the gradle plugin documentation to recommend the maven publish plugin over the maven plugin the documentation doesn t make any recommendation at the moment other than documenting the maven plugin before the maven publish plugin we should swap the order over and also explicitly recommend the maven publish plugin to align with gradle s own recommendations and deprecations | 0 |
54,270 | 13,301,362,685 | IssuesEvent | 2020-08-25 12:50:56 | feelpp/feelpp | https://api.github.com/repos/feelpp/feelpp | opened | Add range-v3 to feelpp contrib | module:build type:feature | range-v3 is the basis for std::range in c++20.
in the last release 0.11 range-v3 (which is C++11/14 and 17 compliant) got closer to the C++20 standard.
see https://github.com/feelpp/range-v3 for more details
see https://en.cppreference.com/w/cpp/header/ranges for more details about std::ranges in the c++20 standard | 1.0 | Add range-v3 to feelpp contrib - range-v3 is the basis for std::range in c++20.
in the last release 0.11 range-v3 (which is C++11/14 and 17 compliant) got closer to the C++20 standard.
see https://github.com/feelpp/range-v3 for more details
see https://en.cppreference.com/w/cpp/header/ranges for more details about std::ranges in the c++20 standard | non_priority | add range to feelpp contrib range is the basis for std range in c in the last release range which is c and compliant got closer to the c standard see for more details see for more details about std ranges in the c standard | 0 |
128,730 | 18,070,107,195 | IssuesEvent | 2021-09-21 01:12:04 | dmyers87/matrix-react-sdk | https://api.github.com/repos/dmyers87/matrix-react-sdk | opened | CVE-2021-3803 (Medium) detected in nth-check-1.0.2.tgz | security vulnerability | ## CVE-2021-3803 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nth-check-1.0.2.tgz</b></p></summary>
<p>performant nth-check parser & compiler</p>
<p>Library home page: <a href="https://registry.npmjs.org/nth-check/-/nth-check-1.0.2.tgz">https://registry.npmjs.org/nth-check/-/nth-check-1.0.2.tgz</a></p>
<p>Path to dependency file: matrix-react-sdk/node_modules/nth-check/package.json</p>
<p>Path to vulnerable library: matrix-react-sdk/node_modules/nth-check/package.json</p>
<p>
Dependency Hierarchy:
- enzyme-3.11.0.tgz (Root Library)
- cheerio-1.0.0-rc.3.tgz
- css-select-1.2.0.tgz
- :x: **nth-check-1.0.2.tgz** (Vulnerable Library)
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
nth-check is vulnerable to Inefficient Regular Expression Complexity
<p>Publish Date: 2021-09-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3803>CVE-2021-3803</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/fb55/nth-check/compare/v2.0.0...v2.0.1">https://github.com/fb55/nth-check/compare/v2.0.0...v2.0.1</a></p>
<p>Release Date: 2021-09-17</p>
<p>Fix Resolution: nth-check - v2.0.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"nth-check","packageVersion":"1.0.2","packageFilePaths":["/node_modules/nth-check/package.json"],"isTransitiveDependency":true,"dependencyTree":"enzyme:3.11.0;cheerio:1.0.0-rc.3;css-select:1.2.0;nth-check:1.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"nth-check - v2.0.1"}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2021-3803","vulnerabilityDetails":"nth-check is vulnerable to Inefficient Regular Expression Complexity","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3803","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-3803 (Medium) detected in nth-check-1.0.2.tgz - ## CVE-2021-3803 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nth-check-1.0.2.tgz</b></p></summary>
<p>performant nth-check parser & compiler</p>
<p>Library home page: <a href="https://registry.npmjs.org/nth-check/-/nth-check-1.0.2.tgz">https://registry.npmjs.org/nth-check/-/nth-check-1.0.2.tgz</a></p>
<p>Path to dependency file: matrix-react-sdk/node_modules/nth-check/package.json</p>
<p>Path to vulnerable library: matrix-react-sdk/node_modules/nth-check/package.json</p>
<p>
Dependency Hierarchy:
- enzyme-3.11.0.tgz (Root Library)
- cheerio-1.0.0-rc.3.tgz
- css-select-1.2.0.tgz
- :x: **nth-check-1.0.2.tgz** (Vulnerable Library)
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
nth-check is vulnerable to Inefficient Regular Expression Complexity
<p>Publish Date: 2021-09-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3803>CVE-2021-3803</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/fb55/nth-check/compare/v2.0.0...v2.0.1">https://github.com/fb55/nth-check/compare/v2.0.0...v2.0.1</a></p>
<p>Release Date: 2021-09-17</p>
<p>Fix Resolution: nth-check - v2.0.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"nth-check","packageVersion":"1.0.2","packageFilePaths":["/node_modules/nth-check/package.json"],"isTransitiveDependency":true,"dependencyTree":"enzyme:3.11.0;cheerio:1.0.0-rc.3;css-select:1.2.0;nth-check:1.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"nth-check - v2.0.1"}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2021-3803","vulnerabilityDetails":"nth-check is vulnerable to Inefficient Regular Expression Complexity","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3803","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> --> | non_priority | cve medium detected in nth check tgz cve medium severity vulnerability vulnerable library nth check tgz performant nth check parser compiler library home page a href path to dependency file matrix react sdk node modules nth check package json path to vulnerable library matrix react sdk node modules nth check package json dependency hierarchy enzyme tgz root library cheerio rc tgz css select tgz x nth check tgz vulnerable library found in base branch develop vulnerability details nth check is vulnerable to inefficient regular expression complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution nth check isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree enzyme cheerio rc css select nth check isminimumfixversionavailable true minimumfixversion nth check basebranches vulnerabilityidentifier cve vulnerabilitydetails nth check is vulnerable to inefficient regular expression complexity vulnerabilityurl | 0 |
132,998 | 18,791,793,970 | IssuesEvent | 2021-11-08 17:32:58 | HTTPArchive/almanac.httparchive.org | https://api.github.com/repos/HTTPArchive/almanac.httparchive.org | closed | Chapter illustrations for new 2021 chapters | design | We have 2 new chapters this year:
- Structured Data
- WebAssembly
We have chapter banners images for each chapter created in 2019, and we managed to create three new ones for the new chapters added in 2020 (Capabilities, Jamstack, and Privacy).
So we would need similar ones created for Structured Data and WebAssembly. We have the PSD files to allow you to reuse some of the resources like the characters.
Unfortunately the designers that created these in 2019 and 2020 are unavailable this year.
@HTTPArchive/designers do any of you have the necessary design skills to do this? | 1.0 | Chapter illustrations for new 2021 chapters - We have 2 new chapters this year:
- Structured Data
- WebAssembly
We have chapter banners images for each chapter created in 2019, and we managed to create three new ones for the new chapters added in 2020 (Capabilities, Jamstack, and Privacy).
So we would need similar ones created for Structured Data and WebAssembly. We have the PSD files to allow you to reuse some of the resources like the characters.
Unfortunately the designers that created these in 2019 and 2020 are unavailable this year.
@HTTPArchive/designers do any of you have the necessary design skills to do this? | non_priority | chapter illustrations for new chapters we have new chapters this year structured data webassembly we have chapter banners images for each chapter created in and we managed to create three new ones for the new chapters added in capabilities jamstack and privacy so we would need similar ones created for structured data and webassembly we have the psd files to allow you to reuse some of the resources like the characters unfortunately the designers that created these in and are unavailable this year httparchive designers do any of you have the necessary design skills to do this | 0 |
449,295 | 31,837,617,104 | IssuesEvent | 2023-09-14 14:23:34 | Studio-Yandex-Practicum/RandomCoffeeBotTelegram | https://api.github.com/repos/Studio-Yandex-Practicum/RandomCoffeeBotTelegram | closed | Создать диаграмму состояний бота, оформить всё в drawing | documentation | # Why?
Необходимо для упрощения разработки.
# How To Do?
Реазилуем в диаграмме логику работы бота. НУжно посмотреть в документацию от заказчика и сделать правки в своей диаграмме.
Необходимо в readme добавить ссылку
# Where?
docs
| 1.0 | Создать диаграмму состояний бота, оформить всё в drawing - # Why?
Необходимо для упрощения разработки.
# How To Do?
Реазилуем в диаграмме логику работы бота. НУжно посмотреть в документацию от заказчика и сделать правки в своей диаграмме.
Необходимо в readme добавить ссылку
# Where?
docs
| non_priority | создать диаграмму состояний бота оформить всё в drawing why необходимо для упрощения разработки how to do реазилуем в диаграмме логику работы бота нужно посмотреть в документацию от заказчика и сделать правки в своей диаграмме необходимо в readme добавить ссылку where docs | 0 |
68,584 | 21,717,751,631 | IssuesEvent | 2022-05-10 19:45:46 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | opened | ZFS Receive of encrypted incremental data stream causes a PANIC | Type: Defect | ### System information
Type | Version/Name
--- | ---
Distribution Name | Debian
Distribution Version | Bullseye
Kernel Version | 5.10.109+truenas
Architecture | amd64
OpenZFS Version | zfs-2.1.2-95_g1d2cdd23b zfs-kmod-2.1.2-95_g1d2cdd23b
<!--
Command to find OpenZFS version:
zfs version
Commands to find kernel version:
uname -r # Linux
freebsd-version -r # FreeBSD
-->
### Describe the problem you're observing
During an incremental receive, ZFS caused a panic and a system hangup.
### Describe how to reproduce the problem
It happens randomly.
### Include any warning/errors/backtraces from the system logs
```
[456678.240841] VERIFY3(0 == dmu_object_set_blocksize(rwa->os, drro->drr_object, drro->drr_blksz, drro->drr_indblkshift, tx)) failed (0 == 95)
[456678.243815] PANIC at dmu_recv.c:1776:receive_object()
[456678.245141] Showing stack for process 2936808
[456678.246532] CPU: 10 PID: 2936808 Comm: receive_writer Tainted: P OE 5.10.109+truenas #1
[456678.247840] Hardware name: Supermicro X9QR7-TF+/X9QRi-F+/X9QR7-TF+/X9QRi-F+, BIOS 3.0b 05/20/2015
[456678.249138] Call Trace:
[456678.250421] dump_stack+0x6b/0x83
[456678.251676] spl_panic+0xd4/0xfc [spl]
[456678.253038] ? arc_buf_access+0x14c/0x250 [zfs]
[456678.254276] ? dnode_hold_impl+0x4e9/0xef0 [zfs]
[456678.255493] ? dnode_set_blksz+0x13b/0x300 [zfs]
[456678.256677] ? dnode_rele_and_unlock+0x5c/0xc0 [zfs]
[456678.257846] receive_object+0xc2c/0xca0 [zfs]
[456678.258984] ? dmu_object_next+0xd6/0x120 [zfs]
[456678.260098] ? receive_writer_thread+0xbd/0xad0 [zfs]
[456678.261160] ? kfree+0x40c/0x480
[456678.262202] ? _cond_resched+0x16/0x40
[456678.263244] receive_writer_thread+0x1cc/0xad0 [zfs]
[456678.264280] ? thread_generic_wrapper+0x62/0x80 [spl]
[456678.265252] ? kfree+0x40c/0x480
[456678.266242] ? receive_process_write_record+0x190/0x190 [zfs]
[456678.267177] ? thread_generic_wrapper+0x6f/0x80 [spl]
[456678.268092] thread_generic_wrapper+0x6f/0x80 [spl]
[456678.268988] ? __thread_exit+0x20/0x20 [spl]
[456678.269864] kthread+0x11b/0x140
[456678.270706] ? __kthread_bind_mask+0x60/0x60
[456678.271538] ret_from_fork+0x22/0x30
```
| 1.0 | ZFS Receive of encrypted incremental data stream causes a PANIC - ### System information
Type | Version/Name
--- | ---
Distribution Name | Debian
Distribution Version | Bullseye
Kernel Version | 5.10.109+truenas
Architecture | amd64
OpenZFS Version | zfs-2.1.2-95_g1d2cdd23b zfs-kmod-2.1.2-95_g1d2cdd23b
<!--
Command to find OpenZFS version:
zfs version
Commands to find kernel version:
uname -r # Linux
freebsd-version -r # FreeBSD
-->
### Describe the problem you're observing
During an incremental receive, ZFS caused a panic and a system hangup.
### Describe how to reproduce the problem
It happens randomly.
### Include any warning/errors/backtraces from the system logs
```
[456678.240841] VERIFY3(0 == dmu_object_set_blocksize(rwa->os, drro->drr_object, drro->drr_blksz, drro->drr_indblkshift, tx)) failed (0 == 95)
[456678.243815] PANIC at dmu_recv.c:1776:receive_object()
[456678.245141] Showing stack for process 2936808
[456678.246532] CPU: 10 PID: 2936808 Comm: receive_writer Tainted: P OE 5.10.109+truenas #1
[456678.247840] Hardware name: Supermicro X9QR7-TF+/X9QRi-F+/X9QR7-TF+/X9QRi-F+, BIOS 3.0b 05/20/2015
[456678.249138] Call Trace:
[456678.250421] dump_stack+0x6b/0x83
[456678.251676] spl_panic+0xd4/0xfc [spl]
[456678.253038] ? arc_buf_access+0x14c/0x250 [zfs]
[456678.254276] ? dnode_hold_impl+0x4e9/0xef0 [zfs]
[456678.255493] ? dnode_set_blksz+0x13b/0x300 [zfs]
[456678.256677] ? dnode_rele_and_unlock+0x5c/0xc0 [zfs]
[456678.257846] receive_object+0xc2c/0xca0 [zfs]
[456678.258984] ? dmu_object_next+0xd6/0x120 [zfs]
[456678.260098] ? receive_writer_thread+0xbd/0xad0 [zfs]
[456678.261160] ? kfree+0x40c/0x480
[456678.262202] ? _cond_resched+0x16/0x40
[456678.263244] receive_writer_thread+0x1cc/0xad0 [zfs]
[456678.264280] ? thread_generic_wrapper+0x62/0x80 [spl]
[456678.265252] ? kfree+0x40c/0x480
[456678.266242] ? receive_process_write_record+0x190/0x190 [zfs]
[456678.267177] ? thread_generic_wrapper+0x6f/0x80 [spl]
[456678.268092] thread_generic_wrapper+0x6f/0x80 [spl]
[456678.268988] ? __thread_exit+0x20/0x20 [spl]
[456678.269864] kthread+0x11b/0x140
[456678.270706] ? __kthread_bind_mask+0x60/0x60
[456678.271538] ret_from_fork+0x22/0x30
```
| non_priority | zfs receive of encrypted incremental data stream causes a panic system information type version name distribution name debian distribution version bullseye kernel version truenas architecture openzfs version zfs zfs kmod command to find openzfs version zfs version commands to find kernel version uname r linux freebsd version r freebsd describe the problem you re observing during an incremental receive zfs caused a panic and a system hangup describe how to reproduce the problem it happens randomly include any warning errors backtraces from the system logs dmu object set blocksize rwa os drro drr object drro drr blksz drro drr indblkshift tx failed panic at dmu recv c receive object showing stack for process cpu pid comm receive writer tainted p oe truenas hardware name supermicro tf f tf f bios call trace dump stack spl panic arc buf access dnode hold impl dnode set blksz dnode rele and unlock receive object dmu object next receive writer thread kfree cond resched receive writer thread thread generic wrapper kfree receive process write record thread generic wrapper thread generic wrapper thread exit kthread kthread bind mask ret from fork | 0 |
58,699 | 11,905,066,912 | IssuesEvent | 2020-03-30 17:55:24 | home-assistant/brands | https://api.github.com/repos/home-assistant/brands | opened | August is missing brand images | domain-missing has-codeowner has-config-flow |
## The problem
The August integration does not have brand images in
this repository.
We recently started this Brands repository, to create a centralized storage of all brand-related images. These images are used on our website and the Home Assistant frontend.
The following images are missing and would ideally be added:
- `src/august/icon.png`
- `src/august/logo.png`
- `src/august/icon@2x.png`
- `src/august/logo@2x.png`
For image specifications and requirements, please see [README.md](https://github.com/home-assistant/brands/blob/master/README.md).
## Updating the documentation repository
Our documentation repository already has a logo for this integration, however, it does not meet the image requirements of this new Brands repository.
If adding images to this repository, please open up a PR to the documentation repository as well, removing the `logo: august.png` line from this file:
<https://github.com/home-assistant/home-assistant.io/blob/current/source/_integrations/august.markdown>
**Note**: The documentation PR needs to be opened against the `current` branch.
**Note2**: Please leave the actual logo file in the documentation repository. It will be cleaned up differently.
## Additional information
For more information about this repository, read the [README.md](https://github.com/home-assistant/brands/blob/master/README.md) file of this repository. It contains information on how this repository works, and image specification and requirements.
## Codeowner mention
Hi there, @bdraco! Mind taking a look at this issue as it is with an integration (august) you are listed as a [codeowner](https://github.com/home-assistant/core/blob/dev/homeassistant/components/august/manifest.json) for? Thanks!
Resolving this issue is not limited to codeowners! If you want to help us out, feel free to resolve this issue! Thanks already!
| 1.0 | August is missing brand images -
## The problem
The August integration does not have brand images in
this repository.
We recently started this Brands repository, to create a centralized storage of all brand-related images. These images are used on our website and the Home Assistant frontend.
The following images are missing and would ideally be added:
- `src/august/icon.png`
- `src/august/logo.png`
- `src/august/icon@2x.png`
- `src/august/logo@2x.png`
For image specifications and requirements, please see [README.md](https://github.com/home-assistant/brands/blob/master/README.md).
## Updating the documentation repository
Our documentation repository already has a logo for this integration, however, it does not meet the image requirements of this new Brands repository.
If adding images to this repository, please open up a PR to the documentation repository as well, removing the `logo: august.png` line from this file:
<https://github.com/home-assistant/home-assistant.io/blob/current/source/_integrations/august.markdown>
**Note**: The documentation PR needs to be opened against the `current` branch.
**Note2**: Please leave the actual logo file in the documentation repository. It will be cleaned up differently.
## Additional information
For more information about this repository, read the [README.md](https://github.com/home-assistant/brands/blob/master/README.md) file of this repository. It contains information on how this repository works, and image specification and requirements.
## Codeowner mention
Hi there, @bdraco! Mind taking a look at this issue as it is with an integration (august) you are listed as a [codeowner](https://github.com/home-assistant/core/blob/dev/homeassistant/components/august/manifest.json) for? Thanks!
Resolving this issue is not limited to codeowners! If you want to help us out, feel free to resolve this issue! Thanks already!
| non_priority | august is missing brand images the problem the august integration does not have brand images in this repository we recently started this brands repository to create a centralized storage of all brand related images these images are used on our website and the home assistant frontend the following images are missing and would ideally be added src august icon png src august logo png src august icon png src august logo png for image specifications and requirements please see updating the documentation repository our documentation repository already has a logo for this integration however it does not meet the image requirements of this new brands repository if adding images to this repository please open up a pr to the documentation repository as well removing the logo august png line from this file note the documentation pr needs to be opened against the current branch please leave the actual logo file in the documentation repository it will be cleaned up differently additional information for more information about this repository read the file of this repository it contains information on how this repository works and image specification and requirements codeowner mention hi there bdraco mind taking a look at this issue as it is with an integration august you are listed as a for thanks resolving this issue is not limited to codeowners if you want to help us out feel free to resolve this issue thanks already | 0 |
181,589 | 14,881,932,373 | IssuesEvent | 2021-01-20 11:10:02 | SAP/code-pal-for-abap | https://api.github.com/repos/SAP/code-pal-for-abap | closed | 'Deprecated Key Word' Documentation | documentation | It would be good to have the respective replacement in the check documentation. It can be described in the check documentation or a link to the original system documentation. | 1.0 | 'Deprecated Key Word' Documentation - It would be good to have the respective replacement in the check documentation. It can be described in the check documentation or a link to the original system documentation. | non_priority | deprecated key word documentation it would be good to have the respective replacement in the check documentation it can be described in the check documentation or a link to the original system documentation | 0 |
7,167 | 16,853,483,406 | IssuesEvent | 2021-06-21 00:41:09 | QutEcoacoustics/workbench-client | https://api.github.com/repos/QutEcoacoustics/workbench-client | opened | Enable prerendering | architecture bug env:production env:staging | Currently prerendering of the SSR files is broken for the website, however making it functional would help SEO performance and should be possible. Currently this is broken because we need to read the configuration file for the instance before building. During SSR we perform this step through the express server, however a different solution is needed for prerending to work. The following guide seems promising: https://www.thirdrocktechkno.com/blog/prerendering-in-angular-9/ | 1.0 | Enable prerendering - Currently prerendering of the SSR files is broken for the website, however making it functional would help SEO performance and should be possible. Currently this is broken because we need to read the configuration file for the instance before building. During SSR we perform this step through the express server, however a different solution is needed for prerending to work. The following guide seems promising: https://www.thirdrocktechkno.com/blog/prerendering-in-angular-9/ | non_priority | enable prerendering currently prerendering of the ssr files is broken for the website however making it functional would help seo performance and should be possible currently this is broken because we need to read the configuration file for the instance before building during ssr we perform this step through the express server however a different solution is needed for prerending to work the following guide seems promising | 0 |
55,886 | 8,033,878,420 | IssuesEvent | 2018-07-29 12:11:30 | PaulHancock/Aegean | https://api.github.com/repos/PaulHancock/Aegean | closed | update wiki to be inline with v 2.0.2 | documentation | The wiki is out of date. In particular the help text for `Aegean` and `BANE` needs to be updated, as well as the description of the various command line options. | 1.0 | update wiki to be inline with v 2.0.2 - The wiki is out of date. In particular the help text for `Aegean` and `BANE` needs to be updated, as well as the description of the various command line options. | non_priority | update wiki to be inline with v the wiki is out of date in particular the help text for aegean and bane needs to be updated as well as the description of the various command line options | 0 |
104,639 | 22,711,425,479 | IssuesEvent | 2022-07-05 19:49:25 | microsoft/vscode-jupyter | https://api.github.com/repos/microsoft/vscode-jupyter | closed | Remote widget source provider doesn't work if any security required | bug upstream-vscode ipywidgets vscode-web verification-found | 1. Remove the CDN sources (make them blank)
2. Connect to a remote server
3. Run something that needs a widget
4. Widget won't be downloaded
Root cause is the remote cdn provider doesn't do anything except pass the URI up to the webview for downloading. This doesn't work because the webview doesn't have the auth headers (or token) to make a valid request.
I think the solution here is to refactor the CDN downloading code to support the server as a CDN source, but then it would make the appropriate requests with the correct headers, download from the remote jupyter server, and copy the source files to the same spot it does for CDNs.
Essentially remove this [source provider ](https://github.com/microsoft/vscode-jupyter/blob/9d9ad7d1f4756789ba9374d1a0351a558279ed8b/src/kernels/ipywidgets-message-coordination/remoteWidgetScriptSourceProvider.ts#L14) and have the CDN code also try kernel's baseURL. Meaning the kernel's base URL is always another CDN source. | 2.0 | Remote widget source provider doesn't work if any security required - 1. Remove the CDN sources (make them blank)
2. Connect to a remote server
3. Run something that needs a widget
4. Widget won't be downloaded
Root cause is the remote cdn provider doesn't do anything except pass the URI up to the webview for downloading. This doesn't work because the webview doesn't have the auth headers (or token) to make a valid request.
I think the solution here is to refactor the CDN downloading code to support the server as a CDN source, but then it would make the appropriate requests with the correct headers, download from the remote jupyter server, and copy the source files to the same spot it does for CDNs.
Essentially remove this [source provider ](https://github.com/microsoft/vscode-jupyter/blob/9d9ad7d1f4756789ba9374d1a0351a558279ed8b/src/kernels/ipywidgets-message-coordination/remoteWidgetScriptSourceProvider.ts#L14) and have the CDN code also try kernel's baseURL. Meaning the kernel's base URL is always another CDN source. | non_priority | remote widget source provider doesn t work if any security required remove the cdn sources make them blank connect to a remote server run something that needs a widget widget won t be downloaded root cause is the remote cdn provider doesn t do anything except pass the uri up to the webview for downloading this doesn t work because the webview doesn t have the auth headers or token to make a valid request i think the solution here is to refactor the cdn downloading code to support the server as a cdn source but then it would make the appropriate requests with the correct headers download from the remote jupyter server and copy the source files to the same spot it does for cdns essentially remove this and have the cdn code also try kernel s baseurl meaning the kernel s base url is always another cdn source | 0 |
366,923 | 25,708,751,022 | IssuesEvent | 2022-12-07 04:02:59 | esmero/archipelago-documentation | https://api.github.com/repos/esmero/archipelago-documentation | closed | Add full KeyName Provider Documentation Section | documentation enhancement help wanted | # What?
Key Name Providers are fundamental for searching/filtering/faceting. We should have
- How to documentation on creating new ones and the different Plugins we provide (including the " Is Date" option) and the Entity Reference One
- What/how Archipelago harvests/compiles data using them
- How these are exposed to the Search API but also programmatically as properties of a field (for DEVS)
- Good practices: e.g if the keys you want to expose are all flat and unique you can use the JSONLD Plugin and type them all into the "extra keys" fields instead of generating too many Key Name Providers, each with a KEY
- Some Concrete JMESPATH examples
- Some basics on Discovery/Faceting/use of these and troubleshooting using Solr Directly
@dmer @alliomeria pretty sure we made a Video for Jack a time ago we can use as starting point
| 1.0 | Add full KeyName Provider Documentation Section - # What?
Key Name Providers are fundamental for searching/filtering/faceting. We should have
- How to documentation on creating new ones and the different Plugins we provide (including the " Is Date" option) and the Entity Reference One
- What/how Archipelago harvests/compiles data using them
- How these are exposed to the Search API but also programmatically as properties of a field (for DEVS)
- Good practices: e.g if the keys you want to expose are all flat and unique you can use the JSONLD Plugin and type them all into the "extra keys" fields instead of generating too many Key Name Providers, each with a KEY
- Some Concrete JMESPATH examples
- Some basics on Discovery/Faceting/use of these and troubleshooting using Solr Directly
@dmer @alliomeria pretty sure we made a Video for Jack a time ago we can use as starting point
| non_priority | add full keyname provider documentation section what key name providers are fundamental for searching filtering faceting we should have how to documentation on creating new ones and the different plugins we provide including the is date option and the entity reference one what how archipelago harvests compiles data using them how these are exposed to the search api but also programmatically as properties of a field for devs good practices e g if the keys you want to expose are all flat and unique you can use the jsonld plugin and type them all into the extra keys fields instead of generating too many key name providers each with a key some concrete jmespath examples some basics on discovery faceting use of these and troubleshooting using solr directly dmer alliomeria pretty sure we made a video for jack a time ago we can use as starting point | 0 |
287,299 | 31,834,284,494 | IssuesEvent | 2023-09-14 12:38:40 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | Expose "maxSpaces"-like property via Spaces setup and start contracts on both client- and server-side. | Team:Security Feature:Security/Spaces NeededFor:Core NeededFor:DataDiscovery Project:Serverless MVP | ## Summary
Spaces API consumers also need to determine whether they can show spaces specific UIs. Since the Spaces plugin is enabled even when the UI is hidden, the spaces plugin contract needs to provide a way of communicating whether the UI should be enabled. | True | Expose "maxSpaces"-like property via Spaces setup and start contracts on both client- and server-side. - ## Summary
Spaces API consumers also need to determine whether they can show spaces specific UIs. Since the Spaces plugin is enabled even when the UI is hidden, the spaces plugin contract needs to provide a way of communicating whether the UI should be enabled. | non_priority | expose maxspaces like property via spaces setup and start contracts on both client and server side summary spaces api consumers also need to determine whether they can show spaces specific uis since the spaces plugin is enabled even when the ui is hidden the spaces plugin contract needs to provide a way of communicating whether the ui should be enabled | 0 |
162,120 | 25,482,618,458 | IssuesEvent | 2022-11-26 01:01:44 | CERT-Polska/ursadb | https://api.github.com/repos/CERT-Polska/ursadb | closed | Add `debug parse` command | status:needs more design | We need a way to introspect the running database, or debug some issues without either recompiling the db with debug prints or attaching with gdb.
Especially now that I'll be working on query parsing & optimisation, I have a few ideas in mind. Right now I'd like to have something easy - `debug parse` command:
```
debug parse index "hmm" with [gram3];
```
Will parse the command `index "hmm" with [gram3]` in this case, pretty print the parsed tree and return it to the user. Expected result is for example:
```
index
"hmm"
index_type_list
gram3
```
(or whatever the real query tree looks like).
This can be returned as a structured JSON or as plain text - it's intended for debugging so we don't make any guarantees about the output. | 1.0 | Add `debug parse` command - We need a way to introspect the running database, or debug some issues without either recompiling the db with debug prints or attaching with gdb.
Especially now that I'll be working on query parsing & optimisation, I have a few ideas in mind. Right now I'd like to have something easy - `debug parse` command:
```
debug parse index "hmm" with [gram3];
```
Will parse the command `index "hmm" with [gram3]` in this case, pretty print the parsed tree and return it to the user. Expected result is for example:
```
index
"hmm"
index_type_list
gram3
```
(or whatever the real query tree looks like).
This can be returned as a structured JSON or as plain text - it's intended for debugging so we don't make any guarantees about the output. | non_priority | add debug parse command we need a way to introspect the running database or debug some issues without either recompiling the db with debug prints or attaching with gdb especially now that i ll be working on query parsing optimisation i have a few ideas in mind right now i d like to have something easy debug parse command debug parse index hmm with will parse the command index hmm with in this case pretty print the parsed tree and return it to the user expected result is for example index hmm index type list or whatever the real query tree looks like this can be returned as a structured json or as plain text it s intended for debugging so we don t make any guarantees about the output | 0 |
221,223 | 24,599,697,613 | IssuesEvent | 2022-10-14 11:23:38 | arbindo/mimock | https://api.github.com/repos/arbindo/mimock | closed | postgresql-42.3.3.jar: 1 vulnerabilities (highest severity is: 8.0) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postgresql-42.3.3.jar</b></p></summary>
<p>PostgreSQL JDBC Driver Postgresql</p>
<p>Library home page: <a href="https://jdbc.postgresql.org">https://jdbc.postgresql.org</a></p>
<p>Path to dependency file: /mimock-backend/pom.xml</p>
<p>Path to vulnerable library: /.m2/repository/org/postgresql/postgresql/42.3.3/postgresql-42.3.3.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/arbindo/mimock/commit/8055f41fefa4c77dfb0ed76c4220dbf3b2d33c6f">8055f41fefa4c77dfb0ed76c4220dbf3b2d33c6f</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-31197](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-31197) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.0 | postgresql-42.3.3.jar | Direct | 42.4.1 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-31197</summary>
### Vulnerable Library - <b>postgresql-42.3.3.jar</b></p>
<p>PostgreSQL JDBC Driver Postgresql</p>
<p>Library home page: <a href="https://jdbc.postgresql.org">https://jdbc.postgresql.org</a></p>
<p>Path to dependency file: /mimock-backend/pom.xml</p>
<p>Path to vulnerable library: /.m2/repository/org/postgresql/postgresql/42.3.3/postgresql-42.3.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **postgresql-42.3.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/arbindo/mimock/commit/8055f41fefa4c77dfb0ed76c4220dbf3b2d33c6f">8055f41fefa4c77dfb0ed76c4220dbf3b2d33c6f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
PostgreSQL JDBC Driver (PgJDBC for short) allows Java programs to connect to a PostgreSQL database using standard, database independent Java code. The PGJDBC implementation of the `java.sql.ResultRow.refreshRow()` method is not performing escaping of column names so a malicious column name that contains a statement terminator, e.g. `;`, could lead to SQL injection. This could lead to executing additional SQL commands as the application's JDBC user. User applications that do not invoke the `ResultSet.refreshRow()` method are not impacted. User application that do invoke that method are impacted if the underlying database that they are querying via their JDBC application may be under the control of an attacker. The attack requires the attacker to trick the user into executing SQL against a table name who's column names would contain the malicious SQL and subsequently invoke the `refreshRow()` method on the ResultSet. Note that the application's JDBC user and the schema owner need not be the same. A JDBC application that executes as a privileged user querying database schemas owned by potentially malicious less-privileged users would be vulnerable. In that situation it may be possible for the malicious user to craft a schema that causes the application to execute commands as the privileged user. Patched versions will be released as `42.2.26` and `42.4.1`. Users are advised to upgrade. There are no known workarounds for this issue.
<p>Publish Date: 2022-08-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-31197>CVE-2022-31197</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>8.0</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/pgjdbc/pgjdbc/security/advisories/GHSA-r38f-c4h4-hqq2">https://github.com/pgjdbc/pgjdbc/security/advisories/GHSA-r38f-c4h4-hqq2</a></p>
<p>Release Date: 2022-08-03</p>
<p>Fix Resolution: 42.4.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | True | postgresql-42.3.3.jar: 1 vulnerabilities (highest severity is: 8.0) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postgresql-42.3.3.jar</b></p></summary>
<p>PostgreSQL JDBC Driver Postgresql</p>
<p>Library home page: <a href="https://jdbc.postgresql.org">https://jdbc.postgresql.org</a></p>
<p>Path to dependency file: /mimock-backend/pom.xml</p>
<p>Path to vulnerable library: /.m2/repository/org/postgresql/postgresql/42.3.3/postgresql-42.3.3.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/arbindo/mimock/commit/8055f41fefa4c77dfb0ed76c4220dbf3b2d33c6f">8055f41fefa4c77dfb0ed76c4220dbf3b2d33c6f</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-31197](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-31197) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.0 | postgresql-42.3.3.jar | Direct | 42.4.1 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-31197</summary>
### Vulnerable Library - <b>postgresql-42.3.3.jar</b></p>
<p>PostgreSQL JDBC Driver Postgresql</p>
<p>Library home page: <a href="https://jdbc.postgresql.org">https://jdbc.postgresql.org</a></p>
<p>Path to dependency file: /mimock-backend/pom.xml</p>
<p>Path to vulnerable library: /.m2/repository/org/postgresql/postgresql/42.3.3/postgresql-42.3.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **postgresql-42.3.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/arbindo/mimock/commit/8055f41fefa4c77dfb0ed76c4220dbf3b2d33c6f">8055f41fefa4c77dfb0ed76c4220dbf3b2d33c6f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
PostgreSQL JDBC Driver (PgJDBC for short) allows Java programs to connect to a PostgreSQL database using standard, database independent Java code. The PGJDBC implementation of the `java.sql.ResultRow.refreshRow()` method is not performing escaping of column names so a malicious column name that contains a statement terminator, e.g. `;`, could lead to SQL injection. This could lead to executing additional SQL commands as the application's JDBC user. User applications that do not invoke the `ResultSet.refreshRow()` method are not impacted. User application that do invoke that method are impacted if the underlying database that they are querying via their JDBC application may be under the control of an attacker. The attack requires the attacker to trick the user into executing SQL against a table name who's column names would contain the malicious SQL and subsequently invoke the `refreshRow()` method on the ResultSet. Note that the application's JDBC user and the schema owner need not be the same. A JDBC application that executes as a privileged user querying database schemas owned by potentially malicious less-privileged users would be vulnerable. In that situation it may be possible for the malicious user to craft a schema that causes the application to execute commands as the privileged user. Patched versions will be released as `42.2.26` and `42.4.1`. Users are advised to upgrade. There are no known workarounds for this issue.
<p>Publish Date: 2022-08-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-31197>CVE-2022-31197</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>8.0</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/pgjdbc/pgjdbc/security/advisories/GHSA-r38f-c4h4-hqq2">https://github.com/pgjdbc/pgjdbc/security/advisories/GHSA-r38f-c4h4-hqq2</a></p>
<p>Release Date: 2022-08-03</p>
<p>Fix Resolution: 42.4.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | non_priority | postgresql jar vulnerabilities highest severity is vulnerable library postgresql jar postgresql jdbc driver postgresql library home page a href path to dependency file mimock backend pom xml path to vulnerable library repository org postgresql postgresql postgresql jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high postgresql jar direct details cve vulnerable library postgresql jar postgresql jdbc driver postgresql library home page a href path to dependency file mimock backend pom xml path to vulnerable library repository org postgresql postgresql postgresql jar dependency hierarchy x postgresql jar vulnerable library found in head commit a href found in base branch main vulnerability details postgresql jdbc driver pgjdbc for short allows java programs to connect to a postgresql database using standard database independent java code the pgjdbc implementation of the java sql resultrow refreshrow method is not performing escaping of column names so a malicious column name that contains a statement terminator e g could lead to sql injection this could lead to executing additional sql commands as the application s jdbc user user applications that do not invoke the resultset refreshrow method are not impacted user application that do invoke that method are impacted if the underlying database that they are querying via their jdbc application may be under the control of an attacker the attack requires the attacker to trick the user into executing sql against a table name who s column names would contain the malicious sql and subsequently invoke the refreshrow method on the resultset note that the application s jdbc user and the schema owner need not be the same a jdbc application that executes as a privileged user querying database schemas owned by potentially malicious less privileged users would be vulnerable in that situation it may be possible for the malicious user to craft a schema that causes the application to execute commands as the privileged user patched versions will be released as and users are advised to upgrade there are no known workarounds for this issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
41,079 | 6,890,715,240 | IssuesEvent | 2017-11-22 14:53:21 | pilosa/pilosa | https://api.github.com/repos/pilosa/pilosa | closed | BSI != range operation documentation | documentation | ### Description
The new `Range` query `!=` operation need to be documented.
This operation works on != null, and != _integer_ | 1.0 | BSI != range operation documentation - ### Description
The new `Range` query `!=` operation need to be documented.
This operation works on != null, and != _integer_ | non_priority | bsi range operation documentation description the new range query operation need to be documented this operation works on null and integer | 0 |
335,698 | 30,082,047,459 | IssuesEvent | 2023-06-29 04:57:44 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: failover/system-non-liveness/deadlock/lease=expiration failed | C-test-failure O-robot O-roachtest branch-master release-blocker T-kv | roachtest.failover/system-non-liveness/deadlock/lease=expiration [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestWeeklyBazel/10720111?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestWeeklyBazel/10720111?buildTab=artifacts#/failover/system-non-liveness/deadlock/lease=expiration) on master @ [7fd4c21157221eae9e7d5892d89d2b5a671aba3e](https://github.com/cockroachdb/cockroach/commits/7fd4c21157221eae9e7d5892d89d2b5a671aba3e):
```
(assertions.go:333).Fail:
Error Trace: github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/failover.go:1408
github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/failover.go:1059
main/pkg/cmd/roachtest/monitor.go:105
golang.org/x/sync/errgroup/external/org_golang_x_sync/errgroup/errgroup.go:75
GOROOT/src/runtime/asm_amd64.s:1594
Error: Received unexpected error:
pq: error getting span statistics - number of spans in request payload (1064) exceeds 'server.span_stats.span_batch_limit' cluster setting limit (500)
Test: failover/system-non-liveness/deadlock/lease=expiration
(require.go:1360).NoError: FailNow called
(monitor.go:137).Wait: monitor failure: monitor task failed: t.Fatal() was called
test artifacts and logs in: /artifacts/failover/system-non-liveness/deadlock/lease=expiration/run_1
```
<p>Parameters: <code>ROACHTEST_arch=amd64</code>
, <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=2</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*failover/system-non-liveness/deadlock/lease=expiration.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-29196 | 2.0 | roachtest: failover/system-non-liveness/deadlock/lease=expiration failed - roachtest.failover/system-non-liveness/deadlock/lease=expiration [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestWeeklyBazel/10720111?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestWeeklyBazel/10720111?buildTab=artifacts#/failover/system-non-liveness/deadlock/lease=expiration) on master @ [7fd4c21157221eae9e7d5892d89d2b5a671aba3e](https://github.com/cockroachdb/cockroach/commits/7fd4c21157221eae9e7d5892d89d2b5a671aba3e):
```
(assertions.go:333).Fail:
Error Trace: github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/failover.go:1408
github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/failover.go:1059
main/pkg/cmd/roachtest/monitor.go:105
golang.org/x/sync/errgroup/external/org_golang_x_sync/errgroup/errgroup.go:75
GOROOT/src/runtime/asm_amd64.s:1594
Error: Received unexpected error:
pq: error getting span statistics - number of spans in request payload (1064) exceeds 'server.span_stats.span_batch_limit' cluster setting limit (500)
Test: failover/system-non-liveness/deadlock/lease=expiration
(require.go:1360).NoError: FailNow called
(monitor.go:137).Wait: monitor failure: monitor task failed: t.Fatal() was called
test artifacts and logs in: /artifacts/failover/system-non-liveness/deadlock/lease=expiration/run_1
```
<p>Parameters: <code>ROACHTEST_arch=amd64</code>
, <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=2</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*failover/system-non-liveness/deadlock/lease=expiration.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-29196 | non_priority | roachtest failover system non liveness deadlock lease expiration failed roachtest failover system non liveness deadlock lease expiration with on master assertions go fail error trace github com cockroachdb cockroach pkg cmd roachtest tests failover go github com cockroachdb cockroach pkg cmd roachtest tests failover go main pkg cmd roachtest monitor go golang org x sync errgroup external org golang x sync errgroup errgroup go goroot src runtime asm s error received unexpected error pq error getting span statistics number of spans in request payload exceeds server span stats span batch limit cluster setting limit test failover system non liveness deadlock lease expiration require go noerror failnow called monitor go wait monitor failure monitor task failed t fatal was called test artifacts and logs in artifacts failover system non liveness deadlock lease expiration run parameters roachtest arch roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see cc cockroachdb kv triage jira issue crdb | 0 |
26,686 | 4,239,362,404 | IssuesEvent | 2016-07-06 09:10:53 | e-government-ua/iBP | https://api.github.com/repos/e-government-ua/iBP | closed | Нетішин, Хмельницька обл. - розкрити 7 послуг. Земля | in work ready for release test | Організація ринку, ярмарку
13
Взяття на облік громадян, які потребують поліпшення житлових умов
17
Надання довідок незайнятому населенню про відсутність земельних ділянок для ведення ОСГ
23
Надання інформації, що підтверджує відсутність(наявність) земельної ділянк
657
Видача Акту обстеження земельної ділянки на наявність зелених насаджень
106
Внесення змін до договору оренди землі
1481
Дозвіл на виготовлення технічної документації із землеустрою щодо встановлення меж земельної ділянки
1392
| 1.0 | Нетішин, Хмельницька обл. - розкрити 7 послуг. Земля - Організація ринку, ярмарку
13
Взяття на облік громадян, які потребують поліпшення житлових умов
17
Надання довідок незайнятому населенню про відсутність земельних ділянок для ведення ОСГ
23
Надання інформації, що підтверджує відсутність(наявність) земельної ділянк
657
Видача Акту обстеження земельної ділянки на наявність зелених насаджень
106
Внесення змін до договору оренди землі
1481
Дозвіл на виготовлення технічної документації із землеустрою щодо встановлення меж земельної ділянки
1392
| non_priority | нетішин хмельницька обл розкрити послуг земля організація ринку ярмарку взяття на облік громадян які потребують поліпшення житлових умов надання довідок незайнятому населенню про відсутність земельних ділянок для ведення осг надання інформації що підтверджує відсутність наявність земельної ділянк видача акту обстеження земельної ділянки на наявність зелених насаджень внесення змін до договору оренди землі дозвіл на виготовлення технічної документації із землеустрою щодо встановлення меж земельної ділянки | 0 |
32,678 | 26,900,945,723 | IssuesEvent | 2023-02-06 15:36:18 | evcc-io/evcc | https://api.github.com/repos/evcc-io/evcc | closed | Debian / Ubuntu Service: Separater Nutzer | infrastructure | ### Discussed in https://github.com/evcc-io/evcc/discussions/4840
<div type='discussions-op-text'>
<sup>Originally posted by **pauxus** October 16, 2022</sup>
Gerade nachdem evcc jetzt ja auch state im Dateisystem ablegt:
Aktuell ist der Service Nutzer unter Debian "root", was ja aus Sicherheitsaspekten eher unzweckmäßig ist. Sollte hier nicht sinnvollerweise ein eigener Nutzer (evcc) angelegt werden? Damit wäre dann auch die sqlite db z.B. unter "/var/lib/evcc". </div> | 1.0 | Debian / Ubuntu Service: Separater Nutzer - ### Discussed in https://github.com/evcc-io/evcc/discussions/4840
<div type='discussions-op-text'>
<sup>Originally posted by **pauxus** October 16, 2022</sup>
Gerade nachdem evcc jetzt ja auch state im Dateisystem ablegt:
Aktuell ist der Service Nutzer unter Debian "root", was ja aus Sicherheitsaspekten eher unzweckmäßig ist. Sollte hier nicht sinnvollerweise ein eigener Nutzer (evcc) angelegt werden? Damit wäre dann auch die sqlite db z.B. unter "/var/lib/evcc". </div> | non_priority | debian ubuntu service separater nutzer discussed in originally posted by pauxus october gerade nachdem evcc jetzt ja auch state im dateisystem ablegt aktuell ist der service nutzer unter debian root was ja aus sicherheitsaspekten eher unzweckmäßig ist sollte hier nicht sinnvollerweise ein eigener nutzer evcc angelegt werden damit wäre dann auch die sqlite db z b unter var lib evcc | 0 |
127,526 | 10,474,349,387 | IssuesEvent | 2019-09-23 14:21:25 | status-im/nim-beacon-chain | https://api.github.com/repos/status-im/nim-beacon-chain | opened | hash_tree_root: pathological case in test_fixture_sanity_blocks | performance test suite | When compiling test_fixture_sanity_blocks as a standalone and running it, I manage to block my machine for a couple minutes on what seems to be faststreams memory manager:
```Nim
[Suite] Official - Sanity - Blocks [Preset: mainnet]
[OK] [Invalid] Previous slot block transition (prev_slot_block_transition)
[OK] [Valid] Same slot block transition (same_slot_block_transition)
[OK] [Valid] Empty block transition (empty_block_transition)
[OK] [Valid] Skipped Slots (skipped_slots)
^CTraceback (most recent call last)
/home/beta/Programming/Status/nim-beacon-chain/tests/official/test_fixture_sanity_blocks.nim(83) test_fixture_sanity_blocks
/home/beta/Programming/Status/nim-beacon-chain/tests/official/test_fixture_sanity_blocks.nim(43) testImpl_blck_empty_epoch_transition
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/state_transition.nim(115) state_transition
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/state_transition.nim(63) process_slots
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/state_transition.nim(45) process_slot
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(584) hash_tree_root
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(454) hashTreeRootImpl
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(584) hash_tree_root
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(454) hashTreeRootImpl
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(584) hash_tree_root
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(525) hashTreeRootImpl
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(446) merkelizeSerializedChunks
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(440) merkelizeSerializedChunks
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(426) newSszHashingStream
/home/beta/Programming/Status/nim-beacon-chain/vendor/nim-faststreams/faststreams/output_stream.nim(88) initWithSinglePage
/home/beta/Programming/Status/nim-beacon-chain/vendor/nim-faststreams/faststreams/output_stream.nim(76) addPage
/home/beta/.choosenim/toolchains/nim-0.19.6/lib/system/gc.nim(477) newObj
SIGINT: Interrupted by Ctrl-C.
Traceback (most recent call last)
proxyexe.nim(62) proxyexe
proxyexe.nim(49) main
osproc.nim(1136) waitForExit
SIGINT: Interrupted by Ctrl-C.
``` | 1.0 | hash_tree_root: pathological case in test_fixture_sanity_blocks - When compiling test_fixture_sanity_blocks as a standalone and running it, I manage to block my machine for a couple minutes on what seems to be faststreams memory manager:
```Nim
[Suite] Official - Sanity - Blocks [Preset: mainnet]
[OK] [Invalid] Previous slot block transition (prev_slot_block_transition)
[OK] [Valid] Same slot block transition (same_slot_block_transition)
[OK] [Valid] Empty block transition (empty_block_transition)
[OK] [Valid] Skipped Slots (skipped_slots)
^CTraceback (most recent call last)
/home/beta/Programming/Status/nim-beacon-chain/tests/official/test_fixture_sanity_blocks.nim(83) test_fixture_sanity_blocks
/home/beta/Programming/Status/nim-beacon-chain/tests/official/test_fixture_sanity_blocks.nim(43) testImpl_blck_empty_epoch_transition
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/state_transition.nim(115) state_transition
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/state_transition.nim(63) process_slots
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/state_transition.nim(45) process_slot
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(584) hash_tree_root
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(454) hashTreeRootImpl
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(584) hash_tree_root
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(454) hashTreeRootImpl
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(584) hash_tree_root
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(525) hashTreeRootImpl
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(446) merkelizeSerializedChunks
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(440) merkelizeSerializedChunks
/home/beta/Programming/Status/nim-beacon-chain/beacon_chain/ssz.nim(426) newSszHashingStream
/home/beta/Programming/Status/nim-beacon-chain/vendor/nim-faststreams/faststreams/output_stream.nim(88) initWithSinglePage
/home/beta/Programming/Status/nim-beacon-chain/vendor/nim-faststreams/faststreams/output_stream.nim(76) addPage
/home/beta/.choosenim/toolchains/nim-0.19.6/lib/system/gc.nim(477) newObj
SIGINT: Interrupted by Ctrl-C.
Traceback (most recent call last)
proxyexe.nim(62) proxyexe
proxyexe.nim(49) main
osproc.nim(1136) waitForExit
SIGINT: Interrupted by Ctrl-C.
``` | non_priority | hash tree root pathological case in test fixture sanity blocks when compiling test fixture sanity blocks as a standalone and running it i manage to block my machine for a couple minutes on what seems to be faststreams memory manager nim official sanity blocks previous slot block transition prev slot block transition same slot block transition same slot block transition empty block transition empty block transition skipped slots skipped slots ctraceback most recent call last home beta programming status nim beacon chain tests official test fixture sanity blocks nim test fixture sanity blocks home beta programming status nim beacon chain tests official test fixture sanity blocks nim testimpl blck empty epoch transition home beta programming status nim beacon chain beacon chain state transition nim state transition home beta programming status nim beacon chain beacon chain state transition nim process slots home beta programming status nim beacon chain beacon chain state transition nim process slot home beta programming status nim beacon chain beacon chain ssz nim hash tree root home beta programming status nim beacon chain beacon chain ssz nim hashtreerootimpl home beta programming status nim beacon chain beacon chain ssz nim hash tree root home beta programming status nim beacon chain beacon chain ssz nim hashtreerootimpl home beta programming status nim beacon chain beacon chain ssz nim hash tree root home beta programming status nim beacon chain beacon chain ssz nim hashtreerootimpl home beta programming status nim beacon chain beacon chain ssz nim merkelizeserializedchunks home beta programming status nim beacon chain beacon chain ssz nim merkelizeserializedchunks home beta programming status nim beacon chain beacon chain ssz nim newsszhashingstream home beta programming status nim beacon chain vendor nim faststreams faststreams output stream nim initwithsinglepage home beta programming status nim beacon chain vendor nim faststreams faststreams output stream nim addpage home beta choosenim toolchains nim lib system gc nim newobj sigint interrupted by ctrl c traceback most recent call last proxyexe nim proxyexe proxyexe nim main osproc nim waitforexit sigint interrupted by ctrl c | 0 |
164,479 | 20,364,488,993 | IssuesEvent | 2022-02-21 02:54:01 | arohablue/skill-india-frontend | https://api.github.com/repos/arohablue/skill-india-frontend | closed | CVE-2018-11698 (High) detected in node-sass-4.11.0.tgz, node-sassv4.12.0 - autoclosed | security vulnerability | ## CVE-2018-11698 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.11.0.tgz</b>, <b>node-sassv4.12.0</b></p></summary>
<p>
<details><summary><b>node-sass-4.11.0.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz</a></p>
<p>Path to dependency file: /skill-india-frontend/package.json</p>
<p>Path to vulnerable library: skill-india-frontend/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.13.1.tgz (Root Library)
- :x: **node-sass-4.11.0.tgz** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in LibSass through 3.5.4. An out-of-bounds read of a memory region was found in the function Sass::handle_error which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service.
<p>Publish Date: 2018-06-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11698>CVE-2018-11698</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/sass/libsass/releases/tag/3.5.5">https://github.com/sass/libsass/releases/tag/3.5.5</a></p>
<p>Release Date: 2018-06-04</p>
<p>Fix Resolution: libsass - 3.5.5;node-sass - 4.14.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-11698 (High) detected in node-sass-4.11.0.tgz, node-sassv4.12.0 - autoclosed - ## CVE-2018-11698 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.11.0.tgz</b>, <b>node-sassv4.12.0</b></p></summary>
<p>
<details><summary><b>node-sass-4.11.0.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz</a></p>
<p>Path to dependency file: /skill-india-frontend/package.json</p>
<p>Path to vulnerable library: skill-india-frontend/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.13.1.tgz (Root Library)
- :x: **node-sass-4.11.0.tgz** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in LibSass through 3.5.4. An out-of-bounds read of a memory region was found in the function Sass::handle_error which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service.
<p>Publish Date: 2018-06-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11698>CVE-2018-11698</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/sass/libsass/releases/tag/3.5.5">https://github.com/sass/libsass/releases/tag/3.5.5</a></p>
<p>Release Date: 2018-06-04</p>
<p>Fix Resolution: libsass - 3.5.5;node-sass - 4.14.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in node sass tgz node autoclosed cve high severity vulnerability vulnerable libraries node sass tgz node node sass tgz wrapper around libsass library home page a href path to dependency file skill india frontend package json path to vulnerable library skill india frontend node modules node sass package json dependency hierarchy build angular tgz root library x node sass tgz vulnerable library vulnerability details an issue was discovered in libsass through an out of bounds read of a memory region was found in the function sass handle error which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass node sass step up your open source security game with whitesource | 0 |
185,019 | 21,785,052,467 | IssuesEvent | 2022-05-14 02:18:17 | Yash-Handa/GitHub-Org-Geographics | https://api.github.com/repos/Yash-Handa/GitHub-Org-Geographics | closed | WS-2019-0369 (Medium) detected in handlebars-4.1.1.tgz - autoclosed | security vulnerability | ## WS-2019-0369 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.1.1.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.1.1.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.1.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/GitHub-Org-Geographics/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/GitHub-Org-Geographics/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- karma-coverage-istanbul-reporter-2.0.5.tgz (Root Library)
- istanbul-api-2.1.1.tgz
- istanbul-reports-2.1.1.tgz
- :x: **handlebars-4.1.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Yash-Handa/GitHub-Org-Geographics/commit/69e1362228223135794f86d35da5d76214470dda">69e1362228223135794f86d35da5d76214470dda</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype Pollution vulnerability found in handlebars.js before 4.5.3. Attacker may use Remote-Code-Execution exploits.
<p>Publish Date: 2020-01-08
<p>URL: <a href=https://github.com/RetireJS/retire.js/commit/f07a7557d3fc1c26b86fe11a5b33cb1b8f3dcf2f>WS-2019-0369</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/wycats/handlebars.js/blob/master/release-notes.md#v453---november-18th-2019">https://github.com/wycats/handlebars.js/blob/master/release-notes.md#v453---november-18th-2019</a></p>
<p>Release Date: 2020-01-08</p>
<p>Fix Resolution: handlebars - 4.5.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2019-0369 (Medium) detected in handlebars-4.1.1.tgz - autoclosed - ## WS-2019-0369 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.1.1.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.1.1.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.1.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/GitHub-Org-Geographics/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/GitHub-Org-Geographics/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- karma-coverage-istanbul-reporter-2.0.5.tgz (Root Library)
- istanbul-api-2.1.1.tgz
- istanbul-reports-2.1.1.tgz
- :x: **handlebars-4.1.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Yash-Handa/GitHub-Org-Geographics/commit/69e1362228223135794f86d35da5d76214470dda">69e1362228223135794f86d35da5d76214470dda</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype Pollution vulnerability found in handlebars.js before 4.5.3. Attacker may use Remote-Code-Execution exploits.
<p>Publish Date: 2020-01-08
<p>URL: <a href=https://github.com/RetireJS/retire.js/commit/f07a7557d3fc1c26b86fe11a5b33cb1b8f3dcf2f>WS-2019-0369</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/wycats/handlebars.js/blob/master/release-notes.md#v453---november-18th-2019">https://github.com/wycats/handlebars.js/blob/master/release-notes.md#v453---november-18th-2019</a></p>
<p>Release Date: 2020-01-08</p>
<p>Fix Resolution: handlebars - 4.5.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | ws medium detected in handlebars tgz autoclosed ws medium severity vulnerability vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file tmp ws scm github org geographics package json path to vulnerable library tmp ws scm github org geographics node modules handlebars package json dependency hierarchy karma coverage istanbul reporter tgz root library istanbul api tgz istanbul reports tgz x handlebars tgz vulnerable library found in head commit a href vulnerability details prototype pollution vulnerability found in handlebars js before attacker may use remote code execution exploits publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution handlebars step up your open source security game with whitesource | 0 |
43,118 | 23,122,180,219 | IssuesEvent | 2022-07-27 23:07:36 | open-policy-agent/gatekeeper | https://api.github.com/repos/open-policy-agent/gatekeeper | opened | Do not deserialize admission request object more than once | Performance | With the new Expansion feature in https://github.com/open-policy-agent/gatekeeper/pull/2062, we are deserializing the request object in the validation webhook twice:
- once when we expand the object (new code)
- again when we actually match the object ([existing code](https://github.com/open-policy-agent/gatekeeper/blob/841e10000acf3e7fc42e35a0ab1be5dfbe0138ae/pkg/target/matcher.go#L67-L87))
We should somehow pass the deserialized object down the chain such that it does not have to be deserialized again when matched.
| True | Do not deserialize admission request object more than once - With the new Expansion feature in https://github.com/open-policy-agent/gatekeeper/pull/2062, we are deserializing the request object in the validation webhook twice:
- once when we expand the object (new code)
- again when we actually match the object ([existing code](https://github.com/open-policy-agent/gatekeeper/blob/841e10000acf3e7fc42e35a0ab1be5dfbe0138ae/pkg/target/matcher.go#L67-L87))
We should somehow pass the deserialized object down the chain such that it does not have to be deserialized again when matched.
| non_priority | do not deserialize admission request object more than once with the new expansion feature in we are deserializing the request object in the validation webhook twice once when we expand the object new code again when we actually match the object we should somehow pass the deserialized object down the chain such that it does not have to be deserialized again when matched | 0 |
223,026 | 17,101,165,779 | IssuesEvent | 2021-07-09 11:27:27 | bitcoin-s/bitcoin-s | https://api.github.com/repos/bitcoin-s/bitcoin-s | closed | Update secp256k1 build docs | bug documentation | **Describe the bug**
https://bitcoin-s.org/docs/secp256k1/jni-modify contains "You will need to go to the bitcoin-s/secp256k1" which should now be "bitcoin-s/secp256k1-zkp"
May also want to add steps to build on Mac M1 architecture
https://github.com/bitcoin-s/secp256k1-zkp/commit/35d09472e1647510f54ee9ebe34fa338a7e6e446 | 1.0 | Update secp256k1 build docs - **Describe the bug**
https://bitcoin-s.org/docs/secp256k1/jni-modify contains "You will need to go to the bitcoin-s/secp256k1" which should now be "bitcoin-s/secp256k1-zkp"
May also want to add steps to build on Mac M1 architecture
https://github.com/bitcoin-s/secp256k1-zkp/commit/35d09472e1647510f54ee9ebe34fa338a7e6e446 | non_priority | update build docs describe the bug contains you will need to go to the bitcoin s which should now be bitcoin s zkp may also want to add steps to build on mac architecture | 0 |
208,141 | 16,104,974,179 | IssuesEvent | 2021-04-27 13:59:04 | GuillaumeFalourd/developers-tips-and-tricks-resources | https://api.github.com/repos/GuillaumeFalourd/developers-tips-and-tricks-resources | opened | Add new cheat-sheets contents | documentation enhancement | ## What is expected?
Add new cheat-sheets to the [related section](https://github.com/GuillaumeFalourd/developers-tips-and-tricks-resources/tree/master/cheat-sheets) respecting the [repository guidelines](https://github.com/GuillaumeFalourd/developers-tips-and-tricks-resources/blob/master/CONTRIBUTION.md) | 1.0 | Add new cheat-sheets contents - ## What is expected?
Add new cheat-sheets to the [related section](https://github.com/GuillaumeFalourd/developers-tips-and-tricks-resources/tree/master/cheat-sheets) respecting the [repository guidelines](https://github.com/GuillaumeFalourd/developers-tips-and-tricks-resources/blob/master/CONTRIBUTION.md) | non_priority | add new cheat sheets contents what is expected add new cheat sheets to the respecting the | 0 |
31,198 | 25,408,425,391 | IssuesEvent | 2022-11-22 16:55:35 | SonarSource/sonarlint-visualstudio | https://api.github.com/repos/SonarSource/sonarlint-visualstudio | opened | [Infra] Auto-regenerate the list of Sonar Roslyn rules that can be suppressed on build | Infrastructure | ### Description
The list of issues C#/VB rules that can be suppress in Connected Mode is generated using a [T4 template](https://github.com/SonarSource/sonarlint-visualstudio/blob/69de0102f4155f7f45aafebbc41b0b79f2958af2/src/Roslyn.Suppressions/Roslyn.Suppressions/SupportedSuppressionBuilder.tt#L1).
At present the code is only regenerated manually i.e. we have to remember to do it each time we update the Roslyn analyzers.
We should use the T4 build integration to cause the file to be updated automatically on build. | 1.0 | [Infra] Auto-regenerate the list of Sonar Roslyn rules that can be suppressed on build - ### Description
The list of issues C#/VB rules that can be suppress in Connected Mode is generated using a [T4 template](https://github.com/SonarSource/sonarlint-visualstudio/blob/69de0102f4155f7f45aafebbc41b0b79f2958af2/src/Roslyn.Suppressions/Roslyn.Suppressions/SupportedSuppressionBuilder.tt#L1).
At present the code is only regenerated manually i.e. we have to remember to do it each time we update the Roslyn analyzers.
We should use the T4 build integration to cause the file to be updated automatically on build. | non_priority | auto regenerate the list of sonar roslyn rules that can be suppressed on build description the list of issues c vb rules that can be suppress in connected mode is generated using a at present the code is only regenerated manually i e we have to remember to do it each time we update the roslyn analyzers we should use the build integration to cause the file to be updated automatically on build | 0 |
167,426 | 20,726,076,551 | IssuesEvent | 2022-03-14 02:09:35 | hitokoto-osc/hitokoto-api | https://api.github.com/repos/hitokoto-osc/hitokoto-api | closed | commitizen-4.1.2.tgz: 4 vulnerabilities (highest severity is: 9.8) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commitizen-4.1.2.tgz</b></p></summary>
<p></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/commitizen/node_modules/lodash/package.json</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2020-28499](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | merge-1.2.1.tgz | Transitive | 4.2.4 | ❌ |
| [CVE-2020-8203](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.4 | lodash-4.17.15.tgz | Transitive | 4.1.3 | ❌ |
| [CVE-2021-23337](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.2 | lodash-4.17.15.tgz | Transitive | 4.1.3 | ❌ |
| [CVE-2020-28500](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | lodash-4.17.15.tgz | Transitive | 4.1.3 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-28499</summary>
### Vulnerable Library - <b>merge-1.2.1.tgz</b></p>
<p>Merge multiple objects into one, optionally creating a new cloned object. Similar to the jQuery.extend but more flexible. Works in Node.js and the browser.</p>
<p>Library home page: <a href="https://registry.npmjs.org/merge/-/merge-1.2.1.tgz">https://registry.npmjs.org/merge/-/merge-1.2.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/merge/package.json</p>
<p>
Dependency Hierarchy:
- commitizen-4.1.2.tgz (Root Library)
- find-node-modules-2.0.0.tgz
- :x: **merge-1.2.1.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
All versions of package merge are vulnerable to Prototype Pollution via _recursiveMerge .
<p>Publish Date: 2021-02-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499>CVE-2020-28499</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/yeikos/js.merge/pull/38">https://github.com/yeikos/js.merge/pull/38</a></p>
<p>Release Date: 2021-02-18</p>
<p>Fix Resolution (merge): 2.1.1</p>
<p>Direct dependency fix Resolution (commitizen): 4.2.4</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-8203</summary>
### Vulnerable Library - <b>lodash-4.17.15.tgz</b></p>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/commitizen/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- commitizen-4.1.2.tgz (Root Library)
- :x: **lodash-4.17.15.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20.
<p>Publish Date: 2020-07-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203>CVE-2020-8203</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.4</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1523">https://www.npmjs.com/advisories/1523</a></p>
<p>Release Date: 2020-07-15</p>
<p>Fix Resolution (lodash): 4.17.19</p>
<p>Direct dependency fix Resolution (commitizen): 4.1.3</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-23337</summary>
### Vulnerable Library - <b>lodash-4.17.15.tgz</b></p>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/commitizen/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- commitizen-4.1.2.tgz (Root Library)
- :x: **lodash-4.17.15.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.2</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution (lodash): 4.17.21</p>
<p>Direct dependency fix Resolution (commitizen): 4.1.3</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-28500</summary>
### Vulnerable Library - <b>lodash-4.17.15.tgz</b></p>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/commitizen/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- commitizen-4.1.2.tgz (Root Library)
- :x: **lodash-4.17.15.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions.
WhiteSource Note: After conducting further research, WhiteSource has determined that CVE-2020-28500 only affects environments with versions 4.0.0 to 4.17.20 of Lodash.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500>CVE-2020-28500</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution (lodash): 4.17.21</p>
<p>Direct dependency fix Resolution (commitizen): 4.1.3</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
<!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"commitizen","packageVersion":"4.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"commitizen:4.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.2.4","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-28499","vulnerabilityDetails":"All versions of package merge are vulnerable to Prototype Pollution via _recursiveMerge .","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"commitizen","packageVersion":"4.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"commitizen:4.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.1.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-8203","vulnerabilityDetails":"Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203","cvss3Severity":"high","cvss3Score":"7.4","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"commitizen","packageVersion":"4.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"commitizen:4.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.1.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23337","vulnerabilityDetails":"Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337","cvss3Severity":"high","cvss3Score":"7.2","cvss3Metrics":{"A":"High","AC":"Low","PR":"High","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"commitizen","packageVersion":"4.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"commitizen:4.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.1.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-28500","vulnerabilityDetails":"Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions.\n WhiteSource Note: After conducting further research, WhiteSource has determined that CVE-2020-28500 only affects environments with versions 4.0.0 to 4.17.20 of Lodash.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}]</REMEDIATE> --> | True | commitizen-4.1.2.tgz: 4 vulnerabilities (highest severity is: 9.8) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commitizen-4.1.2.tgz</b></p></summary>
<p></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/commitizen/node_modules/lodash/package.json</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2020-28499](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | merge-1.2.1.tgz | Transitive | 4.2.4 | ❌ |
| [CVE-2020-8203](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.4 | lodash-4.17.15.tgz | Transitive | 4.1.3 | ❌ |
| [CVE-2021-23337](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.2 | lodash-4.17.15.tgz | Transitive | 4.1.3 | ❌ |
| [CVE-2020-28500](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | lodash-4.17.15.tgz | Transitive | 4.1.3 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-28499</summary>
### Vulnerable Library - <b>merge-1.2.1.tgz</b></p>
<p>Merge multiple objects into one, optionally creating a new cloned object. Similar to the jQuery.extend but more flexible. Works in Node.js and the browser.</p>
<p>Library home page: <a href="https://registry.npmjs.org/merge/-/merge-1.2.1.tgz">https://registry.npmjs.org/merge/-/merge-1.2.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/merge/package.json</p>
<p>
Dependency Hierarchy:
- commitizen-4.1.2.tgz (Root Library)
- find-node-modules-2.0.0.tgz
- :x: **merge-1.2.1.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
All versions of package merge are vulnerable to Prototype Pollution via _recursiveMerge .
<p>Publish Date: 2021-02-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499>CVE-2020-28499</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/yeikos/js.merge/pull/38">https://github.com/yeikos/js.merge/pull/38</a></p>
<p>Release Date: 2021-02-18</p>
<p>Fix Resolution (merge): 2.1.1</p>
<p>Direct dependency fix Resolution (commitizen): 4.2.4</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-8203</summary>
### Vulnerable Library - <b>lodash-4.17.15.tgz</b></p>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/commitizen/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- commitizen-4.1.2.tgz (Root Library)
- :x: **lodash-4.17.15.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20.
<p>Publish Date: 2020-07-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203>CVE-2020-8203</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.4</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1523">https://www.npmjs.com/advisories/1523</a></p>
<p>Release Date: 2020-07-15</p>
<p>Fix Resolution (lodash): 4.17.19</p>
<p>Direct dependency fix Resolution (commitizen): 4.1.3</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-23337</summary>
### Vulnerable Library - <b>lodash-4.17.15.tgz</b></p>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/commitizen/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- commitizen-4.1.2.tgz (Root Library)
- :x: **lodash-4.17.15.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.2</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution (lodash): 4.17.21</p>
<p>Direct dependency fix Resolution (commitizen): 4.1.3</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-28500</summary>
### Vulnerable Library - <b>lodash-4.17.15.tgz</b></p>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/commitizen/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- commitizen-4.1.2.tgz (Root Library)
- :x: **lodash-4.17.15.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions.
WhiteSource Note: After conducting further research, WhiteSource has determined that CVE-2020-28500 only affects environments with versions 4.0.0 to 4.17.20 of Lodash.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500>CVE-2020-28500</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28500</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution (lodash): 4.17.21</p>
<p>Direct dependency fix Resolution (commitizen): 4.1.3</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
<!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"commitizen","packageVersion":"4.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"commitizen:4.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.2.4","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-28499","vulnerabilityDetails":"All versions of package merge are vulnerable to Prototype Pollution via _recursiveMerge .","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"commitizen","packageVersion":"4.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"commitizen:4.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.1.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-8203","vulnerabilityDetails":"Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203","cvss3Severity":"high","cvss3Score":"7.4","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"commitizen","packageVersion":"4.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"commitizen:4.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.1.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23337","vulnerabilityDetails":"Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337","cvss3Severity":"high","cvss3Score":"7.2","cvss3Metrics":{"A":"High","AC":"Low","PR":"High","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"commitizen","packageVersion":"4.1.2","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"commitizen:4.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.1.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-28500","vulnerabilityDetails":"Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions.\n WhiteSource Note: After conducting further research, WhiteSource has determined that CVE-2020-28500 only affects environments with versions 4.0.0 to 4.17.20 of Lodash.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}]</REMEDIATE> --> | non_priority | commitizen tgz vulnerabilities highest severity is vulnerable library commitizen tgz path to dependency file package json path to vulnerable library node modules commitizen node modules lodash package json vulnerabilities cve severity cvss dependency type fixed in remediation available high merge tgz transitive high lodash tgz transitive high lodash tgz transitive medium lodash tgz transitive details cve vulnerable library merge tgz merge multiple objects into one optionally creating a new cloned object similar to the jquery extend but more flexible works in node js and the browser library home page a href path to dependency file package json path to vulnerable library node modules merge package json dependency hierarchy commitizen tgz root library find node modules tgz x merge tgz vulnerable library found in base branch master vulnerability details all versions of package merge are vulnerable to prototype pollution via recursivemerge publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution merge direct dependency fix resolution commitizen step up your open source security game with whitesource cve vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules commitizen node modules lodash package json dependency hierarchy commitizen tgz root library x lodash tgz vulnerable library found in base branch master vulnerability details prototype pollution attack when using zipobjectdeep in lodash before publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash direct dependency fix resolution commitizen step up your open source security game with whitesource cve vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules commitizen node modules lodash package json dependency hierarchy commitizen tgz root library x lodash tgz vulnerable library found in base branch master vulnerability details lodash versions prior to are vulnerable to command injection via the template function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash direct dependency fix resolution commitizen step up your open source security game with whitesource cve vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules commitizen node modules lodash package json dependency hierarchy commitizen tgz root library x lodash tgz vulnerable library found in base branch master vulnerability details lodash versions prior to are vulnerable to regular expression denial of service redos via the tonumber trim and trimend functions whitesource note after conducting further research whitesource has determined that cve only affects environments with versions to of lodash publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash direct dependency fix resolution commitizen step up your open source security game with whitesource istransitivedependency false dependencytree commitizen isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails all versions of package merge are vulnerable to prototype pollution via recursivemerge vulnerabilityurl istransitivedependency false dependencytree commitizen isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails prototype pollution attack when using zipobjectdeep in lodash before vulnerabilityurl istransitivedependency false dependencytree commitizen isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails lodash versions prior to are vulnerable to command injection via the template function vulnerabilityurl istransitivedependency false dependencytree commitizen isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails lodash versions prior to are vulnerable to regular expression denial of service redos via the tonumber trim and trimend functions n whitesource note after conducting further research whitesource has determined that cve only affects environments with versions to of lodash vulnerabilityurl | 0 |
2,537 | 4,760,262,236 | IssuesEvent | 2016-10-25 02:15:16 | CactusDev/CactusBot | https://api.github.com/repos/CactusDev/CactusBot | opened | Discord service errors if response length too long | bug service:discord | The maximum length for Discord messages is 2,000 characters. If the response is too long, the bot should split up the message and send individually. | 1.0 | Discord service errors if response length too long - The maximum length for Discord messages is 2,000 characters. If the response is too long, the bot should split up the message and send individually. | non_priority | discord service errors if response length too long the maximum length for discord messages is characters if the response is too long the bot should split up the message and send individually | 0 |
122,114 | 17,687,357,294 | IssuesEvent | 2021-08-24 04:47:11 | panasalap/sqlite-3.22.0 | https://api.github.com/repos/panasalap/sqlite-3.22.0 | opened | CVE-2019-9937 (High) detected in multiple libraries | security vulnerability | ## CVE-2019-9937 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>sqliteversion-3.22.0</b>, <b>sqliteversion-3.22.0</b>, <b>sqliteversion-3.22.0</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In SQLite 3.27.2, interleaving reads and writes in a single transaction with an fts5 virtual table will lead to a NULL Pointer Dereference in fts5ChunkIterate in sqlite3.c. This is related to ext/fts5/fts5_hash.c and ext/fts5/fts5_index.c.
<p>Publish Date: 2019-03-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-9937>CVE-2019-9937</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.sqlite.org/releaselog/3_28_0.html">https://www.sqlite.org/releaselog/3_28_0.html</a></p>
<p>Release Date: 2019-03-22</p>
<p>Fix Resolution: 3.28.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-9937 (High) detected in multiple libraries - ## CVE-2019-9937 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>sqliteversion-3.22.0</b>, <b>sqliteversion-3.22.0</b>, <b>sqliteversion-3.22.0</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In SQLite 3.27.2, interleaving reads and writes in a single transaction with an fts5 virtual table will lead to a NULL Pointer Dereference in fts5ChunkIterate in sqlite3.c. This is related to ext/fts5/fts5_hash.c and ext/fts5/fts5_index.c.
<p>Publish Date: 2019-03-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-9937>CVE-2019-9937</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.sqlite.org/releaselog/3_28_0.html">https://www.sqlite.org/releaselog/3_28_0.html</a></p>
<p>Release Date: 2019-03-22</p>
<p>Fix Resolution: 3.28.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries sqliteversion sqliteversion sqliteversion vulnerability details in sqlite interleaving reads and writes in a single transaction with an virtual table will lead to a null pointer dereference in in c this is related to ext hash c and ext index c publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
72,722 | 19,430,764,592 | IssuesEvent | 2021-12-21 11:40:12 | openego/eGon-data | https://api.github.com/repos/openego/eGon-data | closed | Insert cost assumptions from NEP and technology-data to scenario_parameters | :building_construction: integration | Fuel and CO2 prices will be added to the scenario parameters table according to table 6 in NEP 2035, Version 2021, 1. Entwurf (S.39).
Other marginal and capital costs are taken from technology-data repository | 1.0 | Insert cost assumptions from NEP and technology-data to scenario_parameters - Fuel and CO2 prices will be added to the scenario parameters table according to table 6 in NEP 2035, Version 2021, 1. Entwurf (S.39).
Other marginal and capital costs are taken from technology-data repository | non_priority | insert cost assumptions from nep and technology data to scenario parameters fuel and prices will be added to the scenario parameters table according to table in nep version entwurf s other marginal and capital costs are taken from technology data repository | 0 |
97,575 | 16,236,393,319 | IssuesEvent | 2021-05-07 01:37:41 | michaeldotson/mini-capstone | https://api.github.com/repos/michaeldotson/mini-capstone | opened | CVE-2020-7595 (High) detected in nokogiri-1.10.3.gem | security vulnerability | ## CVE-2020-7595 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nokogiri-1.10.3.gem</b></p></summary>
<p>Nokogiri (鋸) is an HTML, XML, SAX, and Reader parser. Among
Nokogiri's many features is the ability to search documents via XPath
or CSS3 selectors.</p>
<p>Library home page: <a href="https://rubygems.org/gems/nokogiri-1.10.3.gem">https://rubygems.org/gems/nokogiri-1.10.3.gem</a></p>
<p>Path to dependency file: /mini-capstone/Gemfile.lock</p>
<p>Path to vulnerable library: /var/lib/gems/2.3.0/cache/nokogiri-1.10.3.gem</p>
<p>
Dependency Hierarchy:
- sass-rails-5.0.7.gem (Root Library)
- sprockets-rails-3.2.1.gem
- actionpack-5.2.2.gem
- rails-html-sanitizer-1.0.4.gem
- loofah-2.2.3.gem
- :x: **nokogiri-1.10.3.gem** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
xmlStringLenDecodeEntities in parser.c in libxml2 2.9.10 has an infinite loop in a certain end-of-file situation.
<p>Publish Date: 2020-01-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7595>CVE-2020-7595</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://security.gentoo.org/glsa/202010-04">https://security.gentoo.org/glsa/202010-04</a></p>
<p>Fix Resolution: All libxml2 users should upgrade to the latest version # emerge --sync
# emerge --ask --oneshot --verbose >=dev-libs/libxml2-2.9.10 >= </p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-7595 (High) detected in nokogiri-1.10.3.gem - ## CVE-2020-7595 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nokogiri-1.10.3.gem</b></p></summary>
<p>Nokogiri (鋸) is an HTML, XML, SAX, and Reader parser. Among
Nokogiri's many features is the ability to search documents via XPath
or CSS3 selectors.</p>
<p>Library home page: <a href="https://rubygems.org/gems/nokogiri-1.10.3.gem">https://rubygems.org/gems/nokogiri-1.10.3.gem</a></p>
<p>Path to dependency file: /mini-capstone/Gemfile.lock</p>
<p>Path to vulnerable library: /var/lib/gems/2.3.0/cache/nokogiri-1.10.3.gem</p>
<p>
Dependency Hierarchy:
- sass-rails-5.0.7.gem (Root Library)
- sprockets-rails-3.2.1.gem
- actionpack-5.2.2.gem
- rails-html-sanitizer-1.0.4.gem
- loofah-2.2.3.gem
- :x: **nokogiri-1.10.3.gem** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
xmlStringLenDecodeEntities in parser.c in libxml2 2.9.10 has an infinite loop in a certain end-of-file situation.
<p>Publish Date: 2020-01-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7595>CVE-2020-7595</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://security.gentoo.org/glsa/202010-04">https://security.gentoo.org/glsa/202010-04</a></p>
<p>Fix Resolution: All libxml2 users should upgrade to the latest version # emerge --sync
# emerge --ask --oneshot --verbose >=dev-libs/libxml2-2.9.10 >= </p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in nokogiri gem cve high severity vulnerability vulnerable library nokogiri gem nokogiri 鋸 is an html xml sax and reader parser among nokogiri s many features is the ability to search documents via xpath or selectors library home page a href path to dependency file mini capstone gemfile lock path to vulnerable library var lib gems cache nokogiri gem dependency hierarchy sass rails gem root library sprockets rails gem actionpack gem rails html sanitizer gem loofah gem x nokogiri gem vulnerable library vulnerability details xmlstringlendecodeentities in parser c in has an infinite loop in a certain end of file situation publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href fix resolution all users should upgrade to the latest version emerge sync emerge ask oneshot verbose dev libs step up your open source security game with whitesource | 0 |
26,943 | 4,839,165,210 | IssuesEvent | 2016-11-09 08:21:20 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | closed | Timeline inside tabView not working | 6.0.8 defect | Reported by Pro User;
> We are using timeline component and it's working. Now, I need to use a <p:tabView, and when I moved the timeline's code inside this component, then the timeline's event stops working (the call to the method in the backend bean never happens) and when the oncomplete is processed, args are null and the process stops. I need the listener method to be executed before in order to load the specific data and updates the dialog with it.
Note: I want the tabView on demand loading.
Testing code:
xhtml:
```
<script type="text/javascript">
/* <![CDATA[ */
function handleLoadOfferData(xhr, status, args) {
alert('args = ' || args);
PF('statusDialog').hide();
if (args.notValid || args.validationFailed || !args.dataOk) {
return;
}
PF('detailsTLWdgt').show();
}
/* ]]> */
</script>
<h:form id="hForm">
<p:tabView id="timelineDetail" dynamic="true" cache="true">
<p:tab title="Tab1">
<h:panelGroup layout="block" styleClass="infoBox" rendered="#{bean.tlModel eq null}">
<span class="ui-icon ui-icon-alert" style="float: left;"></span>
<h:outputText value="Model without elements"/>
</h:panelGroup>
<p:timeline id="timelineNew" value="#{bean.tlModel}"
var="element" varGroup="timelineGroup"
height="400px" axisOnTop="true" eventStyle="dot" rendered="#{not empty bean.tlModel != null}">
<p:ajax event="select" listener="#{bean.onSelect}" process="@form"
onstart="PF('statusDialog').show();" update=":hForm:timelineDetail:detailsTLDlg"
oncomplete="handleLoadOfferData(xhr,status,args)"/>
<h:panelGrid columns="1">
<h:outputText value="#{element.description}"/>
</h:panelGrid>
</p:timeline>
<!-- Offer Details Modal -->
<p:dialog id="detailsTLDlg" widgetVar="detailsTLWdgt"
header="Dialog with details"
modal="true" resizable="false" draggable="false" width="650" height="400" rendered="#{bean.tlModel != null}">
Data loaded in onSelect method
<p:commandButton type="button" value="Close"
onclick="PF('detailsTLWdgt').hide()" icon="ui-icon-closethick"/>
</p:dialog>
</p:tab>
<p:tab title="Tab 2">
dfafdsa
</p:tab>
</p:tabView>
</h:form>
```
Backend bean:
```
@ManagedBean
@ViewScoped
public class Bean implements Serializable {
..........
private TimelineModel tlModel;
private MpanTimeLineVO detail;
..........
public void onSelect(TimelineSelectEvent e) throws Exception {
LOG.debug("onSelect");
TimelineEvent timelineEvent = e.getTimelineEvent();
if (timelineEvent.getData() == null) {
throw new RuntimeException("No data selected");
}
TimelineBasicVO data = (TimelineBasicVO) timelineEvent.getData();
detail = getDataDetail(data.getOfferId());
JSFHelper.getRequestContext().addCallbackParam("dataOk", Boolean.TRUE);
}
.......
}
``` | 1.0 | Timeline inside tabView not working - Reported by Pro User;
> We are using timeline component and it's working. Now, I need to use a <p:tabView, and when I moved the timeline's code inside this component, then the timeline's event stops working (the call to the method in the backend bean never happens) and when the oncomplete is processed, args are null and the process stops. I need the listener method to be executed before in order to load the specific data and updates the dialog with it.
Note: I want the tabView on demand loading.
Testing code:
xhtml:
```
<script type="text/javascript">
/* <![CDATA[ */
function handleLoadOfferData(xhr, status, args) {
alert('args = ' || args);
PF('statusDialog').hide();
if (args.notValid || args.validationFailed || !args.dataOk) {
return;
}
PF('detailsTLWdgt').show();
}
/* ]]> */
</script>
<h:form id="hForm">
<p:tabView id="timelineDetail" dynamic="true" cache="true">
<p:tab title="Tab1">
<h:panelGroup layout="block" styleClass="infoBox" rendered="#{bean.tlModel eq null}">
<span class="ui-icon ui-icon-alert" style="float: left;"></span>
<h:outputText value="Model without elements"/>
</h:panelGroup>
<p:timeline id="timelineNew" value="#{bean.tlModel}"
var="element" varGroup="timelineGroup"
height="400px" axisOnTop="true" eventStyle="dot" rendered="#{not empty bean.tlModel != null}">
<p:ajax event="select" listener="#{bean.onSelect}" process="@form"
onstart="PF('statusDialog').show();" update=":hForm:timelineDetail:detailsTLDlg"
oncomplete="handleLoadOfferData(xhr,status,args)"/>
<h:panelGrid columns="1">
<h:outputText value="#{element.description}"/>
</h:panelGrid>
</p:timeline>
<!-- Offer Details Modal -->
<p:dialog id="detailsTLDlg" widgetVar="detailsTLWdgt"
header="Dialog with details"
modal="true" resizable="false" draggable="false" width="650" height="400" rendered="#{bean.tlModel != null}">
Data loaded in onSelect method
<p:commandButton type="button" value="Close"
onclick="PF('detailsTLWdgt').hide()" icon="ui-icon-closethick"/>
</p:dialog>
</p:tab>
<p:tab title="Tab 2">
dfafdsa
</p:tab>
</p:tabView>
</h:form>
```
Backend bean:
```
@ManagedBean
@ViewScoped
public class Bean implements Serializable {
..........
private TimelineModel tlModel;
private MpanTimeLineVO detail;
..........
public void onSelect(TimelineSelectEvent e) throws Exception {
LOG.debug("onSelect");
TimelineEvent timelineEvent = e.getTimelineEvent();
if (timelineEvent.getData() == null) {
throw new RuntimeException("No data selected");
}
TimelineBasicVO data = (TimelineBasicVO) timelineEvent.getData();
detail = getDataDetail(data.getOfferId());
JSFHelper.getRequestContext().addCallbackParam("dataOk", Boolean.TRUE);
}
.......
}
``` | non_priority | timeline inside tabview not working reported by pro user we are using timeline component and it s working now i need to use a p tabview and when i moved the timeline s code inside this component then the timeline s event stops working the call to the method in the backend bean never happens and when the oncomplete is processed args are null and the process stops i need the listener method to be executed before in order to load the specific data and updates the dialog with it note i want the tabview on demand loading testing code xhtml cdata function handleloadofferdata xhr status args alert args args pf statusdialog hide if args notvalid args validationfailed args dataok return pf detailstlwdgt show p timeline id timelinenew value bean tlmodel var element vargroup timelinegroup height axisontop true eventstyle dot rendered not empty bean tlmodel null p ajax event select listener bean onselect process form onstart pf statusdialog show update hform timelinedetail detailstldlg oncomplete handleloadofferdata xhr status args p dialog id detailstldlg widgetvar detailstlwdgt header dialog with details modal true resizable false draggable false width height rendered bean tlmodel null data loaded in onselect method p commandbutton type button value close onclick pf detailstlwdgt hide icon ui icon closethick dfafdsa backend bean managedbean viewscoped public class bean implements serializable private timelinemodel tlmodel private mpantimelinevo detail public void onselect timelineselectevent e throws exception log debug onselect timelineevent timelineevent e gettimelineevent if timelineevent getdata null throw new runtimeexception no data selected timelinebasicvo data timelinebasicvo timelineevent getdata detail getdatadetail data getofferid jsfhelper getrequestcontext addcallbackparam dataok boolean true | 0 |
25,699 | 7,741,320,151 | IssuesEvent | 2018-05-29 04:58:46 | ShaikASK/Testing | https://api.github.com/repos/ShaikASK/Testing | closed | HR Admin & HR user : Forgot Password : User unable to access the forgot password page | Defect Forgot Password HR Admin Module HR User Module P1 Release#2 Build #1 | Steps to Replicate:
1. Launch the url :
2. Navigate to "signin" page
3. Click on "Forgot your Password" link
4. Enter valid email id in the email id field
5. Click on "Submit" button
Experienced Behavior : Observed that user is navigated to apache tomcat page
Expected Behavior : Ensure that user should be able to update the password using "forgot password page" | 1.0 | HR Admin & HR user : Forgot Password : User unable to access the forgot password page - Steps to Replicate:
1. Launch the url :
2. Navigate to "signin" page
3. Click on "Forgot your Password" link
4. Enter valid email id in the email id field
5. Click on "Submit" button
Experienced Behavior : Observed that user is navigated to apache tomcat page
Expected Behavior : Ensure that user should be able to update the password using "forgot password page" | non_priority | hr admin hr user forgot password user unable to access the forgot password page steps to replicate launch the url navigate to signin page click on forgot your password link enter valid email id in the email id field click on submit button experienced behavior observed that user is navigated to apache tomcat page expected behavior ensure that user should be able to update the password using forgot password page | 0 |
70,210 | 15,057,546,885 | IssuesEvent | 2021-02-03 21:51:54 | department-of-veterans-affairs/va.gov-team | https://api.github.com/repos/department-of-veterans-affairs/va.gov-team | opened | Security - iterate to improve the management of user access | needs-grooming operations security | ## Problem Statement
Maintaining credentials is difficult, and we lack the ability to grant fine-grained access to specific platform resources, without granting broad administrative access which is outside of the scope of a developers given role. This is caused by permission redundancy between groups and there being no unified strategy to granting permissions. This results in over-privileged users that have more access than needed to perform their role / job function. This also results in an expectation of needing elevated privileges that are outside of the scope of day-to-day work in order to carry on one's tasks. There is also inconsistency in the existing design of various security-related resources.
How might we add easily manage credentials and access to platform resources in a way that grants only the access which is required for a developers given job function?
How might we integrate this new credentials management workflow and associated tools into ops-owned applications and services?
How might we reduce administrative overhead by automating these processes and workflows for granting access to resources?
## Hypothesis or Bet
Our hypothesis is that by implementing a solution for better management of credentials, we can improve platform security, and the administration thereof.
## We will know we're done when... ("Definition of Done")
We have migrated permissions from credstash to Amazon's parameter store for all platform resources, ensuring that only required access is granted.
There is a clear, documented process in place for managing credentials in a way that is easy to do and requires minimal manual administration.
There is plan in place for how applications will use only SSM integration (and can be migrated completely off of cred-stash).
When we've considerably reduced policy sprawl (ie clear, straight-forward, easy-to-understand policy in place).
When we have a plan for how to automate the securing of non-compliant credentials.
Maybe: When we have a plan for how to notify users of non-compliant credentials.
## Known Blockers/Dependencies
*List any blockers or dependencies for this work to be completed*
## Projected Launch Date
By the end of Q1 2021, we expect to have new permissions management system in place.
## Launch Checklist
### Is this service / tool / feature...
### ... tested?
- [ ] Usability test (_TODO: link_) has been performed, to validate that new changes enable users to do what was intended and that these changes don't worsen quality elsewhere. If usability test isn't relevant for this change, document the reason for skipping it.
- [ ] ... and issues discovered in usability testing have been addressed.
* _Note on skipping: metrics that show the impact of before/after can be a substitute for usability testing._
- [ ] End-to-end [manual QA](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/quality-assurance/README.md) or [UAT](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/research/planning/what-is-uat.md) is complete, to validate there are no high-severity issues before launching
- [ ] _(if applicable)_ New functionality has thorough, automated tests running in CI/CD
### ... documented?
- [ ] New documentation is written pursuant to our [documentation style guide](https://github.com/department-of-veterans-affairs/va.gov-team/tree/master/platform/documentation/style-guide)
- [ ] Product is included in the [List of VSP Products](https://docs.google.com/spreadsheets/d/1Fn2lD419WE3sTZJtN2Ensrjqaz0jH3WvLaBtn812Wjo/edit#gid=0)
* _List the existing product that this initiative fits within, or add a new product to this list._
- [ ] Internal-facing: there's a [Product Outline](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/teams/vsp/product-management/product-outline-template.md) checked into [`products/platform/PRODUCT_NAME/`](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/products/platform/)
* _Note: the Product Directory Name should match 1:1 with the List of VSP Products_
- [ ] External-facing: a [VFS-facing README](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/teams/vsp/product-management/product-readme-template.md) exists for this product/feature tool
- [ ] ... and should be located at `platform/PRODUCT_NAME/README.md`
- [ ] External-facing: a [User Guide](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/teams/vsp/product-management/writing-user-guides.md) exists for this product/feature/tool, and is updated for changes from this initiative
- [ ] ... and should be linked from the VFS-facing README for your product
- [ ] ... and should be located within `platform/PRODUCT_NAME/`, unless you already have another location for it
- [ ] _(if applicable)_... and post to [#vsp-content-ia](https://dsva.slack.com/channels/vsp-content-ia) about whether this should be added to the [Documentation homepage](https://department-of-veterans-affairs.github.io/va.gov-team/)
- [ ] _(if applicable)_ Post to [#vsp-service-design](https://dsva.slack.com/channels/vsp-service-design) for external communication about this change (e.g. VSP Newsletter, customer-facing meetings)
### ... measurable
- [ ] _(if applicable)_ This change has clearly-defined success metrics, with instrumentation of those analytics where possible, or a reason documented for skipping it.
* For help, see: [Analytics team](https://github.com/department-of-veterans-affairs/va.gov-team/tree/master/platform/analytics)
- [ ] This change has an accompanying [VSP Initiative Release Plan](https://github.com/department-of-veterans-affairs/va.gov-team/issues/new/choose).
## Required Artifacts
### Documentation
* **`PRODUCT_NAME`**: _directory name used for your product documentation_
* **Product Outline**: _link to Product Outline_
* **README**: _link to VFS-facing README for your product_
* **User Guide**: _link to User Guide_
### Testing
* **Usability test**: _link to GitHub issue, or provide reason for skipping_
* **Manual QA**: _link to GitHub issue or documented results_
* **Automated tests**: _link to tests, or "N/A"_
### Measurement
* **Success metrics**:
- developer onboarding time ~24 hours (match current request SLA)
- no empty groups in IAM
- only operations team members have administrative privileges (unless service account and/or unless justifiable, documented reason for enhanced privileges for a given user)
- full accounting / audit of existing policies
* **Release plan**: _link to Release Plan ticket_
## TODOs
- [x] Convert this issue to an epic
- [x] Add your team's label to this epic
| True | Security - iterate to improve the management of user access - ## Problem Statement
Maintaining credentials is difficult, and we lack the ability to grant fine-grained access to specific platform resources, without granting broad administrative access which is outside of the scope of a developers given role. This is caused by permission redundancy between groups and there being no unified strategy to granting permissions. This results in over-privileged users that have more access than needed to perform their role / job function. This also results in an expectation of needing elevated privileges that are outside of the scope of day-to-day work in order to carry on one's tasks. There is also inconsistency in the existing design of various security-related resources.
How might we add easily manage credentials and access to platform resources in a way that grants only the access which is required for a developers given job function?
How might we integrate this new credentials management workflow and associated tools into ops-owned applications and services?
How might we reduce administrative overhead by automating these processes and workflows for granting access to resources?
## Hypothesis or Bet
Our hypothesis is that by implementing a solution for better management of credentials, we can improve platform security, and the administration thereof.
## We will know we're done when... ("Definition of Done")
We have migrated permissions from credstash to Amazon's parameter store for all platform resources, ensuring that only required access is granted.
There is a clear, documented process in place for managing credentials in a way that is easy to do and requires minimal manual administration.
There is plan in place for how applications will use only SSM integration (and can be migrated completely off of cred-stash).
When we've considerably reduced policy sprawl (ie clear, straight-forward, easy-to-understand policy in place).
When we have a plan for how to automate the securing of non-compliant credentials.
Maybe: When we have a plan for how to notify users of non-compliant credentials.
## Known Blockers/Dependencies
*List any blockers or dependencies for this work to be completed*
## Projected Launch Date
By the end of Q1 2021, we expect to have new permissions management system in place.
## Launch Checklist
### Is this service / tool / feature...
### ... tested?
- [ ] Usability test (_TODO: link_) has been performed, to validate that new changes enable users to do what was intended and that these changes don't worsen quality elsewhere. If usability test isn't relevant for this change, document the reason for skipping it.
- [ ] ... and issues discovered in usability testing have been addressed.
* _Note on skipping: metrics that show the impact of before/after can be a substitute for usability testing._
- [ ] End-to-end [manual QA](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/quality-assurance/README.md) or [UAT](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/research/planning/what-is-uat.md) is complete, to validate there are no high-severity issues before launching
- [ ] _(if applicable)_ New functionality has thorough, automated tests running in CI/CD
### ... documented?
- [ ] New documentation is written pursuant to our [documentation style guide](https://github.com/department-of-veterans-affairs/va.gov-team/tree/master/platform/documentation/style-guide)
- [ ] Product is included in the [List of VSP Products](https://docs.google.com/spreadsheets/d/1Fn2lD419WE3sTZJtN2Ensrjqaz0jH3WvLaBtn812Wjo/edit#gid=0)
* _List the existing product that this initiative fits within, or add a new product to this list._
- [ ] Internal-facing: there's a [Product Outline](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/teams/vsp/product-management/product-outline-template.md) checked into [`products/platform/PRODUCT_NAME/`](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/products/platform/)
* _Note: the Product Directory Name should match 1:1 with the List of VSP Products_
- [ ] External-facing: a [VFS-facing README](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/teams/vsp/product-management/product-readme-template.md) exists for this product/feature tool
- [ ] ... and should be located at `platform/PRODUCT_NAME/README.md`
- [ ] External-facing: a [User Guide](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/teams/vsp/product-management/writing-user-guides.md) exists for this product/feature/tool, and is updated for changes from this initiative
- [ ] ... and should be linked from the VFS-facing README for your product
- [ ] ... and should be located within `platform/PRODUCT_NAME/`, unless you already have another location for it
- [ ] _(if applicable)_... and post to [#vsp-content-ia](https://dsva.slack.com/channels/vsp-content-ia) about whether this should be added to the [Documentation homepage](https://department-of-veterans-affairs.github.io/va.gov-team/)
- [ ] _(if applicable)_ Post to [#vsp-service-design](https://dsva.slack.com/channels/vsp-service-design) for external communication about this change (e.g. VSP Newsletter, customer-facing meetings)
### ... measurable
- [ ] _(if applicable)_ This change has clearly-defined success metrics, with instrumentation of those analytics where possible, or a reason documented for skipping it.
* For help, see: [Analytics team](https://github.com/department-of-veterans-affairs/va.gov-team/tree/master/platform/analytics)
- [ ] This change has an accompanying [VSP Initiative Release Plan](https://github.com/department-of-veterans-affairs/va.gov-team/issues/new/choose).
## Required Artifacts
### Documentation
* **`PRODUCT_NAME`**: _directory name used for your product documentation_
* **Product Outline**: _link to Product Outline_
* **README**: _link to VFS-facing README for your product_
* **User Guide**: _link to User Guide_
### Testing
* **Usability test**: _link to GitHub issue, or provide reason for skipping_
* **Manual QA**: _link to GitHub issue or documented results_
* **Automated tests**: _link to tests, or "N/A"_
### Measurement
* **Success metrics**:
- developer onboarding time ~24 hours (match current request SLA)
- no empty groups in IAM
- only operations team members have administrative privileges (unless service account and/or unless justifiable, documented reason for enhanced privileges for a given user)
- full accounting / audit of existing policies
* **Release plan**: _link to Release Plan ticket_
## TODOs
- [x] Convert this issue to an epic
- [x] Add your team's label to this epic
| non_priority | security iterate to improve the management of user access problem statement maintaining credentials is difficult and we lack the ability to grant fine grained access to specific platform resources without granting broad administrative access which is outside of the scope of a developers given role this is caused by permission redundancy between groups and there being no unified strategy to granting permissions this results in over privileged users that have more access than needed to perform their role job function this also results in an expectation of needing elevated privileges that are outside of the scope of day to day work in order to carry on one s tasks there is also inconsistency in the existing design of various security related resources how might we add easily manage credentials and access to platform resources in a way that grants only the access which is required for a developers given job function how might we integrate this new credentials management workflow and associated tools into ops owned applications and services how might we reduce administrative overhead by automating these processes and workflows for granting access to resources hypothesis or bet our hypothesis is that by implementing a solution for better management of credentials we can improve platform security and the administration thereof we will know we re done when definition of done we have migrated permissions from credstash to amazon s parameter store for all platform resources ensuring that only required access is granted there is a clear documented process in place for managing credentials in a way that is easy to do and requires minimal manual administration there is plan in place for how applications will use only ssm integration and can be migrated completely off of cred stash when we ve considerably reduced policy sprawl ie clear straight forward easy to understand policy in place when we have a plan for how to automate the securing of non compliant credentials maybe when we have a plan for how to notify users of non compliant credentials known blockers dependencies list any blockers or dependencies for this work to be completed projected launch date by the end of we expect to have new permissions management system in place launch checklist is this service tool feature tested usability test todo link has been performed to validate that new changes enable users to do what was intended and that these changes don t worsen quality elsewhere if usability test isn t relevant for this change document the reason for skipping it and issues discovered in usability testing have been addressed note on skipping metrics that show the impact of before after can be a substitute for usability testing end to end or is complete to validate there are no high severity issues before launching if applicable new functionality has thorough automated tests running in ci cd documented new documentation is written pursuant to our product is included in the list the existing product that this initiative fits within or add a new product to this list internal facing there s a checked into note the product directory name should match with the list of vsp products external facing a exists for this product feature tool and should be located at platform product name readme md external facing a exists for this product feature tool and is updated for changes from this initiative and should be linked from the vfs facing readme for your product and should be located within platform product name unless you already have another location for it if applicable and post to about whether this should be added to the if applicable post to for external communication about this change e g vsp newsletter customer facing meetings measurable if applicable this change has clearly defined success metrics with instrumentation of those analytics where possible or a reason documented for skipping it for help see this change has an accompanying required artifacts documentation product name directory name used for your product documentation product outline link to product outline readme link to vfs facing readme for your product user guide link to user guide testing usability test link to github issue or provide reason for skipping manual qa link to github issue or documented results automated tests link to tests or n a measurement success metrics developer onboarding time hours match current request sla no empty groups in iam only operations team members have administrative privileges unless service account and or unless justifiable documented reason for enhanced privileges for a given user full accounting audit of existing policies release plan link to release plan ticket todos convert this issue to an epic add your team s label to this epic | 0 |
31,262 | 6,477,744,173 | IssuesEvent | 2017-08-18 05:06:02 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | ICache.getAndRemove() doesn't invalidate the Near Cache | Team: Core Type: Defect | `ICache.getAndRemove()` doesn't invalidate the Near Cache when called from the Near Cache side. I noticed this while adding more variants of the `remove()` and `delete()` methods to the Near Cache tests. The failures are all on the `onNearCacheAdapter` method and in the `ClientNearCacheBasicTest`. That's a good indication that a call to the `NearCachedClientCacheProxy` is missing, which invalidates the Near Cache locally. | 1.0 | ICache.getAndRemove() doesn't invalidate the Near Cache - `ICache.getAndRemove()` doesn't invalidate the Near Cache when called from the Near Cache side. I noticed this while adding more variants of the `remove()` and `delete()` methods to the Near Cache tests. The failures are all on the `onNearCacheAdapter` method and in the `ClientNearCacheBasicTest`. That's a good indication that a call to the `NearCachedClientCacheProxy` is missing, which invalidates the Near Cache locally. | non_priority | icache getandremove doesn t invalidate the near cache icache getandremove doesn t invalidate the near cache when called from the near cache side i noticed this while adding more variants of the remove and delete methods to the near cache tests the failures are all on the onnearcacheadapter method and in the clientnearcachebasictest that s a good indication that a call to the nearcachedclientcacheproxy is missing which invalidates the near cache locally | 0 |
86,603 | 17,027,780,169 | IssuesEvent | 2021-07-03 22:59:46 | TeamAmaze/AmazeFileManager | https://api.github.com/repos/TeamAmaze/AmazeFileManager | opened | Convert all AsyncTasks to Callables called from ReactiveX | Area-CodeHealth Issue-Discussion Issue-Task (low) | `public class ReadFileTask extends AsyncTask<Void, Void, ReturnedValueOnReadFile> {`
to
`public class ReadFileTask implements Callable<ReturnedValueOnReadFile> {`
and used as
```java
final ReadFileTask task = new ReadFileTask(getContentResolver(), viewModel.getFile(), getExternalCacheDir(), isRootExplorer());
final Consumer<ReturnedValueOnReadFile> onAsyncTaskFinished = (data) -> {
//operations
}
final Consumer<? super Throwable> onError = error -> {
error.printStackTrace();
}
Flowable.fromCallable(task)
.subscribeOn(Schedulers.io())
.observeOn(Schedulers.single())
.subscribe(onAsyncTaskFinished, onError);
``` | 1.0 | Convert all AsyncTasks to Callables called from ReactiveX - `public class ReadFileTask extends AsyncTask<Void, Void, ReturnedValueOnReadFile> {`
to
`public class ReadFileTask implements Callable<ReturnedValueOnReadFile> {`
and used as
```java
final ReadFileTask task = new ReadFileTask(getContentResolver(), viewModel.getFile(), getExternalCacheDir(), isRootExplorer());
final Consumer<ReturnedValueOnReadFile> onAsyncTaskFinished = (data) -> {
//operations
}
final Consumer<? super Throwable> onError = error -> {
error.printStackTrace();
}
Flowable.fromCallable(task)
.subscribeOn(Schedulers.io())
.observeOn(Schedulers.single())
.subscribe(onAsyncTaskFinished, onError);
``` | non_priority | convert all asynctasks to callables called from reactivex public class readfiletask extends asynctask to public class readfiletask implements callable and used as java final readfiletask task new readfiletask getcontentresolver viewmodel getfile getexternalcachedir isrootexplorer final consumer onasynctaskfinished data operations final consumer onerror error error printstacktrace flowable fromcallable task subscribeon schedulers io observeon schedulers single subscribe onasynctaskfinished onerror | 0 |
181,268 | 21,658,345,552 | IssuesEvent | 2022-05-06 16:17:42 | snowdensb/microsoft-teams-apps-company-communicator | https://api.github.com/repos/snowdensb/microsoft-teams-apps-company-communicator | opened | WS-2020-0438 (Medium) detected in i18next-19.8.2.tgz | security vulnerability | ## WS-2020-0438 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>i18next-19.8.2.tgz</b></p></summary>
<p>i18next internationalization framework</p>
<p>Library home page: <a href="https://registry.npmjs.org/i18next/-/i18next-19.8.2.tgz">https://registry.npmjs.org/i18next/-/i18next-19.8.2.tgz</a></p>
<p>Path to dependency file: /Source/CompanyCommunicator/ClientApp/package.json</p>
<p>Path to vulnerable library: /Source/CompanyCommunicator/ClientApp/node_modules/i18next/package.json</p>
<p>
Dependency Hierarchy:
- :x: **i18next-19.8.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/snowdensb/microsoft-teams-apps-company-communicator/commit/32bf8559b65e6695a5e727d0fdf93e7ae135d32a">32bf8559b65e6695a5e727d0fdf93e7ae135d32a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In i18next in versions v19.6.0 to v19.8.2 is vulnerable to prototype pollution, it allows to modify the prototype of a base object, which may result in DoS, XSS, RCE, etc.
<p>Publish Date: 2020-02-12
<p>URL: <a href=https://github.com/i18next/i18next/commit/ed6169fa21b67d8e73b20d82579aff6f9b808a69>WS-2020-0438</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://hackerone.com/reports/968355">https://hackerone.com/reports/968355</a></p>
<p>Release Date: 2020-02-12</p>
<p>Fix Resolution: 19.8.3</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"i18next","packageVersion":"19.8.2","packageFilePaths":["/Source/CompanyCommunicator/ClientApp/package.json"],"isTransitiveDependency":false,"dependencyTree":"i18next:19.8.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"19.8.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2020-0438","vulnerabilityDetails":"In i18next in versions v19.6.0 to v19.8.2 is vulnerable to prototype pollution, it allows to modify the prototype of a base object, which may result in DoS, XSS, RCE, etc.","vulnerabilityUrl":"https://github.com/i18next/i18next/commit/ed6169fa21b67d8e73b20d82579aff6f9b808a69","cvss3Severity":"medium","cvss3Score":"4.8","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | True | WS-2020-0438 (Medium) detected in i18next-19.8.2.tgz - ## WS-2020-0438 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>i18next-19.8.2.tgz</b></p></summary>
<p>i18next internationalization framework</p>
<p>Library home page: <a href="https://registry.npmjs.org/i18next/-/i18next-19.8.2.tgz">https://registry.npmjs.org/i18next/-/i18next-19.8.2.tgz</a></p>
<p>Path to dependency file: /Source/CompanyCommunicator/ClientApp/package.json</p>
<p>Path to vulnerable library: /Source/CompanyCommunicator/ClientApp/node_modules/i18next/package.json</p>
<p>
Dependency Hierarchy:
- :x: **i18next-19.8.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/snowdensb/microsoft-teams-apps-company-communicator/commit/32bf8559b65e6695a5e727d0fdf93e7ae135d32a">32bf8559b65e6695a5e727d0fdf93e7ae135d32a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In i18next in versions v19.6.0 to v19.8.2 is vulnerable to prototype pollution, it allows to modify the prototype of a base object, which may result in DoS, XSS, RCE, etc.
<p>Publish Date: 2020-02-12
<p>URL: <a href=https://github.com/i18next/i18next/commit/ed6169fa21b67d8e73b20d82579aff6f9b808a69>WS-2020-0438</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://hackerone.com/reports/968355">https://hackerone.com/reports/968355</a></p>
<p>Release Date: 2020-02-12</p>
<p>Fix Resolution: 19.8.3</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"i18next","packageVersion":"19.8.2","packageFilePaths":["/Source/CompanyCommunicator/ClientApp/package.json"],"isTransitiveDependency":false,"dependencyTree":"i18next:19.8.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"19.8.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2020-0438","vulnerabilityDetails":"In i18next in versions v19.6.0 to v19.8.2 is vulnerable to prototype pollution, it allows to modify the prototype of a base object, which may result in DoS, XSS, RCE, etc.","vulnerabilityUrl":"https://github.com/i18next/i18next/commit/ed6169fa21b67d8e73b20d82579aff6f9b808a69","cvss3Severity":"medium","cvss3Score":"4.8","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | non_priority | ws medium detected in tgz ws medium severity vulnerability vulnerable library tgz internationalization framework library home page a href path to dependency file source companycommunicator clientapp package json path to vulnerable library source companycommunicator clientapp node modules package json dependency hierarchy x tgz vulnerable library found in head commit a href found in base branch master vulnerability details in in versions to is vulnerable to prototype pollution it allows to modify the prototype of a base object which may result in dos xss rce etc publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier ws vulnerabilitydetails in in versions to is vulnerable to prototype pollution it allows to modify the prototype of a base object which may result in dos xss rce etc vulnerabilityurl | 0 |
137,214 | 18,752,659,704 | IssuesEvent | 2021-11-05 05:45:46 | madhans23/linux-4.15 | https://api.github.com/repos/madhans23/linux-4.15 | opened | CVE-2018-20669 (High) detected in multiple libraries | security vulnerability | ## CVE-2018-20669 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxv4.15</b>, <b>linuxv4.15</b>, <b>linuxv4.15</b>, <b>linuxv4.15</b>, <b>linuxv4.15</b>, <b>linuxv4.15</b>, <b>linux-yocto-devv5.4</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue where a provided address with access_ok() is not checked was discovered in i915_gem_execbuffer2_ioctl in drivers/gpu/drm/i915/i915_gem_execbuffer.c in the Linux kernel through 4.19.13. A local attacker can craft a malicious IOCTL function call to overwrite arbitrary kernel memory, resulting in a Denial of Service or privilege escalation.
<p>Publish Date: 2019-03-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20669>CVE-2018-20669</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20669">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20669</a></p>
<p>Release Date: 2019-03-21</p>
<p>Fix Resolution: v5.0-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-20669 (High) detected in multiple libraries - ## CVE-2018-20669 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxv4.15</b>, <b>linuxv4.15</b>, <b>linuxv4.15</b>, <b>linuxv4.15</b>, <b>linuxv4.15</b>, <b>linuxv4.15</b>, <b>linux-yocto-devv5.4</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue where a provided address with access_ok() is not checked was discovered in i915_gem_execbuffer2_ioctl in drivers/gpu/drm/i915/i915_gem_execbuffer.c in the Linux kernel through 4.19.13. A local attacker can craft a malicious IOCTL function call to overwrite arbitrary kernel memory, resulting in a Denial of Service or privilege escalation.
<p>Publish Date: 2019-03-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20669>CVE-2018-20669</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20669">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20669</a></p>
<p>Release Date: 2019-03-21</p>
<p>Fix Resolution: v5.0-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries linux yocto vulnerability details an issue where a provided address with access ok is not checked was discovered in gem ioctl in drivers gpu drm gem execbuffer c in the linux kernel through a local attacker can craft a malicious ioctl function call to overwrite arbitrary kernel memory resulting in a denial of service or privilege escalation publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
255,964 | 27,539,554,629 | IssuesEvent | 2023-03-07 07:27:45 | istio/istio | https://api.github.com/repos/istio/istio | closed | Github wiki docs: Broken link regarding mTLS troubleshoot | kind/docs area/security | ### Bug Description
Hey guys!
https://github.com/istio/istio/wiki/Mutual-TLS-TroubleShooting
This link have multiple broken links that would help us to fix our issues regarding mTLS / HTTPS / Certificates.
### Version
```prose
istioctl version
client version: 1.17.1
control plane version: 1.17.1
data plane version: 1.17.1 (9 proxies)
kubectl version --short
Flag --short has been deprecated, and will be removed in the future. The --short output will become the default.
Client Version: v1.26.1
Kustomize Version: v4.5.7
Server Version: v1.25.3
helm version --short
```
### Additional Information
not relevant | True | Github wiki docs: Broken link regarding mTLS troubleshoot - ### Bug Description
Hey guys!
https://github.com/istio/istio/wiki/Mutual-TLS-TroubleShooting
This link have multiple broken links that would help us to fix our issues regarding mTLS / HTTPS / Certificates.
### Version
```prose
istioctl version
client version: 1.17.1
control plane version: 1.17.1
data plane version: 1.17.1 (9 proxies)
kubectl version --short
Flag --short has been deprecated, and will be removed in the future. The --short output will become the default.
Client Version: v1.26.1
Kustomize Version: v4.5.7
Server Version: v1.25.3
helm version --short
```
### Additional Information
not relevant | non_priority | github wiki docs broken link regarding mtls troubleshoot bug description hey guys this link have multiple broken links that would help us to fix our issues regarding mtls https certificates version prose istioctl version client version control plane version data plane version proxies kubectl version short flag short has been deprecated and will be removed in the future the short output will become the default client version kustomize version server version helm version short additional information not relevant | 0 |
276,282 | 23,982,831,069 | IssuesEvent | 2022-09-13 16:21:51 | pitt-crc/bank | https://api.github.com/repos/pitt-crc/bank | opened | Make sure SLURM account locking does not bar access to investment hardware | testing | We should check that our locking mechanism (`GrpTresRunMins = 0`) does not interfere with users ability to submit to hardware they've invested in.
A possible proxy test for this is to see if a locked account can still submit to `scavenger`, as the billing weights _should_ be set similarly. | 1.0 | Make sure SLURM account locking does not bar access to investment hardware - We should check that our locking mechanism (`GrpTresRunMins = 0`) does not interfere with users ability to submit to hardware they've invested in.
A possible proxy test for this is to see if a locked account can still submit to `scavenger`, as the billing weights _should_ be set similarly. | non_priority | make sure slurm account locking does not bar access to investment hardware we should check that our locking mechanism grptresrunmins does not interfere with users ability to submit to hardware they ve invested in a possible proxy test for this is to see if a locked account can still submit to scavenger as the billing weights should be set similarly | 0 |
44,175 | 5,765,120,524 | IssuesEvent | 2017-04-27 00:57:24 | red-eight/Tactics | https://api.github.com/repos/red-eight/Tactics | closed | Unit::PerformNextAction() vs ActionController::ProcessNextAction() | Design | A unified naming convention should be used. | 1.0 | Unit::PerformNextAction() vs ActionController::ProcessNextAction() - A unified naming convention should be used. | non_priority | unit performnextaction vs actioncontroller processnextaction a unified naming convention should be used | 0 |
8,879 | 4,347,265,059 | IssuesEvent | 2016-07-29 18:53:33 | mapbox/mapbox-gl-native | https://api.github.com/repos/mapbox/mapbox-gl-native | closed | CocoaPods: Unable to satisfy the following requirements... | build iOS | Hi,
I try to install Mapbox iOS SDK 3.2 through Cocoapods (0.39.0) but i have this error :
[!] Unable to satisfy the following requirements:
- `Mapbox-iOS-SDK (~> 3.2.0)` required by `Podfile`
Have you any idea of the problem?
Do i need to install the pre-version 1.0.0.beta.6 of Cocoapods to use your last version of Mapbox ? | 1.0 | CocoaPods: Unable to satisfy the following requirements... - Hi,
I try to install Mapbox iOS SDK 3.2 through Cocoapods (0.39.0) but i have this error :
[!] Unable to satisfy the following requirements:
- `Mapbox-iOS-SDK (~> 3.2.0)` required by `Podfile`
Have you any idea of the problem?
Do i need to install the pre-version 1.0.0.beta.6 of Cocoapods to use your last version of Mapbox ? | non_priority | cocoapods unable to satisfy the following requirements hi i try to install mapbox ios sdk through cocoapods but i have this error unable to satisfy the following requirements mapbox ios sdk required by podfile have you any idea of the problem do i need to install the pre version beta of cocoapods to use your last version of mapbox | 0 |
200,636 | 15,116,632,971 | IssuesEvent | 2021-02-09 07:06:47 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | kv/kvserver: TestReplicateRogueRemovedNode failed | C-test-failure O-robot branch-master | [(kv/kvserver).TestReplicateRogueRemovedNode failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2657499&tab=buildLog) on [master@7853fd32de8b6dea869f2a2a92dcd7506f4a8998](https://github.com/cockroachdb/cockroach/commits/7853fd32de8b6dea869f2a2a92dcd7506f4a8998):
Fatal error:
```
panic: unexpected error: [NotLeaseHolderError] refusing to acquire lease on follower; r36: replica (n3,s3):3 not lease holder; current lease is repl=(n1,s1):1 seq=0 start=0,0 exp=<nil>
```
Stack:
```
goroutine 341335 [running]:
github.com/cockroachdb/cockroach/pkg/kv/kvserver_test.TestReplicateRogueRemovedNode.func2(0xc0027b5470, 0xd, 0x10, 0xc0073edd60, 0xc0073edd70, 0xc00468c000, 0xc00e16ad80, 0x300000003, 0x3, 0x0)
/go/src/github.com/cockroachdb/cockroach/pkg/kv/kvserver/client_raft_test.go:3338 +0x3a5
created by github.com/cockroachdb/cockroach/pkg/kv/kvserver_test.TestReplicateRogueRemovedNode
/go/src/github.com/cockroachdb/cockroach/pkg/kv/kvserver/client_raft_test.go:3322 +0xea5
```
<details><summary>Log preceding fatal error</summary><p>
```
=== RUN TestReplicateRogueRemovedNode
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestReplicateRogueRemovedNode993083353
test_log_scope.go:74: use -show-logs to present logs inline
```
</p></details>
<details><summary>More</summary><p>
Parameters:
- TAGS=
- GOFLAGS=-parallel=4
```
make stressrace TESTS=TestReplicateRogueRemovedNode PKG=./pkg/kv/kvserver TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestReplicateRogueRemovedNode.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
| 1.0 | kv/kvserver: TestReplicateRogueRemovedNode failed - [(kv/kvserver).TestReplicateRogueRemovedNode failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2657499&tab=buildLog) on [master@7853fd32de8b6dea869f2a2a92dcd7506f4a8998](https://github.com/cockroachdb/cockroach/commits/7853fd32de8b6dea869f2a2a92dcd7506f4a8998):
Fatal error:
```
panic: unexpected error: [NotLeaseHolderError] refusing to acquire lease on follower; r36: replica (n3,s3):3 not lease holder; current lease is repl=(n1,s1):1 seq=0 start=0,0 exp=<nil>
```
Stack:
```
goroutine 341335 [running]:
github.com/cockroachdb/cockroach/pkg/kv/kvserver_test.TestReplicateRogueRemovedNode.func2(0xc0027b5470, 0xd, 0x10, 0xc0073edd60, 0xc0073edd70, 0xc00468c000, 0xc00e16ad80, 0x300000003, 0x3, 0x0)
/go/src/github.com/cockroachdb/cockroach/pkg/kv/kvserver/client_raft_test.go:3338 +0x3a5
created by github.com/cockroachdb/cockroach/pkg/kv/kvserver_test.TestReplicateRogueRemovedNode
/go/src/github.com/cockroachdb/cockroach/pkg/kv/kvserver/client_raft_test.go:3322 +0xea5
```
<details><summary>Log preceding fatal error</summary><p>
```
=== RUN TestReplicateRogueRemovedNode
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestReplicateRogueRemovedNode993083353
test_log_scope.go:74: use -show-logs to present logs inline
```
</p></details>
<details><summary>More</summary><p>
Parameters:
- TAGS=
- GOFLAGS=-parallel=4
```
make stressrace TESTS=TestReplicateRogueRemovedNode PKG=./pkg/kv/kvserver TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestReplicateRogueRemovedNode.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
| non_priority | kv kvserver testreplicaterogueremovednode failed on fatal error panic unexpected error refusing to acquire lease on follower replica not lease holder current lease is repl seq start exp stack goroutine github com cockroachdb cockroach pkg kv kvserver test testreplicaterogueremovednode go src github com cockroachdb cockroach pkg kv kvserver client raft test go created by github com cockroachdb cockroach pkg kv kvserver test testreplicaterogueremovednode go src github com cockroachdb cockroach pkg kv kvserver client raft test go log preceding fatal error run testreplicaterogueremovednode test log scope go test logs captured to go src github com cockroachdb cockroach artifacts test log scope go use show logs to present logs inline more parameters tags goflags parallel make stressrace tests testreplicaterogueremovednode pkg pkg kv kvserver testtimeout stressflags timeout powered by | 0 |
211,962 | 16,384,572,381 | IssuesEvent | 2021-05-17 08:47:20 | wazuh/wazuh-qa | https://api.github.com/repos/wazuh/wazuh-qa | opened | Migrate system tests fixes from master to 4.2 | bug framework/api system-tests | ## Description
Hi team!
There were many changes to the cluster in the master branch of Wazuh. These changes implied the update of some tests in QA, which were merged in QA master:
- https://github.com/wazuh/wazuh-qa/pull/1188
- https://github.com/wazuh/wazuh-qa/pull/1232
- https://github.com/wazuh/wazuh-qa/pull/1237
However, we migrated most of the cluster improvements that were planned for 4.3.0 to 4.2.0 in this PR: https://github.com/wazuh/wazuh/pull/8674
Therefore, all the fixes that were merged in QA master branch need to be moved to 4.2 too. Otherwise, the tests will fail.
Regards,
Selu. | 1.0 | Migrate system tests fixes from master to 4.2 - ## Description
Hi team!
There were many changes to the cluster in the master branch of Wazuh. These changes implied the update of some tests in QA, which were merged in QA master:
- https://github.com/wazuh/wazuh-qa/pull/1188
- https://github.com/wazuh/wazuh-qa/pull/1232
- https://github.com/wazuh/wazuh-qa/pull/1237
However, we migrated most of the cluster improvements that were planned for 4.3.0 to 4.2.0 in this PR: https://github.com/wazuh/wazuh/pull/8674
Therefore, all the fixes that were merged in QA master branch need to be moved to 4.2 too. Otherwise, the tests will fail.
Regards,
Selu. | non_priority | migrate system tests fixes from master to description hi team there were many changes to the cluster in the master branch of wazuh these changes implied the update of some tests in qa which were merged in qa master however we migrated most of the cluster improvements that were planned for to in this pr therefore all the fixes that were merged in qa master branch need to be moved to too otherwise the tests will fail regards selu | 0 |
1,370 | 2,595,499,384 | IssuesEvent | 2015-02-20 14:44:11 | docker/fig | https://api.github.com/repos/docker/fig | opened | Installation instructions for Compose should point and Docker installation instructions | documentation | It seems odd that we're repeating the installations instructions for Docker in the same documentation: https://github.com/docker/fig/blob/master/docs/install.md | 1.0 | Installation instructions for Compose should point and Docker installation instructions - It seems odd that we're repeating the installations instructions for Docker in the same documentation: https://github.com/docker/fig/blob/master/docs/install.md | non_priority | installation instructions for compose should point and docker installation instructions it seems odd that we re repeating the installations instructions for docker in the same documentation | 0 |
5,580 | 27,958,345,405 | IssuesEvent | 2023-03-24 13:59:38 | software-mansion/react-native-reanimated | https://api.github.com/repos/software-mansion/react-native-reanimated | closed | ☂️ Deadlock/ANR in performOperations | Platform: Android Platform: iOS Bug Maintainer issue | ### Description
This is an umbrella issue for ANRs/deadlocks on Android/iOS in NodesManager.performOperations.
The bug was introduced in #1215.
Android:
- #2251
- #3062
iOS:
- #3180
- #3862
- #3946
PRs trying to solve this issue:
- #3082
- #3194
### Repro
We don't have a repro yet but it needs to use modal or datetime picker as well as animate layout props using Reanimated.
### Reanimated version
\>= 2.0.0, >= 3.0.0
### Platforms
Android, iOS | True | ☂️ Deadlock/ANR in performOperations - ### Description
This is an umbrella issue for ANRs/deadlocks on Android/iOS in NodesManager.performOperations.
The bug was introduced in #1215.
Android:
- #2251
- #3062
iOS:
- #3180
- #3862
- #3946
PRs trying to solve this issue:
- #3082
- #3194
### Repro
We don't have a repro yet but it needs to use modal or datetime picker as well as animate layout props using Reanimated.
### Reanimated version
\>= 2.0.0, >= 3.0.0
### Platforms
Android, iOS | non_priority | ☂️ deadlock anr in performoperations description this is an umbrella issue for anrs deadlocks on android ios in nodesmanager performoperations the bug was introduced in android ios prs trying to solve this issue repro we don t have a repro yet but it needs to use modal or datetime picker as well as animate layout props using reanimated reanimated version platforms android ios | 0 |
12,049 | 9,558,136,101 | IssuesEvent | 2019-05-03 13:32:17 | elastic/kibana | https://api.github.com/repos/elastic/kibana | opened | [Logs UI] Add documentation for column customization | :Logs UI :infrastructure [zube]: Ready docs v7.2.0 v8.0.0 | The new column customization UI introduced in #34916 should be explained in the documentation. Aspects that are potentially worth mentioning:
* Influences the main log view
* Configured via the source configuration flyout, 2nd tab
* Three column types:
* built-in timestamp
* built-in message
* custom field from underlying document | 1.0 | [Logs UI] Add documentation for column customization - The new column customization UI introduced in #34916 should be explained in the documentation. Aspects that are potentially worth mentioning:
* Influences the main log view
* Configured via the source configuration flyout, 2nd tab
* Three column types:
* built-in timestamp
* built-in message
* custom field from underlying document | non_priority | add documentation for column customization the new column customization ui introduced in should be explained in the documentation aspects that are potentially worth mentioning influences the main log view configured via the source configuration flyout tab three column types built in timestamp built in message custom field from underlying document | 0 |
8,116 | 7,242,961,049 | IssuesEvent | 2018-02-14 09:58:10 | stiftungswo/Dime | https://api.github.com/repos/stiftungswo/Dime | opened | Frontend Migration | in progress infrastructure | Ansprechperson: Andi
AngularDart v1 ist unmaintained und schlecht dokumentiert. Es soll eine Auffrischung der Technologie für das Dime Frontend evaluiert werden. Die verschiedenen Varianten sollen evaluiert und der zeitliche Aufwand geschätzt werden.
Kandidaten:
## AngularDart v4
### Motivation
Die Sprache bleibt dieselbe, aber die Frameworkversion wird auf die aktuelle und weiterentwickelte Version upgegradet. Somit profitiert man von den aktuellen neuen Dokumentationen und neuen Features, die entwickelt werden. Bestehender Code und Projektstruktur können weitgehend wiederverwertet werden.
### Evaluation
Die Hammock Library musste angepasst werden, funktioniert aber jetzt soweit. Login, ServiceOverview, ServiceEdit funktionieren soweit. Die Struktur von Code und Templates konnte fast 1:1 übernommen werden.
Code auf Branch https://github.com/stiftungswo/Dime/tree/angulardart-4-evaluation
Voraussichtlicher Zeitaufwand: 4 Wochen
## Typescript Angular
### Motivation
Die Technologie bleibt ähnlich, aber anstatt der Dart Version wird die standard Typescript Version des Frameworks verwendet, das eine grössere Community und somit mehr Support und Zukunftssicherheit bietet.
## React/Inferno
### Motivation
Der Vorteil hierbei wäre eine Vereinigung der Frontend Technologien, die in der Stiftung SWO entwickelt werden, da iZivi momentan mit Inferno läuft. Die Einstiegshürde für neue Zivis würde somit verringert. | 1.0 | Frontend Migration - Ansprechperson: Andi
AngularDart v1 ist unmaintained und schlecht dokumentiert. Es soll eine Auffrischung der Technologie für das Dime Frontend evaluiert werden. Die verschiedenen Varianten sollen evaluiert und der zeitliche Aufwand geschätzt werden.
Kandidaten:
## AngularDart v4
### Motivation
Die Sprache bleibt dieselbe, aber die Frameworkversion wird auf die aktuelle und weiterentwickelte Version upgegradet. Somit profitiert man von den aktuellen neuen Dokumentationen und neuen Features, die entwickelt werden. Bestehender Code und Projektstruktur können weitgehend wiederverwertet werden.
### Evaluation
Die Hammock Library musste angepasst werden, funktioniert aber jetzt soweit. Login, ServiceOverview, ServiceEdit funktionieren soweit. Die Struktur von Code und Templates konnte fast 1:1 übernommen werden.
Code auf Branch https://github.com/stiftungswo/Dime/tree/angulardart-4-evaluation
Voraussichtlicher Zeitaufwand: 4 Wochen
## Typescript Angular
### Motivation
Die Technologie bleibt ähnlich, aber anstatt der Dart Version wird die standard Typescript Version des Frameworks verwendet, das eine grössere Community und somit mehr Support und Zukunftssicherheit bietet.
## React/Inferno
### Motivation
Der Vorteil hierbei wäre eine Vereinigung der Frontend Technologien, die in der Stiftung SWO entwickelt werden, da iZivi momentan mit Inferno läuft. Die Einstiegshürde für neue Zivis würde somit verringert. | non_priority | frontend migration ansprechperson andi angulardart ist unmaintained und schlecht dokumentiert es soll eine auffrischung der technologie für das dime frontend evaluiert werden die verschiedenen varianten sollen evaluiert und der zeitliche aufwand geschätzt werden kandidaten angulardart motivation die sprache bleibt dieselbe aber die frameworkversion wird auf die aktuelle und weiterentwickelte version upgegradet somit profitiert man von den aktuellen neuen dokumentationen und neuen features die entwickelt werden bestehender code und projektstruktur können weitgehend wiederverwertet werden evaluation die hammock library musste angepasst werden funktioniert aber jetzt soweit login serviceoverview serviceedit funktionieren soweit die struktur von code und templates konnte fast übernommen werden code auf branch voraussichtlicher zeitaufwand wochen typescript angular motivation die technologie bleibt ähnlich aber anstatt der dart version wird die standard typescript version des frameworks verwendet das eine grössere community und somit mehr support und zukunftssicherheit bietet react inferno motivation der vorteil hierbei wäre eine vereinigung der frontend technologien die in der stiftung swo entwickelt werden da izivi momentan mit inferno läuft die einstiegshürde für neue zivis würde somit verringert | 0 |
326,532 | 27,998,548,035 | IssuesEvent | 2023-03-27 10:06:38 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | pkg/ccl/spanconfigccl/spanconfigsqlwatcherccl/spanconfigsqlwatcherccl_test: TestSQLWatcherOnEventError failed | C-test-failure O-robot X-stale T-kv no-test-failure-activity branch-release-22.2 | pkg/ccl/spanconfigccl/spanconfigsqlwatcherccl/spanconfigsqlwatcherccl_test.TestSQLWatcherOnEventError [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/8751541?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/8751541?buildTab=artifacts#/) on release-22.2 @ [31eb3ce6ceb18123fd2e7eac464b25e203d7de0f](https://github.com/cockroachdb/cockroach/commits/31eb3ce6ceb18123fd2e7eac464b25e203d7de0f):
```
I230218 08:33:09.958975 890758 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r51/1:/Table/5{1-2}] 507 + 0.619ms 0.619ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:682 [n1,replicate,s1,r51/1:/Table/5{1-2}] add voter - missing voter need=3, have=1, priority=10001.00
I230218 08:33:09.958975 890758 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r51/1:/Table/5{1-2}] 507 + 0.704ms 0.084ms event:kv/kvserver/replicate_queue.go:866 [n1,replicate,s1,r51/1:/Table/5{1-2}] next replica action: add voter
I230218 08:33:09.958975 890758 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r51/1:/Table/5{1-2}] 507 + 0.861ms 0.157ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:1037 [n1,replicate,s1,r51/1:/Table/5{1-2}] allocate voter: []
I230218 08:33:09.961667 885282 sql/syntheticprivilegecache/cache.go:206 [n1] 508 warmed privileges for virtual tables in 60.47079ms
W230218 08:33:09.965828 811817 kv/kvserver/replica_proposal.go:490 [n1,s1,r4/1:/System{/tsd-tse},raft] 509 lease <empty> expired before being followed by lease repl=(n1,s1):1 seq=1 start=0,0 epo=1 pro=1676709189.964821364,0; foreground traffic may have been impacted
I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 error processing replica: 0 of 1 live stores are able to take a new replica for the range (1 already has a voter, 0 already have a non-voter); likely not enough nodes in cluster
I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 +trace:
I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 + 0.000ms 0.000ms === operation:process replica _unfinished:1 _verbose:1 node:1 replicate: store:1 range:4/1:/System{/tsd-tse}
I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 + 0.250ms 0.250ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:682 [n1,replicate,s1,r4/1:/System{/tsd-tse}] add voter - missing voter need=3, have=1, priority=10001.00
I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 + 0.282ms 0.032ms event:kv/kvserver/replicate_queue.go:866 [n1,replicate,s1,r4/1:/System{/tsd-tse}] next replica action: add voter
I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 + 0.367ms 0.084ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:1037 [n1,replicate,s1,r4/1:/System{/tsd-tse}] allocate voter: []
I230218 08:33:09.972991 811813 3@pebble/event.go:678 [n1,pebble,s1] 511 [JOB 6] WAL created 000005
I230218 08:33:09.976140 805484 1@server/server_sql.go:1451 [n1] 512 serving sql connections
I230218 08:33:09.977398 805484 testutils/testcluster/testcluster.go:1339 [-] 513 WaitForFullReplication
I230218 08:33:09.977537 805484 testutils/testcluster/testcluster.go:1343 [-] 514 WaitForFullReplication took: 368ns
I230218 08:33:09.977715 805484 2@rpc/context.go:1774 [n1,rnode=1,raddr=127.0.0.1:34123,class=default] 515 dialing
I230218 08:33:10.050656 894362 util/log/event_log.go:32 [n1,client=127.0.0.1:37074,user=root] 516 ={"Timestamp":1676709190028595263,"EventType":"set_cluster_setting","Statement":"SET CLUSTER SETTING \"kv.rangefeed.enabled\" = true","Tag":"SET CLUSTER SETTING","User":"root","SettingName":"kv.rangefeed.enabled","Value":"true"}
I230218 08:33:10.094869 894362 util/log/event_log.go:32 [n1,client=127.0.0.1:37074,user=root] 517 ={"Timestamp":1676709190058971980,"EventType":"set_cluster_setting","Statement":"SET CLUSTER SETTING \"kv.closed_timestamp.target_duration\" = '100ms'","Tag":"SET CLUSTER SETTING","User":"root","SettingName":"kv.closed_timestamp.target_duration","Value":"00:00:00.1"}
W230218 08:33:10.169386 811829 kv/kvserver/replica_proposal.go:490 [n1,s1,r23/1:/Table/2{2-3},raft] 518 lease <empty> expired before being followed by lease repl=(n1,s1):1 seq=1 start=0,0 epo=1 pro=1676709190.167501234,0; foreground traffic may have been impacted
I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 error processing replica: 0 of 1 live stores are able to take a new replica for the range (1 already has a voter, 0 already have a non-voter); likely not enough nodes in cluster
I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 +trace:
I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 + 0.000ms 0.000ms === operation:process replica _unfinished:1 _verbose:1 node:1 replicate: store:1 range:23/1:/Table/2{2-3}
I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 + 0.194ms 0.194ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:682 [n1,replicate,s1,r23/1:/Table/2{2-3}] add voter - missing voter need=3, have=1, priority=10001.00
I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 + 0.239ms 0.044ms event:kv/kvserver/replicate_queue.go:866 [n1,replicate,s1,r23/1:/Table/2{2-3}] next replica action: add voter
I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 + 0.333ms 0.094ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:1037 [n1,replicate,s1,r23/1:/Table/2{2-3}] allocate voter: []
I230218 08:33:10.220490 894362 5@util/log/event_log.go:32 [n1,client=127.0.0.1:37074,user=root] 520 ={"Timestamp":1676709190117699231,"EventType":"create_table","Statement":"CREATE TABLE defaultdb.public.t ()","Tag":"CREATE TABLE","User":"root","DescriptorID":104,"TableName":"defaultdb.public.t"}
I230218 08:33:10.222227 805484 spanconfig/spanconfigsqlwatcher/sqlwatcher.go:244 [-] 521 established range feed over system.descriptors starting at time 1676709190.095893637,0
I230218 08:33:10.222410 805484 spanconfig/spanconfigsqlwatcher/sqlwatcher.go:303 [-] 522 established range feed over system.zones starting at time 1676709190.095893637,0
I230218 08:33:10.224092 805484 spanconfig/spanconfigsqlwatcher/sqlwatcher.go:399 [-] 523 established range feed over system.protected_ts_records starting at time 1676709190.095893637,0
sqlwatcher_test.go:460: handler should never run
I230218 08:33:10.234792 805484 testutils/testcluster/testcluster.go:128 [-] 524 TestCluster quiescing nodes
W230218 08:33:10.235023 811565 kv/kvserver/replica_proposal.go:490 [n1,s1,r33/1:/Table/3{2-3},raft] 525 lease <empty> expired before being followed by lease repl=(n1,s1):1 seq=1 start=0,0 epo=1 pro=1676709190.233308221,0; foreground traffic may have been impacted
I230218 08:33:10.236189 893316 2@rpc/context.go:1393 [n1] 526 closing [{n1: 127.0.0.1:34123 (default)} {n0: 127.0.0.1:34123 (default)}]
W230218 08:33:10.245611 816472 sql/sqlliveness/slinstance/slinstance.go:242 [n1] 527 exiting heartbeat loop
W230218 08:33:10.246151 866759 kv/kvserver/intentresolver/intent_resolver.go:799 [-] 528 failed to gc transaction record: could not GC completed transaction anchored at /Table/30/1/100/0/"public"/4/1: node unavailable; try another peer
W230218 08:33:10.246541 811565 kv/kvserver/store.go:2585 [n1,s1,r33/1:/Table/3{2-3},raft] 529 unable to gossip on capacity change: node unavailable; try another peer
W230218 08:33:10.246541 811565 kv/kvserver/store.go:2585 [n1,s1,r33/1:/Table/3{2-3},raft] 529 +(1) node unavailable; try another peer
W230218 08:33:10.246541 811565 kv/kvserver/store.go:2585 [n1,s1,r33/1:/Table/3{2-3},raft] 529 +Error types: (1) *roachpb.NodeUnavailableError
W230218 08:33:10.247802 817426 kv/txn.go:705 [n1] 530 failure aborting transaction: node unavailable; try another peer; abort caused by: expire-sessions: failed to read query result: context canceled
E230218 08:33:10.250288 817426 jobs/registry.go:744 [n1] 531 error expiring job sessions: expire-sessions: failed to read query result: context canceled
W230218 08:33:10.250467 817426 kv/txn.go:705 [n1] 532 failure aborting transaction: node unavailable; try another peer; abort caused by: txn exec: context canceled
E230218 08:33:10.250544 817426 jobs/registry.go:752 [n1] 533 failed to serve pause and cancel requests: txn exec: context canceled
W230218 08:33:10.250604 817426 jobs/registry.go:821 [n1] 534 canceling all adopted jobs due to stopper quiescing
W230218 08:33:10.253153 903355 kv/txn.go:705 [n1,intExec=select-running/get-claimed-jobs] 535 failure aborting transaction: node unavailable; try another peer; abort caused by: connExecutor closing
E230218 08:33:10.253580 817877 jobs/registry.go:805 [n1] 536 error processing claimed jobs: could not query for claimed jobs: select-running/get-claimed-jobs: failed to read query result: context canceled
E230218 08:33:10.265287 816760 sql/stats/automatic_stats.go:580 [n1] 537 failed to get tables for automatic stats: get-tables: node unavailable; try another peer
I230218 08:33:10.265586 816760 sql/stats/automatic_stats.go:509 [n1] 538 quiescing auto stats refresher
W230218 08:33:10.266359 904143 kv/txn.go:705 [n1,client=127.0.0.1:37074,user=root] 539 failure aborting transaction: node unavailable; try another peer; abort caused by: result is ambiguous: server shutdown
W230218 08:33:10.266541 904143 1@sql/event_log.go:641 [n1,client=127.0.0.1:37074,user=root] 540 unable to save 1 entries to system.eventlog: result is ambiguous: server shutdown
--- FAIL: TestSQLWatcherOnEventError (1.43s)
```
<p>Parameters: <code>TAGS=bazel,gss,deadlock</code>
</p>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #82031 pkg/ccl/spanconfigccl/spanconfigsqlwatcherccl/spanconfigsqlwatcherccl_test: TestSQLWatcherOnEventError failed [C-test-failure O-robot branch-release-22.1 sync-me sync-me-5]
- #76765 pkg/ccl/spanconfigccl/spanconfigsqlwatcherccl/spanconfigsqlwatcherccl_test: TestSQLWatcherOnEventError failed [C-test-failure O-robot branch-master sync-me-8]
</p>
</details>
/cc @cockroachdb/kv
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestSQLWatcherOnEventError.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-24633 | 2.0 | pkg/ccl/spanconfigccl/spanconfigsqlwatcherccl/spanconfigsqlwatcherccl_test: TestSQLWatcherOnEventError failed - pkg/ccl/spanconfigccl/spanconfigsqlwatcherccl/spanconfigsqlwatcherccl_test.TestSQLWatcherOnEventError [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/8751541?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/8751541?buildTab=artifacts#/) on release-22.2 @ [31eb3ce6ceb18123fd2e7eac464b25e203d7de0f](https://github.com/cockroachdb/cockroach/commits/31eb3ce6ceb18123fd2e7eac464b25e203d7de0f):
```
I230218 08:33:09.958975 890758 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r51/1:/Table/5{1-2}] 507 + 0.619ms 0.619ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:682 [n1,replicate,s1,r51/1:/Table/5{1-2}] add voter - missing voter need=3, have=1, priority=10001.00
I230218 08:33:09.958975 890758 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r51/1:/Table/5{1-2}] 507 + 0.704ms 0.084ms event:kv/kvserver/replicate_queue.go:866 [n1,replicate,s1,r51/1:/Table/5{1-2}] next replica action: add voter
I230218 08:33:09.958975 890758 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r51/1:/Table/5{1-2}] 507 + 0.861ms 0.157ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:1037 [n1,replicate,s1,r51/1:/Table/5{1-2}] allocate voter: []
I230218 08:33:09.961667 885282 sql/syntheticprivilegecache/cache.go:206 [n1] 508 warmed privileges for virtual tables in 60.47079ms
W230218 08:33:09.965828 811817 kv/kvserver/replica_proposal.go:490 [n1,s1,r4/1:/System{/tsd-tse},raft] 509 lease <empty> expired before being followed by lease repl=(n1,s1):1 seq=1 start=0,0 epo=1 pro=1676709189.964821364,0; foreground traffic may have been impacted
I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 error processing replica: 0 of 1 live stores are able to take a new replica for the range (1 already has a voter, 0 already have a non-voter); likely not enough nodes in cluster
I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 +trace:
I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 + 0.000ms 0.000ms === operation:process replica _unfinished:1 _verbose:1 node:1 replicate: store:1 range:4/1:/System{/tsd-tse}
I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 + 0.250ms 0.250ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:682 [n1,replicate,s1,r4/1:/System{/tsd-tse}] add voter - missing voter need=3, have=1, priority=10001.00
I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 + 0.282ms 0.032ms event:kv/kvserver/replicate_queue.go:866 [n1,replicate,s1,r4/1:/System{/tsd-tse}] next replica action: add voter
I230218 08:33:09.966958 891663 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r4/1:/System{/tsd-tse}] 510 + 0.367ms 0.084ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:1037 [n1,replicate,s1,r4/1:/System{/tsd-tse}] allocate voter: []
I230218 08:33:09.972991 811813 3@pebble/event.go:678 [n1,pebble,s1] 511 [JOB 6] WAL created 000005
I230218 08:33:09.976140 805484 1@server/server_sql.go:1451 [n1] 512 serving sql connections
I230218 08:33:09.977398 805484 testutils/testcluster/testcluster.go:1339 [-] 513 WaitForFullReplication
I230218 08:33:09.977537 805484 testutils/testcluster/testcluster.go:1343 [-] 514 WaitForFullReplication took: 368ns
I230218 08:33:09.977715 805484 2@rpc/context.go:1774 [n1,rnode=1,raddr=127.0.0.1:34123,class=default] 515 dialing
I230218 08:33:10.050656 894362 util/log/event_log.go:32 [n1,client=127.0.0.1:37074,user=root] 516 ={"Timestamp":1676709190028595263,"EventType":"set_cluster_setting","Statement":"SET CLUSTER SETTING \"kv.rangefeed.enabled\" = true","Tag":"SET CLUSTER SETTING","User":"root","SettingName":"kv.rangefeed.enabled","Value":"true"}
I230218 08:33:10.094869 894362 util/log/event_log.go:32 [n1,client=127.0.0.1:37074,user=root] 517 ={"Timestamp":1676709190058971980,"EventType":"set_cluster_setting","Statement":"SET CLUSTER SETTING \"kv.closed_timestamp.target_duration\" = '100ms'","Tag":"SET CLUSTER SETTING","User":"root","SettingName":"kv.closed_timestamp.target_duration","Value":"00:00:00.1"}
W230218 08:33:10.169386 811829 kv/kvserver/replica_proposal.go:490 [n1,s1,r23/1:/Table/2{2-3},raft] 518 lease <empty> expired before being followed by lease repl=(n1,s1):1 seq=1 start=0,0 epo=1 pro=1676709190.167501234,0; foreground traffic may have been impacted
I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 error processing replica: 0 of 1 live stores are able to take a new replica for the range (1 already has a voter, 0 already have a non-voter); likely not enough nodes in cluster
I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 +trace:
I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 + 0.000ms 0.000ms === operation:process replica _unfinished:1 _verbose:1 node:1 replicate: store:1 range:23/1:/Table/2{2-3}
I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 + 0.194ms 0.194ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:682 [n1,replicate,s1,r23/1:/Table/2{2-3}] add voter - missing voter need=3, have=1, priority=10001.00
I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 + 0.239ms 0.044ms event:kv/kvserver/replicate_queue.go:866 [n1,replicate,s1,r23/1:/Table/2{2-3}] next replica action: add voter
I230218 08:33:10.176968 901700 13@kv/kvserver/replicate_queue.go:817 [n1,replicate,s1,r23/1:/Table/2{2-3}] 519 + 0.333ms 0.094ms event:kv/kvserver/allocator/allocatorimpl/allocator.go:1037 [n1,replicate,s1,r23/1:/Table/2{2-3}] allocate voter: []
I230218 08:33:10.220490 894362 5@util/log/event_log.go:32 [n1,client=127.0.0.1:37074,user=root] 520 ={"Timestamp":1676709190117699231,"EventType":"create_table","Statement":"CREATE TABLE defaultdb.public.t ()","Tag":"CREATE TABLE","User":"root","DescriptorID":104,"TableName":"defaultdb.public.t"}
I230218 08:33:10.222227 805484 spanconfig/spanconfigsqlwatcher/sqlwatcher.go:244 [-] 521 established range feed over system.descriptors starting at time 1676709190.095893637,0
I230218 08:33:10.222410 805484 spanconfig/spanconfigsqlwatcher/sqlwatcher.go:303 [-] 522 established range feed over system.zones starting at time 1676709190.095893637,0
I230218 08:33:10.224092 805484 spanconfig/spanconfigsqlwatcher/sqlwatcher.go:399 [-] 523 established range feed over system.protected_ts_records starting at time 1676709190.095893637,0
sqlwatcher_test.go:460: handler should never run
I230218 08:33:10.234792 805484 testutils/testcluster/testcluster.go:128 [-] 524 TestCluster quiescing nodes
W230218 08:33:10.235023 811565 kv/kvserver/replica_proposal.go:490 [n1,s1,r33/1:/Table/3{2-3},raft] 525 lease <empty> expired before being followed by lease repl=(n1,s1):1 seq=1 start=0,0 epo=1 pro=1676709190.233308221,0; foreground traffic may have been impacted
I230218 08:33:10.236189 893316 2@rpc/context.go:1393 [n1] 526 closing [{n1: 127.0.0.1:34123 (default)} {n0: 127.0.0.1:34123 (default)}]
W230218 08:33:10.245611 816472 sql/sqlliveness/slinstance/slinstance.go:242 [n1] 527 exiting heartbeat loop
W230218 08:33:10.246151 866759 kv/kvserver/intentresolver/intent_resolver.go:799 [-] 528 failed to gc transaction record: could not GC completed transaction anchored at /Table/30/1/100/0/"public"/4/1: node unavailable; try another peer
W230218 08:33:10.246541 811565 kv/kvserver/store.go:2585 [n1,s1,r33/1:/Table/3{2-3},raft] 529 unable to gossip on capacity change: node unavailable; try another peer
W230218 08:33:10.246541 811565 kv/kvserver/store.go:2585 [n1,s1,r33/1:/Table/3{2-3},raft] 529 +(1) node unavailable; try another peer
W230218 08:33:10.246541 811565 kv/kvserver/store.go:2585 [n1,s1,r33/1:/Table/3{2-3},raft] 529 +Error types: (1) *roachpb.NodeUnavailableError
W230218 08:33:10.247802 817426 kv/txn.go:705 [n1] 530 failure aborting transaction: node unavailable; try another peer; abort caused by: expire-sessions: failed to read query result: context canceled
E230218 08:33:10.250288 817426 jobs/registry.go:744 [n1] 531 error expiring job sessions: expire-sessions: failed to read query result: context canceled
W230218 08:33:10.250467 817426 kv/txn.go:705 [n1] 532 failure aborting transaction: node unavailable; try another peer; abort caused by: txn exec: context canceled
E230218 08:33:10.250544 817426 jobs/registry.go:752 [n1] 533 failed to serve pause and cancel requests: txn exec: context canceled
W230218 08:33:10.250604 817426 jobs/registry.go:821 [n1] 534 canceling all adopted jobs due to stopper quiescing
W230218 08:33:10.253153 903355 kv/txn.go:705 [n1,intExec=select-running/get-claimed-jobs] 535 failure aborting transaction: node unavailable; try another peer; abort caused by: connExecutor closing
E230218 08:33:10.253580 817877 jobs/registry.go:805 [n1] 536 error processing claimed jobs: could not query for claimed jobs: select-running/get-claimed-jobs: failed to read query result: context canceled
E230218 08:33:10.265287 816760 sql/stats/automatic_stats.go:580 [n1] 537 failed to get tables for automatic stats: get-tables: node unavailable; try another peer
I230218 08:33:10.265586 816760 sql/stats/automatic_stats.go:509 [n1] 538 quiescing auto stats refresher
W230218 08:33:10.266359 904143 kv/txn.go:705 [n1,client=127.0.0.1:37074,user=root] 539 failure aborting transaction: node unavailable; try another peer; abort caused by: result is ambiguous: server shutdown
W230218 08:33:10.266541 904143 1@sql/event_log.go:641 [n1,client=127.0.0.1:37074,user=root] 540 unable to save 1 entries to system.eventlog: result is ambiguous: server shutdown
--- FAIL: TestSQLWatcherOnEventError (1.43s)
```
<p>Parameters: <code>TAGS=bazel,gss,deadlock</code>
</p>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #82031 pkg/ccl/spanconfigccl/spanconfigsqlwatcherccl/spanconfigsqlwatcherccl_test: TestSQLWatcherOnEventError failed [C-test-failure O-robot branch-release-22.1 sync-me sync-me-5]
- #76765 pkg/ccl/spanconfigccl/spanconfigsqlwatcherccl/spanconfigsqlwatcherccl_test: TestSQLWatcherOnEventError failed [C-test-failure O-robot branch-master sync-me-8]
</p>
</details>
/cc @cockroachdb/kv
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestSQLWatcherOnEventError.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-24633 | non_priority | pkg ccl spanconfigccl spanconfigsqlwatcherccl spanconfigsqlwatcherccl test testsqlwatcheroneventerror failed pkg ccl spanconfigccl spanconfigsqlwatcherccl spanconfigsqlwatcherccl test testsqlwatcheroneventerror with on release kv kvserver replicate queue go event kv kvserver allocator allocatorimpl allocator go add voter missing voter need have priority kv kvserver replicate queue go event kv kvserver replicate queue go next replica action add voter kv kvserver replicate queue go event kv kvserver allocator allocatorimpl allocator go allocate voter sql syntheticprivilegecache cache go warmed privileges for virtual tables in kv kvserver replica proposal go lease expired before being followed by lease repl seq start epo pro foreground traffic may have been impacted kv kvserver replicate queue go error processing replica of live stores are able to take a new replica for the range already has a voter already have a non voter likely not enough nodes in cluster kv kvserver replicate queue go trace kv kvserver replicate queue go operation process replica unfinished verbose node replicate store range system tsd tse kv kvserver replicate queue go event kv kvserver allocator allocatorimpl allocator go add voter missing voter need have priority kv kvserver replicate queue go event kv kvserver replicate queue go next replica action add voter kv kvserver replicate queue go event kv kvserver allocator allocatorimpl allocator go allocate voter pebble event go wal created server server sql go serving sql connections testutils testcluster testcluster go waitforfullreplication testutils testcluster testcluster go waitforfullreplication took rpc context go dialing util log event log go timestamp eventtype set cluster setting statement set cluster setting kv rangefeed enabled true tag set cluster setting user root settingname kv rangefeed enabled value true util log event log go timestamp eventtype set cluster setting statement set cluster setting kv closed timestamp target duration tag set cluster setting user root settingname kv closed timestamp target duration value kv kvserver replica proposal go lease expired before being followed by lease repl seq start epo pro foreground traffic may have been impacted kv kvserver replicate queue go error processing replica of live stores are able to take a new replica for the range already has a voter already have a non voter likely not enough nodes in cluster kv kvserver replicate queue go trace kv kvserver replicate queue go operation process replica unfinished verbose node replicate store range table kv kvserver replicate queue go event kv kvserver allocator allocatorimpl allocator go add voter missing voter need have priority kv kvserver replicate queue go event kv kvserver replicate queue go next replica action add voter kv kvserver replicate queue go event kv kvserver allocator allocatorimpl allocator go allocate voter util log event log go timestamp eventtype create table statement create table defaultdb public t tag create table user root descriptorid tablename defaultdb public t spanconfig spanconfigsqlwatcher sqlwatcher go established range feed over system descriptors starting at time spanconfig spanconfigsqlwatcher sqlwatcher go established range feed over system zones starting at time spanconfig spanconfigsqlwatcher sqlwatcher go established range feed over system protected ts records starting at time sqlwatcher test go handler should never run testutils testcluster testcluster go testcluster quiescing nodes kv kvserver replica proposal go lease expired before being followed by lease repl seq start epo pro foreground traffic may have been impacted rpc context go closing sql sqlliveness slinstance slinstance go exiting heartbeat loop kv kvserver intentresolver intent resolver go failed to gc transaction record could not gc completed transaction anchored at table public node unavailable try another peer kv kvserver store go unable to gossip on capacity change node unavailable try another peer kv kvserver store go node unavailable try another peer kv kvserver store go error types roachpb nodeunavailableerror kv txn go failure aborting transaction node unavailable try another peer abort caused by expire sessions failed to read query result context canceled jobs registry go error expiring job sessions expire sessions failed to read query result context canceled kv txn go failure aborting transaction node unavailable try another peer abort caused by txn exec context canceled jobs registry go failed to serve pause and cancel requests txn exec context canceled jobs registry go canceling all adopted jobs due to stopper quiescing kv txn go failure aborting transaction node unavailable try another peer abort caused by connexecutor closing jobs registry go error processing claimed jobs could not query for claimed jobs select running get claimed jobs failed to read query result context canceled sql stats automatic stats go failed to get tables for automatic stats get tables node unavailable try another peer sql stats automatic stats go quiescing auto stats refresher kv txn go failure aborting transaction node unavailable try another peer abort caused by result is ambiguous server shutdown sql event log go unable to save entries to system eventlog result is ambiguous server shutdown fail testsqlwatcheroneventerror parameters tags bazel gss deadlock help see also same failure on other branches pkg ccl spanconfigccl spanconfigsqlwatcherccl spanconfigsqlwatcherccl test testsqlwatcheroneventerror failed pkg ccl spanconfigccl spanconfigsqlwatcherccl spanconfigsqlwatcherccl test testsqlwatcheroneventerror failed cc cockroachdb kv jira issue crdb | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.