Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
496,804
14,355,386,991
IssuesEvent
2020-11-30 10:01:58
zeebe-io/zeebe
https://api.github.com/repos/zeebe-io/zeebe
closed
Add checksum to AppendEntries request
Impact: Data Priority: High Scope: broker Status: Needs Review Type: Maintenance
**Description** We should add checksums to AppendEntries requests in order to ensure that followers don't write corrupted entries.
1.0
Add checksum to AppendEntries request - **Description** We should add checksums to AppendEntries requests in order to ensure that followers don't write corrupted entries.
non_test
add checksum to appendentries request description we should add checksums to appendentries requests in order to ensure that followers don t write corrupted entries
0
351,327
25,022,488,342
IssuesEvent
2022-11-04 03:06:29
AY2223S1-CS2103T-W15-2/tp
https://api.github.com/repos/AY2223S1-CS2103T-W15-2/tp
closed
[PE-D][Tester B] Unexpected behavior for addO command
documentation priority.High
**Steps to produce:** 1. Enter addO id/abc a/John Doe o/2000000 as suggested by the UG. 2. Enter add0 id/1 a/Betsy Crowe o/20 as suggested by the UG. ![image.png](https://raw.githubusercontent.com/wweqg/ped/main/files/8b276cd4-a2a6-49b1-88fa-56ad4eec0ea4.png) **Expected:** 1. Added 2 new orders **Actual:** 1. Received "Invalid Command" for first command. ![image.png](https://raw.githubusercontent.com/wweqg/ped/main/files/cc0f967b-0fd8-4ba2-b611-760bd7493d28.png) 2. Received "Unknown Command" for the second. ![image.png](https://raw.githubusercontent.com/wweqg/ped/main/files/c5202ea8-b6c9-42ad-b16b-ddc2a53379ce.png) <!--session: 1666944217695-35af3a60-58e4-4a68-9c43-9e45d7bd947a--> <!--Version: Web v3.4.4--> ------------- Labels: `type.DocumentationBug` `severity.Low` original: wweqg/ped#5
1.0
[PE-D][Tester B] Unexpected behavior for addO command - **Steps to produce:** 1. Enter addO id/abc a/John Doe o/2000000 as suggested by the UG. 2. Enter add0 id/1 a/Betsy Crowe o/20 as suggested by the UG. ![image.png](https://raw.githubusercontent.com/wweqg/ped/main/files/8b276cd4-a2a6-49b1-88fa-56ad4eec0ea4.png) **Expected:** 1. Added 2 new orders **Actual:** 1. Received "Invalid Command" for first command. ![image.png](https://raw.githubusercontent.com/wweqg/ped/main/files/cc0f967b-0fd8-4ba2-b611-760bd7493d28.png) 2. Received "Unknown Command" for the second. ![image.png](https://raw.githubusercontent.com/wweqg/ped/main/files/c5202ea8-b6c9-42ad-b16b-ddc2a53379ce.png) <!--session: 1666944217695-35af3a60-58e4-4a68-9c43-9e45d7bd947a--> <!--Version: Web v3.4.4--> ------------- Labels: `type.DocumentationBug` `severity.Low` original: wweqg/ped#5
non_test
unexpected behavior for addo command steps to produce enter addo id abc a john doe o as suggested by the ug enter id a betsy crowe o as suggested by the ug expected added new orders actual received invalid command for first command received unknown command for the second labels type documentationbug severity low original wweqg ped
0
28,018
4,350,480,181
IssuesEvent
2016-07-31 08:49:12
EasyRPG/Player
https://api.github.com/repos/EasyRPG/Player
closed
3DS crash on Moby Housekeeper
3DS Crash Duplicate Testcase available
__Name of the game__: Moby Housekeeper __Player platform__: 3DS, .cia version (Edit: Using the continuous build) __Describe the issue in detail and how to reproduce it__: Load the game Moby Housekeeper from https://rpgmaker.net/games/23/ . Play it on a 3DS (I used the English translated RTP--using a Japanese RTP is problematic because of the FAT32 on the SD card). Start a new game, see the E rating screen, then it either crashes or locks up. It does not do this on a PC. [easyrpg_log.txt](https://github.com/EasyRPG/Player/files/391667/easyrpg_log.txt)
1.0
3DS crash on Moby Housekeeper - __Name of the game__: Moby Housekeeper __Player platform__: 3DS, .cia version (Edit: Using the continuous build) __Describe the issue in detail and how to reproduce it__: Load the game Moby Housekeeper from https://rpgmaker.net/games/23/ . Play it on a 3DS (I used the English translated RTP--using a Japanese RTP is problematic because of the FAT32 on the SD card). Start a new game, see the E rating screen, then it either crashes or locks up. It does not do this on a PC. [easyrpg_log.txt](https://github.com/EasyRPG/Player/files/391667/easyrpg_log.txt)
test
crash on moby housekeeper name of the game moby housekeeper player platform cia version edit using the continuous build describe the issue in detail and how to reproduce it load the game moby housekeeper from play it on a i used the english translated rtp using a japanese rtp is problematic because of the on the sd card start a new game see the e rating screen then it either crashes or locks up it does not do this on a pc
1
155,415
12,255,038,045
IssuesEvent
2020-05-06 09:30:51
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
Failures on Allow `NoSuchMethodError.withInvocation` to accept any `Invocation`.
area-test gardening
There are new test failures on [Allow `NoSuchMethodError.withInvocation` to accept any `Invocation`.](https://github.com/dart-lang/sdk/commit/2ca3555c44ac7b98bb366e19ff891c4b4c4940e7). The tests ``` standalone_2/no_such_method_error_with_invocation_test RuntimeError (expected Pass) ``` are failing on configurations ``` dartkp-obfuscate-linux-release-x64 ``` This test should either be modified to work with obfuscation or SkipByDesign'd in the status file.
1.0
Failures on Allow `NoSuchMethodError.withInvocation` to accept any `Invocation`. - There are new test failures on [Allow `NoSuchMethodError.withInvocation` to accept any `Invocation`.](https://github.com/dart-lang/sdk/commit/2ca3555c44ac7b98bb366e19ff891c4b4c4940e7). The tests ``` standalone_2/no_such_method_error_with_invocation_test RuntimeError (expected Pass) ``` are failing on configurations ``` dartkp-obfuscate-linux-release-x64 ``` This test should either be modified to work with obfuscation or SkipByDesign'd in the status file.
test
failures on allow nosuchmethoderror withinvocation to accept any invocation there are new test failures on the tests standalone no such method error with invocation test runtimeerror expected pass are failing on configurations dartkp obfuscate linux release this test should either be modified to work with obfuscation or skipbydesign d in the status file
1
246,943
20,945,962,468
IssuesEvent
2022-03-26 00:08:50
microsoft/msquic
https://api.github.com/repos/microsoft/msquic
opened
Server Initial Read Key not discarded on Compatible Version Negotiation
Bug: Core Area: Core Bug: Test/Tool
### Describe the bug The QUIC version 2 draft says that a server which upgrades a client to Version 2 via Compatible Version Negotiation (CVN) > The server MUST NOT discard its original version Initial receive keys until it successfully processes a packet with the negotiated version. [Source](https://www.ietf.org/archive/id/draft-ietf-quic-v2-01.html#section-4.1-5) This was not implemented yet, and is difficult to test ### Affected OS - [ ] All - [ ] Windows Server 2022 - [ ] Windows 11 - [ ] Windows Insider Preview (specify affected build below) - [ ] Ubuntu - [ ] Debian - [ ] Other (specify below) ### Additional OS information _No response_ ### MsQuic version main ### Steps taken to reproduce bug 1. Client sends an Initial packet with all the crypto payload in it. Client VersionInfo starts with Version 1, but supports Version 2. 2. Server processes client Initial packet and uses CVN to upgrade client to Version 2. Server discards both read and write initial keys for Version 1, and replaces them with Initial keys for Version 2. 3. Client sends a second Initial packet with padding(??), before processing the Server's flight. ### Expected behavior The Server should still be able to read the Client's second Initial packet until the Client acknowledges the CVN by sending a Handshake flight with the new version. ### Actual outcome Server fails to decrypt client's second Initial packet and drops it. ### Additional details _No response_
1.0
Server Initial Read Key not discarded on Compatible Version Negotiation - ### Describe the bug The QUIC version 2 draft says that a server which upgrades a client to Version 2 via Compatible Version Negotiation (CVN) > The server MUST NOT discard its original version Initial receive keys until it successfully processes a packet with the negotiated version. [Source](https://www.ietf.org/archive/id/draft-ietf-quic-v2-01.html#section-4.1-5) This was not implemented yet, and is difficult to test ### Affected OS - [ ] All - [ ] Windows Server 2022 - [ ] Windows 11 - [ ] Windows Insider Preview (specify affected build below) - [ ] Ubuntu - [ ] Debian - [ ] Other (specify below) ### Additional OS information _No response_ ### MsQuic version main ### Steps taken to reproduce bug 1. Client sends an Initial packet with all the crypto payload in it. Client VersionInfo starts with Version 1, but supports Version 2. 2. Server processes client Initial packet and uses CVN to upgrade client to Version 2. Server discards both read and write initial keys for Version 1, and replaces them with Initial keys for Version 2. 3. Client sends a second Initial packet with padding(??), before processing the Server's flight. ### Expected behavior The Server should still be able to read the Client's second Initial packet until the Client acknowledges the CVN by sending a Handshake flight with the new version. ### Actual outcome Server fails to decrypt client's second Initial packet and drops it. ### Additional details _No response_
test
server initial read key not discarded on compatible version negotiation describe the bug the quic version draft says that a server which upgrades a client to version via compatible version negotiation cvn the server must not discard its original version initial receive keys until it successfully processes a packet with the negotiated version this was not implemented yet and is difficult to test affected os all windows server windows windows insider preview specify affected build below ubuntu debian other specify below additional os information no response msquic version main steps taken to reproduce bug client sends an initial packet with all the crypto payload in it client versioninfo starts with version but supports version server processes client initial packet and uses cvn to upgrade client to version server discards both read and write initial keys for version and replaces them with initial keys for version client sends a second initial packet with padding before processing the server s flight expected behavior the server should still be able to read the client s second initial packet until the client acknowledges the cvn by sending a handshake flight with the new version actual outcome server fails to decrypt client s second initial packet and drops it additional details no response
1
3,152
4,105,936,007
IssuesEvent
2016-06-06 05:58:05
KhronosGroup/glslang
https://api.github.com/repos/KhronosGroup/glslang
closed
On windows, debug and release builds are installed to the same location with the same file names
Infrastructure
This makes it problematic to embed this project's libraries into a larger CMake project that needs to build in both debug and release.
1.0
On windows, debug and release builds are installed to the same location with the same file names - This makes it problematic to embed this project's libraries into a larger CMake project that needs to build in both debug and release.
non_test
on windows debug and release builds are installed to the same location with the same file names this makes it problematic to embed this project s libraries into a larger cmake project that needs to build in both debug and release
0
63,574
6,850,513,045
IssuesEvent
2017-11-14 03:43:58
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
teamcity: failed tests on 20019: Jepsen/Jepsen: JepsenBank: JepsenBank/split, Jepsen/Jepsen: JepsenBank-multitable: JepsenBank-multitable/start-kill-2, Jepsen/Jepsen: JepsenSets: JepsenSets/subcritical-skews+start-kill-2
Robot test-failure
The following tests appear to have failed: [#408054](https://teamcity.cockroachdb.com/viewLog.html?buildId=408054): ``` --- FAIL: Jepsen/Jepsen: JepsenBank: JepsenBank/split (130.492s) None --- FAIL: Jepsen/Jepsen: JepsenBank-multitable: JepsenBank-multitable/start-kill-2 (467.398s) None --- FAIL: Jepsen/Jepsen: JepsenSets: JepsenSets/subcritical-skews+start-kill-2 (129.548s) None ``` Please assign, take a look and update the issue accordingly.
1.0
teamcity: failed tests on 20019: Jepsen/Jepsen: JepsenBank: JepsenBank/split, Jepsen/Jepsen: JepsenBank-multitable: JepsenBank-multitable/start-kill-2, Jepsen/Jepsen: JepsenSets: JepsenSets/subcritical-skews+start-kill-2 - The following tests appear to have failed: [#408054](https://teamcity.cockroachdb.com/viewLog.html?buildId=408054): ``` --- FAIL: Jepsen/Jepsen: JepsenBank: JepsenBank/split (130.492s) None --- FAIL: Jepsen/Jepsen: JepsenBank-multitable: JepsenBank-multitable/start-kill-2 (467.398s) None --- FAIL: Jepsen/Jepsen: JepsenSets: JepsenSets/subcritical-skews+start-kill-2 (129.548s) None ``` Please assign, take a look and update the issue accordingly.
test
teamcity failed tests on jepsen jepsen jepsenbank jepsenbank split jepsen jepsen jepsenbank multitable jepsenbank multitable start kill jepsen jepsen jepsensets jepsensets subcritical skews start kill the following tests appear to have failed fail jepsen jepsen jepsenbank jepsenbank split none fail jepsen jepsen jepsenbank multitable jepsenbank multitable start kill none fail jepsen jepsen jepsensets jepsensets subcritical skews start kill none please assign take a look and update the issue accordingly
1
302,764
26,161,119,176
IssuesEvent
2022-12-31 14:51:22
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: unoptimized-query-oracle/disable-rules=half failed
C-test-failure O-robot O-roachtest branch-release-22.2
roachtest.unoptimized-query-oracle/disable-rules=half [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=artifacts#/unoptimized-query-oracle/disable-rules=half) on release-22.2 @ [07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1](https://github.com/cockroachdb/cockroach/commits/07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1): ``` test artifacts and logs in: /artifacts/unoptimized-query-oracle/disable-rules=half/run_1 (test_impl.go:286).Fatal: pq: Use of partitions requires an enterprise license. Your evaluation license expired on December 30, 2022. If you're interested in getting a new license, please contact subscriptions@cockroachlabs.com and we can help you out. ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_fs=ext4</code> , <code>ROACHTEST_localSSD=true</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #94557 roachtest: unoptimized-query-oracle/disable-rules=half/seed-multi-region failed [C-test-failure O-roachtest O-robot T-sql-queries branch-master release-blocker] </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*unoptimized-query-oracle/disable-rules=half.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: unoptimized-query-oracle/disable-rules=half failed - roachtest.unoptimized-query-oracle/disable-rules=half [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=artifacts#/unoptimized-query-oracle/disable-rules=half) on release-22.2 @ [07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1](https://github.com/cockroachdb/cockroach/commits/07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1): ``` test artifacts and logs in: /artifacts/unoptimized-query-oracle/disable-rules=half/run_1 (test_impl.go:286).Fatal: pq: Use of partitions requires an enterprise license. Your evaluation license expired on December 30, 2022. If you're interested in getting a new license, please contact subscriptions@cockroachlabs.com and we can help you out. ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_fs=ext4</code> , <code>ROACHTEST_localSSD=true</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #94557 roachtest: unoptimized-query-oracle/disable-rules=half/seed-multi-region failed [C-test-failure O-roachtest O-robot T-sql-queries branch-master release-blocker] </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*unoptimized-query-oracle/disable-rules=half.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
roachtest unoptimized query oracle disable rules half failed roachtest unoptimized query oracle disable rules half with on release test artifacts and logs in artifacts unoptimized query oracle disable rules half run test impl go fatal pq use of partitions requires an enterprise license your evaluation license expired on december if you re interested in getting a new license please contact subscriptions cockroachlabs com and we can help you out parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest fs roachtest localssd true roachtest ssd help see see same failure on other branches roachtest unoptimized query oracle disable rules half seed multi region failed cc cockroachdb sql queries
1
205,241
15,598,147,127
IssuesEvent
2021-03-18 17:44:44
scikit-hep/pyhf
https://api.github.com/repos/scikit-hep/pyhf
closed
Regression in sbottom regression tests
bug tests
# Description As of 2021-03-11 there is a regression in `tests/test_regression.py` that is arising from the following change (as noted by comapring outputs from the minimum supported dependencies workflow ``` $ diff pass_requirements.txt fail_requirements.txt 84,85c84,85 < prompt-toolkit 3.0.16 < protobuf 3.15.5 --- > prompt-toolkit 3.0.17 > protobuf 3.15.6 ``` # Expected Behavior CI passes # Actual Behavior ```pytb =================================== FAILURES =================================== _______________________ test_sbottom_regionA_1300_205_60 _______________________ sbottom_likelihoods_download = <tarfile.TarFile object at 0x7fd1b420add0> get_json_from_tarfile = <function get_json_from_tarfile.<locals>._get_json_from_tarfile at 0x7fd1d05925f0> def test_sbottom_regionA_1300_205_60( sbottom_likelihoods_download, get_json_from_tarfile ): sbottom_regionA_bkgonly_json = get_json_from_tarfile( sbottom_likelihoods_download, "RegionA/BkgOnly.json" ) sbottom_regionA_1300_205_60_patch_json = get_json_from_tarfile( sbottom_likelihoods_download, "RegionA/patch.sbottom_1300_205_60.json" ) CLs_obs, CLs_exp = calculate_CLs( sbottom_regionA_bkgonly_json, sbottom_regionA_1300_205_60_patch_json ) assert CLs_obs == pytest.approx(0.24443627759085326, rel=1e-5) > assert np.all( np.isclose( np.array(CLs_exp), np.array( [ 0.09022509053507759, 0.1937839194960632, 0.38432344933992, 0.6557757334303531, 0.8910420971601081, ] ), rtol=1e-5, ) ) E assert False E + where False = <function all at 0x7fd2c139d0e0>(array([False, True, True, True, True])) E + where <function all at 0x7fd2c139d0e0> = np.all E + and array([False, True, True, True, True]) = <function isclose at 0x7fd2c13af4d0>(array([0.09022391, 0.19378211, 0.38432119, 0.65577379, 0.89104121]), array([0.09022509, 0.19378392, 0.38432345, 0.65577573, 0.8910421 ]), rtol=1e-05) E + where <function isclose at 0x7fd2c13af4d0> = np.isclose E + and array([0.09022391, 0.19378211, 0.38432119, 0.65577379, 0.89104121]) = <built-in function array>([array(0.09022391), array(0.19378211), array(0.38432119), array(0.65577379), array(0.89104121)]) E + where <built-in function array> = np.array E + and array([0.09022509, 0.19378392, 0.38432345, 0.65577573, 0.8910421 ]) = <built-in function array>([0.09022509053507759, 0.1937839194960632, 0.38432344933992, 0.6557757334303531, 0.8910420971601081]) E + where <built-in function array> = np.array tests/test_regression.py:47: AssertionError ``` # Steps to Reproduce Run CI # Checklist - [x] Run `git fetch` to get the most up to date version of `master` - [x] Searched through existing Issues to confirm this is not a duplicate issue - [x] Filled out the Description, Expected Behavior, Actual Behavior, and Steps to Reproduce sections above or have edited/removed them in a way that fully describes the issue
1.0
Regression in sbottom regression tests - # Description As of 2021-03-11 there is a regression in `tests/test_regression.py` that is arising from the following change (as noted by comapring outputs from the minimum supported dependencies workflow ``` $ diff pass_requirements.txt fail_requirements.txt 84,85c84,85 < prompt-toolkit 3.0.16 < protobuf 3.15.5 --- > prompt-toolkit 3.0.17 > protobuf 3.15.6 ``` # Expected Behavior CI passes # Actual Behavior ```pytb =================================== FAILURES =================================== _______________________ test_sbottom_regionA_1300_205_60 _______________________ sbottom_likelihoods_download = <tarfile.TarFile object at 0x7fd1b420add0> get_json_from_tarfile = <function get_json_from_tarfile.<locals>._get_json_from_tarfile at 0x7fd1d05925f0> def test_sbottom_regionA_1300_205_60( sbottom_likelihoods_download, get_json_from_tarfile ): sbottom_regionA_bkgonly_json = get_json_from_tarfile( sbottom_likelihoods_download, "RegionA/BkgOnly.json" ) sbottom_regionA_1300_205_60_patch_json = get_json_from_tarfile( sbottom_likelihoods_download, "RegionA/patch.sbottom_1300_205_60.json" ) CLs_obs, CLs_exp = calculate_CLs( sbottom_regionA_bkgonly_json, sbottom_regionA_1300_205_60_patch_json ) assert CLs_obs == pytest.approx(0.24443627759085326, rel=1e-5) > assert np.all( np.isclose( np.array(CLs_exp), np.array( [ 0.09022509053507759, 0.1937839194960632, 0.38432344933992, 0.6557757334303531, 0.8910420971601081, ] ), rtol=1e-5, ) ) E assert False E + where False = <function all at 0x7fd2c139d0e0>(array([False, True, True, True, True])) E + where <function all at 0x7fd2c139d0e0> = np.all E + and array([False, True, True, True, True]) = <function isclose at 0x7fd2c13af4d0>(array([0.09022391, 0.19378211, 0.38432119, 0.65577379, 0.89104121]), array([0.09022509, 0.19378392, 0.38432345, 0.65577573, 0.8910421 ]), rtol=1e-05) E + where <function isclose at 0x7fd2c13af4d0> = np.isclose E + and array([0.09022391, 0.19378211, 0.38432119, 0.65577379, 0.89104121]) = <built-in function array>([array(0.09022391), array(0.19378211), array(0.38432119), array(0.65577379), array(0.89104121)]) E + where <built-in function array> = np.array E + and array([0.09022509, 0.19378392, 0.38432345, 0.65577573, 0.8910421 ]) = <built-in function array>([0.09022509053507759, 0.1937839194960632, 0.38432344933992, 0.6557757334303531, 0.8910420971601081]) E + where <built-in function array> = np.array tests/test_regression.py:47: AssertionError ``` # Steps to Reproduce Run CI # Checklist - [x] Run `git fetch` to get the most up to date version of `master` - [x] Searched through existing Issues to confirm this is not a duplicate issue - [x] Filled out the Description, Expected Behavior, Actual Behavior, and Steps to Reproduce sections above or have edited/removed them in a way that fully describes the issue
test
regression in sbottom regression tests description as of there is a regression in tests test regression py that is arising from the following change as noted by comapring outputs from the minimum supported dependencies workflow diff pass requirements txt fail requirements txt prompt toolkit protobuf prompt toolkit protobuf expected behavior ci passes actual behavior pytb failures test sbottom regiona sbottom likelihoods download get json from tarfile get json from tarfile at def test sbottom regiona sbottom likelihoods download get json from tarfile sbottom regiona bkgonly json get json from tarfile sbottom likelihoods download regiona bkgonly json sbottom regiona patch json get json from tarfile sbottom likelihoods download regiona patch sbottom json cls obs cls exp calculate cls sbottom regiona bkgonly json sbottom regiona patch json assert cls obs pytest approx rel assert np all np isclose np array cls exp np array rtol e assert false e where false array e where np all e and array array array rtol e where np isclose e and array e where np array e and array e where np array tests test regression py assertionerror steps to reproduce run ci checklist run git fetch to get the most up to date version of master searched through existing issues to confirm this is not a duplicate issue filled out the description expected behavior actual behavior and steps to reproduce sections above or have edited removed them in a way that fully describes the issue
1
38,105
5,166,527,367
IssuesEvent
2017-01-17 16:26:11
glitchassassin/lackey
https://api.github.com/repos/glitchassassin/lackey
closed
TypeError: bytes or integer address expected instead of str instance
bug fix in testing
Hello, i'm using python 3.5.x, and windows 7 x64. When i use this piece of code: ``` import lackey pattern = lackey.Pattern(r'C:\Users\rainman\Documents\cam\my.png') screen = lackey.Screen() screen.click(pattern) # or screen.find(patter) or screen.capture() ``` Error happens: ``` hdc = self._gdi32.CreateDCA(ctypes.c_char_p(device_name), 0, 0, 0) TypeError: bytes or integer address expected instead of str instance ```
1.0
TypeError: bytes or integer address expected instead of str instance - Hello, i'm using python 3.5.x, and windows 7 x64. When i use this piece of code: ``` import lackey pattern = lackey.Pattern(r'C:\Users\rainman\Documents\cam\my.png') screen = lackey.Screen() screen.click(pattern) # or screen.find(patter) or screen.capture() ``` Error happens: ``` hdc = self._gdi32.CreateDCA(ctypes.c_char_p(device_name), 0, 0, 0) TypeError: bytes or integer address expected instead of str instance ```
test
typeerror bytes or integer address expected instead of str instance hello i m using python x and windows when i use this piece of code import lackey pattern lackey pattern r c users rainman documents cam my png screen lackey screen screen click pattern or screen find patter or screen capture error happens hdc self createdca ctypes c char p device name typeerror bytes or integer address expected instead of str instance
1
805,895
29,736,098,320
IssuesEvent
2023-06-14 01:07:38
longhorn/longhorn
https://api.github.com/repos/longhorn/longhorn
closed
[BUG] After migration of Longhorn from Rancher old UI to dashboard, the csi-plugin doesn't update
kind/bug reproduce/always priority/0 area/csi severity/4
## Describe the bug Do a migration from Rancher old UI to dashboard using https://longhorn.io/kb/how-to-migrate-longhorn-chart-installed-in-old-rancher-ui-to-the-chart-in-new-rancher-ui/, the csi-plugin doesn't update and remains with the longhornio image instead of Rancher mirrored image. ## To Reproduce Steps to reproduce the behavior: 1. Set up a cluster of Kubernetes 1.20. 1. Access old rancher UI by navigating to `<your-rancher-url>/g`. 1. Install Longhorn 1.0.2. 1. Create/attach some volumes. Create a few recurring snapshot/backup job that run every minutes. 1. Upgrade Longhorn to v1.2.4. 1. Migrate Longhorn to new chart in new rancher UI https://longhorn.io/kb/how-to-migrate-longhorn-chart-installed-in-old-rancher-ui-to-the-chart-in-new-rancher-ui/. 1. Check the csi plugin image ## Expected behavior The CSI plugin image should also be updated to rancher mirrored image ## Log or Support bundle <img width="1537" alt="Screen Shot 2022-09-01 at 3 10 54 PM" src="https://user-images.githubusercontent.com/60111667/188244752-1451fbf6-8b1c-4662-993c-1f7be0179dbb.png"> ## Environment - Longhorn version: v1.2.4 - Installation method (e.g. Rancher Catalog App/Helm/Kubectl): Rancher App - Kubernetes distro (e.g. RKE/K3s/EKS/OpenShift) and version: v1.20
1.0
[BUG] After migration of Longhorn from Rancher old UI to dashboard, the csi-plugin doesn't update - ## Describe the bug Do a migration from Rancher old UI to dashboard using https://longhorn.io/kb/how-to-migrate-longhorn-chart-installed-in-old-rancher-ui-to-the-chart-in-new-rancher-ui/, the csi-plugin doesn't update and remains with the longhornio image instead of Rancher mirrored image. ## To Reproduce Steps to reproduce the behavior: 1. Set up a cluster of Kubernetes 1.20. 1. Access old rancher UI by navigating to `<your-rancher-url>/g`. 1. Install Longhorn 1.0.2. 1. Create/attach some volumes. Create a few recurring snapshot/backup job that run every minutes. 1. Upgrade Longhorn to v1.2.4. 1. Migrate Longhorn to new chart in new rancher UI https://longhorn.io/kb/how-to-migrate-longhorn-chart-installed-in-old-rancher-ui-to-the-chart-in-new-rancher-ui/. 1. Check the csi plugin image ## Expected behavior The CSI plugin image should also be updated to rancher mirrored image ## Log or Support bundle <img width="1537" alt="Screen Shot 2022-09-01 at 3 10 54 PM" src="https://user-images.githubusercontent.com/60111667/188244752-1451fbf6-8b1c-4662-993c-1f7be0179dbb.png"> ## Environment - Longhorn version: v1.2.4 - Installation method (e.g. Rancher Catalog App/Helm/Kubectl): Rancher App - Kubernetes distro (e.g. RKE/K3s/EKS/OpenShift) and version: v1.20
non_test
after migration of longhorn from rancher old ui to dashboard the csi plugin doesn t update describe the bug do a migration from rancher old ui to dashboard using the csi plugin doesn t update and remains with the longhornio image instead of rancher mirrored image to reproduce steps to reproduce the behavior set up a cluster of kubernetes access old rancher ui by navigating to g install longhorn create attach some volumes create a few recurring snapshot backup job that run every minutes upgrade longhorn to migrate longhorn to new chart in new rancher ui check the csi plugin image expected behavior the csi plugin image should also be updated to rancher mirrored image log or support bundle img width alt screen shot at pm src environment longhorn version installation method e g rancher catalog app helm kubectl rancher app kubernetes distro e g rke eks openshift and version
0
680,390
23,268,643,182
IssuesEvent
2022-08-04 20:07:13
intel/cve-bin-tool
https://api.github.com/repos/intel/cve-bin-tool
closed
Feature request: Filters for component view (HTML reports)
enhancement higher priority
Feature request I received by email: >In component view – it will be great to have opportunity to filter what is: New, Confirmed, Mitigated, Unexplored, Ignored.
1.0
Feature request: Filters for component view (HTML reports) - Feature request I received by email: >In component view – it will be great to have opportunity to filter what is: New, Confirmed, Mitigated, Unexplored, Ignored.
non_test
feature request filters for component view html reports feature request i received by email in component view – it will be great to have opportunity to filter what is new confirmed mitigated unexplored ignored
0
332,206
29,190,661,874
IssuesEvent
2023-05-19 19:38:25
ValveSoftware/steam-for-linux
https://api.github.com/repos/ValveSoftware/steam-for-linux
closed
Steam guard in big picture cannot enter code no imput field
Big Picture Need Retest
#### Your system information * Steam client version (build number or date): Steam beta * Distribution (e.g. Ubuntu): Gamer-os / arch * Opted into Steam client beta?: [Yes/No] Yes * Have you checked for system updates?: [Yes/No] Yes #### Please describe your issue in as much detail as possible: Describe what you _expected_ should happen and what _did_ happen. Please link any large code pastes as a [Github Gist](https://gist.github.com/) I removed my steam machine from the trusted devices so that I have to enter the steam guard code again. Unfortunately I cannot enter the code as no input field is rendered in bpm. #### Steps for reproducing this issue: 1. Remove steam machine from trusted devices 2. Next system start shows pop-up "you need to enter validation code" mail gets sent 3. After pressing "OK" there is nothing rendered no input field just blue background and select and done at the bottom
1.0
Steam guard in big picture cannot enter code no imput field - #### Your system information * Steam client version (build number or date): Steam beta * Distribution (e.g. Ubuntu): Gamer-os / arch * Opted into Steam client beta?: [Yes/No] Yes * Have you checked for system updates?: [Yes/No] Yes #### Please describe your issue in as much detail as possible: Describe what you _expected_ should happen and what _did_ happen. Please link any large code pastes as a [Github Gist](https://gist.github.com/) I removed my steam machine from the trusted devices so that I have to enter the steam guard code again. Unfortunately I cannot enter the code as no input field is rendered in bpm. #### Steps for reproducing this issue: 1. Remove steam machine from trusted devices 2. Next system start shows pop-up "you need to enter validation code" mail gets sent 3. After pressing "OK" there is nothing rendered no input field just blue background and select and done at the bottom
test
steam guard in big picture cannot enter code no imput field your system information steam client version build number or date steam beta distribution e g ubuntu gamer os arch opted into steam client beta yes have you checked for system updates yes please describe your issue in as much detail as possible describe what you expected should happen and what did happen please link any large code pastes as a i removed my steam machine from the trusted devices so that i have to enter the steam guard code again unfortunately i cannot enter the code as no input field is rendered in bpm steps for reproducing this issue remove steam machine from trusted devices next system start shows pop up you need to enter validation code mail gets sent after pressing ok there is nothing rendered no input field just blue background and select and done at the bottom
1
176,431
13,641,547,029
IssuesEvent
2020-09-25 14:18:52
tracim/tracim
https://api.github.com/repos/tracim/tracim
closed
Keep only chosen TLMs in notification panel
add to changelog frontend manually tested
## Feature description and goals For now, every TLM is displayed as a notification in Tracim. After some integration tests, our goal is to keep only the notifications that we believe will be important for the user. We didn't try filtering them sooner because we want to try this first to avoid writing unnecessary code. TLM types/categories to remove from the notification display: - [X] ~~TLMs whose author is the logged user~~ done in https://github.com/tracim/tracim/issues/3523 - [X] create i18n text for notifications that are not handled yet - [x] check that the portuguese translation works well on small screens - [X] add a .ini parameter which is the list of TLMs which won't generate a notification - [X] add this parameter to the tracim config API so that frontend has access to it - [X] choose an icon and add wording for all current notifications - [X] implement in frontend the mechanism which ignores blacklisted TLMs - [X] keep the current mechanism which shows unknown notifications as "raw notifications" - [x] Display an INFO flash message "Only an administrator can see this user's profile" will be displayed on a notification related to a user" - [x] Display a WARNING flash message "This notification does not have an associated content" will be displayed for the default "unhandled" notification if it does not have any content Default blacklist (configuration web.notifications.excluded in development.ini): - user.created - user.modified - user.deleted - user.undeleted - workspace.modified - workspace.deleted - workspace.undeleted - workspace_member.modified - content.modified ### Current TLMs that are processed notifications - workspace_member.created - with a difference between "added you to" and "added {{user}} to" - content.comment.created - content.(file/html-document/thread/folder.)created - content.(file/html-document/thread/folder.)modified - with a difference between status update and the other modification - mention.created - with a difference between mention in a document and in a comment ### Current TLMs that are notifications but not processed, i.e. don't have a specific text and icon - user.created - user.modified - user.deleted - user.undeleted - workspace.created - workspace.modified - workspace.deleted - workspace.undeleted - workspace_member.modified - workspace_member.deleted - content.(file/html-document/thread/folder.)deleted - content.(file/html-document/thread/folder.)undeleted ### Current TLMs that are not notifications - TLMs whose author is the logged user - content.comment.modified - content.comment.deleted - content.comment.undeleted ## Discussed Which notifications should we filter out? _18-09-2020_ Decision: https://github.com/tracim/tracim/issues/3477#issuecomment-694749770 ## Translations ||English|French|Portuguese| |-|-|-|-| |**user.created**| {{author}} created {{user}}'s profile | {{author}} a créé le profil de {{user}} | {{author}} criou o perfil de {{user}} | |**user.modified**| {{author}} updated {{user}}'s profile | {{author}} a mis à jour le profil de {{user}} | {{author}} actualizou o perfil de {{user}} | |**user.deleted**| {{author}} deleted {{user}}'s profile | {{author}} a supprimé le profil de {{user}} | {{author}} eliminou o perfil de {{user}} | |**user.undeleted**| {{author}} restored {{user}}'s profile | {{author}} a restauré le profil de {{user}} | {{author}} restaurou o perfil de {{user}} | |**workspace.created**| {{author}} created the space {{space}} | {{author}} a créé l'espace {{space}} | {{author}} criou o espaço {{space}} | |**workspace.modified**| {{author}} updated the space {{space}} | {{author}} a mis à jour l'espace {{space}} | {{author}} actualizou o espaço {{space}} | |**workspace.deleted**| {{author}} deleted the space {{space}} | {{author}} a supprimé l'espace {{space}} | {{author}} eliminou o espaço {{space}} | |**workspace.undeleted**| {{author}} restored the space {{space}} | {{author}} a restauré l'espace {{space}} | {{author}} restaurou o espaço {{space}} | |**workspace_member.modified**| {{author}} updated member {{member}} in {{space}} | {{author}} a mis à jour le membre {{member}} dans {{space}} | {{author}} actualizou o membro {{member}} no espaço {{space}} | |**workspace_member.deleted**| {{author}} deleted member {{member}} from {{space}} | {{author}} a supprimé le membre {{member}} de {{space}} | {{author}} eliminou o membro {{member}} de {{space}} | |**content.deleted**| {{author}} deleted {{content}} from {{space}} | {{author}} a supprimé {{content}} de {{space}} | {{author}} eliminou {{content}} de {{space}} | |**content.undeleted**| {{author}} restored {{content}} in {{space}} | {{author}} a restauré {{content}} dans {{space}} | {{author}} restaurou {{content}} no espaço {{space}} | ## Icons ||Icon| |-|-| |**user.created**|user-plus| |**user.modified**|user + history (use ComposedIcon component)| |**user.deleted**|user-times| |**user.undeleted**|user + undo (use ComposedIcon component)| |**workspace.created**|university + plus (use ComposedIcon component)| |**workspace.modified**|university + history (use ComposedIcon component)| |**workspace.deleted**|university + times (use ComposedIcon component)| |**workspace.undeleted**|university + undo (use ComposedIcon component)| |**workspace_member.modified**|user-o + history (use ComposedIcon component)| |**workspace_member.deleted**|user-o + times (use ComposedIcon component)| |**content.(file/html-document/thread/folder.)deleted**|file-o + times (use ComposedIcon component)| |**content.(file/html-document/thread/folder.)undeleted**|file-o + undo (use ComposedIcon component)| ## Redirections ||Redirect to| |-|-| |**user.created**|if logged user is admin `/ui/admin/user/{{user_id}}` else `/ui`| |**user.modified**|if logged user is admin `/ui/admin/user/{{user_id}}` else `/ui`| |**user.deleted**|if logged user is admin `/ui/admin/user/{{user_id}}` else `/ui`| |**user.undeleted**|if logged user is admin `/ui/admin/user/{{user_id}}` else `/ui`| |**workspace.created**|`/ui/workspaces/{{workspace_id}}/dashboard`| |**workspace.modified**|`/ui/workspaces/{{workspace_id}}/dashboard`| |**workspace.deleted**|`/ui`| |**workspace.undeleted**|`/ui/workspaces/{{workspace_id}}/dashboard`| |**workspace_member.modified**|`/ui/workspaces/{{workspace_id}}/dashboard`| |**workspace_member.deleted**|`/ui/workspaces/{{workspace_id}}/dashboard`| |**content.deleted**|`/ui/workspaces/{{workspace_id}}/contents/{{content_type}}/{{content_id}}`| |**content.undeleted**|`/ui/workspaces/{{workspace_id}}/contents/{{content_type}}/{{content_id}}`| ___ _2020-09-16: added the idem about text string to write_ _2020-09-21: added translations, icons and redirections_
1.0
Keep only chosen TLMs in notification panel - ## Feature description and goals For now, every TLM is displayed as a notification in Tracim. After some integration tests, our goal is to keep only the notifications that we believe will be important for the user. We didn't try filtering them sooner because we want to try this first to avoid writing unnecessary code. TLM types/categories to remove from the notification display: - [X] ~~TLMs whose author is the logged user~~ done in https://github.com/tracim/tracim/issues/3523 - [X] create i18n text for notifications that are not handled yet - [x] check that the portuguese translation works well on small screens - [X] add a .ini parameter which is the list of TLMs which won't generate a notification - [X] add this parameter to the tracim config API so that frontend has access to it - [X] choose an icon and add wording for all current notifications - [X] implement in frontend the mechanism which ignores blacklisted TLMs - [X] keep the current mechanism which shows unknown notifications as "raw notifications" - [x] Display an INFO flash message "Only an administrator can see this user's profile" will be displayed on a notification related to a user" - [x] Display a WARNING flash message "This notification does not have an associated content" will be displayed for the default "unhandled" notification if it does not have any content Default blacklist (configuration web.notifications.excluded in development.ini): - user.created - user.modified - user.deleted - user.undeleted - workspace.modified - workspace.deleted - workspace.undeleted - workspace_member.modified - content.modified ### Current TLMs that are processed notifications - workspace_member.created - with a difference between "added you to" and "added {{user}} to" - content.comment.created - content.(file/html-document/thread/folder.)created - content.(file/html-document/thread/folder.)modified - with a difference between status update and the other modification - mention.created - with a difference between mention in a document and in a comment ### Current TLMs that are notifications but not processed, i.e. don't have a specific text and icon - user.created - user.modified - user.deleted - user.undeleted - workspace.created - workspace.modified - workspace.deleted - workspace.undeleted - workspace_member.modified - workspace_member.deleted - content.(file/html-document/thread/folder.)deleted - content.(file/html-document/thread/folder.)undeleted ### Current TLMs that are not notifications - TLMs whose author is the logged user - content.comment.modified - content.comment.deleted - content.comment.undeleted ## Discussed Which notifications should we filter out? _18-09-2020_ Decision: https://github.com/tracim/tracim/issues/3477#issuecomment-694749770 ## Translations ||English|French|Portuguese| |-|-|-|-| |**user.created**| {{author}} created {{user}}'s profile | {{author}} a créé le profil de {{user}} | {{author}} criou o perfil de {{user}} | |**user.modified**| {{author}} updated {{user}}'s profile | {{author}} a mis à jour le profil de {{user}} | {{author}} actualizou o perfil de {{user}} | |**user.deleted**| {{author}} deleted {{user}}'s profile | {{author}} a supprimé le profil de {{user}} | {{author}} eliminou o perfil de {{user}} | |**user.undeleted**| {{author}} restored {{user}}'s profile | {{author}} a restauré le profil de {{user}} | {{author}} restaurou o perfil de {{user}} | |**workspace.created**| {{author}} created the space {{space}} | {{author}} a créé l'espace {{space}} | {{author}} criou o espaço {{space}} | |**workspace.modified**| {{author}} updated the space {{space}} | {{author}} a mis à jour l'espace {{space}} | {{author}} actualizou o espaço {{space}} | |**workspace.deleted**| {{author}} deleted the space {{space}} | {{author}} a supprimé l'espace {{space}} | {{author}} eliminou o espaço {{space}} | |**workspace.undeleted**| {{author}} restored the space {{space}} | {{author}} a restauré l'espace {{space}} | {{author}} restaurou o espaço {{space}} | |**workspace_member.modified**| {{author}} updated member {{member}} in {{space}} | {{author}} a mis à jour le membre {{member}} dans {{space}} | {{author}} actualizou o membro {{member}} no espaço {{space}} | |**workspace_member.deleted**| {{author}} deleted member {{member}} from {{space}} | {{author}} a supprimé le membre {{member}} de {{space}} | {{author}} eliminou o membro {{member}} de {{space}} | |**content.deleted**| {{author}} deleted {{content}} from {{space}} | {{author}} a supprimé {{content}} de {{space}} | {{author}} eliminou {{content}} de {{space}} | |**content.undeleted**| {{author}} restored {{content}} in {{space}} | {{author}} a restauré {{content}} dans {{space}} | {{author}} restaurou {{content}} no espaço {{space}} | ## Icons ||Icon| |-|-| |**user.created**|user-plus| |**user.modified**|user + history (use ComposedIcon component)| |**user.deleted**|user-times| |**user.undeleted**|user + undo (use ComposedIcon component)| |**workspace.created**|university + plus (use ComposedIcon component)| |**workspace.modified**|university + history (use ComposedIcon component)| |**workspace.deleted**|university + times (use ComposedIcon component)| |**workspace.undeleted**|university + undo (use ComposedIcon component)| |**workspace_member.modified**|user-o + history (use ComposedIcon component)| |**workspace_member.deleted**|user-o + times (use ComposedIcon component)| |**content.(file/html-document/thread/folder.)deleted**|file-o + times (use ComposedIcon component)| |**content.(file/html-document/thread/folder.)undeleted**|file-o + undo (use ComposedIcon component)| ## Redirections ||Redirect to| |-|-| |**user.created**|if logged user is admin `/ui/admin/user/{{user_id}}` else `/ui`| |**user.modified**|if logged user is admin `/ui/admin/user/{{user_id}}` else `/ui`| |**user.deleted**|if logged user is admin `/ui/admin/user/{{user_id}}` else `/ui`| |**user.undeleted**|if logged user is admin `/ui/admin/user/{{user_id}}` else `/ui`| |**workspace.created**|`/ui/workspaces/{{workspace_id}}/dashboard`| |**workspace.modified**|`/ui/workspaces/{{workspace_id}}/dashboard`| |**workspace.deleted**|`/ui`| |**workspace.undeleted**|`/ui/workspaces/{{workspace_id}}/dashboard`| |**workspace_member.modified**|`/ui/workspaces/{{workspace_id}}/dashboard`| |**workspace_member.deleted**|`/ui/workspaces/{{workspace_id}}/dashboard`| |**content.deleted**|`/ui/workspaces/{{workspace_id}}/contents/{{content_type}}/{{content_id}}`| |**content.undeleted**|`/ui/workspaces/{{workspace_id}}/contents/{{content_type}}/{{content_id}}`| ___ _2020-09-16: added the idem about text string to write_ _2020-09-21: added translations, icons and redirections_
test
keep only chosen tlms in notification panel feature description and goals for now every tlm is displayed as a notification in tracim after some integration tests our goal is to keep only the notifications that we believe will be important for the user we didn t try filtering them sooner because we want to try this first to avoid writing unnecessary code tlm types categories to remove from the notification display tlms whose author is the logged user done in create text for notifications that are not handled yet check that the portuguese translation works well on small screens add a ini parameter which is the list of tlms which won t generate a notification add this parameter to the tracim config api so that frontend has access to it choose an icon and add wording for all current notifications implement in frontend the mechanism which ignores blacklisted tlms keep the current mechanism which shows unknown notifications as raw notifications display an info flash message only an administrator can see this user s profile will be displayed on a notification related to a user display a warning flash message this notification does not have an associated content will be displayed for the default unhandled notification if it does not have any content default blacklist configuration web notifications excluded in development ini user created user modified user deleted user undeleted workspace modified workspace deleted workspace undeleted workspace member modified content modified current tlms that are processed notifications workspace member created with a difference between added you to and added user to content comment created content file html document thread folder created content file html document thread folder modified with a difference between status update and the other modification mention created with a difference between mention in a document and in a comment current tlms that are notifications but not processed i e don t have a specific text and icon user created user modified user deleted user undeleted workspace created workspace modified workspace deleted workspace undeleted workspace member modified workspace member deleted content file html document thread folder deleted content file html document thread folder undeleted current tlms that are not notifications tlms whose author is the logged user content comment modified content comment deleted content comment undeleted discussed which notifications should we filter out decision translations english french portuguese user created author created user s profile author a créé le profil de user author criou o perfil de user user modified author updated user s profile author a mis à jour le profil de user author actualizou o perfil de user user deleted author deleted user s profile author a supprimé le profil de user author eliminou o perfil de user user undeleted author restored user s profile author a restauré le profil de user author restaurou o perfil de user workspace created author created the space space author a créé l espace space author criou o espaço space workspace modified author updated the space space author a mis à jour l espace space author actualizou o espaço space workspace deleted author deleted the space space author a supprimé l espace space author eliminou o espaço space workspace undeleted author restored the space space author a restauré l espace space author restaurou o espaço space workspace member modified author updated member member in space author a mis à jour le membre member dans space author actualizou o membro member no espaço space workspace member deleted author deleted member member from space author a supprimé le membre member de space author eliminou o membro member de space content deleted author deleted content from space author a supprimé content de space author eliminou content de space content undeleted author restored content in space author a restauré content dans space author restaurou content no espaço space icons icon user created user plus user modified user history use composedicon component user deleted user times user undeleted user undo use composedicon component workspace created university plus use composedicon component workspace modified university history use composedicon component workspace deleted university times use composedicon component workspace undeleted university undo use composedicon component workspace member modified user o history use composedicon component workspace member deleted user o times use composedicon component content file html document thread folder deleted file o times use composedicon component content file html document thread folder undeleted file o undo use composedicon component redirections redirect to user created if logged user is admin ui admin user user id else ui user modified if logged user is admin ui admin user user id else ui user deleted if logged user is admin ui admin user user id else ui user undeleted if logged user is admin ui admin user user id else ui workspace created ui workspaces workspace id dashboard workspace modified ui workspaces workspace id dashboard workspace deleted ui workspace undeleted ui workspaces workspace id dashboard workspace member modified ui workspaces workspace id dashboard workspace member deleted ui workspaces workspace id dashboard content deleted ui workspaces workspace id contents content type content id content undeleted ui workspaces workspace id contents content type content id added the idem about text string to write added translations icons and redirections
1
285,491
24,670,299,789
IssuesEvent
2022-10-18 13:21:16
NationalSecurityAgency/skills-service
https://api.github.com/repos/NationalSecurityAgency/skills-service
closed
Changing a Skill from Approval Required to Honor System incorrectly reports pending requests
bug test
If a self report approval required skill is edited and changed to honor system, the edit skill dialog incorrectly reports previously approved requests as pending requests that will be automatically approved. To replicate 1. create a self report approval required skill 2. request points for that skill 3. approve the requested points (ensure that there are no other pending requests for that skill) 4. edit the skill and change from approval required to honor system ![image](https://user-images.githubusercontent.com/61209506/194347854-dd812445-1ff2-4ed5-ad39-0b797112bd92.png)
1.0
Changing a Skill from Approval Required to Honor System incorrectly reports pending requests - If a self report approval required skill is edited and changed to honor system, the edit skill dialog incorrectly reports previously approved requests as pending requests that will be automatically approved. To replicate 1. create a self report approval required skill 2. request points for that skill 3. approve the requested points (ensure that there are no other pending requests for that skill) 4. edit the skill and change from approval required to honor system ![image](https://user-images.githubusercontent.com/61209506/194347854-dd812445-1ff2-4ed5-ad39-0b797112bd92.png)
test
changing a skill from approval required to honor system incorrectly reports pending requests if a self report approval required skill is edited and changed to honor system the edit skill dialog incorrectly reports previously approved requests as pending requests that will be automatically approved to replicate create a self report approval required skill request points for that skill approve the requested points ensure that there are no other pending requests for that skill edit the skill and change from approval required to honor system
1
1,809
3,129,977,994
IssuesEvent
2015-09-09 06:27:18
SpriteStudio/SS5PlayerForUnity
https://api.github.com/repos/SpriteStudio/SS5PlayerForUnity
closed
シェーダファイルを1つにする
improvement performance
Unity4.3 から BlendFunc を動的に指定できるようになったので1つにできそうだ。 結果として、1テクスチャ:1マテリアルになる。 Drawcall 減少には寄与しないが、以下のメリットがあるだろう。 - アセットサイズの削減 - カラー指定などマテリアルに対する操作を行いたい時、1つのマテリアルだけ留意すればいい。 参考 http://answers.unity3d.com/questions/161945/is-it-possible-to-change-blend-mode-in-shader-at-r.html
True
シェーダファイルを1つにする - Unity4.3 から BlendFunc を動的に指定できるようになったので1つにできそうだ。 結果として、1テクスチャ:1マテリアルになる。 Drawcall 減少には寄与しないが、以下のメリットがあるだろう。 - アセットサイズの削減 - カラー指定などマテリアルに対する操作を行いたい時、1つのマテリアルだけ留意すればいい。 参考 http://answers.unity3d.com/questions/161945/is-it-possible-to-change-blend-mode-in-shader-at-r.html
non_test
から blendfunc 。 結果として、 : 。 drawcall 減少には寄与しないが、以下のメリットがあるだろう。 アセットサイズの削減 カラー指定などマテリアルに対する操作を行いたい時、 。 参考
0
258,316
22,302,314,362
IssuesEvent
2022-06-13 09:49:16
perlang-org/perlang
https://api.github.com/repos/perlang-org/perlang
closed
Add data-driven tests for binary operators
tests
#230 added data-driven tests for comparison operators. Similar work could be done elsewhere as well: ### Data driven tests with test data for all types currently supported in the language - [x] `Addition` and `Subtraction`: #317 - [x] `AdditionAssignment` and `SubtractionAssignment`: #319 - [x] `Exponential`: #323 - [x] `Modulo`: #324 - [x] `Multiplication` and `Division`: #318 - [x] `Equal` and `NotEqual`: #320 - [x] `ShiftLeft` and `ShiftRight`: #325 - [x] `Less` and `LessEqual`: #321 - [x] `Greater` and `GreaterEqual`: #322 ### Non-binary operators which would also be good to cover at some point - [ ] `PostfixDecrement` - [ ] `PostfixIncrement`
1.0
Add data-driven tests for binary operators - #230 added data-driven tests for comparison operators. Similar work could be done elsewhere as well: ### Data driven tests with test data for all types currently supported in the language - [x] `Addition` and `Subtraction`: #317 - [x] `AdditionAssignment` and `SubtractionAssignment`: #319 - [x] `Exponential`: #323 - [x] `Modulo`: #324 - [x] `Multiplication` and `Division`: #318 - [x] `Equal` and `NotEqual`: #320 - [x] `ShiftLeft` and `ShiftRight`: #325 - [x] `Less` and `LessEqual`: #321 - [x] `Greater` and `GreaterEqual`: #322 ### Non-binary operators which would also be good to cover at some point - [ ] `PostfixDecrement` - [ ] `PostfixIncrement`
test
add data driven tests for binary operators added data driven tests for comparison operators similar work could be done elsewhere as well data driven tests with test data for all types currently supported in the language addition and subtraction additionassignment and subtractionassignment exponential modulo multiplication and division equal and notequal shiftleft and shiftright less and lessequal greater and greaterequal non binary operators which would also be good to cover at some point postfixdecrement postfixincrement
1
438,952
12,664,052,252
IssuesEvent
2020-06-18 03:21:55
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
Add support for &, |, ^, <<, >>, >>> operators in compound-assignment statement
Component/Parser Priority/High Type/NewFeature
Add support for &, |, ^, <<, >>, >>> operators for compound-assignment-stmt := lvexpr CompoundAssignmentOperator action-or-expr ; CompoundAssignmentOperator := BinaryOperator = BinaryOperator := + | - | * | / | & | | | ^ | << | >> | >>> in incremental parser
1.0
Add support for &, |, ^, <<, >>, >>> operators in compound-assignment statement - Add support for &, |, ^, <<, >>, >>> operators for compound-assignment-stmt := lvexpr CompoundAssignmentOperator action-or-expr ; CompoundAssignmentOperator := BinaryOperator = BinaryOperator := + | - | * | / | & | | | ^ | << | >> | >>> in incremental parser
non_test
add support for operators in compound assignment statement add support for operators for compound assignment stmt lvexpr compoundassignmentoperator action or expr compoundassignmentoperator binaryoperator binaryoperator in incremental parser
0
87,530
8,093,476,264
IssuesEvent
2018-08-10 01:06:45
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: tpccbench/nodes=3/cpu=4 failed on master
C-test-failure O-robot
SHA: https://github.com/cockroachdb/cockroach/commits/bf76db84cb64dc90f65d8b2e129c75028127cda2 Parameters: Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=822738&tab=buildLog ``` test.go:494,cluster.go:1095,tpcc.go:563,tpcc.go:234: unexpected node event: 1: dead ```
1.0
roachtest: tpccbench/nodes=3/cpu=4 failed on master - SHA: https://github.com/cockroachdb/cockroach/commits/bf76db84cb64dc90f65d8b2e129c75028127cda2 Parameters: Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=822738&tab=buildLog ``` test.go:494,cluster.go:1095,tpcc.go:563,tpcc.go:234: unexpected node event: 1: dead ```
test
roachtest tpccbench nodes cpu failed on master sha parameters failed test test go cluster go tpcc go tpcc go unexpected node event dead
1
321,285
27,520,241,665
IssuesEvent
2023-03-06 14:36:33
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix tensor.test_torch_instance_sort
PyTorch Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
1.0
Fix tensor.test_torch_instance_sort - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
test
fix tensor test torch instance sort tensorflow img src torch img src numpy img src jax img src
1
604,803
18,719,223,727
IssuesEvent
2021-11-03 09:50:56
google/ExoPlayer
https://api.github.com/repos/google/ExoPlayer
closed
ENDED event is not fired by ExoPlayer if seek after pause media
bug low priority
Hi, Observed on Demo app 2.7.0 We try to seek to END after selecting pause on this media ( HLS - Apple 16x9) (This media is multi audio) https://devstreaming-cdn.apple.com/videos/streaming/examples/bipbop_16x9/bipbop_16x9_variant.m3u8 But Ended event is not fired For (HLS - ID3 Metadata) media the ended is fired for same scenario when pause and then seek to END http://devimages.apple.com/samplecode/adDemo/ad.m3u8 (HLS ID3 Metadata) What can be the reason? 10x Gilad.
1.0
ENDED event is not fired by ExoPlayer if seek after pause media - Hi, Observed on Demo app 2.7.0 We try to seek to END after selecting pause on this media ( HLS - Apple 16x9) (This media is multi audio) https://devstreaming-cdn.apple.com/videos/streaming/examples/bipbop_16x9/bipbop_16x9_variant.m3u8 But Ended event is not fired For (HLS - ID3 Metadata) media the ended is fired for same scenario when pause and then seek to END http://devimages.apple.com/samplecode/adDemo/ad.m3u8 (HLS ID3 Metadata) What can be the reason? 10x Gilad.
non_test
ended event is not fired by exoplayer if seek after pause media hi observed on demo app we try to seek to end after selecting pause on this media hls apple this media is multi audio but ended event is not fired for hls metadata media the ended is fired for same scenario when pause and then seek to end hls metadata what can be the reason gilad
0
66,457
7,001,039,136
IssuesEvent
2017-12-18 08:41:11
chamilo/chamilo-lms
https://api.github.com/repos/chamilo/chamilo-lms
closed
Error al corregir ejercicios con preguntas de combinación exacta
Requires testing/validation
### Current behavior / Resultado actual / Résultat actuel En un ejercicio con preguntas de combinación exacta, cuando el profesor accede al resultado de un alumno, las calificaciones mostradas son erróneas y aparecen preguntas como incorrectas aunque no lo sean. Tanto en la lista de intentos como cuando el alumno accede a ver su intento la puntuación se muestra bien. ### Expected behavior / Resultado esperado / Résultat attendu El profesor debería ver las preguntas con las respuestas que ha dado el alumno, y no es asi ### Steps to reproduce / Pasos para reproducir / Étapes pour reproduire -Crear un ejercicio con preguntas de combinación exacta, con un par de opciones como correctas. -Acceder como alumno, realizar el ejercicio correctamente. -Acceder como profesor a revisar el intento, aparecen preguntas incorrectas. Se puede probar aqui: https://11.chamilo.org/main/exercise/exercise_report.php?cidReq=123333&id_session=0&gidReq=0&gradebook=0&origin=&exerciseId=736 ### Chamilo Version / Versión de Chamilo / Version de Chamilo 1.11.x
1.0
Error al corregir ejercicios con preguntas de combinación exacta - ### Current behavior / Resultado actual / Résultat actuel En un ejercicio con preguntas de combinación exacta, cuando el profesor accede al resultado de un alumno, las calificaciones mostradas son erróneas y aparecen preguntas como incorrectas aunque no lo sean. Tanto en la lista de intentos como cuando el alumno accede a ver su intento la puntuación se muestra bien. ### Expected behavior / Resultado esperado / Résultat attendu El profesor debería ver las preguntas con las respuestas que ha dado el alumno, y no es asi ### Steps to reproduce / Pasos para reproducir / Étapes pour reproduire -Crear un ejercicio con preguntas de combinación exacta, con un par de opciones como correctas. -Acceder como alumno, realizar el ejercicio correctamente. -Acceder como profesor a revisar el intento, aparecen preguntas incorrectas. Se puede probar aqui: https://11.chamilo.org/main/exercise/exercise_report.php?cidReq=123333&id_session=0&gidReq=0&gradebook=0&origin=&exerciseId=736 ### Chamilo Version / Versión de Chamilo / Version de Chamilo 1.11.x
test
error al corregir ejercicios con preguntas de combinación exacta current behavior resultado actual résultat actuel en un ejercicio con preguntas de combinación exacta cuando el profesor accede al resultado de un alumno las calificaciones mostradas son erróneas y aparecen preguntas como incorrectas aunque no lo sean tanto en la lista de intentos como cuando el alumno accede a ver su intento la puntuación se muestra bien expected behavior resultado esperado résultat attendu el profesor debería ver las preguntas con las respuestas que ha dado el alumno y no es asi steps to reproduce pasos para reproducir étapes pour reproduire crear un ejercicio con preguntas de combinación exacta con un par de opciones como correctas acceder como alumno realizar el ejercicio correctamente acceder como profesor a revisar el intento aparecen preguntas incorrectas se puede probar aqui chamilo version versión de chamilo version de chamilo x
1
211,559
23,833,151,826
IssuesEvent
2022-09-06 01:08:29
RG4421/atlasdb
https://api.github.com/repos/RG4421/atlasdb
opened
CVE-2022-38749 (Medium) detected in multiple libraries
security vulnerability
## CVE-2022-38749 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>snakeyaml-1.23.jar</b>, <b>snakeyaml-1.24.jar</b>, <b>snakeyaml-1.26.jar</b></p></summary> <p> <details><summary><b>snakeyaml-1.23.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /lock-impl/build.gradle</p> <p>Path to vulnerable library: /20210226193332_TENYLC/downloadResource_RSJUCV/20210226194911/snakeyaml-1.23.jar</p> <p> Dependency Hierarchy: - :x: **snakeyaml-1.23.jar** (Vulnerable Library) </details> <details><summary><b>snakeyaml-1.24.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to vulnerable library: /timelock-api/build/conjureCompiler/lib/snakeyaml-1.24.jar</p> <p> Dependency Hierarchy: - :x: **snakeyaml-1.24.jar** (Vulnerable Library) </details> <details><summary><b>snakeyaml-1.26.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /atlasdb-service/build.gradle</p> <p>Path to vulnerable library: /timelock-api/build/conjureJava/lib/snakeyaml-1.26.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.yaml/snakeyaml/1.26/a78a8747147d2c5807683e76ec2b633e95c14fe9/snakeyaml-1.26.jar,/canner/.gradle/caches/modules-2/files-2.1/org.yaml/snakeyaml/1.26/a78a8747147d2c5807683e76ec2b633e95c14fe9/snakeyaml-1.26.jar</p> <p> Dependency Hierarchy: - :x: **snakeyaml-1.26.jar** (Vulnerable Library) </details> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow. <p>Publish Date: 2022-09-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-38749>CVE-2022-38749</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bitbucket.org/snakeyaml/snakeyaml/issues/525/got-stackoverflowerror-for-many-open">https://bitbucket.org/snakeyaml/snakeyaml/issues/525/got-stackoverflowerror-for-many-open</a></p> <p>Release Date: 2022-09-05</p> <p>Fix Resolution: org.yaml:snakeyaml:1.31</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
True
CVE-2022-38749 (Medium) detected in multiple libraries - ## CVE-2022-38749 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>snakeyaml-1.23.jar</b>, <b>snakeyaml-1.24.jar</b>, <b>snakeyaml-1.26.jar</b></p></summary> <p> <details><summary><b>snakeyaml-1.23.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /lock-impl/build.gradle</p> <p>Path to vulnerable library: /20210226193332_TENYLC/downloadResource_RSJUCV/20210226194911/snakeyaml-1.23.jar</p> <p> Dependency Hierarchy: - :x: **snakeyaml-1.23.jar** (Vulnerable Library) </details> <details><summary><b>snakeyaml-1.24.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to vulnerable library: /timelock-api/build/conjureCompiler/lib/snakeyaml-1.24.jar</p> <p> Dependency Hierarchy: - :x: **snakeyaml-1.24.jar** (Vulnerable Library) </details> <details><summary><b>snakeyaml-1.26.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /atlasdb-service/build.gradle</p> <p>Path to vulnerable library: /timelock-api/build/conjureJava/lib/snakeyaml-1.26.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.yaml/snakeyaml/1.26/a78a8747147d2c5807683e76ec2b633e95c14fe9/snakeyaml-1.26.jar,/canner/.gradle/caches/modules-2/files-2.1/org.yaml/snakeyaml/1.26/a78a8747147d2c5807683e76ec2b633e95c14fe9/snakeyaml-1.26.jar</p> <p> Dependency Hierarchy: - :x: **snakeyaml-1.26.jar** (Vulnerable Library) </details> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow. <p>Publish Date: 2022-09-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-38749>CVE-2022-38749</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bitbucket.org/snakeyaml/snakeyaml/issues/525/got-stackoverflowerror-for-many-open">https://bitbucket.org/snakeyaml/snakeyaml/issues/525/got-stackoverflowerror-for-many-open</a></p> <p>Release Date: 2022-09-05</p> <p>Fix Resolution: org.yaml:snakeyaml:1.31</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
non_test
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries snakeyaml jar snakeyaml jar snakeyaml jar snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file lock impl build gradle path to vulnerable library tenylc downloadresource rsjucv snakeyaml jar dependency hierarchy x snakeyaml jar vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to vulnerable library timelock api build conjurecompiler lib snakeyaml jar dependency hierarchy x snakeyaml jar vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file atlasdb service build gradle path to vulnerable library timelock api build conjurejava lib snakeyaml jar home wss scanner gradle caches modules files org yaml snakeyaml snakeyaml jar canner gradle caches modules files org yaml snakeyaml snakeyaml jar dependency hierarchy x snakeyaml jar vulnerable library found in base branch develop vulnerability details using snakeyaml to parse untrusted yaml files may be vulnerable to denial of service attacks dos if the parser is running on user supplied input an attacker may supply content that causes the parser to crash by stackoverflow publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org yaml snakeyaml rescue worker helmet automatic remediation is available for this issue
0
64,746
14,677,231,954
IssuesEvent
2020-12-30 22:36:06
GooseWSS/ksa
https://api.github.com/repos/GooseWSS/ksa
opened
CVE-2018-14040 (Medium) detected in bootstrap-2.1.0.js
security vulnerability
## CVE-2018-14040 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-2.1.0.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/2.1.0/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/2.1.0/bootstrap.js</a></p> <p>Path to vulnerable library: ksa/ksa-web-root/ksa-web/target/ROOT/rs/bootstrap/js/bootstrap.js</p> <p> Dependency Hierarchy: - :x: **bootstrap-2.1.0.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/GooseWSS/ksa/commit/92b4fc1a7755c8d454d53e9ae803447b86a9521a">92b4fc1a7755c8d454d53e9ae803447b86a9521a</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap before 4.1.2, XSS is possible in the collapse data-parent attribute. <p>Publish Date: 2018-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14040>CVE-2018-14040</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/twbs/bootstrap/pull/26630">https://github.com/twbs/bootstrap/pull/26630</a></p> <p>Release Date: 2018-07-13</p> <p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"twitter-bootstrap","packageVersion":"2.1.0","isTransitiveDependency":false,"dependencyTree":"twitter-bootstrap:2.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0"}],"vulnerabilityIdentifier":"CVE-2018-14040","vulnerabilityDetails":"In Bootstrap before 4.1.2, XSS is possible in the collapse data-parent attribute.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14040","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2018-14040 (Medium) detected in bootstrap-2.1.0.js - ## CVE-2018-14040 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-2.1.0.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/2.1.0/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/2.1.0/bootstrap.js</a></p> <p>Path to vulnerable library: ksa/ksa-web-root/ksa-web/target/ROOT/rs/bootstrap/js/bootstrap.js</p> <p> Dependency Hierarchy: - :x: **bootstrap-2.1.0.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/GooseWSS/ksa/commit/92b4fc1a7755c8d454d53e9ae803447b86a9521a">92b4fc1a7755c8d454d53e9ae803447b86a9521a</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap before 4.1.2, XSS is possible in the collapse data-parent attribute. <p>Publish Date: 2018-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14040>CVE-2018-14040</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/twbs/bootstrap/pull/26630">https://github.com/twbs/bootstrap/pull/26630</a></p> <p>Release Date: 2018-07-13</p> <p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"twitter-bootstrap","packageVersion":"2.1.0","isTransitiveDependency":false,"dependencyTree":"twitter-bootstrap:2.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0"}],"vulnerabilityIdentifier":"CVE-2018-14040","vulnerabilityDetails":"In Bootstrap before 4.1.2, XSS is possible in the collapse data-parent attribute.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14040","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_test
cve medium detected in bootstrap js cve medium severity vulnerability vulnerable library bootstrap js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to vulnerable library ksa ksa web root ksa web target root rs bootstrap js bootstrap js dependency hierarchy x bootstrap js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the collapse data parent attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org webjars npm bootstrap org webjars bootstrap isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails in bootstrap before xss is possible in the collapse data parent attribute vulnerabilityurl
0
23,995
10,962,192,653
IssuesEvent
2019-11-27 16:45:43
TukuToi/MinimaX1
https://api.github.com/repos/TukuToi/MinimaX1
closed
Secure theme from direct file attacks
enhancement security
This is not a direct problem but it will clog your error logs in case some smartass starts attacking your site with direct-file attacks. Let's say for example someone tries to call the 404 PHP file directly, it would fail on get_header() with a Fatal error because WordPress is not loaded yet. Hence, all that happens is an error logged, no real safety issue, but imagine a robot calling your 404 PHP a 1000 times a day, what will happen to your error log is it will grow to the GB size. That will eat up your server space and it is simply avoidable by doing `if (!defined('ABSPATH')) exit;` or/and **silence is golden** index files. This theme does not include such security measures, and I will add them with priority ASAP.
True
Secure theme from direct file attacks - This is not a direct problem but it will clog your error logs in case some smartass starts attacking your site with direct-file attacks. Let's say for example someone tries to call the 404 PHP file directly, it would fail on get_header() with a Fatal error because WordPress is not loaded yet. Hence, all that happens is an error logged, no real safety issue, but imagine a robot calling your 404 PHP a 1000 times a day, what will happen to your error log is it will grow to the GB size. That will eat up your server space and it is simply avoidable by doing `if (!defined('ABSPATH')) exit;` or/and **silence is golden** index files. This theme does not include such security measures, and I will add them with priority ASAP.
non_test
secure theme from direct file attacks this is not a direct problem but it will clog your error logs in case some smartass starts attacking your site with direct file attacks let s say for example someone tries to call the php file directly it would fail on get header with a fatal error because wordpress is not loaded yet hence all that happens is an error logged no real safety issue but imagine a robot calling your php a times a day what will happen to your error log is it will grow to the gb size that will eat up your server space and it is simply avoidable by doing if defined abspath exit or and silence is golden index files this theme does not include such security measures and i will add them with priority asap
0
64,065
6,892,085,039
IssuesEvent
2017-11-22 19:30:52
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
opened
[job failure] gce-1.8-master-upgrade-cluster
kind/bug priority/critical-urgent priority/failing-test sig/cluster-lifecycle status/approved-for-milestone
/priority critical-urgent /priority failing-test /kind bug /status approved-for-milestone @kubernetes/sig-cluster-lifecycle-test-failures This job has been failing since at least 2017-11-08. It's on the [sig-release-master-upgrade dashboard](https://k8s-testgrid.appspot.com/sig-release-master-upgrade), and prevents us from cutting [v1.9.0-beta.1] (kubernetes/sig-release#34). Is there work ongoing to bring this job back to green? https://k8s-testgrid.appspot.com/sig-release-master-upgrade#gce-1.8-master-upgrade-cluster - none of the upgrade jobs are passing https://k8s-testgrid.appspot.com/sig-release-master-upgrade - last UpgradeTest green: https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-new-master-upgrade-cluster/1593 - first UpgradeTest red: https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-new-master-upgrade-cluster/1594 - latest failure: https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-new-master-upgrade-cluster/1608 - suspect commit range: https://github.com/kubernetes/kubernetes/compare/dc4994169...c1d1b9268
1.0
[job failure] gce-1.8-master-upgrade-cluster - /priority critical-urgent /priority failing-test /kind bug /status approved-for-milestone @kubernetes/sig-cluster-lifecycle-test-failures This job has been failing since at least 2017-11-08. It's on the [sig-release-master-upgrade dashboard](https://k8s-testgrid.appspot.com/sig-release-master-upgrade), and prevents us from cutting [v1.9.0-beta.1] (kubernetes/sig-release#34). Is there work ongoing to bring this job back to green? https://k8s-testgrid.appspot.com/sig-release-master-upgrade#gce-1.8-master-upgrade-cluster - none of the upgrade jobs are passing https://k8s-testgrid.appspot.com/sig-release-master-upgrade - last UpgradeTest green: https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-new-master-upgrade-cluster/1593 - first UpgradeTest red: https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-new-master-upgrade-cluster/1594 - latest failure: https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-new-master-upgrade-cluster/1608 - suspect commit range: https://github.com/kubernetes/kubernetes/compare/dc4994169...c1d1b9268
test
gce master upgrade cluster priority critical urgent priority failing test kind bug status approved for milestone kubernetes sig cluster lifecycle test failures this job has been failing since at least it s on the and prevents us from cutting kubernetes sig release is there work ongoing to bring this job back to green none of the upgrade jobs are passing last upgradetest green first upgradetest red latest failure suspect commit range
1
283,549
21,317,824,348
IssuesEvent
2022-04-16 15:43:59
WerewolfCustoms/ModsStudio2IssueBoard
https://api.github.com/repos/WerewolfCustoms/ModsStudio2IssueBoard
closed
Update Web Site content (Tutorial Videos, Tutorial Documentation)
📚 documentation 🌍 web site
Add Game take videos to the web site :) Add old "How to create XML Template documentation" from old web site to new web site.
1.0
Update Web Site content (Tutorial Videos, Tutorial Documentation) - Add Game take videos to the web site :) Add old "How to create XML Template documentation" from old web site to new web site.
non_test
update web site content tutorial videos tutorial documentation add game take videos to the web site add old how to create xml template documentation from old web site to new web site
0
206,610
15,761,287,441
IssuesEvent
2021-03-31 09:50:07
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: lib/pq failed
C-test-failure O-roachtest O-robot branch-release-21.1 release-blocker
[(roachtest).lib/pq failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2832008&tab=buildLog) on [release-21.1@e6d07fa0b8fb13c447805709e2eadcceb783d661](https://github.com/cockroachdb/cockroach/commits/e6d07fa0b8fb13c447805709e2eadcceb783d661): ``` The test failed on branch=release-21.1, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/lib/pq/run_1 orm_helpers.go:228,orm_helpers.go:154,java_helpers.go:216,libpq.go:99,test_runner.go:767: Tests run on Cockroach v21.1.0-beta.1-248-ge6d07fa0b8 Tests run against lib/pq v1.10.0 122 Total Tests Run 112 tests passed 10 tests failed 2 tests skipped 0 tests ignored 0 tests passed unexpectedly 0 tests failed unexpectedly 0 tests expected failed but skipped 25 tests expected failed but not run --- For a full summary look at the lib/pq artifacts An updated blocklist (libPQBlocklist21_1) is available in the artifacts' lib/pq log ``` <details><summary>More</summary><p> Artifacts: [/lib/pq](https://teamcity.cockroachdb.com/viewLog.html?buildId=2832008&tab=artifacts#/lib/pq) Related: - #57855 roachtest: lib/pq failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) - #56486 roachtest: lib/pq failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.2) - #54338 roachtest: lib/pq failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.1) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Alib%2Fpq.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
2.0
roachtest: lib/pq failed - [(roachtest).lib/pq failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2832008&tab=buildLog) on [release-21.1@e6d07fa0b8fb13c447805709e2eadcceb783d661](https://github.com/cockroachdb/cockroach/commits/e6d07fa0b8fb13c447805709e2eadcceb783d661): ``` The test failed on branch=release-21.1, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/lib/pq/run_1 orm_helpers.go:228,orm_helpers.go:154,java_helpers.go:216,libpq.go:99,test_runner.go:767: Tests run on Cockroach v21.1.0-beta.1-248-ge6d07fa0b8 Tests run against lib/pq v1.10.0 122 Total Tests Run 112 tests passed 10 tests failed 2 tests skipped 0 tests ignored 0 tests passed unexpectedly 0 tests failed unexpectedly 0 tests expected failed but skipped 25 tests expected failed but not run --- For a full summary look at the lib/pq artifacts An updated blocklist (libPQBlocklist21_1) is available in the artifacts' lib/pq log ``` <details><summary>More</summary><p> Artifacts: [/lib/pq](https://teamcity.cockroachdb.com/viewLog.html?buildId=2832008&tab=artifacts#/lib/pq) Related: - #57855 roachtest: lib/pq failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) - #56486 roachtest: lib/pq failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.2) - #54338 roachtest: lib/pq failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.1) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Alib%2Fpq.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
test
roachtest lib pq failed on the test failed on branch release cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts lib pq run orm helpers go orm helpers go java helpers go libpq go test runner go tests run on cockroach beta tests run against lib pq total tests run tests passed tests failed tests skipped tests ignored tests passed unexpectedly tests failed unexpectedly tests expected failed but skipped tests expected failed but not run for a full summary look at the lib pq artifacts an updated blocklist is available in the artifacts lib pq log more artifacts related roachtest lib pq failed roachtest lib pq failed roachtest lib pq failed powered by
1
86,645
10,776,207,342
IssuesEvent
2019-11-03 19:16:49
The-Squadratics/openMRS_security_project
https://api.github.com/repos/The-Squadratics/openMRS_security_project
closed
Designing for SSE - Task 1 - Level 0 Data Flow Diagrams
Design
### Designing for SSE - Task 1 - Level 0 Data Flow Diagrams The intent of this task is the creation of the level 0 DFDs for the Use Cases created in previous stages of the project. ### Deliverables 1. Develop the level 0 DFDs for each use case in LucidChart 2. Add diagrams to the `Designing for SSE.md` document
1.0
Designing for SSE - Task 1 - Level 0 Data Flow Diagrams - ### Designing for SSE - Task 1 - Level 0 Data Flow Diagrams The intent of this task is the creation of the level 0 DFDs for the Use Cases created in previous stages of the project. ### Deliverables 1. Develop the level 0 DFDs for each use case in LucidChart 2. Add diagrams to the `Designing for SSE.md` document
non_test
designing for sse task level data flow diagrams designing for sse task level data flow diagrams the intent of this task is the creation of the level dfds for the use cases created in previous stages of the project deliverables develop the level dfds for each use case in lucidchart add diagrams to the designing for sse md document
0
421,177
12,254,825,259
IssuesEvent
2020-05-06 09:09:03
canonical-web-and-design/deployment-configs
https://api.github.com/repos/canonical-web-and-design/deployment-configs
closed
Confirm the ideal Squid configuration
Priority: Medium
When playing around with the squid installation created in https://portal.admin.canonical.com/C109653/, we discovered that it didn't cache anything with a `max-age` less than 1 minute. This is not acceptable. We should check that the Squid cache behaves as we'd expect in that it will: - cache large objects - cache objects with a `max-age` any greater than 0 - refresh the cache in the background (don't delay the response - honours `stale-if-error` and `stale-while-revalidate` - honours `private` To check this we should experiment with a local squid cache with similar rules. For this purpose, I set up https://github.com/nottrobin/squid-caching-experiment
1.0
Confirm the ideal Squid configuration - When playing around with the squid installation created in https://portal.admin.canonical.com/C109653/, we discovered that it didn't cache anything with a `max-age` less than 1 minute. This is not acceptable. We should check that the Squid cache behaves as we'd expect in that it will: - cache large objects - cache objects with a `max-age` any greater than 0 - refresh the cache in the background (don't delay the response - honours `stale-if-error` and `stale-while-revalidate` - honours `private` To check this we should experiment with a local squid cache with similar rules. For this purpose, I set up https://github.com/nottrobin/squid-caching-experiment
non_test
confirm the ideal squid configuration when playing around with the squid installation created in we discovered that it didn t cache anything with a max age less than minute this is not acceptable we should check that the squid cache behaves as we d expect in that it will cache large objects cache objects with a max age any greater than refresh the cache in the background don t delay the response honours stale if error and stale while revalidate honours private to check this we should experiment with a local squid cache with similar rules for this purpose i set up
0
179,506
13,884,557,397
IssuesEvent
2020-10-18 16:37:56
ayumi-cloud/oc2-security-module
https://api.github.com/repos/ayumi-cloud/oc2-security-module
closed
Add list of fake bots to blacklist definition files
Add to Blacklist Enhancement FINSIHED Firewall Priority: Low Testing - Passed
### Enhancement idea - [x] Add list of fake bots to blacklist definition files.
1.0
Add list of fake bots to blacklist definition files - ### Enhancement idea - [x] Add list of fake bots to blacklist definition files.
test
add list of fake bots to blacklist definition files enhancement idea add list of fake bots to blacklist definition files
1
38,217
2,842,252,199
IssuesEvent
2015-05-28 08:11:15
soi-toolkit/soi-toolkit-mule
https://api.github.com/repos/soi-toolkit/soi-toolkit-mule
closed
Introduce Catch/Rollback exception strategies for improved logging and fault handling
AffectsVersion-v0.6.0 BackwardCompatibility-MinorChange Component-tools-templates Milestone-Release0.7.0 Priority-Medium Type-Review
Original [issue 359](https://code.google.com/p/soi-toolkit/issues/detail?id=359) created by soi-toolkit on 2013-11-09T09:17:16.000Z: The current ServiceExceptionStrategy-implementation (generated at the bottom of flows): &lt;custom-exception-strategy class=&quot;org.soitoolkit.commons.mule.error.ServiceExceptionStrategy&quot;/&gt; lacks features: 1. Control over retry-handling: for which kind of exceptions should processing be retried/aborted? Note: retry-handling currently falls back on individual transports, typically using JMS-inbound with retry-parameters. 2. Access to MuleMessage when exceptions occur: needed for logging error-context like message-headers, specifically correlationId for a flow. Note: in current ServiceExceptionStrategy the MuleMessage is not available in all cases, like when a TransformerException occurs.
1.0
Introduce Catch/Rollback exception strategies for improved logging and fault handling - Original [issue 359](https://code.google.com/p/soi-toolkit/issues/detail?id=359) created by soi-toolkit on 2013-11-09T09:17:16.000Z: The current ServiceExceptionStrategy-implementation (generated at the bottom of flows): &lt;custom-exception-strategy class=&quot;org.soitoolkit.commons.mule.error.ServiceExceptionStrategy&quot;/&gt; lacks features: 1. Control over retry-handling: for which kind of exceptions should processing be retried/aborted? Note: retry-handling currently falls back on individual transports, typically using JMS-inbound with retry-parameters. 2. Access to MuleMessage when exceptions occur: needed for logging error-context like message-headers, specifically correlationId for a flow. Note: in current ServiceExceptionStrategy the MuleMessage is not available in all cases, like when a TransformerException occurs.
non_test
introduce catch rollback exception strategies for improved logging and fault handling original created by soi toolkit on the current serviceexceptionstrategy implementation generated at the bottom of flows lt custom exception strategy class quot org soitoolkit commons mule error serviceexceptionstrategy quot gt lacks features control over retry handling for which kind of exceptions should processing be retried aborted note retry handling currently falls back on individual transports typically using jms inbound with retry parameters access to mulemessage when exceptions occur needed for logging error context like message headers specifically correlationid for a flow note in current serviceexceptionstrategy the mulemessage is not available in all cases like when a transformerexception occurs
0
20,189
2,622,191,847
IssuesEvent
2015-03-04 00:23:27
byzhang/cudpp
https://api.github.com/repos/byzhang/cudpp
closed
Multiple scans to exploit fermi's concurrent kernel execution feature
auto-migrated Priority-Low Type-Other
``` Suppose I have multiple scan jobs to do. Each has different length so I can't use the cudppMultiScan API. Is it possible to add a stream interface so that I can take advantage of the concurrent kernel execution in fermi? Of course, it is only beneficial if each scan job is relatively small and can't fully use the gpu resource. Your response is highly appreciated! ``` Original issue reported on code.google.com by `zhang.yo...@gmail.com` on 30 Mar 2011 at 6:12
1.0
Multiple scans to exploit fermi's concurrent kernel execution feature - ``` Suppose I have multiple scan jobs to do. Each has different length so I can't use the cudppMultiScan API. Is it possible to add a stream interface so that I can take advantage of the concurrent kernel execution in fermi? Of course, it is only beneficial if each scan job is relatively small and can't fully use the gpu resource. Your response is highly appreciated! ``` Original issue reported on code.google.com by `zhang.yo...@gmail.com` on 30 Mar 2011 at 6:12
non_test
multiple scans to exploit fermi s concurrent kernel execution feature suppose i have multiple scan jobs to do each has different length so i can t use the cudppmultiscan api is it possible to add a stream interface so that i can take advantage of the concurrent kernel execution in fermi of course it is only beneficial if each scan job is relatively small and can t fully use the gpu resource your response is highly appreciated original issue reported on code google com by zhang yo gmail com on mar at
0
343,911
30,699,267,561
IssuesEvent
2023-07-26 21:22:39
eclipse-openj9/openj9
https://api.github.com/repos/eclipse-openj9/openj9
opened
TestJava9AttachAPI_0_FAILED testAllowAttachSelfBareEnable AssertionError: wrong exit status expected [13] but found [-1]
test failure
Failure link ------------ https://openj9-jenkins.osuosl.org/job/Test_openjdk11_j9_extended.functional_s390x_linux_Nightly/586/tapResults/(`ub20-390-5`) ``` 22:35:24 openjdk version "11.0.20-internal" 2023-07-18 22:35:24 OpenJDK Runtime Environment (build 11.0.20-internal+0-adhoc.****.BuildJDK11s390xlinuxNightly) 22:35:24 Eclipse OpenJ9 VM (build master-63a90ea2642, JRE 11 Linux s390x-64-Bit Compressed References 20230724_594 (JIT enabled, AOT enabled) 22:35:24 OpenJ9 - 63a90ea2642 22:35:24 OMR - 605ec7ca239 22:35:24 JCL - 7ef5802f8c7 based on jdk-11.0.20+8) ``` [Rerun in Grinder](https://openj9-jenkins.osuosl.org/job/Grinder/parambuild/?SDK_RESOURCE=customized&TARGET=testList+TESTLIST=TestJava9AttachAPI_0&TEST_FLAG=&UPSTREAM_TEST_JOB_NAME=&DOCKER_REQUIRED=false&ACTIVE_NODE_TIMEOUT=&VENDOR_TEST_DIRS=&EXTRA_DOCKER_ARGS=&TKG_OWNER_BRANCH=adoptium%3Amaster&OPENJ9_SYSTEMTEST_OWNER_BRANCH=eclipse%3Amaster&PLATFORM=s390x_linux&GENERATE_JOBS=true&KEEP_REPORTDIR=false&PERSONAL_BUILD=false&DOCKER_REGISTRY_DIR=&ADOPTOPENJDK_REPO=https%3A%2F%2Fgithub.com%2Fadoptium%2Faqa-tests.git&RERUN_ITERATIONS=0&SETUP_JCK_RUN=false&DOCKER_REGISTRY_URL_CREDENTIAL_ID=&LABEL=&EXTRA_OPTIONS=&BUILD_IDENTIFIER=Nightly&CUSTOMIZED_SDK_URL=https%3A%2F%2Fopenj9-artifactory.osuosl.org%2Fartifactory%2Fci-openj9%2FBuild_JDK11_s390x_linux_Nightly%2F594%2FOpenJ9-JDK11-s390x_linux-20230724-213858.tar.gz+https%3A%2F%2Fopenj9-artifactory.osuosl.org%2Fartifactory%2Fci-openj9%2FBuild_JDK11_s390x_linux_Nightly%2F594%2Ftest-images.tar.gz&JENKINS_KEY=&ADOPTOPENJDK_BRANCH=master&LIGHT_WEIGHT_CHECKOUT=true&USE_JRE=false&ARTIFACTORY_SERVER=ci-eclipse-openj9&KEEP_WORKSPACE=false&USER_CREDENTIALS_ID=&JDK_VERSION=11&DOCKER_REGISTRY_URL=&ITERATIONS=1&VENDOR_TEST_REPOS=&JDK_REPO=https%3A%2F%2Fgithub.com%2Fibmruntimes%2Fopenj9-openjdk-jdk11.git&JCK_GIT_BRANCH=master&OPENJ9_BRANCH=master&OPENJ9_SHA=63a90ea264222f664cc8880116af8e20e03ac0fb&JCK_GIT_REPO=&VENDOR_TEST_BRANCHES=&OPENJ9_REPO=https%3A%2F%2Fgithub.com%2Feclipse-openj9%2Fopenj9.git&UPSTREAM_JOB_NAME=&CLOUD_PROVIDER=&CUSTOM_TARGET=&VENDOR_TEST_SHAS=&JDK_BRANCH=openj9&LABEL_ADDITION=%21sw.os.rhel.8&ARTIFACTORY_REPO=ci-openj9&ARTIFACTORY_ROOT_DIR=Test&UPSTREAM_TEST_JOB_NUMBER=&DOCKERIMAGE_TAG=&JDK_IMPL=openj9&TEST_TIME=&SSH_AGENT_CREDENTIAL=&AUTO_DETECT=true&SLACK_CHANNEL=&DYNAMIC_COMPILE=true&RELATED_NODES=&ADOPTOPENJDK_SYSTEMTEST_OWNER_BRANCH=adoptium%3Amaster&APPLICATION_OPTIONS=&CUSTOMIZED_SDK_URL_CREDENTIAL_ID=e047f50a-26c4-4b82-8525-6da17ec68450&ARCHIVE_TEST_RESULTS=false&NUM_MACHINES=2&OPENJDK_SHA=&TRSS_URL=&USE_TESTENV_PROPERTIES=false&BUILD_LIST=functional&ADDITIONAL_ARTIFACTS_REQUIRED=&UPSTREAM_JOB_NUMBER=&STF_OWNER_BRANCH=adoptium%3Amaster&TIME_LIMIT=10&JVM_OPTIONS=&PARALLEL=Dynamic) - Change TARGET to run only the failed test targets. Optional info ------------- Failure output (captured from console output) --------------------------------------------- ``` variation: NoOptions JVM_OPTIONS: [TargetManager] [ERROR] java.io.IOException: Broken pipe at java.base/java.io.FileOutputStream.write(FileOutputStream.java:354) at java.base/java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:81) at java.base/java.io.BufferedOutputStream.flush(BufferedOutputStream.java:142) at java.base/sun.nio.cs.StreamEncoder.implFlush(StreamEncoder.java:318) at java.base/sun.nio.cs.StreamEncoder.flush(StreamEncoder.java:153) at java.base/java.io.OutputStreamWriter.flush(OutputStreamWriter.java:251) at java.base/java.io.BufferedWriter.flush(BufferedWriter.java:257) at org.openj9.test.attachAPI.TargetManager.launchTarget(TargetManager.java:325) at org.openj9.test.attachAPI.TargetManager.<init>(TargetManager.java:245) at org.openj9.test.attachAPI.TestJava9AttachAPI.testAllowAttachSelfBareEnable(TestJava9AttachAPI.java:71) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:124) at org.testng.internal.Invoker.invokeMethod(Invoker.java:580) at org.testng.internal.Invoker.invokeTestMethod(Invoker.java:716) at org.testng.internal.Invoker.invokeTestMethods(Invoker.java:988) at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:125) at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:109) at org.testng.TestRunner.privateRun(TestRunner.java:648) at org.testng.TestRunner.run(TestRunner.java:505) at org.testng.SuiteRunner.runTest(SuiteRunner.java:455) at org.testng.SuiteRunner.runSequentially(SuiteRunner.java:450) at org.testng.SuiteRunner.privateRun(SuiteRunner.java:415) at org.testng.SuiteRunner.run(SuiteRunner.java:364) at org.testng.SuiteRunnerWorker.runSuite(SuiteRunnerWorker.java:52) at org.testng.SuiteRunnerWorker.run(SuiteRunnerWorker.java:84) at org.testng.TestNG.runSuitesSequentially(TestNG.java:1208) at org.testng.TestNG.runSuitesLocally(TestNG.java:1137) at org.testng.TestNG.runSuites(TestNG.java:1049) at org.testng.TestNG.run(TestNG.java:1017) at org.testng.TestNG.privateMain(TestNG.java:1354) at org.testng.TestNG.main(TestNG.java:1323) FAILED: testAllowAttachSelfBareEnable java.lang.AssertionError: wrong exit status expected [13] but found [-1] at org.testng.Assert.fail(Assert.java:96) at org.testng.Assert.failNotEquals(Assert.java:776) at org.testng.Assert.assertEqualsImpl(Assert.java:137) at org.testng.Assert.assertEquals(Assert.java:118) at org.testng.Assert.assertEquals(Assert.java:652) at org.openj9.test.attachAPI.TestJava9AttachAPI.testAllowAttachSelfBareEnable(TestJava9AttachAPI.java:74) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:124) at org.testng.internal.Invoker.invokeMethod(Invoker.java:580) at org.testng.internal.Invoker.invokeTestMethod(Invoker.java:716) at org.testng.internal.Invoker.invokeTestMethods(Invoker.java:988) at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:125) at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:109) at org.testng.TestRunner.privateRun(TestRunner.java:648) at org.testng.TestRunner.run(TestRunner.java:505) at org.testng.SuiteRunner.runTest(SuiteRunner.java:455) at org.testng.SuiteRunner.runSequentially(SuiteRunner.java:450) at org.testng.SuiteRunner.privateRun(SuiteRunner.java:415) at org.testng.SuiteRunner.run(SuiteRunner.java:364) at org.testng.SuiteRunnerWorker.runSuite(SuiteRunnerWorker.java:52) at org.testng.SuiteRunnerWorker.run(SuiteRunnerWorker.java:84) at org.testng.TestNG.runSuitesSequentially(TestNG.java:1208) at org.testng.TestNG.runSuitesLocally(TestNG.java:1137) at org.testng.TestNG.runSuites(TestNG.java:1049) at org.testng.TestNG.run(TestNG.java:1017) at org.testng.TestNG.privateMain(TestNG.java:1354) at org.testng.TestNG.main(TestNG.java:1323) =============================================== TestJava9AttachAPI Tests run: 4, Failures: 1, Skips: 0 =============================================== =============================================== Java9andUp suite Total tests run: 4, Failures: 1, Skips: 0 =============================================== ----------------------------------- TestJava9AttachAPI_0_FAILED ``` It seems a machine/network issue.
1.0
TestJava9AttachAPI_0_FAILED testAllowAttachSelfBareEnable AssertionError: wrong exit status expected [13] but found [-1] - Failure link ------------ https://openj9-jenkins.osuosl.org/job/Test_openjdk11_j9_extended.functional_s390x_linux_Nightly/586/tapResults/(`ub20-390-5`) ``` 22:35:24 openjdk version "11.0.20-internal" 2023-07-18 22:35:24 OpenJDK Runtime Environment (build 11.0.20-internal+0-adhoc.****.BuildJDK11s390xlinuxNightly) 22:35:24 Eclipse OpenJ9 VM (build master-63a90ea2642, JRE 11 Linux s390x-64-Bit Compressed References 20230724_594 (JIT enabled, AOT enabled) 22:35:24 OpenJ9 - 63a90ea2642 22:35:24 OMR - 605ec7ca239 22:35:24 JCL - 7ef5802f8c7 based on jdk-11.0.20+8) ``` [Rerun in Grinder](https://openj9-jenkins.osuosl.org/job/Grinder/parambuild/?SDK_RESOURCE=customized&TARGET=testList+TESTLIST=TestJava9AttachAPI_0&TEST_FLAG=&UPSTREAM_TEST_JOB_NAME=&DOCKER_REQUIRED=false&ACTIVE_NODE_TIMEOUT=&VENDOR_TEST_DIRS=&EXTRA_DOCKER_ARGS=&TKG_OWNER_BRANCH=adoptium%3Amaster&OPENJ9_SYSTEMTEST_OWNER_BRANCH=eclipse%3Amaster&PLATFORM=s390x_linux&GENERATE_JOBS=true&KEEP_REPORTDIR=false&PERSONAL_BUILD=false&DOCKER_REGISTRY_DIR=&ADOPTOPENJDK_REPO=https%3A%2F%2Fgithub.com%2Fadoptium%2Faqa-tests.git&RERUN_ITERATIONS=0&SETUP_JCK_RUN=false&DOCKER_REGISTRY_URL_CREDENTIAL_ID=&LABEL=&EXTRA_OPTIONS=&BUILD_IDENTIFIER=Nightly&CUSTOMIZED_SDK_URL=https%3A%2F%2Fopenj9-artifactory.osuosl.org%2Fartifactory%2Fci-openj9%2FBuild_JDK11_s390x_linux_Nightly%2F594%2FOpenJ9-JDK11-s390x_linux-20230724-213858.tar.gz+https%3A%2F%2Fopenj9-artifactory.osuosl.org%2Fartifactory%2Fci-openj9%2FBuild_JDK11_s390x_linux_Nightly%2F594%2Ftest-images.tar.gz&JENKINS_KEY=&ADOPTOPENJDK_BRANCH=master&LIGHT_WEIGHT_CHECKOUT=true&USE_JRE=false&ARTIFACTORY_SERVER=ci-eclipse-openj9&KEEP_WORKSPACE=false&USER_CREDENTIALS_ID=&JDK_VERSION=11&DOCKER_REGISTRY_URL=&ITERATIONS=1&VENDOR_TEST_REPOS=&JDK_REPO=https%3A%2F%2Fgithub.com%2Fibmruntimes%2Fopenj9-openjdk-jdk11.git&JCK_GIT_BRANCH=master&OPENJ9_BRANCH=master&OPENJ9_SHA=63a90ea264222f664cc8880116af8e20e03ac0fb&JCK_GIT_REPO=&VENDOR_TEST_BRANCHES=&OPENJ9_REPO=https%3A%2F%2Fgithub.com%2Feclipse-openj9%2Fopenj9.git&UPSTREAM_JOB_NAME=&CLOUD_PROVIDER=&CUSTOM_TARGET=&VENDOR_TEST_SHAS=&JDK_BRANCH=openj9&LABEL_ADDITION=%21sw.os.rhel.8&ARTIFACTORY_REPO=ci-openj9&ARTIFACTORY_ROOT_DIR=Test&UPSTREAM_TEST_JOB_NUMBER=&DOCKERIMAGE_TAG=&JDK_IMPL=openj9&TEST_TIME=&SSH_AGENT_CREDENTIAL=&AUTO_DETECT=true&SLACK_CHANNEL=&DYNAMIC_COMPILE=true&RELATED_NODES=&ADOPTOPENJDK_SYSTEMTEST_OWNER_BRANCH=adoptium%3Amaster&APPLICATION_OPTIONS=&CUSTOMIZED_SDK_URL_CREDENTIAL_ID=e047f50a-26c4-4b82-8525-6da17ec68450&ARCHIVE_TEST_RESULTS=false&NUM_MACHINES=2&OPENJDK_SHA=&TRSS_URL=&USE_TESTENV_PROPERTIES=false&BUILD_LIST=functional&ADDITIONAL_ARTIFACTS_REQUIRED=&UPSTREAM_JOB_NUMBER=&STF_OWNER_BRANCH=adoptium%3Amaster&TIME_LIMIT=10&JVM_OPTIONS=&PARALLEL=Dynamic) - Change TARGET to run only the failed test targets. Optional info ------------- Failure output (captured from console output) --------------------------------------------- ``` variation: NoOptions JVM_OPTIONS: [TargetManager] [ERROR] java.io.IOException: Broken pipe at java.base/java.io.FileOutputStream.write(FileOutputStream.java:354) at java.base/java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:81) at java.base/java.io.BufferedOutputStream.flush(BufferedOutputStream.java:142) at java.base/sun.nio.cs.StreamEncoder.implFlush(StreamEncoder.java:318) at java.base/sun.nio.cs.StreamEncoder.flush(StreamEncoder.java:153) at java.base/java.io.OutputStreamWriter.flush(OutputStreamWriter.java:251) at java.base/java.io.BufferedWriter.flush(BufferedWriter.java:257) at org.openj9.test.attachAPI.TargetManager.launchTarget(TargetManager.java:325) at org.openj9.test.attachAPI.TargetManager.<init>(TargetManager.java:245) at org.openj9.test.attachAPI.TestJava9AttachAPI.testAllowAttachSelfBareEnable(TestJava9AttachAPI.java:71) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:124) at org.testng.internal.Invoker.invokeMethod(Invoker.java:580) at org.testng.internal.Invoker.invokeTestMethod(Invoker.java:716) at org.testng.internal.Invoker.invokeTestMethods(Invoker.java:988) at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:125) at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:109) at org.testng.TestRunner.privateRun(TestRunner.java:648) at org.testng.TestRunner.run(TestRunner.java:505) at org.testng.SuiteRunner.runTest(SuiteRunner.java:455) at org.testng.SuiteRunner.runSequentially(SuiteRunner.java:450) at org.testng.SuiteRunner.privateRun(SuiteRunner.java:415) at org.testng.SuiteRunner.run(SuiteRunner.java:364) at org.testng.SuiteRunnerWorker.runSuite(SuiteRunnerWorker.java:52) at org.testng.SuiteRunnerWorker.run(SuiteRunnerWorker.java:84) at org.testng.TestNG.runSuitesSequentially(TestNG.java:1208) at org.testng.TestNG.runSuitesLocally(TestNG.java:1137) at org.testng.TestNG.runSuites(TestNG.java:1049) at org.testng.TestNG.run(TestNG.java:1017) at org.testng.TestNG.privateMain(TestNG.java:1354) at org.testng.TestNG.main(TestNG.java:1323) FAILED: testAllowAttachSelfBareEnable java.lang.AssertionError: wrong exit status expected [13] but found [-1] at org.testng.Assert.fail(Assert.java:96) at org.testng.Assert.failNotEquals(Assert.java:776) at org.testng.Assert.assertEqualsImpl(Assert.java:137) at org.testng.Assert.assertEquals(Assert.java:118) at org.testng.Assert.assertEquals(Assert.java:652) at org.openj9.test.attachAPI.TestJava9AttachAPI.testAllowAttachSelfBareEnable(TestJava9AttachAPI.java:74) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:124) at org.testng.internal.Invoker.invokeMethod(Invoker.java:580) at org.testng.internal.Invoker.invokeTestMethod(Invoker.java:716) at org.testng.internal.Invoker.invokeTestMethods(Invoker.java:988) at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:125) at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:109) at org.testng.TestRunner.privateRun(TestRunner.java:648) at org.testng.TestRunner.run(TestRunner.java:505) at org.testng.SuiteRunner.runTest(SuiteRunner.java:455) at org.testng.SuiteRunner.runSequentially(SuiteRunner.java:450) at org.testng.SuiteRunner.privateRun(SuiteRunner.java:415) at org.testng.SuiteRunner.run(SuiteRunner.java:364) at org.testng.SuiteRunnerWorker.runSuite(SuiteRunnerWorker.java:52) at org.testng.SuiteRunnerWorker.run(SuiteRunnerWorker.java:84) at org.testng.TestNG.runSuitesSequentially(TestNG.java:1208) at org.testng.TestNG.runSuitesLocally(TestNG.java:1137) at org.testng.TestNG.runSuites(TestNG.java:1049) at org.testng.TestNG.run(TestNG.java:1017) at org.testng.TestNG.privateMain(TestNG.java:1354) at org.testng.TestNG.main(TestNG.java:1323) =============================================== TestJava9AttachAPI Tests run: 4, Failures: 1, Skips: 0 =============================================== =============================================== Java9andUp suite Total tests run: 4, Failures: 1, Skips: 0 =============================================== ----------------------------------- TestJava9AttachAPI_0_FAILED ``` It seems a machine/network issue.
test
failed testallowattachselfbareenable assertionerror wrong exit status expected but found failure link openjdk version internal openjdk runtime environment build internal adhoc eclipse vm build master jre linux bit compressed references jit enabled aot enabled omr jcl based on jdk change target to run only the failed test targets optional info failure output captured from console output variation nooptions jvm options java io ioexception broken pipe at java base java io fileoutputstream write fileoutputstream java at java base java io bufferedoutputstream flushbuffer bufferedoutputstream java at java base java io bufferedoutputstream flush bufferedoutputstream java at java base sun nio cs streamencoder implflush streamencoder java at java base sun nio cs streamencoder flush streamencoder java at java base java io outputstreamwriter flush outputstreamwriter java at java base java io bufferedwriter flush bufferedwriter java at org test attachapi targetmanager launchtarget targetmanager java at org test attachapi targetmanager targetmanager java at org test attachapi testallowattachselfbareenable java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org testng internal methodinvocationhelper invokemethod methodinvocationhelper java at org testng internal invoker invokemethod invoker java at org testng internal invoker invoketestmethod invoker java at org testng internal invoker invoketestmethods invoker java at org testng internal testmethodworker invoketestmethods testmethodworker java at org testng internal testmethodworker run testmethodworker java at org testng testrunner privaterun testrunner java at org testng testrunner run testrunner java at org testng suiterunner runtest suiterunner java at org testng suiterunner runsequentially suiterunner java at org testng suiterunner privaterun suiterunner java at org testng suiterunner run suiterunner java at org testng suiterunnerworker runsuite suiterunnerworker java at org testng suiterunnerworker run suiterunnerworker java at org testng testng runsuitessequentially testng java at org testng testng runsuiteslocally testng java at org testng testng runsuites testng java at org testng testng run testng java at org testng testng privatemain testng java at org testng testng main testng java failed testallowattachselfbareenable java lang assertionerror wrong exit status expected but found at org testng assert fail assert java at org testng assert failnotequals assert java at org testng assert assertequalsimpl assert java at org testng assert assertequals assert java at org testng assert assertequals assert java at org test attachapi testallowattachselfbareenable java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org testng internal methodinvocationhelper invokemethod methodinvocationhelper java at org testng internal invoker invokemethod invoker java at org testng internal invoker invoketestmethod invoker java at org testng internal invoker invoketestmethods invoker java at org testng internal testmethodworker invoketestmethods testmethodworker java at org testng internal testmethodworker run testmethodworker java at org testng testrunner privaterun testrunner java at org testng testrunner run testrunner java at org testng suiterunner runtest suiterunner java at org testng suiterunner runsequentially suiterunner java at org testng suiterunner privaterun suiterunner java at org testng suiterunner run suiterunner java at org testng suiterunnerworker runsuite suiterunnerworker java at org testng suiterunnerworker run suiterunnerworker java at org testng testng runsuitessequentially testng java at org testng testng runsuiteslocally testng java at org testng testng runsuites testng java at org testng testng run testng java at org testng testng privatemain testng java at org testng testng main testng java tests run failures skips suite total tests run failures skips failed it seems a machine network issue
1
78,111
14,949,039,558
IssuesEvent
2021-01-26 10:56:14
Genuitec/CodeTogether
https://api.github.com/repos/Genuitec/CodeTogether
closed
[VSCode][Client] Quick Open does not work
bug vscode
**Describe the bug** Quick Open does not work when a session is joined from the VSCode client. **To Reproduce** Steps to reproduce the behavior: 1. Join a session from VSCode. 2. Press CTRL+p. 3. Type file name. 4. Nothing happens. **Expected behavior** Filtered filenames are shown. **IDE Participant** - IDE: VSCode
1.0
[VSCode][Client] Quick Open does not work - **Describe the bug** Quick Open does not work when a session is joined from the VSCode client. **To Reproduce** Steps to reproduce the behavior: 1. Join a session from VSCode. 2. Press CTRL+p. 3. Type file name. 4. Nothing happens. **Expected behavior** Filtered filenames are shown. **IDE Participant** - IDE: VSCode
non_test
quick open does not work describe the bug quick open does not work when a session is joined from the vscode client to reproduce steps to reproduce the behavior join a session from vscode press ctrl p type file name nothing happens expected behavior filtered filenames are shown ide participant ide vscode
0
182,079
6,666,642,882
IssuesEvent
2017-10-03 09:09:56
PlasmaPy/PlasmaPy
https://api.github.com/repos/PlasmaPy/PlasmaPy
opened
Add bindings to functions from `physics` or `atomic` in `Plasma` and `Species` classes
Effort: high Priority: low Programming
Eventually we'd probably like to be able to define a plasma, run some simulations on it or whatever, then do `plasma.Alfven_speed()` instead of having to go through `physics.Alfven_speed(plasma.B, plasma.rho)`. This is pretty far off as `Plasma` and `Species` are still crystallizing, but I wanted to note it down for later.
1.0
Add bindings to functions from `physics` or `atomic` in `Plasma` and `Species` classes - Eventually we'd probably like to be able to define a plasma, run some simulations on it or whatever, then do `plasma.Alfven_speed()` instead of having to go through `physics.Alfven_speed(plasma.B, plasma.rho)`. This is pretty far off as `Plasma` and `Species` are still crystallizing, but I wanted to note it down for later.
non_test
add bindings to functions from physics or atomic in plasma and species classes eventually we d probably like to be able to define a plasma run some simulations on it or whatever then do plasma alfven speed instead of having to go through physics alfven speed plasma b plasma rho this is pretty far off as plasma and species are still crystallizing but i wanted to note it down for later
0
42,659
5,452,320,107
IssuesEvent
2017-03-08 02:30:03
chartjs/Chart.js
https://api.github.com/repos/chartjs/Chart.js
closed
Error when canvas is null using Chart.js in Angular 2
Category: Bug Help wanted Needs test case Priority: p2
Using TypeScript/Angular 2, I have an issue with Chart.js when the data of the chart is changing. To be more precise, Chart.js does not like the fact that I temporarily hide the canvas while data is loading. You might wonder why I do that. Well, in Angular 1/2 it is quite "normal" to use conditionals as follows: `<canvas *ngIf="chart.hasValidData" ...></canvas>` I found out that Chart.js has problems with that because Angular 2 is suddenly removing the canvas from the DOM. Here is one of the errors that I see: ``` Uncaught TypeError: Cannot set property 'width' of null at Chart.Controller.resize (eval at webpackJsonp.1314.module.exports (addScript.js:9), <anonymous>:8398:30) at eval (eval at webpackJsonp.1314.module.exports (addScript.js:9), <anonymous>:8332:24) at eval (eval at webpackJsonp.1314.module.exports (addScript.js:9), <anonymous>:10308:36) ``` The error was raised, but the chart was still working even later. I was able to mute the errors by the following workarounds: ``` resize: function(silent) { [...] if (chart.width === newWidth && chart.height === newHeight) { return; } else if (canvas === null) { // new check to prevent problems return; } ``` ``` helpers.getMaximumWidth = function(domNode) { if (domNode === null) return 0; // new check to prevent problems [...] }; ``` ``` helpers.getMaximumHeight = function(domNode) { if (domNode === null) return 0; // new check to prevent problems [...] }; ``` It would be nice if you considered including this fix in a later release.
1.0
Error when canvas is null using Chart.js in Angular 2 - Using TypeScript/Angular 2, I have an issue with Chart.js when the data of the chart is changing. To be more precise, Chart.js does not like the fact that I temporarily hide the canvas while data is loading. You might wonder why I do that. Well, in Angular 1/2 it is quite "normal" to use conditionals as follows: `<canvas *ngIf="chart.hasValidData" ...></canvas>` I found out that Chart.js has problems with that because Angular 2 is suddenly removing the canvas from the DOM. Here is one of the errors that I see: ``` Uncaught TypeError: Cannot set property 'width' of null at Chart.Controller.resize (eval at webpackJsonp.1314.module.exports (addScript.js:9), <anonymous>:8398:30) at eval (eval at webpackJsonp.1314.module.exports (addScript.js:9), <anonymous>:8332:24) at eval (eval at webpackJsonp.1314.module.exports (addScript.js:9), <anonymous>:10308:36) ``` The error was raised, but the chart was still working even later. I was able to mute the errors by the following workarounds: ``` resize: function(silent) { [...] if (chart.width === newWidth && chart.height === newHeight) { return; } else if (canvas === null) { // new check to prevent problems return; } ``` ``` helpers.getMaximumWidth = function(domNode) { if (domNode === null) return 0; // new check to prevent problems [...] }; ``` ``` helpers.getMaximumHeight = function(domNode) { if (domNode === null) return 0; // new check to prevent problems [...] }; ``` It would be nice if you considered including this fix in a later release.
test
error when canvas is null using chart js in angular using typescript angular i have an issue with chart js when the data of the chart is changing to be more precise chart js does not like the fact that i temporarily hide the canvas while data is loading you might wonder why i do that well in angular it is quite normal to use conditionals as follows i found out that chart js has problems with that because angular is suddenly removing the canvas from the dom here is one of the errors that i see uncaught typeerror cannot set property width of null at chart controller resize eval at webpackjsonp module exports addscript js at eval eval at webpackjsonp module exports addscript js at eval eval at webpackjsonp module exports addscript js the error was raised but the chart was still working even later i was able to mute the errors by the following workarounds resize function silent if chart width newwidth chart height newheight return else if canvas null new check to prevent problems return helpers getmaximumwidth function domnode if domnode null return new check to prevent problems helpers getmaximumheight function domnode if domnode null return new check to prevent problems it would be nice if you considered including this fix in a later release
1
214,460
16,591,841,290
IssuesEvent
2021-06-01 08:40:33
dzhw/zofar
https://api.github.com/repos/dzhw/zofar
closed
MDM-Export: page_number vs. question_number
status: testing
im json-Export question_number und page_number differentieren (page_number hinzufügen): 1. Spalte für page_number (Beispiel: A08 --> ohne .1) 2. Spalte für question_number (Beispiel: 1 / 2)
1.0
MDM-Export: page_number vs. question_number - im json-Export question_number und page_number differentieren (page_number hinzufügen): 1. Spalte für page_number (Beispiel: A08 --> ohne .1) 2. Spalte für question_number (Beispiel: 1 / 2)
test
mdm export page number vs question number im json export question number und page number differentieren page number hinzufügen spalte für page number beispiel ohne spalte für question number beispiel
1
267,299
23,290,959,298
IssuesEvent
2022-08-05 22:50:50
danbudris/vulnerabilityProcessor
https://api.github.com/repos/danbudris/vulnerabilityProcessor
opened
HIGH vulnerability ALAS2-2021-1710 - ca-certificates affecting 1 resources
hey there test severity/HIGH
Issue auto cut by Vulnerability Processor Processor Version: `v0.0.0-dev` Message Source: `EventBridge` Finding Source: `inspectorV2` HIGH vulnerability ALAS2-2021-1710 detected in 1 resources - arn:aws:ecr:us-west-2:338155784195:repository/test-inspector/sha256:7585bd31388fb7584260436e613c871868fd1509a728bf0c60bfe3f792e43aff Affected Packages: - ca-certificates Associated Pull Requests: - https://github.com/danbudris/vulnerabilityProcessor/pull/1243
1.0
HIGH vulnerability ALAS2-2021-1710 - ca-certificates affecting 1 resources - Issue auto cut by Vulnerability Processor Processor Version: `v0.0.0-dev` Message Source: `EventBridge` Finding Source: `inspectorV2` HIGH vulnerability ALAS2-2021-1710 detected in 1 resources - arn:aws:ecr:us-west-2:338155784195:repository/test-inspector/sha256:7585bd31388fb7584260436e613c871868fd1509a728bf0c60bfe3f792e43aff Affected Packages: - ca-certificates Associated Pull Requests: - https://github.com/danbudris/vulnerabilityProcessor/pull/1243
test
high vulnerability ca certificates affecting resources issue auto cut by vulnerability processor processor version dev message source eventbridge finding source high vulnerability detected in resources arn aws ecr us west repository test inspector affected packages ca certificates associated pull requests
1
232,072
18,843,100,675
IssuesEvent
2021-11-11 11:57:35
Roukys/HHauto
https://api.github.com/repos/Roukys/HHauto
closed
Enhancement : Function collect auto missions
enhancement to be tested
Hello to all, I think that the "Collect" function in activities is not working properly anymore. It collects the missions but not from 1:30 pm.
1.0
Enhancement : Function collect auto missions - Hello to all, I think that the "Collect" function in activities is not working properly anymore. It collects the missions but not from 1:30 pm.
test
enhancement function collect auto missions hello to all i think that the collect function in activities is not working properly anymore it collects the missions but not from pm
1
6,022
2,802,986,554
IssuesEvent
2015-05-14 00:15:54
d3athrow/vgstation13
https://api.github.com/repos/d3athrow/vgstation13
closed
Diona can't be controlled after taking monkey SE
Bug Needs Moar Testing
I evolved into a full-grown Diona, and wanted to turn back into a nymph, so I injected myself with a monkey SE. It worked, but I was unable to move, and when I tried throwing items, I dropped them instead. No idea what happened.
1.0
Diona can't be controlled after taking monkey SE - I evolved into a full-grown Diona, and wanted to turn back into a nymph, so I injected myself with a monkey SE. It worked, but I was unable to move, and when I tried throwing items, I dropped them instead. No idea what happened.
test
diona can t be controlled after taking monkey se i evolved into a full grown diona and wanted to turn back into a nymph so i injected myself with a monkey se it worked but i was unable to move and when i tried throwing items i dropped them instead no idea what happened
1
126,821
10,436,731,642
IssuesEvent
2019-09-17 20:14:45
ansible/awx
https://api.github.com/repos/ansible/awx
closed
Management Jobs shown in two columns instead of three
component:ui priority:low state:needs_test type:bug
##### ISSUE TYPE - Bug Report ##### SUMMARY <img width="1558" alt="Screen Shot 2019-08-22 at 10 26 36 AM" src="https://user-images.githubusercontent.com/9889020/63523057-727f6b80-c4c7-11e9-9852-469a677053f9.png"> ##### ENVIRONMENT * AWX version: 6.1.0 ##### STEPS TO REPRODUCE 1. Go to Management Jobs 2. Observe ##### EXPECTED RESULTS 3 columns ##### ACTUAL RESULTS 2 columns
1.0
Management Jobs shown in two columns instead of three - ##### ISSUE TYPE - Bug Report ##### SUMMARY <img width="1558" alt="Screen Shot 2019-08-22 at 10 26 36 AM" src="https://user-images.githubusercontent.com/9889020/63523057-727f6b80-c4c7-11e9-9852-469a677053f9.png"> ##### ENVIRONMENT * AWX version: 6.1.0 ##### STEPS TO REPRODUCE 1. Go to Management Jobs 2. Observe ##### EXPECTED RESULTS 3 columns ##### ACTUAL RESULTS 2 columns
test
management jobs shown in two columns instead of three issue type bug report summary img width alt screen shot at am src environment awx version steps to reproduce go to management jobs observe expected results columns actual results columns
1
114,786
9,755,129,782
IssuesEvent
2019-06-04 13:17:01
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: gossip/chaos/nodes=9 failed
C-test-failure O-roachtest O-robot
SHA: https://github.com/cockroachdb/cockroach/commits/99306ec3e9fcbba01c05431cbf496e8b5b8954b4 Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=gossip/chaos/nodes=9 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1260033&tab=buildLog ``` The test failed on master: cluster.go:1349,gossip.go:117,gossip.go:125,test.go:1245: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod start teamcity-1260033-gossip-chaos-nodes-9:9 returned: stderr: stdout: teamcity-1260033-gossip-chaos-nodes-9: starting........................................................................................................................ 0: exit status 255 ~ ./cockroach version github.com/cockroachdb/cockroach/pkg/cmd/roachprod/install.getCockroachVersion /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/install/cockroach.go:95 github.com/cockroachdb/cockroach/pkg/cmd/roachprod/install.Cockroach.Start.func7 /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/install/cockroach.go:289 github.com/cockroachdb/cockroach/pkg/cmd/roachprod/install.(*SyncedCluster).Parallel.func1.1 /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/install/cluster_synced.go:1441 runtime.goexit /usr/local/go/src/runtime/asm_amd64.s:1333: I190425 07:42:47.437015 1 cluster_synced.go:1523 command failed : exit status 1 cluster.go:1016,context.go:89,cluster.go:1005,asm_amd64.s:522,panic.go:397,test.go:790,test.go:796,cluster.go:1349,gossip.go:117,gossip.go:125,test.go:1245: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-1260033-gossip-chaos-nodes-9 --oneshot --ignore-empty-nodes: exit status 1 1: 5186 5: 5066 2: 5435 6: 5048 7: 6837 3: 4717 4: 4759 9: dead 8: 5048 Error: 9: dead ```
2.0
roachtest: gossip/chaos/nodes=9 failed - SHA: https://github.com/cockroachdb/cockroach/commits/99306ec3e9fcbba01c05431cbf496e8b5b8954b4 Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=gossip/chaos/nodes=9 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1260033&tab=buildLog ``` The test failed on master: cluster.go:1349,gossip.go:117,gossip.go:125,test.go:1245: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod start teamcity-1260033-gossip-chaos-nodes-9:9 returned: stderr: stdout: teamcity-1260033-gossip-chaos-nodes-9: starting........................................................................................................................ 0: exit status 255 ~ ./cockroach version github.com/cockroachdb/cockroach/pkg/cmd/roachprod/install.getCockroachVersion /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/install/cockroach.go:95 github.com/cockroachdb/cockroach/pkg/cmd/roachprod/install.Cockroach.Start.func7 /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/install/cockroach.go:289 github.com/cockroachdb/cockroach/pkg/cmd/roachprod/install.(*SyncedCluster).Parallel.func1.1 /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/install/cluster_synced.go:1441 runtime.goexit /usr/local/go/src/runtime/asm_amd64.s:1333: I190425 07:42:47.437015 1 cluster_synced.go:1523 command failed : exit status 1 cluster.go:1016,context.go:89,cluster.go:1005,asm_amd64.s:522,panic.go:397,test.go:790,test.go:796,cluster.go:1349,gossip.go:117,gossip.go:125,test.go:1245: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-1260033-gossip-chaos-nodes-9 --oneshot --ignore-empty-nodes: exit status 1 1: 5186 5: 5066 2: 5435 6: 5048 7: 6837 3: 4717 4: 4759 9: dead 8: 5048 Error: 9: dead ```
test
roachtest gossip chaos nodes failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach stdbuf ol el make stressrace tests gossip chaos nodes pkg roachtest testtimeout stressflags maxtime timeout tee tmp stress log failed test the test failed on master cluster go gossip go gossip go test go home agent work go src github com cockroachdb cockroach bin roachprod start teamcity gossip chaos nodes returned stderr stdout teamcity gossip chaos nodes starting exit status cockroach version github com cockroachdb cockroach pkg cmd roachprod install getcockroachversion home agent work go src github com cockroachdb cockroach pkg cmd roachprod install cockroach go github com cockroachdb cockroach pkg cmd roachprod install cockroach start home agent work go src github com cockroachdb cockroach pkg cmd roachprod install cockroach go github com cockroachdb cockroach pkg cmd roachprod install syncedcluster parallel home agent work go src github com cockroachdb cockroach pkg cmd roachprod install cluster synced go runtime goexit usr local go src runtime asm s cluster synced go command failed exit status cluster go context go cluster go asm s panic go test go test go cluster go gossip go gossip go test go dead node detection home agent work go src github com cockroachdb cockroach bin roachprod monitor teamcity gossip chaos nodes oneshot ignore empty nodes exit status dead error dead
1
88,087
15,800,729,521
IssuesEvent
2021-04-03 01:01:32
jgeraigery/hadoop
https://api.github.com/repos/jgeraigery/hadoop
opened
WS-2019-0509 (Medium) detected in bcprov-jdk15on-1.60.jar
security vulnerability
## WS-2019-0509 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bcprov-jdk15on-1.60.jar</b></p></summary> <p>The Bouncy Castle Crypto package is a Java implementation of cryptographic algorithms. This jar contains JCE provider and lightweight API for the Bouncy Castle Cryptography APIs for JDK 1.5 to JDK 1.8.</p> <p>Library home page: <a href="http://www.bouncycastle.org/java.html">http://www.bouncycastle.org/java.html</a></p> <p>Path to dependency file: hadoop/hadoop-minicluster/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.60/bcprov-jdk15on-1.60.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.60/bcprov-jdk15on-1.60.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.60/bcprov-jdk15on-1.60.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.60/bcprov-jdk15on-1.60.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.60/bcprov-jdk15on-1.60.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.60/bcprov-jdk15on-1.60.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.60/bcprov-jdk15on-1.60.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.60/bcprov-jdk15on-1.60.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.60/bcprov-jdk15on-1.60.jar,hadoop/hadoop-tools/hadoop-aws/target/lib/bcprov-jdk15on-1.60.jar</p> <p> Dependency Hierarchy: - :x: **bcprov-jdk15on-1.60.jar** (Vulnerable Library) <p>Found in base branch: <b>trunk</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Bouncy Castle through 1.68 is vulnerable to Denial of Service (DoS). The Dump.class file utilizes the FileInputStream object when reading from user-provided files, but doesn't close these streams properly. <p>Publish Date: 2019-12-24 <p>URL: <a href=https://github.com/bcgit/bc-java/issues/634>WS-2019-0509</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.bouncycastle","packageName":"bcprov-jdk15on","packageVersion":"1.60","packageFilePaths":["/hadoop-minicluster/pom.xml","/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/pom.xml","/hadoop-tools/hadoop-dynamometer/hadoop-dynamometer-dist/pom.xml","/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-distributedshell/pom.xml","/hadoop-tools/hadoop-resourceestimator/pom.xml","/hadoop-tools/hadoop-tools-dist/pom.xml","/hadoop-tools/hadoop-archive-logs/pom.xml","/hadoop-client-modules/hadoop-client-minicluster/pom.xml","/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.bouncycastle:bcprov-jdk15on:1.60","isMinimumFixVersionAvailable":false}],"baseBranches":["trunk"],"vulnerabilityIdentifier":"WS-2019-0509","vulnerabilityDetails":"Bouncy Castle through 1.68 is vulnerable to Denial of Service (DoS). The Dump.class file utilizes the FileInputStream object when reading from user-provided files, but doesn\u0027t close these streams properly.","vulnerabilityUrl":"https://github.com/bcgit/bc-java/issues/634","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
WS-2019-0509 (Medium) detected in bcprov-jdk15on-1.60.jar - ## WS-2019-0509 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bcprov-jdk15on-1.60.jar</b></p></summary> <p>The Bouncy Castle Crypto package is a Java implementation of cryptographic algorithms. This jar contains JCE provider and lightweight API for the Bouncy Castle Cryptography APIs for JDK 1.5 to JDK 1.8.</p> <p>Library home page: <a href="http://www.bouncycastle.org/java.html">http://www.bouncycastle.org/java.html</a></p> <p>Path to dependency file: hadoop/hadoop-minicluster/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.60/bcprov-jdk15on-1.60.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.60/bcprov-jdk15on-1.60.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.60/bcprov-jdk15on-1.60.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.60/bcprov-jdk15on-1.60.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.60/bcprov-jdk15on-1.60.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.60/bcprov-jdk15on-1.60.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.60/bcprov-jdk15on-1.60.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.60/bcprov-jdk15on-1.60.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.60/bcprov-jdk15on-1.60.jar,hadoop/hadoop-tools/hadoop-aws/target/lib/bcprov-jdk15on-1.60.jar</p> <p> Dependency Hierarchy: - :x: **bcprov-jdk15on-1.60.jar** (Vulnerable Library) <p>Found in base branch: <b>trunk</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Bouncy Castle through 1.68 is vulnerable to Denial of Service (DoS). The Dump.class file utilizes the FileInputStream object when reading from user-provided files, but doesn't close these streams properly. <p>Publish Date: 2019-12-24 <p>URL: <a href=https://github.com/bcgit/bc-java/issues/634>WS-2019-0509</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.bouncycastle","packageName":"bcprov-jdk15on","packageVersion":"1.60","packageFilePaths":["/hadoop-minicluster/pom.xml","/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/pom.xml","/hadoop-tools/hadoop-dynamometer/hadoop-dynamometer-dist/pom.xml","/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-distributedshell/pom.xml","/hadoop-tools/hadoop-resourceestimator/pom.xml","/hadoop-tools/hadoop-tools-dist/pom.xml","/hadoop-tools/hadoop-archive-logs/pom.xml","/hadoop-client-modules/hadoop-client-minicluster/pom.xml","/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.bouncycastle:bcprov-jdk15on:1.60","isMinimumFixVersionAvailable":false}],"baseBranches":["trunk"],"vulnerabilityIdentifier":"WS-2019-0509","vulnerabilityDetails":"Bouncy Castle through 1.68 is vulnerable to Denial of Service (DoS). The Dump.class file utilizes the FileInputStream object when reading from user-provided files, but doesn\u0027t close these streams properly.","vulnerabilityUrl":"https://github.com/bcgit/bc-java/issues/634","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_test
ws medium detected in bcprov jar ws medium severity vulnerability vulnerable library bcprov jar the bouncy castle crypto package is a java implementation of cryptographic algorithms this jar contains jce provider and lightweight api for the bouncy castle cryptography apis for jdk to jdk library home page a href path to dependency file hadoop hadoop minicluster pom xml path to vulnerable library home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar hadoop hadoop tools hadoop aws target lib bcprov jar dependency hierarchy x bcprov jar vulnerable library found in base branch trunk vulnerability details bouncy castle through is vulnerable to denial of service dos the dump class file utilizes the fileinputstream object when reading from user provided files but doesn t close these streams properly publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree org bouncycastle bcprov isminimumfixversionavailable false basebranches vulnerabilityidentifier ws vulnerabilitydetails bouncy castle through is vulnerable to denial of service dos the dump class file utilizes the fileinputstream object when reading from user provided files but doesn close these streams properly vulnerabilityurl
0
70,816
23,328,056,378
IssuesEvent
2022-08-09 00:10:31
primefaces/primeng
https://api.github.com/repos/primefaces/primeng
opened
Component: File upload
defect
### Describe the bug if you choose files more than fileLimit you see an alert message, but when you remove a file and the fileLimit is correct now, the alert message stays.It should go but it is not.Can someone look into it? ### Environment primeng ### Reproducer _No response_ ### Angular version 13.1.0 ### PrimeNG version 13.1.0 ### Build / Runtime Angular CLI App ### Language TypeScript ### Node version (for AoT issues node --version) 14.18.1 ### Browser(s) _No response_ ### Steps to reproduce the behavior 1. use advanced p-fileupload. Use following in component.html file.Set fileLimit to 3 <p-fileUpload #fileUpload [fileLimit]="3" class="center" name="demoA" (onUpload)="onUpload($event)" [multiple]="true" accept="image/*" [maxFileSize]="1000000"> </p-fileUpload> 2.select 4 files.You will see alert message. Maximum number of files exceeded,limit is 3 at most. 3.Now remove any one file by clicking on X button next to it. 4. Alert message should go of because now the no of selected file is 3.However alert message stays. ### Expected behavior Alert message should go off
1.0
Component: File upload - ### Describe the bug if you choose files more than fileLimit you see an alert message, but when you remove a file and the fileLimit is correct now, the alert message stays.It should go but it is not.Can someone look into it? ### Environment primeng ### Reproducer _No response_ ### Angular version 13.1.0 ### PrimeNG version 13.1.0 ### Build / Runtime Angular CLI App ### Language TypeScript ### Node version (for AoT issues node --version) 14.18.1 ### Browser(s) _No response_ ### Steps to reproduce the behavior 1. use advanced p-fileupload. Use following in component.html file.Set fileLimit to 3 <p-fileUpload #fileUpload [fileLimit]="3" class="center" name="demoA" (onUpload)="onUpload($event)" [multiple]="true" accept="image/*" [maxFileSize]="1000000"> </p-fileUpload> 2.select 4 files.You will see alert message. Maximum number of files exceeded,limit is 3 at most. 3.Now remove any one file by clicking on X button next to it. 4. Alert message should go of because now the no of selected file is 3.However alert message stays. ### Expected behavior Alert message should go off
non_test
component file upload describe the bug if you choose files more than filelimit you see an alert message but when you remove a file and the filelimit is correct now the alert message stays it should go but it is not can someone look into it environment primeng reproducer no response angular version primeng version build runtime angular cli app language typescript node version for aot issues node version browser s no response steps to reproduce the behavior use advanced p fileupload use following in component html file set filelimit to p fileupload fileupload class center name demoa onupload onupload event true accept image select files you will see alert message maximum number of files exceeded limit is at most now remove any one file by clicking on x button next to it alert message should go of because now the no of selected file is however alert message stays expected behavior alert message should go off
0
70,869
13,541,861,259
IssuesEvent
2020-09-16 16:27:35
concrete-utopia/utopia
https://api.github.com/repos/concrete-utopia/utopia
closed
CSS importing is tied to eager evaluation
Code / Parsing
Follow on from #142. Blocked by #249. Right now CSS files imported by other files in the project will be slipped into the DOM due to eager evaluation. As well as this, removing those imports will then _not_ unload that CSS. We need to either not eagerly load this CSS files transitively, or clear them from the DOM when opening a canvas file. On top of this, the _only_ reason we are even loading CSS files that are imported in the canvas file is because of the `getExportValuesFromAllModules` evaluating the currently open file after a change is made to it.
1.0
CSS importing is tied to eager evaluation - Follow on from #142. Blocked by #249. Right now CSS files imported by other files in the project will be slipped into the DOM due to eager evaluation. As well as this, removing those imports will then _not_ unload that CSS. We need to either not eagerly load this CSS files transitively, or clear them from the DOM when opening a canvas file. On top of this, the _only_ reason we are even loading CSS files that are imported in the canvas file is because of the `getExportValuesFromAllModules` evaluating the currently open file after a change is made to it.
non_test
css importing is tied to eager evaluation follow on from blocked by right now css files imported by other files in the project will be slipped into the dom due to eager evaluation as well as this removing those imports will then not unload that css we need to either not eagerly load this css files transitively or clear them from the dom when opening a canvas file on top of this the only reason we are even loading css files that are imported in the canvas file is because of the getexportvaluesfromallmodules evaluating the currently open file after a change is made to it
0
26,957
4,265,609,098
IssuesEvent
2016-07-12 11:50:40
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
stress: failed test in cockroach/storage/storage.test: TestStoreRangeSplitStats
Robot test-failure
Binary: cockroach/static-tests.tar.gz sha: https://github.com/cockroachdb/cockroach/commits/2cc5a467f6c280e97e49a22b884b649fd7790273 Stress build found a failed test: ``` === RUN TestStoreRangeSplitStats I160517 07:24:34.574483 storage/engine/rocksdb.go:137 opening in memory rocksdb instance W160517 07:24:34.575211 gossip/gossip.go:893 not connected to cluster; use --join to specify a connected node I160517 07:24:34.577161 storage/replica_command.go:1470 range 1: new leader lease replica {1 1 1} 1970-01-01 00:00:00 +0000 UTC 1.000000123s following replica {0 0 0} 1970-01-01 00:00:00 +0000 UTC 0 [physicalTime=1970-01-01 00:00:00.000000123 +0000 UTC] I160517 07:24:34.578130 storage/replica_command.go:2011 initiating a split of range=1 [/Min-/Max) at key /Table/50 I160517 07:24:34.580548 storage/replica_command.go:1470 range 2: new leader lease replica {1 1 1} 1970-01-01 00:00:00 +0000 UTC 1.000000123s following replica {0 0 0} 1970-01-01 00:00:00 +0000 UTC 0 [physicalTime=1970-01-01 00:00:00.000000123 +0000 UTC] I160517 07:24:34.605112 storage/replica_command.go:2011 initiating a split of range=2 [/Table/50-/Max) at key /Table/50/0/"Z" W160517 07:24:34.605402 storage/stores.go:200 range not contained in one range: [/Meta2/Table/50/0/"Z",/Table/50/0/"Z\x00"), but have [/Min,/Table/50) I160517 07:24:34.608896 stopper.go:352 draining; tasks left: 1 storage/intent_resolver.go:338 1 storage/intent_resolver.go:299 I160517 07:24:34.617668 stopper.go:352 draining; tasks left: 1 storage/intent_resolver.go:299 W160517 07:25:34.577609 storage/store.go:995 [node=1,store=1] error gossiping first range data: storage/replica.go:2126: node is stopping W160517 07:25:34.577663 storage/store.go:1014 [node=1,store=1] error gossiping system config: storage/replica.go:2126: node is stopping W160517 07:26:34.577591 storage/store.go:995 [node=1,store=1] error gossiping first range data: storage/replica.go:2126: node is stopping W160517 07:26:34.577636 storage/store.go:1014 [node=1,store=1] error gossiping system config: storage/replica.go:2126: node is stopping W160517 07:27:34.577608 storage/store.go:995 [node=1,store=1] error gossiping first range data: storage/replica.go:2126: node is stopping W160517 07:27:34.577663 storage/store.go:1014 [node=1,store=1] error gossiping system config: storage/replica.go:2126: node is stopping W160517 07:28:34.577613 storage/store.go:995 [node=1,store=1] error gossiping first range data: storage/replica.go:2126: node is stopping W160517 07:28:34.577664 storage/store.go:1014 [node=1,store=1] error gossiping system config: storage/replica.go:2126: node is stopping W160517 07:29:34.577718 storage/store.go:995 [node=1,store=1] error gossiping first range data: storage/replica.go:2126: node is stopping W160517 07:29:34.577810 storage/store.go:1014 [node=1,store=1] error gossiping system config: storage/replica.go:2126: node is stopping W160517 07:30:34.577627 storage/store.go:995 [node=1,store=1] error gossiping first range data: storage/replica.go:2126: node is stopping W160517 07:30:34.577646 storage/store.go:1014 [node=1,store=1] error gossiping system config: storage/replica.go:2126: node is stopping W160517 07:31:34.577610 storage/store.go:995 [node=1,store=1] error gossiping first range data: storage/replica.go:2126: node is stopping W160517 07:31:34.577756 storage/store.go:1014 [node=1,store=1] error gossiping system config: storage/replica.go:2126: node is stopping W160517 07:32:34.577595 storage/store.go:995 [node=1,store=1] error gossiping first range data: storage/replica.go:2126: node is stopping W160517 07:32:34.577779 storage/store.go:1014 [node=1,store=1] error gossiping system config: storage/replica.go:2126: node is stopping SIGABRT: abort PC=0x62de21 m=0 goroutine 0 [idle]: runtime.futex(0x2801188, 0x0, 0x0, 0x0, 0x0, 0x2800010, 0x0, 0x0, 0x5dc6f4, 0x2801188, ...) /usr/local/go/src/runtime/sys_linux_amd64.s:302 +0x21 runtime.futexsleep(0x2801188, 0x0, 0xffffffffffffffff) /usr/local/go/src/runtime/os1_linux.go:40 +0x53 runtime.notesleep(0x2801188) /usr/local/go/src/runtime/lock_futex.go:145 +0xa4 runtime.stopm() /usr/local/go/src/runtime/proc.go:1535 +0x10b runtime.findrunnable(0xc82001aa00, 0x0) /usr/local/go/src/runtime/proc.go:1973 +0x739 runtime.schedule() /usr/local/go/src/runtime/proc.go:2072 +0x24f runtime.park_m(0xc82006ea80) /usr/local/go/src/runtime/proc.go:2137 +0x18b runtime.mcall(0x7ffe96854c10) /usr/local/go/src/runtime/asm_amd64.s:233 +0x5b goroutine 1 [chan receive, 8 minutes]: testing.RunTests(0x1d7d610, 0x25283a0, 0x10f, 0x10f, 0xc82012bd01) /usr/local/go/src/testing/testing.go:583 +0x8d2 testing.(*M).Run(0xc8200efef8, 0xc820307c98) /usr/local/go/src/testing/testing.go:515 +0x81 github.com/cockroachdb/cockroach/storage_test.TestMain(0xc8200efef8) /go/src/github.com/cockroachdb/cockroach/storage/main_test.go:54 +0x1de main.main() github.com/cockroachdb/cockroach/storage/_test/_testmain.go:604 +0x114 goroutine 17 [syscall, 9 minutes, locked to thread]: runtime.goexit() /usr/local/go/src/runtime/asm_amd64.s:1998 +0x1 goroutine 35 [chan receive]: github.com/cockroachdb/cockroach/util/log.(*loggingT).flushDaemon(0x27fff20) /go/src/github.com/cockroachdb/cockroach/util/log/clog.go:1011 +0x64 created by github.com/cockroachdb/cockroach/util/log.init.1 /go/src/github.com/cockroachdb/cockroach/util/log/clog.go:598 +0x8a goroutine 15663 [semacquire, 8 minutes]: sync.runtime_Syncsemacquire(0xc82113e5d0) /usr/local/go/src/runtime/sema.go:241 +0x201 sync.(*Cond).Wait(0xc82113e5c0) /usr/local/go/src/sync/cond.go:63 +0x9b github.com/cockroachdb/cockroach/util/stop.(*Stopper).Quiesce(0xc82144e4d0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:354 +0x1eb github.com/cockroachdb/cockroach/util/stop.(*Stopper).Stop(0xc82144e4d0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:300 +0x1a7 github.com/cockroachdb/cockroach/storage_test.TestStoreRangeSplitStats(0xc8212c0750) /go/src/github.com/cockroachdb/cockroach/storage/client_split_test.go:494 +0x1e08 testing.tRunner(0xc8212c0750, 0x2529bd0) /usr/local/go/src/testing/testing.go:473 +0x98 created by testing.RunTests /usr/local/go/src/testing/testing.go:582 +0x892 goroutine 15735 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*Store).Start.func4() /go/src/github.com/cockroachdb/cockroach/storage/store.go:921 +0x13d github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc82112b340) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15723 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc8210a4c80) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15722 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*replicaScanner).waitAndProcess(0xc82144fc70, 0xececcbe32, 0x2262c243, 0x27ff340, 0xc82113e600, 0xc82144e4d0, 0xc820898f00, 0xc820545e98) /go/src/github.com/cockroachdb/cockroach/storage/scanner.go:161 +0x568 github.com/cockroachdb/cockroach/storage.(*replicaScanner).scanLoop.func1.1(0xc820898f00, 0xc821b86930) /go/src/github.com/cockroachdb/cockroach/storage/scanner.go:203 +0x78 github.com/cockroachdb/cockroach/storage.(*storeRangeSet).Visit(0xc821b86cc0, 0xc8215aca00) /go/src/github.com/cockroachdb/cockroach/storage/store.go:231 +0x246 github.com/cockroachdb/cockroach/storage.(*replicaScanner).scanLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/scanner.go:205 +0x21e github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc8210a4c40) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15724 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc8210a4ca0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15733 [select]: github.com/cockroachdb/cockroach/storage.(*Store).processRaft.func1() /go/src/github.com/cockroachdb/cockroach/storage/store.go:2059 +0xbf1 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc8203639f0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15728 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc8210a4d20) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15737 [select]: github.com/cockroachdb/cockroach/storage.(*Store).startGossip.func2() /go/src/github.com/cockroachdb/cockroach/storage/store.go:1011 +0x3cf github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc82112b380) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15679 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*idAllocator).start.func1() /go/src/github.com/cockroachdb/cockroach/storage/id_alloc.go:123 +0x86a github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc821b68050) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15732 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*StorePool).start.func1() /go/src/github.com/cockroachdb/cockroach/storage/store_pool.go:225 +0x23f github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc82112a460) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15736 [select]: github.com/cockroachdb/cockroach/storage.(*Store).startGossip.func1() /go/src/github.com/cockroachdb/cockroach/storage/store.go:992 +0x3cf github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc82112b360) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15727 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc8210a4d00) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15746 [runnable]: time.absDate(0x7ffffffe7214923f, 0x2698f01, 0x27ff340, 0x7ffffffe7214923f, 0x11, 0x89) /usr/local/go/src/time/time.go:678 time.Time.date(0xececcc03f, 0xc802698f8b, 0x27ff340, 0x8d4901, 0x7e0, 0x5, 0x11, 0x89) /usr/local/go/src/time/time.go:674 +0x4a time.Time.Day(0xececcc03f, 0x2698f8b, 0x27ff340, 0x11) /usr/local/go/src/time/time.go:309 +0x38 golang.org/x/net/trace.(*trace).delta(0xc8202763c0, 0xececcc03f, 0x2698f8b, 0x27ff340, 0x1d779dc, 0x1d779dd) /go/src/golang.org/x/net/trace/trace.go:729 +0x15e golang.org/x/net/trace.(*trace).addEvent(0xc8202763c0, 0x17c7da0, 0xc821b192c0, 0x15e0000) /go/src/golang.org/x/net/trace/trace.go:755 +0x308 golang.org/x/net/trace.(*trace).LazyPrintf(0xc8202763c0, 0x19bd558, 0x2, 0xc8219ca010, 0x1, 0x1) /go/src/golang.org/x/net/trace/trace.go:790 +0xa2 github.com/opentracing/basictracer-go.glob.func2.1(0x1862220, 0xc8218baa80) /go/src/github.com/opentracing/basictracer-go/event_nettrace.go:19 +0x343 github.com/opentracing/basictracer-go.(*spanImpl).onLog(0xc8214a6360, 0x0, 0x0, 0x0, 0xc820abd7a0, 0x1e, 0x0, 0x0) /go/src/github.com/opentracing/basictracer-go/event.go:40 +0xc4 github.com/opentracing/basictracer-go.(*spanImpl).Log(0xc8214a6360, 0x0, 0x0, 0x0, 0xc820abd7a0, 0x1e, 0x0, 0x0) /go/src/github.com/opentracing/basictracer-go/span.go:109 +0x114 github.com/opentracing/basictracer-go.(*spanImpl).LogEvent(0xc8214a6360, 0xc820abd7a0, 0x1e) /go/src/github.com/opentracing/basictracer-go/span.go:94 +0x7d github.com/cockroachdb/cockroach/util/log.Trace(0x7f4c1bb69938, 0xc821b19140, 0xc820abd7a0, 0x1e) /go/src/github.com/cockroachdb/cockroach/util/log/trace.go:29 +0x63 github.com/cockroachdb/cockroach/storage.(*Store).Send(0xc820e1af00, 0x7f4c1bb69938, 0xc821b19140, 0xdf, 0x209fce6, 0x100000001, 0x1, 0x3, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/storage/store.go:1824 +0x10e6 github.com/cockroachdb/cockroach/storage.(*Stores).Send(0xc82113e740, 0x7f4c1bb69938, 0xc821b190e0, 0x0, 0x0, 0x100000001, 0x1, 0x3, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/storage/stores.go:178 +0x501 github.com/cockroachdb/cockroach/kv.(*senderTransport).SendNext(0xc8202dc1b0, 0xc8219320c0) /go/src/github.com/cockroachdb/cockroach/kv/transport.go:290 +0x3cc github.com/cockroachdb/cockroach/kv.(*DistSender).sendToReplicas(0xc8212c0990, 0x7f4c1bb69938, 0xc82166e030, 0x1, 0x2540be400, 0xb2d05e00, 0xc821b864b0, 0x2, 0xc8210680e0, 0x1, ...) /go/src/github.com/cockroachdb/cockroach/kv/dist_sender.go:1007 +0x8eb github.com/cockroachdb/cockroach/kv.(*DistSender).sendRPC(0xc8212c0990, 0x7f4c1bb69938, 0xc82166e030, 0x2, 0xc8210680e0, 0x1, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/kv/dist_sender.go:337 +0x230 github.com/cockroachdb/cockroach/kv.(*DistSender).sendSingleRange(0xc8212c0990, 0x7f4c1bb69938, 0xc82166e030, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/kv/dist_sender.go:452 +0x206 github.com/cockroachdb/cockroach/kv.(*DistSender).sendChunk.func2(0xc8215ddf40, 0xc8215ddbf0, 0xc8215de6d0, 0xc8212c0990, 0xc8215de6c0, 0x4, 0xd) /go/src/github.com/cockroachdb/cockroach/kv/dist_sender.go:711 +0x448 github.com/cockroachdb/cockroach/kv.(*DistSender).sendChunk(0xc8212c0990, 0x7f4c1bb69938, 0xc82166e030, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/kv/dist_sender.go:712 +0x9e1 github.com/cockroachdb/cockroach/kv.(*DistSender).Send(0xc8212c0990, 0x7f4c1bb69938, 0xc82166e030, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/kv/dist_sender.go:556 +0x6d5 github.com/cockroachdb/cockroach/kv.(*TxnCoordSender).Send(0xc821140f00, 0x7f4c1bb69938, 0xc82166e030, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/kv/txn_coord_sender.go:367 +0xa5d github.com/cockroachdb/cockroach/client.(*DB).send(0xc820dc22d0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/client/db.go:471 +0x3ee github.com/cockroachdb/cockroach/client.(*DB).(github.com/cockroachdb/cockroach/client.send)-fm(0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/client/db.go:418 +0x37 github.com/cockroachdb/cockroach/client.sendAndFill(0xc8215df508, 0xc8212fe000, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/client/db.go:400 +0x200 github.com/cockroachdb/cockroach/client.(*DB).Run(0xc820dc22d0, 0xc8212fe000, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/client/db.go:418 +0x85 github.com/cockroachdb/cockroach/storage.(*intentResolver).maybePushTransactions(0xc82112a540, 0x7f4c1baef610, 0xc82166e000, 0xc82096f440, 0x2, 0x2, 0xdf, 0xd, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/storage/intent_resolver.go:223 +0x783 github.com/cockroachdb/cockroach/storage.(*intentResolver).processIntentsAsync.func1() /go/src/github.com/cockroachdb/cockroach/storage/intent_resolver.go:275 +0x174 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunLimitedAsyncTask.func1(0xc82144e4d0, 0x222d823, 0x1a, 0x12b, 0xc8211410e0, 0xc8217a8d40) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:218 +0x8d created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunLimitedAsyncTask /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:219 +0x1e2 goroutine 15721 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc8210a4c20) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15726 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc8210a4ce0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15731 [select]: github.com/cockroachdb/cockroach/kv.(*TxnCoordSender).startStats(0xc821140f00) /go/src/github.com/cockroachdb/cockroach/kv/txn_coord_sender.go:203 +0xd21 github.com/cockroachdb/cockroach/kv.(*TxnCoordSender).(github.com/cockroachdb/cockroach/kv.startStats)-fm() /go/src/github.com/cockroachdb/cockroach/kv/txn_coord_sender.go:188 +0x20 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc8203629c0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15725 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc8210a4cc0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15664 [chan receive, 8 minutes]: github.com/cockroachdb/cockroach/storage/engine.(*RocksDB).Open.func1(0xc821140300) /go/src/github.com/cockroachdb/cockroach/storage/engine/rocksdb.go:164 +0x3a created by github.com/cockroachdb/cockroach/storage/engine.(*RocksDB).Open /go/src/github.com/cockroachdb/cockroach/storage/engine/rocksdb.go:165 +0x8f4 rax 0xca rbx 0x0 rcx 0x62de23 rdx 0x0 rdi 0x2801188 rsi 0x0 rbp 0x1 rsp 0x7ffe96854a70 r8 0x0 r9 0x0 r10 0x0 r11 0x286 r12 0x14d r13 0x1d779de r14 0x1 r15 0x2c rip 0x62de21 rflags 0x286 cs 0x33 fs 0x0 gs 0x0 ERROR: exit status 2 ``` Run Details: ``` 0 runs so far, 0 failures, over 5s 0 runs so far, 0 failures, over 10s 0 runs so far, 0 failures, over 15s 0 runs so far, 0 failures, over 20s 0 runs so far, 0 failures, over 25s 0 runs so far, 0 failures, over 30s 0 runs so far, 0 failures, over 35s 16 runs so far, 0 failures, over 40s 16 runs so far, 0 failures, over 45s 16 runs so far, 0 failures, over 50s 16 runs so far, 0 failures, over 55s 16 runs so far, 0 failures, over 1m0s 16 runs so far, 0 failures, over 1m5s 16 runs so far, 0 failures, over 1m10s 19 runs so far, 0 failures, over 1m15s 32 runs so far, 0 failures, over 1m20s 32 runs so far, 0 failures, over 1m25s 32 runs so far, 0 failures, over 1m30s 32 runs so far, 0 failures, over 1m35s 32 runs so far, 0 failures, over 1m40s 32 runs so far, 0 failures, over 1m45s 32 runs so far, 0 failures, over 1m50s 47 runs so far, 0 failures, over 1m55s 47 runs so far, 0 failures, over 2m0s 47 runs so far, 0 failures, over 2m5s 47 runs so far, 0 failures, over 2m10s 47 runs so far, 0 failures, over 2m15s 47 runs so far, 0 failures, over 2m20s 48 runs so far, 0 failures, over 2m25s 53 runs so far, 0 failures, over 2m30s 62 runs so far, 0 failures, over 2m35s 62 runs so far, 0 failures, over 2m40s 62 runs so far, 0 failures, over 2m45s 62 runs so far, 0 failures, over 2m50s 62 runs so far, 0 failures, over 2m55s 62 runs so far, 0 failures, over 3m0s 66 runs so far, 0 failures, over 3m5s 76 runs so far, 0 failures, over 3m10s 77 runs so far, 0 failures, over 3m15s 77 runs so far, 0 failures, over 3m20s 77 runs so far, 0 failures, over 3m25s 77 runs so far, 0 failures, over 3m30s 77 runs so far, 0 failures, over 3m35s 79 runs so far, 0 failures, over 3m40s 88 runs so far, 0 failures, over 3m45s 91 runs so far, 0 failures, over 3m50s 91 runs so far, 0 failures, over 3m55s 91 runs so far, 0 failures, over 4m0s 91 runs so far, 0 failures, over 4m5s 91 runs so far, 0 failures, over 4m10s 93 runs so far, 0 failures, over 4m15s 99 runs so far, 0 failures, over 4m20s 105 runs so far, 0 failures, over 4m25s 105 runs so far, 0 failures, over 4m30s 105 runs so far, 0 failures, over 4m35s 105 runs so far, 0 failures, over 4m40s 106 runs so far, 0 failures, over 4m45s 108 runs so far, 0 failures, over 4m50s 114 runs so far, 0 failures, over 4m55s 120 runs so far, 0 failures, over 5m0s 120 runs so far, 0 failures, over 5m5s 120 runs so far, 0 failures, over 5m10s 120 runs so far, 0 failures, over 5m15s 121 runs so far, 0 failures, over 5m20s 123 runs so far, 0 failures, over 5m25s 128 runs so far, 0 failures, over 5m30s 132 runs so far, 0 failures, over 5m35s 134 runs so far, 0 failures, over 5m40s 134 runs so far, 0 failures, over 5m45s 134 runs so far, 0 failures, over 5m50s 136 runs so far, 0 failures, over 5m55s 137 runs so far, 0 failures, over 6m0s 141 runs so far, 0 failures, over 6m5s 145 runs so far, 0 failures, over 6m10s 148 runs so far, 0 failures, over 6m15s 148 runs so far, 0 failures, over 6m20s 148 runs so far, 0 failures, over 6m25s 150 runs so far, 0 failures, over 6m30s 152 runs so far, 0 failures, over 6m35s 156 runs so far, 0 failures, over 6m40s 160 runs so far, 0 failures, over 6m45s 163 runs so far, 0 failures, over 6m50s 163 runs so far, 0 failures, over 6m55s 163 runs so far, 0 failures, over 7m0s 165 runs so far, 0 failures, over 7m5s 166 runs so far, 0 failures, over 7m10s 171 runs so far, 0 failures, over 7m15s 174 runs so far, 0 failures, over 7m20s 178 runs so far, 0 failures, over 7m25s 178 runs so far, 0 failures, over 7m30s 178 runs so far, 0 failures, over 7m35s 180 runs so far, 0 failures, over 7m40s 181 runs so far, 0 failures, over 7m45s 185 runs so far, 0 failures, over 7m50s 188 runs so far, 0 failures, over 7m55s 192 runs so far, 0 failures, over 8m0s 192 runs so far, 0 failures, over 8m5s 193 runs so far, 0 failures, over 8m10s 194 runs so far, 0 failures, over 8m15s 195 runs so far, 0 failures, over 8m20s 199 runs so far, 0 failures, over 8m25s 201 runs so far, 0 failures, over 8m30s 206 runs so far, 0 failures, over 8m35s 207 runs so far, 0 failures, over 8m40s 207 runs so far, 0 failures, over 8m45s 208 runs so far, 0 failures, over 8m50s 211 runs so far, 0 failures, over 8m55s 214 runs so far, 0 failures, over 9m0s 216 runs so far, 0 failures, over 9m5s 221 runs so far, 0 failures, over 9m10s 221 runs so far, 0 failures, over 9m15s 222 runs so far, 0 failures, over 9m20s 223 runs so far, 0 failures, over 9m25s 226 runs so far, 0 failures, over 9m30s 228 runs so far, 0 failures, over 9m35s 230 runs so far, 0 failures, over 9m40s 235 runs so far, 0 failures, over 9m45s 235 runs so far, 0 failures, over 9m50s 236 runs so far, 0 failures, over 9m55s 237 runs so far, 0 failures, over 10m0s 240 runs so far, 0 failures, over 10m5s 242 runs so far, 0 failures, over 10m10s 244 runs so far, 0 failures, over 10m15s 249 runs so far, 0 failures, over 10m20s 249 runs so far, 0 failures, over 10m25s 250 runs so far, 0 failures, over 10m30s 252 runs so far, 0 failures, over 10m35s 254 runs so far, 0 failures, over 10m40s 257 runs so far, 0 failures, over 10m45s 259 runs so far, 0 failures, over 10m50s 264 runs so far, 0 failures, over 10m55s 264 runs so far, 0 failures, over 11m0s 265 runs so far, 0 failures, over 11m5s 267 runs so far, 0 failures, over 11m10s 269 runs so far, 0 failures, over 11m15s 271 runs completed, 1 failures, over 11m16s FAIL ``` Please assign, take a look and update the issue accordingly.
1.0
stress: failed test in cockroach/storage/storage.test: TestStoreRangeSplitStats - Binary: cockroach/static-tests.tar.gz sha: https://github.com/cockroachdb/cockroach/commits/2cc5a467f6c280e97e49a22b884b649fd7790273 Stress build found a failed test: ``` === RUN TestStoreRangeSplitStats I160517 07:24:34.574483 storage/engine/rocksdb.go:137 opening in memory rocksdb instance W160517 07:24:34.575211 gossip/gossip.go:893 not connected to cluster; use --join to specify a connected node I160517 07:24:34.577161 storage/replica_command.go:1470 range 1: new leader lease replica {1 1 1} 1970-01-01 00:00:00 +0000 UTC 1.000000123s following replica {0 0 0} 1970-01-01 00:00:00 +0000 UTC 0 [physicalTime=1970-01-01 00:00:00.000000123 +0000 UTC] I160517 07:24:34.578130 storage/replica_command.go:2011 initiating a split of range=1 [/Min-/Max) at key /Table/50 I160517 07:24:34.580548 storage/replica_command.go:1470 range 2: new leader lease replica {1 1 1} 1970-01-01 00:00:00 +0000 UTC 1.000000123s following replica {0 0 0} 1970-01-01 00:00:00 +0000 UTC 0 [physicalTime=1970-01-01 00:00:00.000000123 +0000 UTC] I160517 07:24:34.605112 storage/replica_command.go:2011 initiating a split of range=2 [/Table/50-/Max) at key /Table/50/0/"Z" W160517 07:24:34.605402 storage/stores.go:200 range not contained in one range: [/Meta2/Table/50/0/"Z",/Table/50/0/"Z\x00"), but have [/Min,/Table/50) I160517 07:24:34.608896 stopper.go:352 draining; tasks left: 1 storage/intent_resolver.go:338 1 storage/intent_resolver.go:299 I160517 07:24:34.617668 stopper.go:352 draining; tasks left: 1 storage/intent_resolver.go:299 W160517 07:25:34.577609 storage/store.go:995 [node=1,store=1] error gossiping first range data: storage/replica.go:2126: node is stopping W160517 07:25:34.577663 storage/store.go:1014 [node=1,store=1] error gossiping system config: storage/replica.go:2126: node is stopping W160517 07:26:34.577591 storage/store.go:995 [node=1,store=1] error gossiping first range data: storage/replica.go:2126: node is stopping W160517 07:26:34.577636 storage/store.go:1014 [node=1,store=1] error gossiping system config: storage/replica.go:2126: node is stopping W160517 07:27:34.577608 storage/store.go:995 [node=1,store=1] error gossiping first range data: storage/replica.go:2126: node is stopping W160517 07:27:34.577663 storage/store.go:1014 [node=1,store=1] error gossiping system config: storage/replica.go:2126: node is stopping W160517 07:28:34.577613 storage/store.go:995 [node=1,store=1] error gossiping first range data: storage/replica.go:2126: node is stopping W160517 07:28:34.577664 storage/store.go:1014 [node=1,store=1] error gossiping system config: storage/replica.go:2126: node is stopping W160517 07:29:34.577718 storage/store.go:995 [node=1,store=1] error gossiping first range data: storage/replica.go:2126: node is stopping W160517 07:29:34.577810 storage/store.go:1014 [node=1,store=1] error gossiping system config: storage/replica.go:2126: node is stopping W160517 07:30:34.577627 storage/store.go:995 [node=1,store=1] error gossiping first range data: storage/replica.go:2126: node is stopping W160517 07:30:34.577646 storage/store.go:1014 [node=1,store=1] error gossiping system config: storage/replica.go:2126: node is stopping W160517 07:31:34.577610 storage/store.go:995 [node=1,store=1] error gossiping first range data: storage/replica.go:2126: node is stopping W160517 07:31:34.577756 storage/store.go:1014 [node=1,store=1] error gossiping system config: storage/replica.go:2126: node is stopping W160517 07:32:34.577595 storage/store.go:995 [node=1,store=1] error gossiping first range data: storage/replica.go:2126: node is stopping W160517 07:32:34.577779 storage/store.go:1014 [node=1,store=1] error gossiping system config: storage/replica.go:2126: node is stopping SIGABRT: abort PC=0x62de21 m=0 goroutine 0 [idle]: runtime.futex(0x2801188, 0x0, 0x0, 0x0, 0x0, 0x2800010, 0x0, 0x0, 0x5dc6f4, 0x2801188, ...) /usr/local/go/src/runtime/sys_linux_amd64.s:302 +0x21 runtime.futexsleep(0x2801188, 0x0, 0xffffffffffffffff) /usr/local/go/src/runtime/os1_linux.go:40 +0x53 runtime.notesleep(0x2801188) /usr/local/go/src/runtime/lock_futex.go:145 +0xa4 runtime.stopm() /usr/local/go/src/runtime/proc.go:1535 +0x10b runtime.findrunnable(0xc82001aa00, 0x0) /usr/local/go/src/runtime/proc.go:1973 +0x739 runtime.schedule() /usr/local/go/src/runtime/proc.go:2072 +0x24f runtime.park_m(0xc82006ea80) /usr/local/go/src/runtime/proc.go:2137 +0x18b runtime.mcall(0x7ffe96854c10) /usr/local/go/src/runtime/asm_amd64.s:233 +0x5b goroutine 1 [chan receive, 8 minutes]: testing.RunTests(0x1d7d610, 0x25283a0, 0x10f, 0x10f, 0xc82012bd01) /usr/local/go/src/testing/testing.go:583 +0x8d2 testing.(*M).Run(0xc8200efef8, 0xc820307c98) /usr/local/go/src/testing/testing.go:515 +0x81 github.com/cockroachdb/cockroach/storage_test.TestMain(0xc8200efef8) /go/src/github.com/cockroachdb/cockroach/storage/main_test.go:54 +0x1de main.main() github.com/cockroachdb/cockroach/storage/_test/_testmain.go:604 +0x114 goroutine 17 [syscall, 9 minutes, locked to thread]: runtime.goexit() /usr/local/go/src/runtime/asm_amd64.s:1998 +0x1 goroutine 35 [chan receive]: github.com/cockroachdb/cockroach/util/log.(*loggingT).flushDaemon(0x27fff20) /go/src/github.com/cockroachdb/cockroach/util/log/clog.go:1011 +0x64 created by github.com/cockroachdb/cockroach/util/log.init.1 /go/src/github.com/cockroachdb/cockroach/util/log/clog.go:598 +0x8a goroutine 15663 [semacquire, 8 minutes]: sync.runtime_Syncsemacquire(0xc82113e5d0) /usr/local/go/src/runtime/sema.go:241 +0x201 sync.(*Cond).Wait(0xc82113e5c0) /usr/local/go/src/sync/cond.go:63 +0x9b github.com/cockroachdb/cockroach/util/stop.(*Stopper).Quiesce(0xc82144e4d0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:354 +0x1eb github.com/cockroachdb/cockroach/util/stop.(*Stopper).Stop(0xc82144e4d0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:300 +0x1a7 github.com/cockroachdb/cockroach/storage_test.TestStoreRangeSplitStats(0xc8212c0750) /go/src/github.com/cockroachdb/cockroach/storage/client_split_test.go:494 +0x1e08 testing.tRunner(0xc8212c0750, 0x2529bd0) /usr/local/go/src/testing/testing.go:473 +0x98 created by testing.RunTests /usr/local/go/src/testing/testing.go:582 +0x892 goroutine 15735 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*Store).Start.func4() /go/src/github.com/cockroachdb/cockroach/storage/store.go:921 +0x13d github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc82112b340) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15723 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc8210a4c80) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15722 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*replicaScanner).waitAndProcess(0xc82144fc70, 0xececcbe32, 0x2262c243, 0x27ff340, 0xc82113e600, 0xc82144e4d0, 0xc820898f00, 0xc820545e98) /go/src/github.com/cockroachdb/cockroach/storage/scanner.go:161 +0x568 github.com/cockroachdb/cockroach/storage.(*replicaScanner).scanLoop.func1.1(0xc820898f00, 0xc821b86930) /go/src/github.com/cockroachdb/cockroach/storage/scanner.go:203 +0x78 github.com/cockroachdb/cockroach/storage.(*storeRangeSet).Visit(0xc821b86cc0, 0xc8215aca00) /go/src/github.com/cockroachdb/cockroach/storage/store.go:231 +0x246 github.com/cockroachdb/cockroach/storage.(*replicaScanner).scanLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/scanner.go:205 +0x21e github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc8210a4c40) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15724 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc8210a4ca0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15733 [select]: github.com/cockroachdb/cockroach/storage.(*Store).processRaft.func1() /go/src/github.com/cockroachdb/cockroach/storage/store.go:2059 +0xbf1 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc8203639f0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15728 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc8210a4d20) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15737 [select]: github.com/cockroachdb/cockroach/storage.(*Store).startGossip.func2() /go/src/github.com/cockroachdb/cockroach/storage/store.go:1011 +0x3cf github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc82112b380) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15679 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*idAllocator).start.func1() /go/src/github.com/cockroachdb/cockroach/storage/id_alloc.go:123 +0x86a github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc821b68050) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15732 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*StorePool).start.func1() /go/src/github.com/cockroachdb/cockroach/storage/store_pool.go:225 +0x23f github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc82112a460) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15736 [select]: github.com/cockroachdb/cockroach/storage.(*Store).startGossip.func1() /go/src/github.com/cockroachdb/cockroach/storage/store.go:992 +0x3cf github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc82112b360) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15727 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc8210a4d00) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15746 [runnable]: time.absDate(0x7ffffffe7214923f, 0x2698f01, 0x27ff340, 0x7ffffffe7214923f, 0x11, 0x89) /usr/local/go/src/time/time.go:678 time.Time.date(0xececcc03f, 0xc802698f8b, 0x27ff340, 0x8d4901, 0x7e0, 0x5, 0x11, 0x89) /usr/local/go/src/time/time.go:674 +0x4a time.Time.Day(0xececcc03f, 0x2698f8b, 0x27ff340, 0x11) /usr/local/go/src/time/time.go:309 +0x38 golang.org/x/net/trace.(*trace).delta(0xc8202763c0, 0xececcc03f, 0x2698f8b, 0x27ff340, 0x1d779dc, 0x1d779dd) /go/src/golang.org/x/net/trace/trace.go:729 +0x15e golang.org/x/net/trace.(*trace).addEvent(0xc8202763c0, 0x17c7da0, 0xc821b192c0, 0x15e0000) /go/src/golang.org/x/net/trace/trace.go:755 +0x308 golang.org/x/net/trace.(*trace).LazyPrintf(0xc8202763c0, 0x19bd558, 0x2, 0xc8219ca010, 0x1, 0x1) /go/src/golang.org/x/net/trace/trace.go:790 +0xa2 github.com/opentracing/basictracer-go.glob.func2.1(0x1862220, 0xc8218baa80) /go/src/github.com/opentracing/basictracer-go/event_nettrace.go:19 +0x343 github.com/opentracing/basictracer-go.(*spanImpl).onLog(0xc8214a6360, 0x0, 0x0, 0x0, 0xc820abd7a0, 0x1e, 0x0, 0x0) /go/src/github.com/opentracing/basictracer-go/event.go:40 +0xc4 github.com/opentracing/basictracer-go.(*spanImpl).Log(0xc8214a6360, 0x0, 0x0, 0x0, 0xc820abd7a0, 0x1e, 0x0, 0x0) /go/src/github.com/opentracing/basictracer-go/span.go:109 +0x114 github.com/opentracing/basictracer-go.(*spanImpl).LogEvent(0xc8214a6360, 0xc820abd7a0, 0x1e) /go/src/github.com/opentracing/basictracer-go/span.go:94 +0x7d github.com/cockroachdb/cockroach/util/log.Trace(0x7f4c1bb69938, 0xc821b19140, 0xc820abd7a0, 0x1e) /go/src/github.com/cockroachdb/cockroach/util/log/trace.go:29 +0x63 github.com/cockroachdb/cockroach/storage.(*Store).Send(0xc820e1af00, 0x7f4c1bb69938, 0xc821b19140, 0xdf, 0x209fce6, 0x100000001, 0x1, 0x3, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/storage/store.go:1824 +0x10e6 github.com/cockroachdb/cockroach/storage.(*Stores).Send(0xc82113e740, 0x7f4c1bb69938, 0xc821b190e0, 0x0, 0x0, 0x100000001, 0x1, 0x3, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/storage/stores.go:178 +0x501 github.com/cockroachdb/cockroach/kv.(*senderTransport).SendNext(0xc8202dc1b0, 0xc8219320c0) /go/src/github.com/cockroachdb/cockroach/kv/transport.go:290 +0x3cc github.com/cockroachdb/cockroach/kv.(*DistSender).sendToReplicas(0xc8212c0990, 0x7f4c1bb69938, 0xc82166e030, 0x1, 0x2540be400, 0xb2d05e00, 0xc821b864b0, 0x2, 0xc8210680e0, 0x1, ...) /go/src/github.com/cockroachdb/cockroach/kv/dist_sender.go:1007 +0x8eb github.com/cockroachdb/cockroach/kv.(*DistSender).sendRPC(0xc8212c0990, 0x7f4c1bb69938, 0xc82166e030, 0x2, 0xc8210680e0, 0x1, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/kv/dist_sender.go:337 +0x230 github.com/cockroachdb/cockroach/kv.(*DistSender).sendSingleRange(0xc8212c0990, 0x7f4c1bb69938, 0xc82166e030, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/kv/dist_sender.go:452 +0x206 github.com/cockroachdb/cockroach/kv.(*DistSender).sendChunk.func2(0xc8215ddf40, 0xc8215ddbf0, 0xc8215de6d0, 0xc8212c0990, 0xc8215de6c0, 0x4, 0xd) /go/src/github.com/cockroachdb/cockroach/kv/dist_sender.go:711 +0x448 github.com/cockroachdb/cockroach/kv.(*DistSender).sendChunk(0xc8212c0990, 0x7f4c1bb69938, 0xc82166e030, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/kv/dist_sender.go:712 +0x9e1 github.com/cockroachdb/cockroach/kv.(*DistSender).Send(0xc8212c0990, 0x7f4c1bb69938, 0xc82166e030, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/kv/dist_sender.go:556 +0x6d5 github.com/cockroachdb/cockroach/kv.(*TxnCoordSender).Send(0xc821140f00, 0x7f4c1bb69938, 0xc82166e030, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/kv/txn_coord_sender.go:367 +0xa5d github.com/cockroachdb/cockroach/client.(*DB).send(0xc820dc22d0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/client/db.go:471 +0x3ee github.com/cockroachdb/cockroach/client.(*DB).(github.com/cockroachdb/cockroach/client.send)-fm(0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/client/db.go:418 +0x37 github.com/cockroachdb/cockroach/client.sendAndFill(0xc8215df508, 0xc8212fe000, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/client/db.go:400 +0x200 github.com/cockroachdb/cockroach/client.(*DB).Run(0xc820dc22d0, 0xc8212fe000, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/client/db.go:418 +0x85 github.com/cockroachdb/cockroach/storage.(*intentResolver).maybePushTransactions(0xc82112a540, 0x7f4c1baef610, 0xc82166e000, 0xc82096f440, 0x2, 0x2, 0xdf, 0xd, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/storage/intent_resolver.go:223 +0x783 github.com/cockroachdb/cockroach/storage.(*intentResolver).processIntentsAsync.func1() /go/src/github.com/cockroachdb/cockroach/storage/intent_resolver.go:275 +0x174 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunLimitedAsyncTask.func1(0xc82144e4d0, 0x222d823, 0x1a, 0x12b, 0xc8211410e0, 0xc8217a8d40) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:218 +0x8d created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunLimitedAsyncTask /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:219 +0x1e2 goroutine 15721 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc8210a4c20) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15726 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc8210a4ce0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15731 [select]: github.com/cockroachdb/cockroach/kv.(*TxnCoordSender).startStats(0xc821140f00) /go/src/github.com/cockroachdb/cockroach/kv/txn_coord_sender.go:203 +0xd21 github.com/cockroachdb/cockroach/kv.(*TxnCoordSender).(github.com/cockroachdb/cockroach/kv.startStats)-fm() /go/src/github.com/cockroachdb/cockroach/kv/txn_coord_sender.go:188 +0x20 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc8203629c0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15725 [select, 8 minutes]: github.com/cockroachdb/cockroach/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:356 +0x303 github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc82144e4d0, 0xc8210a4cc0) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 15664 [chan receive, 8 minutes]: github.com/cockroachdb/cockroach/storage/engine.(*RocksDB).Open.func1(0xc821140300) /go/src/github.com/cockroachdb/cockroach/storage/engine/rocksdb.go:164 +0x3a created by github.com/cockroachdb/cockroach/storage/engine.(*RocksDB).Open /go/src/github.com/cockroachdb/cockroach/storage/engine/rocksdb.go:165 +0x8f4 rax 0xca rbx 0x0 rcx 0x62de23 rdx 0x0 rdi 0x2801188 rsi 0x0 rbp 0x1 rsp 0x7ffe96854a70 r8 0x0 r9 0x0 r10 0x0 r11 0x286 r12 0x14d r13 0x1d779de r14 0x1 r15 0x2c rip 0x62de21 rflags 0x286 cs 0x33 fs 0x0 gs 0x0 ERROR: exit status 2 ``` Run Details: ``` 0 runs so far, 0 failures, over 5s 0 runs so far, 0 failures, over 10s 0 runs so far, 0 failures, over 15s 0 runs so far, 0 failures, over 20s 0 runs so far, 0 failures, over 25s 0 runs so far, 0 failures, over 30s 0 runs so far, 0 failures, over 35s 16 runs so far, 0 failures, over 40s 16 runs so far, 0 failures, over 45s 16 runs so far, 0 failures, over 50s 16 runs so far, 0 failures, over 55s 16 runs so far, 0 failures, over 1m0s 16 runs so far, 0 failures, over 1m5s 16 runs so far, 0 failures, over 1m10s 19 runs so far, 0 failures, over 1m15s 32 runs so far, 0 failures, over 1m20s 32 runs so far, 0 failures, over 1m25s 32 runs so far, 0 failures, over 1m30s 32 runs so far, 0 failures, over 1m35s 32 runs so far, 0 failures, over 1m40s 32 runs so far, 0 failures, over 1m45s 32 runs so far, 0 failures, over 1m50s 47 runs so far, 0 failures, over 1m55s 47 runs so far, 0 failures, over 2m0s 47 runs so far, 0 failures, over 2m5s 47 runs so far, 0 failures, over 2m10s 47 runs so far, 0 failures, over 2m15s 47 runs so far, 0 failures, over 2m20s 48 runs so far, 0 failures, over 2m25s 53 runs so far, 0 failures, over 2m30s 62 runs so far, 0 failures, over 2m35s 62 runs so far, 0 failures, over 2m40s 62 runs so far, 0 failures, over 2m45s 62 runs so far, 0 failures, over 2m50s 62 runs so far, 0 failures, over 2m55s 62 runs so far, 0 failures, over 3m0s 66 runs so far, 0 failures, over 3m5s 76 runs so far, 0 failures, over 3m10s 77 runs so far, 0 failures, over 3m15s 77 runs so far, 0 failures, over 3m20s 77 runs so far, 0 failures, over 3m25s 77 runs so far, 0 failures, over 3m30s 77 runs so far, 0 failures, over 3m35s 79 runs so far, 0 failures, over 3m40s 88 runs so far, 0 failures, over 3m45s 91 runs so far, 0 failures, over 3m50s 91 runs so far, 0 failures, over 3m55s 91 runs so far, 0 failures, over 4m0s 91 runs so far, 0 failures, over 4m5s 91 runs so far, 0 failures, over 4m10s 93 runs so far, 0 failures, over 4m15s 99 runs so far, 0 failures, over 4m20s 105 runs so far, 0 failures, over 4m25s 105 runs so far, 0 failures, over 4m30s 105 runs so far, 0 failures, over 4m35s 105 runs so far, 0 failures, over 4m40s 106 runs so far, 0 failures, over 4m45s 108 runs so far, 0 failures, over 4m50s 114 runs so far, 0 failures, over 4m55s 120 runs so far, 0 failures, over 5m0s 120 runs so far, 0 failures, over 5m5s 120 runs so far, 0 failures, over 5m10s 120 runs so far, 0 failures, over 5m15s 121 runs so far, 0 failures, over 5m20s 123 runs so far, 0 failures, over 5m25s 128 runs so far, 0 failures, over 5m30s 132 runs so far, 0 failures, over 5m35s 134 runs so far, 0 failures, over 5m40s 134 runs so far, 0 failures, over 5m45s 134 runs so far, 0 failures, over 5m50s 136 runs so far, 0 failures, over 5m55s 137 runs so far, 0 failures, over 6m0s 141 runs so far, 0 failures, over 6m5s 145 runs so far, 0 failures, over 6m10s 148 runs so far, 0 failures, over 6m15s 148 runs so far, 0 failures, over 6m20s 148 runs so far, 0 failures, over 6m25s 150 runs so far, 0 failures, over 6m30s 152 runs so far, 0 failures, over 6m35s 156 runs so far, 0 failures, over 6m40s 160 runs so far, 0 failures, over 6m45s 163 runs so far, 0 failures, over 6m50s 163 runs so far, 0 failures, over 6m55s 163 runs so far, 0 failures, over 7m0s 165 runs so far, 0 failures, over 7m5s 166 runs so far, 0 failures, over 7m10s 171 runs so far, 0 failures, over 7m15s 174 runs so far, 0 failures, over 7m20s 178 runs so far, 0 failures, over 7m25s 178 runs so far, 0 failures, over 7m30s 178 runs so far, 0 failures, over 7m35s 180 runs so far, 0 failures, over 7m40s 181 runs so far, 0 failures, over 7m45s 185 runs so far, 0 failures, over 7m50s 188 runs so far, 0 failures, over 7m55s 192 runs so far, 0 failures, over 8m0s 192 runs so far, 0 failures, over 8m5s 193 runs so far, 0 failures, over 8m10s 194 runs so far, 0 failures, over 8m15s 195 runs so far, 0 failures, over 8m20s 199 runs so far, 0 failures, over 8m25s 201 runs so far, 0 failures, over 8m30s 206 runs so far, 0 failures, over 8m35s 207 runs so far, 0 failures, over 8m40s 207 runs so far, 0 failures, over 8m45s 208 runs so far, 0 failures, over 8m50s 211 runs so far, 0 failures, over 8m55s 214 runs so far, 0 failures, over 9m0s 216 runs so far, 0 failures, over 9m5s 221 runs so far, 0 failures, over 9m10s 221 runs so far, 0 failures, over 9m15s 222 runs so far, 0 failures, over 9m20s 223 runs so far, 0 failures, over 9m25s 226 runs so far, 0 failures, over 9m30s 228 runs so far, 0 failures, over 9m35s 230 runs so far, 0 failures, over 9m40s 235 runs so far, 0 failures, over 9m45s 235 runs so far, 0 failures, over 9m50s 236 runs so far, 0 failures, over 9m55s 237 runs so far, 0 failures, over 10m0s 240 runs so far, 0 failures, over 10m5s 242 runs so far, 0 failures, over 10m10s 244 runs so far, 0 failures, over 10m15s 249 runs so far, 0 failures, over 10m20s 249 runs so far, 0 failures, over 10m25s 250 runs so far, 0 failures, over 10m30s 252 runs so far, 0 failures, over 10m35s 254 runs so far, 0 failures, over 10m40s 257 runs so far, 0 failures, over 10m45s 259 runs so far, 0 failures, over 10m50s 264 runs so far, 0 failures, over 10m55s 264 runs so far, 0 failures, over 11m0s 265 runs so far, 0 failures, over 11m5s 267 runs so far, 0 failures, over 11m10s 269 runs so far, 0 failures, over 11m15s 271 runs completed, 1 failures, over 11m16s FAIL ``` Please assign, take a look and update the issue accordingly.
test
stress failed test in cockroach storage storage test teststorerangesplitstats binary cockroach static tests tar gz sha stress build found a failed test run teststorerangesplitstats storage engine rocksdb go opening in memory rocksdb instance gossip gossip go not connected to cluster use join to specify a connected node storage replica command go range new leader lease replica utc following replica utc storage replica command go initiating a split of range min max at key table storage replica command go range new leader lease replica utc following replica utc storage replica command go initiating a split of range table max at key table z storage stores go range not contained in one range table z table z but have min table stopper go draining tasks left storage intent resolver go storage intent resolver go stopper go draining tasks left storage intent resolver go storage store go error gossiping first range data storage replica go node is stopping storage store go error gossiping system config storage replica go node is stopping storage store go error gossiping first range data storage replica go node is stopping storage store go error gossiping system config storage replica go node is stopping storage store go error gossiping first range data storage replica go node is stopping storage store go error gossiping system config storage replica go node is stopping storage store go error gossiping first range data storage replica go node is stopping storage store go error gossiping system config storage replica go node is stopping storage store go error gossiping first range data storage replica go node is stopping storage store go error gossiping system config storage replica go node is stopping storage store go error gossiping first range data storage replica go node is stopping storage store go error gossiping system config storage replica go node is stopping storage store go error gossiping first range data storage replica go node is stopping storage store go error gossiping system config storage replica go node is stopping storage store go error gossiping first range data storage replica go node is stopping storage store go error gossiping system config storage replica go node is stopping sigabrt abort pc m goroutine runtime futex usr local go src runtime sys linux s runtime futexsleep usr local go src runtime linux go runtime notesleep usr local go src runtime lock futex go runtime stopm usr local go src runtime proc go runtime findrunnable usr local go src runtime proc go runtime schedule usr local go src runtime proc go runtime park m usr local go src runtime proc go runtime mcall usr local go src runtime asm s goroutine testing runtests usr local go src testing testing go testing m run usr local go src testing testing go github com cockroachdb cockroach storage test testmain go src github com cockroachdb cockroach storage main test go main main github com cockroachdb cockroach storage test testmain go goroutine runtime goexit usr local go src runtime asm s goroutine github com cockroachdb cockroach util log loggingt flushdaemon go src github com cockroachdb cockroach util log clog go created by github com cockroachdb cockroach util log init go src github com cockroachdb cockroach util log clog go goroutine sync runtime syncsemacquire usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach util stop stopper quiesce go src github com cockroachdb cockroach util stop stopper go github com cockroachdb cockroach util stop stopper stop go src github com cockroachdb cockroach util stop stopper go github com cockroachdb cockroach storage test teststorerangesplitstats go src github com cockroachdb cockroach storage client split test go testing trunner usr local go src testing testing go created by testing runtests usr local go src testing testing go goroutine github com cockroachdb cockroach storage store start go src github com cockroachdb cockroach storage store go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage basequeue processloop go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage replicascanner waitandprocess go src github com cockroachdb cockroach storage scanner go github com cockroachdb cockroach storage replicascanner scanloop go src github com cockroachdb cockroach storage scanner go github com cockroachdb cockroach storage storerangeset visit go src github com cockroachdb cockroach storage store go github com cockroachdb cockroach storage replicascanner scanloop go src github com cockroachdb cockroach storage scanner go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage basequeue processloop go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage store processraft go src github com cockroachdb cockroach storage store go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage basequeue processloop go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage store startgossip go src github com cockroachdb cockroach storage store go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage idallocator start go src github com cockroachdb cockroach storage id alloc go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage storepool start go src github com cockroachdb cockroach storage store pool go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage store startgossip go src github com cockroachdb cockroach storage store go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage basequeue processloop go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine time absdate usr local go src time time go time time date usr local go src time time go time time day usr local go src time time go golang org x net trace trace delta go src golang org x net trace trace go golang org x net trace trace addevent go src golang org x net trace trace go golang org x net trace trace lazyprintf go src golang org x net trace trace go github com opentracing basictracer go glob go src github com opentracing basictracer go event nettrace go github com opentracing basictracer go spanimpl onlog go src github com opentracing basictracer go event go github com opentracing basictracer go spanimpl log go src github com opentracing basictracer go span go github com opentracing basictracer go spanimpl logevent go src github com opentracing basictracer go span go github com cockroachdb cockroach util log trace go src github com cockroachdb cockroach util log trace go github com cockroachdb cockroach storage store send go src github com cockroachdb cockroach storage store go github com cockroachdb cockroach storage stores send go src github com cockroachdb cockroach storage stores go github com cockroachdb cockroach kv sendertransport sendnext go src github com cockroachdb cockroach kv transport go github com cockroachdb cockroach kv distsender sendtoreplicas go src github com cockroachdb cockroach kv dist sender go github com cockroachdb cockroach kv distsender sendrpc go src github com cockroachdb cockroach kv dist sender go github com cockroachdb cockroach kv distsender sendsinglerange go src github com cockroachdb cockroach kv dist sender go github com cockroachdb cockroach kv distsender sendchunk go src github com cockroachdb cockroach kv dist sender go github com cockroachdb cockroach kv distsender sendchunk go src github com cockroachdb cockroach kv dist sender go github com cockroachdb cockroach kv distsender send go src github com cockroachdb cockroach kv dist sender go github com cockroachdb cockroach kv txncoordsender send go src github com cockroachdb cockroach kv txn coord sender go github com cockroachdb cockroach client db send go src github com cockroachdb cockroach client db go github com cockroachdb cockroach client db github com cockroachdb cockroach client send fm go src github com cockroachdb cockroach client db go github com cockroachdb cockroach client sendandfill go src github com cockroachdb cockroach client db go github com cockroachdb cockroach client db run go src github com cockroachdb cockroach client db go github com cockroachdb cockroach storage intentresolver maybepushtransactions go src github com cockroachdb cockroach storage intent resolver go github com cockroachdb cockroach storage intentresolver processintentsasync go src github com cockroachdb cockroach storage intent resolver go github com cockroachdb cockroach util stop stopper runlimitedasynctask go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runlimitedasynctask go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage basequeue processloop go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage basequeue processloop go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach kv txncoordsender startstats go src github com cockroachdb cockroach kv txn coord sender go github com cockroachdb cockroach kv txncoordsender github com cockroachdb cockroach kv startstats fm go src github com cockroachdb cockroach kv txn coord sender go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage basequeue processloop go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage engine rocksdb open go src github com cockroachdb cockroach storage engine rocksdb go created by github com cockroachdb cockroach storage engine rocksdb open go src github com cockroachdb cockroach storage engine rocksdb go rax rbx rcx rdx rdi rsi rbp rsp rip rflags cs fs gs error exit status run details runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs completed failures over fail please assign take a look and update the issue accordingly
1
265,670
23,188,039,813
IssuesEvent
2022-08-01 10:06:11
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: tpce/c=5000/nodes=3 failed
C-test-failure O-robot O-roachtest X-stale branch-master release-blocker T-kv no-test-failure-activity
roachtest.tpce/c=5000/nodes=3 [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5453829?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5453829?buildTab=artifacts#/tpce/c=5000/nodes=3) on master @ [87da966cbc15e058591d80c477c78d6059483a5c](https://github.com/cockroachdb/cockroach/commits/87da966cbc15e058591d80c477c78d6059483a5c): ``` test artifacts and logs in: /artifacts/tpce/c=5000/nodes=3/run_1 tpce.go:97,monitor.go:105,errgroup.go:74: UNCLASSIFIED_PROBLEM: context canceled monitor.go:127,tpce.go:105,tpce.go:122,test_runner.go:884: monitor failure: monitor command failure: unexpected node event: 3: dead (exit status 7) (1) attached stack trace -- stack trace: | main.(*monitorImpl).WaitE | main/pkg/cmd/roachtest/monitor.go:115 | main.(*monitorImpl).Wait | main/pkg/cmd/roachtest/monitor.go:123 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerTPCE.func1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/tpce.go:105 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerTPCE.func2 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/tpce.go:122 | [...repeated from below...] Wraps: (2) monitor failure Wraps: (3) attached stack trace -- stack trace: | main.(*monitorImpl).wait.func3 | main/pkg/cmd/roachtest/monitor.go:202 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1581 Wraps: (4) monitor command failure Wraps: (5) unexpected node event: 3: dead (exit status 7) Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *errors.errorString ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_ssd=1</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*tpce/c=5000/nodes=3.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-16689
3.0
roachtest: tpce/c=5000/nodes=3 failed - roachtest.tpce/c=5000/nodes=3 [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5453829?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5453829?buildTab=artifacts#/tpce/c=5000/nodes=3) on master @ [87da966cbc15e058591d80c477c78d6059483a5c](https://github.com/cockroachdb/cockroach/commits/87da966cbc15e058591d80c477c78d6059483a5c): ``` test artifacts and logs in: /artifacts/tpce/c=5000/nodes=3/run_1 tpce.go:97,monitor.go:105,errgroup.go:74: UNCLASSIFIED_PROBLEM: context canceled monitor.go:127,tpce.go:105,tpce.go:122,test_runner.go:884: monitor failure: monitor command failure: unexpected node event: 3: dead (exit status 7) (1) attached stack trace -- stack trace: | main.(*monitorImpl).WaitE | main/pkg/cmd/roachtest/monitor.go:115 | main.(*monitorImpl).Wait | main/pkg/cmd/roachtest/monitor.go:123 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerTPCE.func1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/tpce.go:105 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerTPCE.func2 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/tpce.go:122 | [...repeated from below...] Wraps: (2) monitor failure Wraps: (3) attached stack trace -- stack trace: | main.(*monitorImpl).wait.func3 | main/pkg/cmd/roachtest/monitor.go:202 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1581 Wraps: (4) monitor command failure Wraps: (5) unexpected node event: 3: dead (exit status 7) Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *errors.errorString ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_ssd=1</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*tpce/c=5000/nodes=3.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-16689
test
roachtest tpce c nodes failed roachtest tpce c nodes with on master test artifacts and logs in artifacts tpce c nodes run tpce go monitor go errgroup go unclassified problem context canceled monitor go tpce go tpce go test runner go monitor failure monitor command failure unexpected node event dead exit status attached stack trace stack trace main monitorimpl waite main pkg cmd roachtest monitor go main monitorimpl wait main pkg cmd roachtest monitor go github com cockroachdb cockroach pkg cmd roachtest tests registertpce github com cockroachdb cockroach pkg cmd roachtest tests tpce go github com cockroachdb cockroach pkg cmd roachtest tests registertpce github com cockroachdb cockroach pkg cmd roachtest tests tpce go wraps monitor failure wraps attached stack trace stack trace main monitorimpl wait main pkg cmd roachtest monitor go runtime goexit goroot src runtime asm s wraps monitor command failure wraps unexpected node event dead exit status error types withstack withstack errutil withprefix withstack withstack errutil withprefix errors errorstring parameters roachtest cloud gce roachtest cpu roachtest ssd help see see cc cockroachdb kv triage jira issue crdb
1
24,511
2,668,242,669
IssuesEvent
2015-03-23 06:29:30
cs2103jan2015-t15-2j/main
https://api.github.com/repos/cs2103jan2015-t15-2j/main
closed
A user can modify previously added tasks
priority.high type.story
...so that the user can have more flexibility using the application.
1.0
A user can modify previously added tasks - ...so that the user can have more flexibility using the application.
non_test
a user can modify previously added tasks so that the user can have more flexibility using the application
0
224,916
17,204,183,970
IssuesEvent
2021-07-17 22:31:06
ungoogled-software/ungoogled-chromium-android
https://api.github.com/repos/ungoogled-software/ungoogled-chromium-android
closed
link for download perbuild sdk ndk zips case error
documentation
in build.sh found link for download perbuild sdk ndk zips but case error. sdk_link="https://android-rebuilds.beuc.net/dl/bundles/android-sdk_eng.11.0.0_r27_linux-x86.zip" sdk_tools_link="https://android-rebuilds.beuc.net/dl/repository/sdk-repo-linux-tools-26.1.1.zip" ndk_link="https://android-rebuilds.beuc.net/dl/repository/android-ndk-r20b-linux-x86_64.tar.bz2" seem link is broken
1.0
link for download perbuild sdk ndk zips case error - in build.sh found link for download perbuild sdk ndk zips but case error. sdk_link="https://android-rebuilds.beuc.net/dl/bundles/android-sdk_eng.11.0.0_r27_linux-x86.zip" sdk_tools_link="https://android-rebuilds.beuc.net/dl/repository/sdk-repo-linux-tools-26.1.1.zip" ndk_link="https://android-rebuilds.beuc.net/dl/repository/android-ndk-r20b-linux-x86_64.tar.bz2" seem link is broken
non_test
link for download perbuild sdk ndk zips case error in build sh found link for download perbuild sdk ndk zips but case error sdk link sdk tools link ndk link seem link is broken
0
192
2,495,303,715
IssuesEvent
2015-01-06 09:40:32
ethereum/ethereumj
https://api.github.com/repos/ethereum/ethereumj
closed
Implement test CALL contract with gas > gasRemaining && gas > Long.MAX_VALUE
tests
When calling another contract, you can specify the amount of gas to reserve for that call. Not only should the amount of gas not exceed the gas remaining, it also shouldn't be larger than java.lang.Long because of the risk of an overflow. See: https://github.com/ethereum/ethereumj/commit/a4b8835bac0a5398355bb345e10ffadd214d7127
1.0
Implement test CALL contract with gas > gasRemaining && gas > Long.MAX_VALUE - When calling another contract, you can specify the amount of gas to reserve for that call. Not only should the amount of gas not exceed the gas remaining, it also shouldn't be larger than java.lang.Long because of the risk of an overflow. See: https://github.com/ethereum/ethereumj/commit/a4b8835bac0a5398355bb345e10ffadd214d7127
test
implement test call contract with gas gasremaining gas long max value when calling another contract you can specify the amount of gas to reserve for that call not only should the amount of gas not exceed the gas remaining it also shouldn t be larger than java lang long because of the risk of an overflow see
1
186,848
14,421,718,582
IssuesEvent
2020-12-05 00:05:24
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
opened
7.10 mixed cluster BWC tests are failing
:Core/Infra/Plugins >test-failure
The introduction of https://github.com/elastic/elasticsearch/pull/65490(earlier today) causes bwc de-serialization when executed in a mixed cluster with 7.10.1-SNAPSHOT ( and any prior 7.x, I assume). BWC with 7.11.0 seems fine. Checkout 7.10 branch `./gradlew ':qa:mixed-cluster:v6.8.14#mixedClusterTest' --info` ignore the errors found in the output for the tests, and look at the server logs (i.e. `cat ./qa/mixed-cluster/build/testclusters/v6.8.14-1/logs/es.stdout.log` ) Note - the actual errors from the tests are misleading. The error manifests in different ways, but in general is the root cause of the errors are due to looking up node version information via node stats a failure of deserilization between 7.10 and 6.8.14. I have only included what i believe are the relevant error from the server, and intentionally left out the numerous misleading test failures stack traces. In a mixed cluster the 7.10.1-SNAPSHOT node errors (note the `v6.8.14-1` is the name of the node and initial version of the node. the mixed cluster tests upgrade the nodes in place and this message happens AFTER the node is upgraded to 7.10.1) ``` [2020-12-04T17:36:35,050][WARN ][o.e.t.InboundHandler ] [v6.8.14-1] Failed to deserialize response from [127.0.0.1/127.0.0.1:36585] org.elasticsearch.transport.TransportSerializationException: Failed to deserialize response from handler [org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler/org.elasticsearch.transport.TransportService$6/[cluster:monitor/nodes/info[n]]:org.elasticsearch.action.support.nodes.TransportNodesAction$AsyncAction$1@5a5466d2] at org.elasticsearch.transport.InboundHandler.handleResponse(InboundHandler.java:237) [elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.InboundHandler.messageReceived(InboundHandler.java:133) [elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.InboundHandler.inboundMessage(InboundHandler.java:89) [elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.TcpTransport.inboundMessage(TcpTransport.java:700) [elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.InboundPipeline.forwardFragments(InboundPipeline.java:142) [elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.InboundPipeline.doHandleBytes(InboundPipeline.java:117) [elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.InboundPipeline.handleBytes(InboundPipeline.java:82) [elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.netty4.Netty4MessageChannelHandler.channelRead(Netty4MessageChannelHandler.java:74) [transport-netty4-client-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:271) [netty-handler-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:714) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:615) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:578) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:493) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) [netty-common-4.1.49.Final.jar:4.1.49.Final] at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) [netty-common-4.1.49.Final.jar:4.1.49.Final] at java.lang.Thread.run(Thread.java:832) [?:?] Caused by: java.io.IOException: Invalid string; unexpected character: 243 hex: f3 at org.elasticsearch.common.io.stream.StreamInput.throwOnBrokenChar(StreamInput.java:532) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.common.io.stream.StreamInput.readString(StreamInput.java:491) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.plugins.PluginInfo.<init>(PluginInfo.java:97) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.common.io.stream.StreamInput.readCollection(StreamInput.java:1220) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.common.io.stream.StreamInput.readList(StreamInput.java:1170) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.action.admin.cluster.node.info.PluginsAndModules.<init>(PluginsAndModules.java:48) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.common.io.stream.StreamInput.readOptionalWriteable(StreamInput.java:1003) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.action.admin.cluster.node.info.NodeInfo.<init>(NodeInfo.java:84) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.action.admin.cluster.node.info.TransportNodesInfoAction.newNodeResponse(TransportNodesInfoAction.java:66) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.action.admin.cluster.node.info.TransportNodesInfoAction.newNodeResponse(TransportNodesInfoAction.java:38) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.action.support.nodes.TransportNodesAction$AsyncAction$1.read(TransportNodesAction.java:221) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.action.support.nodes.TransportNodesAction$AsyncAction$1.read(TransportNodesAction.java:218) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.TransportService$6.read(TransportService.java:650) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.TransportService$6.read(TransportService.java:630) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.read(TransportService.java:1162) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.read(TransportService.java:1149) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.InboundHandler.handleResponse(InboundHandler.java:234) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] ... 26 more ``` PluginInfo.java:97 is the following line: ``` javaVersion = in.readString(); ``` Nothing specifically changed here, however, I believe that this information can be sourced from the plugin descriptor. https://github.com/elastic/elasticsearch/pull/65490 introduces some changes to the plugin infrastructure so that is the leading theory to as the relevant change causing this issue. Once this issue is resolved we should also double check that serialization from 5.x is working properly as well no regressions with 7.11+ Also, chose not to revert the commit in 6.x to avoid complicating this issue since there is supporting code in both master and 7.11.
1.0
7.10 mixed cluster BWC tests are failing - The introduction of https://github.com/elastic/elasticsearch/pull/65490(earlier today) causes bwc de-serialization when executed in a mixed cluster with 7.10.1-SNAPSHOT ( and any prior 7.x, I assume). BWC with 7.11.0 seems fine. Checkout 7.10 branch `./gradlew ':qa:mixed-cluster:v6.8.14#mixedClusterTest' --info` ignore the errors found in the output for the tests, and look at the server logs (i.e. `cat ./qa/mixed-cluster/build/testclusters/v6.8.14-1/logs/es.stdout.log` ) Note - the actual errors from the tests are misleading. The error manifests in different ways, but in general is the root cause of the errors are due to looking up node version information via node stats a failure of deserilization between 7.10 and 6.8.14. I have only included what i believe are the relevant error from the server, and intentionally left out the numerous misleading test failures stack traces. In a mixed cluster the 7.10.1-SNAPSHOT node errors (note the `v6.8.14-1` is the name of the node and initial version of the node. the mixed cluster tests upgrade the nodes in place and this message happens AFTER the node is upgraded to 7.10.1) ``` [2020-12-04T17:36:35,050][WARN ][o.e.t.InboundHandler ] [v6.8.14-1] Failed to deserialize response from [127.0.0.1/127.0.0.1:36585] org.elasticsearch.transport.TransportSerializationException: Failed to deserialize response from handler [org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler/org.elasticsearch.transport.TransportService$6/[cluster:monitor/nodes/info[n]]:org.elasticsearch.action.support.nodes.TransportNodesAction$AsyncAction$1@5a5466d2] at org.elasticsearch.transport.InboundHandler.handleResponse(InboundHandler.java:237) [elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.InboundHandler.messageReceived(InboundHandler.java:133) [elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.InboundHandler.inboundMessage(InboundHandler.java:89) [elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.TcpTransport.inboundMessage(TcpTransport.java:700) [elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.InboundPipeline.forwardFragments(InboundPipeline.java:142) [elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.InboundPipeline.doHandleBytes(InboundPipeline.java:117) [elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.InboundPipeline.handleBytes(InboundPipeline.java:82) [elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.netty4.Netty4MessageChannelHandler.channelRead(Netty4MessageChannelHandler.java:74) [transport-netty4-client-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:271) [netty-handler-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:714) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:615) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:578) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:493) [netty-transport-4.1.49.Final.jar:4.1.49.Final] at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) [netty-common-4.1.49.Final.jar:4.1.49.Final] at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) [netty-common-4.1.49.Final.jar:4.1.49.Final] at java.lang.Thread.run(Thread.java:832) [?:?] Caused by: java.io.IOException: Invalid string; unexpected character: 243 hex: f3 at org.elasticsearch.common.io.stream.StreamInput.throwOnBrokenChar(StreamInput.java:532) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.common.io.stream.StreamInput.readString(StreamInput.java:491) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.plugins.PluginInfo.<init>(PluginInfo.java:97) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.common.io.stream.StreamInput.readCollection(StreamInput.java:1220) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.common.io.stream.StreamInput.readList(StreamInput.java:1170) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.action.admin.cluster.node.info.PluginsAndModules.<init>(PluginsAndModules.java:48) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.common.io.stream.StreamInput.readOptionalWriteable(StreamInput.java:1003) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.action.admin.cluster.node.info.NodeInfo.<init>(NodeInfo.java:84) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.action.admin.cluster.node.info.TransportNodesInfoAction.newNodeResponse(TransportNodesInfoAction.java:66) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.action.admin.cluster.node.info.TransportNodesInfoAction.newNodeResponse(TransportNodesInfoAction.java:38) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.action.support.nodes.TransportNodesAction$AsyncAction$1.read(TransportNodesAction.java:221) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.action.support.nodes.TransportNodesAction$AsyncAction$1.read(TransportNodesAction.java:218) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.TransportService$6.read(TransportService.java:650) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.TransportService$6.read(TransportService.java:630) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.read(TransportService.java:1162) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.read(TransportService.java:1149) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] at org.elasticsearch.transport.InboundHandler.handleResponse(InboundHandler.java:234) ~[elasticsearch-7.10.1-SNAPSHOT.jar:7.10.1-SNAPSHOT] ... 26 more ``` PluginInfo.java:97 is the following line: ``` javaVersion = in.readString(); ``` Nothing specifically changed here, however, I believe that this information can be sourced from the plugin descriptor. https://github.com/elastic/elasticsearch/pull/65490 introduces some changes to the plugin infrastructure so that is the leading theory to as the relevant change causing this issue. Once this issue is resolved we should also double check that serialization from 5.x is working properly as well no regressions with 7.11+ Also, chose not to revert the commit in 6.x to avoid complicating this issue since there is supporting code in both master and 7.11.
test
mixed cluster bwc tests are failing the introduction of today causes bwc de serialization when executed in a mixed cluster with snapshot and any prior x i assume bwc with seems fine checkout branch gradlew qa mixed cluster mixedclustertest info ignore the errors found in the output for the tests and look at the server logs i e cat qa mixed cluster build testclusters logs es stdout log note the actual errors from the tests are misleading the error manifests in different ways but in general is the root cause of the errors are due to looking up node version information via node stats a failure of deserilization between and i have only included what i believe are the relevant error from the server and intentionally left out the numerous misleading test failures stack traces in a mixed cluster the snapshot node errors note the is the name of the node and initial version of the node the mixed cluster tests upgrade the nodes in place and this message happens after the node is upgraded to failed to deserialize response from org elasticsearch transport transportserializationexception failed to deserialize response from handler org elasticsearch action support nodes transportnodesaction asyncaction at org elasticsearch transport inboundhandler handleresponse inboundhandler java at org elasticsearch transport inboundhandler messagereceived inboundhandler java at org elasticsearch transport inboundhandler inboundmessage inboundhandler java at org elasticsearch transport tcptransport inboundmessage tcptransport java at org elasticsearch transport inboundpipeline forwardfragments inboundpipeline java at org elasticsearch transport inboundpipeline dohandlebytes inboundpipeline java at org elasticsearch transport inboundpipeline handlebytes inboundpipeline java at org elasticsearch transport channelread java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler logging logginghandler channelread logginghandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline headcontext channelread defaultchannelpipeline java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java at io netty channel nio abstractniobytechannel niobyteunsafe read abstractniobytechannel java at io netty channel nio nioeventloop processselectedkey nioeventloop java at io netty channel nio nioeventloop processselectedkeysplain nioeventloop java at io netty channel nio nioeventloop processselectedkeys nioeventloop java at io netty channel nio nioeventloop run nioeventloop java at io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java at io netty util internal threadexecutormap run threadexecutormap java at java lang thread run thread java caused by java io ioexception invalid string unexpected character hex at org elasticsearch common io stream streaminput throwonbrokenchar streaminput java at org elasticsearch common io stream streaminput readstring streaminput java at org elasticsearch plugins plugininfo plugininfo java at org elasticsearch common io stream streaminput readcollection streaminput java at org elasticsearch common io stream streaminput readlist streaminput java at org elasticsearch action admin cluster node info pluginsandmodules pluginsandmodules java at org elasticsearch common io stream streaminput readoptionalwriteable streaminput java at org elasticsearch action admin cluster node info nodeinfo nodeinfo java at org elasticsearch action admin cluster node info transportnodesinfoaction newnoderesponse transportnodesinfoaction java at org elasticsearch action admin cluster node info transportnodesinfoaction newnoderesponse transportnodesinfoaction java at org elasticsearch action support nodes transportnodesaction asyncaction read transportnodesaction java at org elasticsearch action support nodes transportnodesaction asyncaction read transportnodesaction java at org elasticsearch transport transportservice read transportservice java at org elasticsearch transport transportservice read transportservice java at org elasticsearch transport transportservice contextrestoreresponsehandler read transportservice java at org elasticsearch transport transportservice contextrestoreresponsehandler read transportservice java at org elasticsearch transport inboundhandler handleresponse inboundhandler java more plugininfo java is the following line javaversion in readstring nothing specifically changed here however i believe that this information can be sourced from the plugin descriptor introduces some changes to the plugin infrastructure so that is the leading theory to as the relevant change causing this issue once this issue is resolved we should also double check that serialization from x is working properly as well no regressions with also chose not to revert the commit in x to avoid complicating this issue since there is supporting code in both master and
1
107,990
23,514,158,984
IssuesEvent
2022-08-18 19:34:17
appsmithorg/appsmith
https://api.github.com/repos/appsmithorg/appsmith
closed
[Bug]: GraphQL plugin: checkbox for different page size in cursor pagination section is not working
Bug Backend QA Needs Triaging BE Coders Pod GraphQL Plugin
### Is there an existing issue for this? - [X] I have searched the existing issues ### Description - checkbox for different page size in cursor pagination section is not working as expected. ### Steps To Reproduce 1. click on the checkbox. ### Public Sample App _No response_ ### Version self hosted
1.0
[Bug]: GraphQL plugin: checkbox for different page size in cursor pagination section is not working - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Description - checkbox for different page size in cursor pagination section is not working as expected. ### Steps To Reproduce 1. click on the checkbox. ### Public Sample App _No response_ ### Version self hosted
non_test
graphql plugin checkbox for different page size in cursor pagination section is not working is there an existing issue for this i have searched the existing issues description checkbox for different page size in cursor pagination section is not working as expected steps to reproduce click on the checkbox public sample app no response version self hosted
0
192,465
14,617,568,382
IssuesEvent
2020-12-22 14:58:18
aces/Loris
https://api.github.com/repos/aces/Loris
closed
[23.0][behavioural_qc] Link does not redirect to conflict resolver
23.0.0-testing Bug
**Describe the bug** Link to conflict resolver does not work. **To Reproduce** Steps to reproduce the behavior (attach screenshots if applicable): 1. Go to `Behavioural Quality Control` module 2. Click on `Data Conflicts` tab 3. Click on any instrument **What did you expect to happen?** Should redirect to the `Conflict Resolver` with filters set to show the specific instrument / candidate you clicked on.
1.0
[23.0][behavioural_qc] Link does not redirect to conflict resolver - **Describe the bug** Link to conflict resolver does not work. **To Reproduce** Steps to reproduce the behavior (attach screenshots if applicable): 1. Go to `Behavioural Quality Control` module 2. Click on `Data Conflicts` tab 3. Click on any instrument **What did you expect to happen?** Should redirect to the `Conflict Resolver` with filters set to show the specific instrument / candidate you clicked on.
test
link does not redirect to conflict resolver describe the bug link to conflict resolver does not work to reproduce steps to reproduce the behavior attach screenshots if applicable go to behavioural quality control module click on data conflicts tab click on any instrument what did you expect to happen should redirect to the conflict resolver with filters set to show the specific instrument candidate you clicked on
1
240,082
18,292,473,309
IssuesEvent
2021-10-05 16:40:26
terrapower/armi
https://api.github.com/repos/terrapower/armi
closed
Some gallery examples are broken
bug documentation
Six of the gallery examples show a "broken" thumbnail. https://terrapower.github.io/armi/gallery/index.html Related to #394; some of these examples are calling a `sefl._log()` with arguments that are consistent with the ARMI logger where a standard python logger is expected. ``` Traceback (most recent call last): File "/home/runner/work/armi/armi/doc/gallery-src/framework/run_fuelManagement.py", line 31, in <module> o, reactor = test_reactors.loadTestReactor(inputFileName="refTestCartesian.yaml") File "/home/runner/work/armi/armi/armi/reactor/tests/test_reactors.py", line 176, in loadTestReactor o = operators.factory(cs) File "/home/runner/work/armi/armi/armi/operators/__init__.py", line 52, in factory return getOperatorClassFromSettings(cs)(cs) File "/home/runner/work/armi/armi/armi/operators/operator.py", line 110, in __init__ reportingUtils.writeWelcomeHeaders(self, cs) File "/home/runner/work/armi/armi/armi/bookkeeping/report/reportingUtils.py", line 201, in writeWelcomeHeaders _writeCaseInformation(o, cs) File "/home/runner/work/armi/armi/armi/bookkeeping/report/reportingUtils.py", line 85, in _writeCaseInformation runLog.header("=========== Case Information ===========") File "/home/runner/work/armi/armi/armi/runLog.py", line 384, in header LOG.log("header", msg, single=single, label=label) File "/home/runner/work/armi/armi/armi/runLog.py", line 163, in log self.logger.log(msgLevel, msg, single=single, label=label) File "/opt/hostedtoolcache/Python/3.8.11/x64/lib/python3.8/logging/__init__.py", line 1512, in log self._log(level, msg, args, **kwargs) TypeError: _log() got an unexpected keyword argument 'single' ```
1.0
Some gallery examples are broken - Six of the gallery examples show a "broken" thumbnail. https://terrapower.github.io/armi/gallery/index.html Related to #394; some of these examples are calling a `sefl._log()` with arguments that are consistent with the ARMI logger where a standard python logger is expected. ``` Traceback (most recent call last): File "/home/runner/work/armi/armi/doc/gallery-src/framework/run_fuelManagement.py", line 31, in <module> o, reactor = test_reactors.loadTestReactor(inputFileName="refTestCartesian.yaml") File "/home/runner/work/armi/armi/armi/reactor/tests/test_reactors.py", line 176, in loadTestReactor o = operators.factory(cs) File "/home/runner/work/armi/armi/armi/operators/__init__.py", line 52, in factory return getOperatorClassFromSettings(cs)(cs) File "/home/runner/work/armi/armi/armi/operators/operator.py", line 110, in __init__ reportingUtils.writeWelcomeHeaders(self, cs) File "/home/runner/work/armi/armi/armi/bookkeeping/report/reportingUtils.py", line 201, in writeWelcomeHeaders _writeCaseInformation(o, cs) File "/home/runner/work/armi/armi/armi/bookkeeping/report/reportingUtils.py", line 85, in _writeCaseInformation runLog.header("=========== Case Information ===========") File "/home/runner/work/armi/armi/armi/runLog.py", line 384, in header LOG.log("header", msg, single=single, label=label) File "/home/runner/work/armi/armi/armi/runLog.py", line 163, in log self.logger.log(msgLevel, msg, single=single, label=label) File "/opt/hostedtoolcache/Python/3.8.11/x64/lib/python3.8/logging/__init__.py", line 1512, in log self._log(level, msg, args, **kwargs) TypeError: _log() got an unexpected keyword argument 'single' ```
non_test
some gallery examples are broken six of the gallery examples show a broken thumbnail related to some of these examples are calling a sefl log with arguments that are consistent with the armi logger where a standard python logger is expected traceback most recent call last file home runner work armi armi doc gallery src framework run fuelmanagement py line in o reactor test reactors loadtestreactor inputfilename reftestcartesian yaml file home runner work armi armi armi reactor tests test reactors py line in loadtestreactor o operators factory cs file home runner work armi armi armi operators init py line in factory return getoperatorclassfromsettings cs cs file home runner work armi armi armi operators operator py line in init reportingutils writewelcomeheaders self cs file home runner work armi armi armi bookkeeping report reportingutils py line in writewelcomeheaders writecaseinformation o cs file home runner work armi armi armi bookkeeping report reportingutils py line in writecaseinformation runlog header case information file home runner work armi armi armi runlog py line in header log log header msg single single label label file home runner work armi armi armi runlog py line in log self logger log msglevel msg single single label label file opt hostedtoolcache python lib logging init py line in log self log level msg args kwargs typeerror log got an unexpected keyword argument single
0
131,149
10,682,767,134
IssuesEvent
2019-10-22 06:37:09
apache/incubator-shardingsphere
https://api.github.com/repos/apache/incubator-shardingsphere
closed
Add more unit test cases for sharding-core-preprocessor module
test
sharding-core-preprocessor module is decouple with Sharding & encrypt. The codes has already stable. Should Add more unit test cases for this module.
1.0
Add more unit test cases for sharding-core-preprocessor module - sharding-core-preprocessor module is decouple with Sharding & encrypt. The codes has already stable. Should Add more unit test cases for this module.
test
add more unit test cases for sharding core preprocessor module sharding core preprocessor module is decouple with sharding encrypt the codes has already stable should add more unit test cases for this module
1
185,564
15,025,901,059
IssuesEvent
2021-02-01 21:47:59
choderalab/perses
https://api.github.com/repos/choderalab/perses
opened
Update Install Instructions
documentation
It looks like the `conda-forge` channel is necessary for installation, without it, we get this really confusing error from conda: ``` ➜ conda install -c omnia perses Collecting package metadata (current_repodata.json): done Solving environment: failed with initial frozen solve. Retrying with flexible solve. Solving environment: failed with repodata from current_repodata.json, will retry with next repodata source. Collecting package metadata (repodata.json): done Solving environment: failed with initial frozen solve. Retrying with flexible solve. Solving environment: | Found conflicts! Looking for incompatible packages. This can take several minutes. Press CTRL-C to abort. failed UnsatisfiableError: The following specifications were found to be incompatible with each other: Output in format: Requested package -> Available versions ``` (I've never seen this error fail so hard that it doesn't even say which package it can't find) When I add `-c conda-forge` it works: <details> <summary>Click to expand!</summary> ``` ➜ conda install -c omnia -c conda-forge perses Collecting package metadata (current_repodata.json): done Solving environment: done ## Package Plan ## environment location: /home/mmh/miniconda3/envs/perses added / updated specs: - perses The following packages will be downloaded: package | build ---------------------------|----------------- alabaster-0.7.12 | py_0 15 KB conda-forge ambertools-18.0 | 0 114.9 MB omnia argon2-cffi-20.1.0 | py37h4abf009_2 47 KB conda-forge astunparse-1.6.3 | pyhd8ed1ab_0 15 KB conda-forge async_generator-1.10 | py_0 18 KB conda-forge attrs-20.3.0 | pyhd3deb0d_0 41 KB conda-forge babel-2.9.0 | pyhd3deb0d_0 6.2 MB conda-forge backports-1.0 | py_2 4 KB conda-forge backports.functools_lru_cache-1.6.1| py_0 8 KB conda-forge blas-1.0 | mkl 6 KB bleach-3.2.3 | pyh44b312d_0 111 KB conda-forge blosc-1.20.1 | hd408876_0 71 KB bokeh-1.4.0 | py37hc8dfbb8_1 13.5 MB conda-forge brotlipy-0.7.0 |py37hb5d75c8_1001 346 KB conda-forge bzip2-1.0.8 | h516909a_3 398 KB conda-forge c-ares-1.17.1 | h36c2ea0_0 111 KB conda-forge ca-certificates-2020.12.5 | ha878542_0 137 KB conda-forge cerberus-1.1 | py_1 37 KB omnia certifi-2020.12.5 | py37h89c1867_1 143 KB conda-forge cffi-1.14.4 | py37h261ae71_0 224 KB cftime-1.2.1 | py37h161383b_1 279 KB conda-forge chardet-4.0.0 | py37h89c1867_1 204 KB conda-forge click-7.1.2 | pyh9f0ad1d_0 64 KB conda-forge cloudpickle-1.6.0 | py_0 22 KB conda-forge clusterutils-0.3.1 | py37_1 17 KB omnia colorama-0.4.4 | pyh9f0ad1d_0 18 KB conda-forge cryptography-3.3.1 | py37h3c74f83_0 565 KB curl-7.71.1 | he644dc0_8 139 KB conda-forge cycler-0.10.0 | py_2 9 KB conda-forge cython-0.29.21 | py37hb892b2f_1 2.1 MB conda-forge cytoolz-0.11.0 | py37h4abf009_1 368 KB conda-forge dask-2021.1.1 | pyhd8ed1ab_0 4 KB conda-forge dask-core-2021.1.1 | pyhd8ed1ab_0 674 KB conda-forge dbus-1.13.6 | he372182_0 602 KB conda-forge decorator-4.4.2 | py_0 11 KB conda-forge defusedxml-0.6.0 | py_0 22 KB conda-forge distributed-2021.1.1 | py37h89c1867_0 1.1 MB conda-forge docopt-0.6.2 | py_1 14 KB conda-forge docutils-0.16 | py37h89c1867_3 739 KB conda-forge entrypoints-0.3 | pyhd8ed1ab_1003 8 KB conda-forge expat-2.2.10 | he6710b0_2 153 KB fftw3f-3.3.4 | 2 1.2 MB omnia fontconfig-2.13.1 | he4413a7_1000 327 KB conda-forge freetype-2.10.4 | h7ca028e_0 912 KB conda-forge fsspec-0.8.5 | pyhd8ed1ab_0 69 KB conda-forge glib-2.66.1 | h92f7085_0 2.9 MB gst-plugins-base-1.14.0 | hbbd80ab_1 4.8 MB gstreamer-1.14.0 | h28cd5cc_2 3.2 MB hdf4-4.2.13 | hf30be14_1003 964 KB conda-forge hdf5-1.10.6 |nompi_h7c3c948_1111 3.1 MB conda-forge heapdict-1.0.1 | py_0 7 KB conda-forge icu-58.2 | hf484d3e_1000 22.6 MB conda-forge idna-2.10 | pyh9f0ad1d_0 52 KB conda-forge imagesize-1.2.0 | py_0 8 KB conda-forge importlib-metadata-3.4.0 | py37h89c1867_0 20 KB conda-forge importlib_metadata-3.4.0 | hd8ed1ab_0 3 KB conda-forge intel-openmp-2020.2 | 254 786 KB ipykernel-5.4.3 | py37h888b3d9_0 165 KB conda-forge ipython-5.8.0 | py37_1 1.1 MB conda-forge ipython_genutils-0.2.0 | py_1 21 KB conda-forge ipywidgets-7.6.3 | pyhd3deb0d_0 101 KB conda-forge jinja2-2.11.3 | pyh44b312d_0 93 KB conda-forge jpeg-9d | h36c2ea0_0 264 KB conda-forge jsonschema-3.2.0 | py_2 45 KB conda-forge jupyter-1.0.0 | py37h89c1867_6 6 KB conda-forge jupyter_client-6.1.11 | pyhd8ed1ab_1 76 KB conda-forge jupyter_console-5.2.0 | py37_1 34 KB conda-forge jupyter_core-4.7.1 | py37h89c1867_0 72 KB conda-forge jupyterlab_pygments-0.1.2 | pyh9f0ad1d_0 8 KB conda-forge jupyterlab_widgets-1.0.0 | pyhd8ed1ab_1 130 KB conda-forge kiwisolver-1.3.1 | py37hc928c03_0 86 KB conda-forge krb5-1.17.2 | h926e7f8_0 1.4 MB conda-forge latexcodec-2.0.1 | pyh9f0ad1d_0 18 KB conda-forge libcurl-7.71.1 | hcdd3856_8 312 KB conda-forge libev-4.33 | h516909a_1 104 KB conda-forge libgcc-7.2.0 | h69d50b8_2 304 KB conda-forge libgfortran-ng-7.5.0 | h14aa051_18 22 KB conda-forge libgfortran4-7.5.0 | h14aa051_18 1.3 MB conda-forge libllvm10-10.0.1 | he513fc3_3 26.4 MB conda-forge libnetcdf-4.7.4 |nompi_h56d31a8_107 1.3 MB conda-forge libnghttp2-1.41.0 | h8cfc5f6_2 774 KB conda-forge libpng-1.6.37 | h21135ba_2 306 KB conda-forge libsodium-1.0.18 | h36c2ea0_1 366 KB conda-forge libssh2-1.9.0 | hab1572f_5 225 KB conda-forge libtiff-4.0.10 | hc3755c2_1005 602 KB conda-forge libuuid-2.32.1 | h14c3975_1000 26 KB conda-forge libxcb-1.13 | h14c3975_1002 396 KB conda-forge libxml2-2.9.10 | hb55368b_3 1.2 MB libxslt-1.1.34 | hc22bd24_0 432 KB llvmlite-0.34.0 | py37h269e1b5_4 297 KB locket-0.2.0 | py_2 6 KB conda-forge lxml-4.6.2 | py37h9120a33_0 1.2 MB lz4-c-1.9.2 | he1b5a44_3 203 KB conda-forge lzo-2.10 | h516909a_1000 314 KB conda-forge markupsafe-1.1.1 | py37hb5d75c8_2 27 KB conda-forge matplotlib-3.3.2 | h06a4308_0 24 KB matplotlib-base-3.3.2 | py37h817c723_0 5.1 MB mdtraj-1.9.5 | py37h113d463_0 1.7 MB conda-forge mistune-0.8.4 |py37h4abf009_1002 54 KB conda-forge mkl-2020.2 | 256 138.3 MB mkl-service-2.3.0 | py37h8f50634_2 54 KB conda-forge mkl_fft-1.2.0 | py37h161383b_1 164 KB conda-forge mkl_random-1.2.0 | py37h9fdb41a_1 313 KB conda-forge mock-4.0.3 | py37h89c1867_1 51 KB conda-forge mpich-3.2 | 4 5.3 MB omnia mpiplus-v0.0.1 |py37hc8dfbb8_1002 23 KB conda-forge msgpack-python-1.0.1 | py37hff7bd54_0 82 KB nbclient-0.5.1 | py_0 60 KB conda-forge nbconvert-6.0.7 | py37h89c1867_3 535 KB conda-forge nbformat-5.1.2 | pyhd8ed1ab_1 66 KB conda-forge nest-asyncio-1.4.3 | pyhd8ed1ab_0 9 KB conda-forge netcdf4-1.5.3 |nompi_py37hdc49583_105 547 KB conda-forge networkx-2.5 | py_0 1.2 MB conda-forge notebook-6.2.0 | py37h89c1867_0 6.2 MB conda-forge numba-0.51.2 | py37h9fdb41a_0 3.6 MB conda-forge numexpr-2.7.2 | py37hb2eb853_0 184 KB numpy-1.19.2 | py37h54aff64_0 22 KB numpy-base-1.19.2 | py37hfa32c7d_0 4.1 MB numpydoc-1.1.0 | py_1 42 KB conda-forge olefile-0.46 | pyh9f0ad1d_1 32 KB conda-forge openmm-7.4.2 |py37_cuda101_rc_1 11.9 MB omnia openmmtools-0.20.0 | py37_0 9.0 MB omnia openmoltools-0.8.5 | py37_0 4.3 MB omnia packaging-20.8 | pyhd3deb0d_0 34 KB conda-forge pandas-1.1.4 | py37h10a2094_0 10.5 MB conda-forge pandoc-2.11.4 | h7f98852_0 17.9 MB conda-forge pandocfilters-1.4.2 | py_1 9 KB conda-forge parmed-3.2.0 | py37_0 1.1 MB omnia partd-1.1.0 | py_0 17 KB conda-forge patsy-0.5.1 | py_0 187 KB conda-forge pcre-8.44 | he1b5a44_0 261 KB conda-forge pdbfixer-1.6 | py_1 167 KB omnia perses-0.5.0 | py37_0 1.6 MB omnia pexpect-4.8.0 | pyh9f0ad1d_2 47 KB conda-forge pickleshare-0.7.5 | py_1003 9 KB conda-forge pillow-6.2.1 | py37h6b7be26_0 637 KB conda-forge progressbar2-3.53.1 | pyh9f0ad1d_0 25 KB conda-forge prometheus_client-0.9.0 | pyhd3deb0d_0 44 KB conda-forge prompt_toolkit-1.0.15 | py_1 154 KB conda-forge psutil-5.4.8 |py37h14c3975_1000 312 KB conda-forge pthread-stubs-0.4 | h36c2ea0_1001 5 KB conda-forge ptyprocess-0.7.0 | pyhd3deb0d_0 16 KB conda-forge pybtex-0.24.0 | py37h89c1867_0 305 KB conda-forge pybtex-docutils-1.0.0 | py37h89c1867_0 10 KB conda-forge pycparser-2.20 | pyh9f0ad1d_2 94 KB conda-forge pygments-2.7.4 | pyhd8ed1ab_0 708 KB conda-forge pymbar-3.0.5 | py37ha21ca33_1 119 KB conda-forge pyopenssl-20.0.1 | pyhd8ed1ab_0 48 KB conda-forge pyparsing-2.4.7 | pyh9f0ad1d_0 60 KB conda-forge pyqt-5.9.2 | py37hcca6a23_4 5.7 MB conda-forge pyrsistent-0.17.3 | py37h4abf009_1 89 KB conda-forge pysocks-1.7.1 | py37h89c1867_3 27 KB conda-forge pytables-3.6.1 | py37he17a9a8_3 1.4 MB conda-forge python-dateutil-2.8.1 | py_0 220 KB conda-forge python-utils-2.5.5 | pyh44b312d_0 15 KB conda-forge python_abi-3.7 | 1_cp37m 4 KB conda-forge pytz-2021.1 | pyhd8ed1ab_0 239 KB conda-forge pyyaml-5.3.1 | py37hb5d75c8_1 185 KB conda-forge pyzmq-20.0.0 | py37h5a562af_1 507 KB conda-forge qt-5.9.7 | h5867ecd_1 68.5 MB qtconsole-5.0.2 | pyhd8ed1ab_0 89 KB conda-forge qtpy-1.9.0 | py_0 34 KB conda-forge requests-2.25.1 | pyhd3deb0d_0 51 KB conda-forge scipy-1.5.2 | py37h0b6359f_0 14.3 MB seaborn-0.11.1 | hd8ed1ab_1 4 KB conda-forge seaborn-base-0.11.1 | pyhd8ed1ab_1 217 KB conda-forge send2trash-1.5.0 | py_0 12 KB conda-forge simplegeneric-0.8.1 | py_1 7 KB conda-forge sip-4.19.8 | py37hf484d3e_0 274 KB six-1.15.0 | pyh9f0ad1d_0 14 KB conda-forge snappy-1.1.8 | he1b5a44_3 32 KB conda-forge snowballstemmer-2.1.0 | pyhd8ed1ab_0 57 KB conda-forge sortedcontainers-2.3.0 | pyhd8ed1ab_0 26 KB conda-forge sphinx-3.4.3 | pyhd8ed1ab_0 1.5 MB conda-forge sphinxcontrib-applehelp-1.0.2| py_0 28 KB conda-forge sphinxcontrib-bibtex-2.1.4 | pyhd8ed1ab_0 17 KB conda-forge sphinxcontrib-devhelp-1.0.2| py_0 22 KB conda-forge sphinxcontrib-htmlhelp-1.0.3| py_0 27 KB conda-forge sphinxcontrib-jsmath-1.0.1 | py_0 7 KB conda-forge sphinxcontrib-qthelp-1.0.3 | py_0 25 KB conda-forge sphinxcontrib-serializinghtml-1.1.4| py_0 24 KB conda-forge statsmodels-0.12.1 | py37ha21ca33_1 11.0 MB conda-forge tblib-1.6.0 | py_0 14 KB conda-forge terminado-0.9.2 | py37h89c1867_0 26 KB conda-forge testpath-0.4.4 | py_0 85 KB conda-forge toolz-0.11.1 | py_0 46 KB conda-forge tornado-6.1 | py37h4abf009_0 645 KB conda-forge traitlets-5.0.5 | py_0 81 KB conda-forge typing_extensions-3.7.4.3 | py_0 25 KB conda-forge urllib3-1.26.3 | pyhd8ed1ab_0 99 KB conda-forge wcwidth-0.2.5 | pyh9f0ad1d_2 33 KB conda-forge webencodings-0.5.1 | py_1 12 KB conda-forge widgetsnbextension-3.5.1 | py37h89c1867_4 1.8 MB conda-forge xorg-libxau-1.0.9 | h14c3975_0 13 KB conda-forge xorg-libxdmcp-1.1.3 | h516909a_0 18 KB conda-forge yaml-0.2.5 | h516909a_0 82 KB conda-forge yank-0.25.2 | py37_0 2.3 MB omnia zeromq-4.3.3 | h58526e2_3 288 KB conda-forge zict-2.0.0 | py_0 10 KB conda-forge zipp-3.4.0 | py_0 11 KB conda-forge zstd-1.4.5 | h9ceee32_0 619 KB ------------------------------------------------------------ Total: 561.6 MB The following NEW packages will be INSTALLED: alabaster conda-forge/noarch::alabaster-0.7.12-py_0 ambertools omnia/linux-64::ambertools-18.0-0 argon2-cffi conda-forge/linux-64::argon2-cffi-20.1.0-py37h4abf009_2 astunparse conda-forge/noarch::astunparse-1.6.3-pyhd8ed1ab_0 async_generator conda-forge/noarch::async_generator-1.10-py_0 attrs conda-forge/noarch::attrs-20.3.0-pyhd3deb0d_0 babel conda-forge/noarch::babel-2.9.0-pyhd3deb0d_0 backports conda-forge/noarch::backports-1.0-py_2 backports.functoo~ conda-forge/noarch::backports.functools_lru_cache-1.6.1-py_0 blas pkgs/main/linux-64::blas-1.0-mkl bleach conda-forge/noarch::bleach-3.2.3-pyh44b312d_0 blosc pkgs/main/linux-64::blosc-1.20.1-hd408876_0 bokeh conda-forge/linux-64::bokeh-1.4.0-py37hc8dfbb8_1 brotlipy conda-forge/linux-64::brotlipy-0.7.0-py37hb5d75c8_1001 bzip2 conda-forge/linux-64::bzip2-1.0.8-h516909a_3 c-ares conda-forge/linux-64::c-ares-1.17.1-h36c2ea0_0 cerberus omnia/noarch::cerberus-1.1-py_1 cffi pkgs/main/linux-64::cffi-1.14.4-py37h261ae71_0 cftime conda-forge/linux-64::cftime-1.2.1-py37h161383b_1 chardet conda-forge/linux-64::chardet-4.0.0-py37h89c1867_1 click conda-forge/noarch::click-7.1.2-pyh9f0ad1d_0 cloudpickle conda-forge/noarch::cloudpickle-1.6.0-py_0 clusterutils omnia/linux-64::clusterutils-0.3.1-py37_1 colorama conda-forge/noarch::colorama-0.4.4-pyh9f0ad1d_0 cryptography pkgs/main/linux-64::cryptography-3.3.1-py37h3c74f83_0 curl conda-forge/linux-64::curl-7.71.1-he644dc0_8 cycler conda-forge/noarch::cycler-0.10.0-py_2 cython conda-forge/linux-64::cython-0.29.21-py37hb892b2f_1 cytoolz conda-forge/linux-64::cytoolz-0.11.0-py37h4abf009_1 dask conda-forge/noarch::dask-2021.1.1-pyhd8ed1ab_0 dask-core conda-forge/noarch::dask-core-2021.1.1-pyhd8ed1ab_0 dbus conda-forge/linux-64::dbus-1.13.6-he372182_0 decorator conda-forge/noarch::decorator-4.4.2-py_0 defusedxml conda-forge/noarch::defusedxml-0.6.0-py_0 distributed conda-forge/linux-64::distributed-2021.1.1-py37h89c1867_0 docopt conda-forge/noarch::docopt-0.6.2-py_1 docutils conda-forge/linux-64::docutils-0.16-py37h89c1867_3 entrypoints conda-forge/noarch::entrypoints-0.3-pyhd8ed1ab_1003 expat pkgs/main/linux-64::expat-2.2.10-he6710b0_2 fftw3f omnia/linux-64::fftw3f-3.3.4-2 fontconfig conda-forge/linux-64::fontconfig-2.13.1-he4413a7_1000 freetype conda-forge/linux-64::freetype-2.10.4-h7ca028e_0 fsspec conda-forge/noarch::fsspec-0.8.5-pyhd8ed1ab_0 glib pkgs/main/linux-64::glib-2.66.1-h92f7085_0 gst-plugins-base pkgs/main/linux-64::gst-plugins-base-1.14.0-hbbd80ab_1 gstreamer pkgs/main/linux-64::gstreamer-1.14.0-h28cd5cc_2 hdf4 conda-forge/linux-64::hdf4-4.2.13-hf30be14_1003 hdf5 conda-forge/linux-64::hdf5-1.10.6-nompi_h7c3c948_1111 heapdict conda-forge/noarch::heapdict-1.0.1-py_0 icu conda-forge/linux-64::icu-58.2-hf484d3e_1000 idna conda-forge/noarch::idna-2.10-pyh9f0ad1d_0 imagesize conda-forge/noarch::imagesize-1.2.0-py_0 importlib-metadata conda-forge/linux-64::importlib-metadata-3.4.0-py37h89c1867_0 importlib_metadata conda-forge/noarch::importlib_metadata-3.4.0-hd8ed1ab_0 intel-openmp pkgs/main/linux-64::intel-openmp-2020.2-254 ipykernel conda-forge/linux-64::ipykernel-5.4.3-py37h888b3d9_0 ipython conda-forge/linux-64::ipython-5.8.0-py37_1 ipython_genutils conda-forge/noarch::ipython_genutils-0.2.0-py_1 ipywidgets conda-forge/noarch::ipywidgets-7.6.3-pyhd3deb0d_0 jinja2 conda-forge/noarch::jinja2-2.11.3-pyh44b312d_0 jpeg conda-forge/linux-64::jpeg-9d-h36c2ea0_0 jsonschema conda-forge/noarch::jsonschema-3.2.0-py_2 jupyter conda-forge/linux-64::jupyter-1.0.0-py37h89c1867_6 jupyter_client conda-forge/noarch::jupyter_client-6.1.11-pyhd8ed1ab_1 jupyter_console conda-forge/linux-64::jupyter_console-5.2.0-py37_1 jupyter_core conda-forge/linux-64::jupyter_core-4.7.1-py37h89c1867_0 jupyterlab_pygmen~ conda-forge/noarch::jupyterlab_pygments-0.1.2-pyh9f0ad1d_0 jupyterlab_widgets conda-forge/noarch::jupyterlab_widgets-1.0.0-pyhd8ed1ab_1 kiwisolver conda-forge/linux-64::kiwisolver-1.3.1-py37hc928c03_0 krb5 conda-forge/linux-64::krb5-1.17.2-h926e7f8_0 latexcodec conda-forge/noarch::latexcodec-2.0.1-pyh9f0ad1d_0 libcurl conda-forge/linux-64::libcurl-7.71.1-hcdd3856_8 libev conda-forge/linux-64::libev-4.33-h516909a_1 libgcc conda-forge/linux-64::libgcc-7.2.0-h69d50b8_2 libgfortran-ng conda-forge/linux-64::libgfortran-ng-7.5.0-h14aa051_18 libgfortran4 conda-forge/linux-64::libgfortran4-7.5.0-h14aa051_18 libllvm10 conda-forge/linux-64::libllvm10-10.0.1-he513fc3_3 libnetcdf conda-forge/linux-64::libnetcdf-4.7.4-nompi_h56d31a8_107 libnghttp2 conda-forge/linux-64::libnghttp2-1.41.0-h8cfc5f6_2 libpng conda-forge/linux-64::libpng-1.6.37-h21135ba_2 libsodium conda-forge/linux-64::libsodium-1.0.18-h36c2ea0_1 libssh2 conda-forge/linux-64::libssh2-1.9.0-hab1572f_5 libtiff conda-forge/linux-64::libtiff-4.0.10-hc3755c2_1005 libuuid conda-forge/linux-64::libuuid-2.32.1-h14c3975_1000 libxcb conda-forge/linux-64::libxcb-1.13-h14c3975_1002 libxml2 pkgs/main/linux-64::libxml2-2.9.10-hb55368b_3 libxslt pkgs/main/linux-64::libxslt-1.1.34-hc22bd24_0 llvmlite pkgs/main/linux-64::llvmlite-0.34.0-py37h269e1b5_4 locket conda-forge/noarch::locket-0.2.0-py_2 lxml pkgs/main/linux-64::lxml-4.6.2-py37h9120a33_0 lz4-c conda-forge/linux-64::lz4-c-1.9.2-he1b5a44_3 lzo conda-forge/linux-64::lzo-2.10-h516909a_1000 markupsafe conda-forge/linux-64::markupsafe-1.1.1-py37hb5d75c8_2 matplotlib pkgs/main/linux-64::matplotlib-3.3.2-h06a4308_0 matplotlib-base pkgs/main/linux-64::matplotlib-base-3.3.2-py37h817c723_0 mdtraj conda-forge/linux-64::mdtraj-1.9.5-py37h113d463_0 mistune conda-forge/linux-64::mistune-0.8.4-py37h4abf009_1002 mkl pkgs/main/linux-64::mkl-2020.2-256 mkl-service conda-forge/linux-64::mkl-service-2.3.0-py37h8f50634_2 mkl_fft conda-forge/linux-64::mkl_fft-1.2.0-py37h161383b_1 mkl_random conda-forge/linux-64::mkl_random-1.2.0-py37h9fdb41a_1 mock conda-forge/linux-64::mock-4.0.3-py37h89c1867_1 mpich omnia/linux-64::mpich-3.2-4 mpiplus conda-forge/linux-64::mpiplus-v0.0.1-py37hc8dfbb8_1002 msgpack-python pkgs/main/linux-64::msgpack-python-1.0.1-py37hff7bd54_0 nbclient conda-forge/noarch::nbclient-0.5.1-py_0 nbconvert conda-forge/linux-64::nbconvert-6.0.7-py37h89c1867_3 nbformat conda-forge/noarch::nbformat-5.1.2-pyhd8ed1ab_1 nest-asyncio conda-forge/noarch::nest-asyncio-1.4.3-pyhd8ed1ab_0 netcdf4 conda-forge/linux-64::netcdf4-1.5.3-nompi_py37hdc49583_105 networkx conda-forge/noarch::networkx-2.5-py_0 notebook conda-forge/linux-64::notebook-6.2.0-py37h89c1867_0 numba conda-forge/linux-64::numba-0.51.2-py37h9fdb41a_0 numexpr pkgs/main/linux-64::numexpr-2.7.2-py37hb2eb853_0 numpy pkgs/main/linux-64::numpy-1.19.2-py37h54aff64_0 numpy-base pkgs/main/linux-64::numpy-base-1.19.2-py37hfa32c7d_0 numpydoc conda-forge/noarch::numpydoc-1.1.0-py_1 olefile conda-forge/noarch::olefile-0.46-pyh9f0ad1d_1 openmm omnia/linux-64::openmm-7.4.2-py37_cuda101_rc_1 openmmtools omnia/linux-64::openmmtools-0.20.0-py37_0 openmoltools omnia/linux-64::openmoltools-0.8.5-py37_0 packaging conda-forge/noarch::packaging-20.8-pyhd3deb0d_0 pandas conda-forge/linux-64::pandas-1.1.4-py37h10a2094_0 pandoc conda-forge/linux-64::pandoc-2.11.4-h7f98852_0 pandocfilters conda-forge/noarch::pandocfilters-1.4.2-py_1 parmed omnia/linux-64::parmed-3.2.0-py37_0 partd conda-forge/noarch::partd-1.1.0-py_0 patsy conda-forge/noarch::patsy-0.5.1-py_0 pcre conda-forge/linux-64::pcre-8.44-he1b5a44_0 pdbfixer omnia/noarch::pdbfixer-1.6-py_1 perses omnia/linux-64::perses-0.5.0-py37_0 pexpect conda-forge/noarch::pexpect-4.8.0-pyh9f0ad1d_2 pickleshare conda-forge/noarch::pickleshare-0.7.5-py_1003 pillow conda-forge/linux-64::pillow-6.2.1-py37h6b7be26_0 progressbar2 conda-forge/noarch::progressbar2-3.53.1-pyh9f0ad1d_0 prometheus_client conda-forge/noarch::prometheus_client-0.9.0-pyhd3deb0d_0 prompt_toolkit conda-forge/noarch::prompt_toolkit-1.0.15-py_1 psutil conda-forge/linux-64::psutil-5.4.8-py37h14c3975_1000 pthread-stubs conda-forge/linux-64::pthread-stubs-0.4-h36c2ea0_1001 ptyprocess conda-forge/noarch::ptyprocess-0.7.0-pyhd3deb0d_0 pybtex conda-forge/linux-64::pybtex-0.24.0-py37h89c1867_0 pybtex-docutils conda-forge/linux-64::pybtex-docutils-1.0.0-py37h89c1867_0 pycparser conda-forge/noarch::pycparser-2.20-pyh9f0ad1d_2 pygments conda-forge/noarch::pygments-2.7.4-pyhd8ed1ab_0 pymbar conda-forge/linux-64::pymbar-3.0.5-py37ha21ca33_1 pyopenssl conda-forge/noarch::pyopenssl-20.0.1-pyhd8ed1ab_0 pyparsing conda-forge/noarch::pyparsing-2.4.7-pyh9f0ad1d_0 pyqt conda-forge/linux-64::pyqt-5.9.2-py37hcca6a23_4 pyrsistent conda-forge/linux-64::pyrsistent-0.17.3-py37h4abf009_1 pysocks conda-forge/linux-64::pysocks-1.7.1-py37h89c1867_3 pytables conda-forge/linux-64::pytables-3.6.1-py37he17a9a8_3 python-dateutil conda-forge/noarch::python-dateutil-2.8.1-py_0 python-utils conda-forge/noarch::python-utils-2.5.5-pyh44b312d_0 python_abi conda-forge/linux-64::python_abi-3.7-1_cp37m pytz conda-forge/noarch::pytz-2021.1-pyhd8ed1ab_0 pyyaml conda-forge/linux-64::pyyaml-5.3.1-py37hb5d75c8_1 pyzmq conda-forge/linux-64::pyzmq-20.0.0-py37h5a562af_1 qt pkgs/main/linux-64::qt-5.9.7-h5867ecd_1 qtconsole conda-forge/noarch::qtconsole-5.0.2-pyhd8ed1ab_0 qtpy conda-forge/noarch::qtpy-1.9.0-py_0 requests conda-forge/noarch::requests-2.25.1-pyhd3deb0d_0 scipy pkgs/main/linux-64::scipy-1.5.2-py37h0b6359f_0 seaborn conda-forge/noarch::seaborn-0.11.1-hd8ed1ab_1 seaborn-base conda-forge/noarch::seaborn-base-0.11.1-pyhd8ed1ab_1 send2trash conda-forge/noarch::send2trash-1.5.0-py_0 simplegeneric conda-forge/noarch::simplegeneric-0.8.1-py_1 sip pkgs/main/linux-64::sip-4.19.8-py37hf484d3e_0 six conda-forge/noarch::six-1.15.0-pyh9f0ad1d_0 snappy conda-forge/linux-64::snappy-1.1.8-he1b5a44_3 snowballstemmer conda-forge/noarch::snowballstemmer-2.1.0-pyhd8ed1ab_0 sortedcontainers conda-forge/noarch::sortedcontainers-2.3.0-pyhd8ed1ab_0 sphinx conda-forge/noarch::sphinx-3.4.3-pyhd8ed1ab_0 sphinxcontrib-app~ conda-forge/noarch::sphinxcontrib-applehelp-1.0.2-py_0 sphinxcontrib-bib~ conda-forge/noarch::sphinxcontrib-bibtex-2.1.4-pyhd8ed1ab_0 sphinxcontrib-dev~ conda-forge/noarch::sphinxcontrib-devhelp-1.0.2-py_0 sphinxcontrib-htm~ conda-forge/noarch::sphinxcontrib-htmlhelp-1.0.3-py_0 sphinxcontrib-jsm~ conda-forge/noarch::sphinxcontrib-jsmath-1.0.1-py_0 sphinxcontrib-qth~ conda-forge/noarch::sphinxcontrib-qthelp-1.0.3-py_0 sphinxcontrib-ser~ conda-forge/noarch::sphinxcontrib-serializinghtml-1.1.4-py_0 statsmodels conda-forge/linux-64::statsmodels-0.12.1-py37ha21ca33_1 tblib conda-forge/noarch::tblib-1.6.0-py_0 terminado conda-forge/linux-64::terminado-0.9.2-py37h89c1867_0 testpath conda-forge/noarch::testpath-0.4.4-py_0 toolz conda-forge/noarch::toolz-0.11.1-py_0 tornado conda-forge/linux-64::tornado-6.1-py37h4abf009_0 traitlets conda-forge/noarch::traitlets-5.0.5-py_0 typing_extensions conda-forge/noarch::typing_extensions-3.7.4.3-py_0 urllib3 conda-forge/noarch::urllib3-1.26.3-pyhd8ed1ab_0 wcwidth conda-forge/noarch::wcwidth-0.2.5-pyh9f0ad1d_2 webencodings conda-forge/noarch::webencodings-0.5.1-py_1 widgetsnbextension conda-forge/linux-64::widgetsnbextension-3.5.1-py37h89c1867_4 xorg-libxau conda-forge/linux-64::xorg-libxau-1.0.9-h14c3975_0 xorg-libxdmcp conda-forge/linux-64::xorg-libxdmcp-1.1.3-h516909a_0 yaml conda-forge/linux-64::yaml-0.2.5-h516909a_0 yank omnia/linux-64::yank-0.25.2-py37_0 zeromq conda-forge/linux-64::zeromq-4.3.3-h58526e2_3 zict conda-forge/noarch::zict-2.0.0-py_0 zipp conda-forge/noarch::zipp-3.4.0-py_0 zstd pkgs/main/linux-64::zstd-1.4.5-h9ceee32_0 The following packages will be UPDATED: certifi pkgs/main::certifi-2020.12.5-py37h06a~ --> conda-forge::certifi-2020.12.5-py37h89c1867_1 The following packages will be SUPERSEDED by a higher-priority channel: ca-certificates pkgs/main::ca-certificates-2021.1.19-~ --> conda-forge::ca-certificates-2020.12.5-ha878542_0 Proceed ([y]/n)? ``` </details> I'll use this issue to document other changes that might need to happen as I get things setup locally
1.0
Update Install Instructions - It looks like the `conda-forge` channel is necessary for installation, without it, we get this really confusing error from conda: ``` ➜ conda install -c omnia perses Collecting package metadata (current_repodata.json): done Solving environment: failed with initial frozen solve. Retrying with flexible solve. Solving environment: failed with repodata from current_repodata.json, will retry with next repodata source. Collecting package metadata (repodata.json): done Solving environment: failed with initial frozen solve. Retrying with flexible solve. Solving environment: | Found conflicts! Looking for incompatible packages. This can take several minutes. Press CTRL-C to abort. failed UnsatisfiableError: The following specifications were found to be incompatible with each other: Output in format: Requested package -> Available versions ``` (I've never seen this error fail so hard that it doesn't even say which package it can't find) When I add `-c conda-forge` it works: <details> <summary>Click to expand!</summary> ``` ➜ conda install -c omnia -c conda-forge perses Collecting package metadata (current_repodata.json): done Solving environment: done ## Package Plan ## environment location: /home/mmh/miniconda3/envs/perses added / updated specs: - perses The following packages will be downloaded: package | build ---------------------------|----------------- alabaster-0.7.12 | py_0 15 KB conda-forge ambertools-18.0 | 0 114.9 MB omnia argon2-cffi-20.1.0 | py37h4abf009_2 47 KB conda-forge astunparse-1.6.3 | pyhd8ed1ab_0 15 KB conda-forge async_generator-1.10 | py_0 18 KB conda-forge attrs-20.3.0 | pyhd3deb0d_0 41 KB conda-forge babel-2.9.0 | pyhd3deb0d_0 6.2 MB conda-forge backports-1.0 | py_2 4 KB conda-forge backports.functools_lru_cache-1.6.1| py_0 8 KB conda-forge blas-1.0 | mkl 6 KB bleach-3.2.3 | pyh44b312d_0 111 KB conda-forge blosc-1.20.1 | hd408876_0 71 KB bokeh-1.4.0 | py37hc8dfbb8_1 13.5 MB conda-forge brotlipy-0.7.0 |py37hb5d75c8_1001 346 KB conda-forge bzip2-1.0.8 | h516909a_3 398 KB conda-forge c-ares-1.17.1 | h36c2ea0_0 111 KB conda-forge ca-certificates-2020.12.5 | ha878542_0 137 KB conda-forge cerberus-1.1 | py_1 37 KB omnia certifi-2020.12.5 | py37h89c1867_1 143 KB conda-forge cffi-1.14.4 | py37h261ae71_0 224 KB cftime-1.2.1 | py37h161383b_1 279 KB conda-forge chardet-4.0.0 | py37h89c1867_1 204 KB conda-forge click-7.1.2 | pyh9f0ad1d_0 64 KB conda-forge cloudpickle-1.6.0 | py_0 22 KB conda-forge clusterutils-0.3.1 | py37_1 17 KB omnia colorama-0.4.4 | pyh9f0ad1d_0 18 KB conda-forge cryptography-3.3.1 | py37h3c74f83_0 565 KB curl-7.71.1 | he644dc0_8 139 KB conda-forge cycler-0.10.0 | py_2 9 KB conda-forge cython-0.29.21 | py37hb892b2f_1 2.1 MB conda-forge cytoolz-0.11.0 | py37h4abf009_1 368 KB conda-forge dask-2021.1.1 | pyhd8ed1ab_0 4 KB conda-forge dask-core-2021.1.1 | pyhd8ed1ab_0 674 KB conda-forge dbus-1.13.6 | he372182_0 602 KB conda-forge decorator-4.4.2 | py_0 11 KB conda-forge defusedxml-0.6.0 | py_0 22 KB conda-forge distributed-2021.1.1 | py37h89c1867_0 1.1 MB conda-forge docopt-0.6.2 | py_1 14 KB conda-forge docutils-0.16 | py37h89c1867_3 739 KB conda-forge entrypoints-0.3 | pyhd8ed1ab_1003 8 KB conda-forge expat-2.2.10 | he6710b0_2 153 KB fftw3f-3.3.4 | 2 1.2 MB omnia fontconfig-2.13.1 | he4413a7_1000 327 KB conda-forge freetype-2.10.4 | h7ca028e_0 912 KB conda-forge fsspec-0.8.5 | pyhd8ed1ab_0 69 KB conda-forge glib-2.66.1 | h92f7085_0 2.9 MB gst-plugins-base-1.14.0 | hbbd80ab_1 4.8 MB gstreamer-1.14.0 | h28cd5cc_2 3.2 MB hdf4-4.2.13 | hf30be14_1003 964 KB conda-forge hdf5-1.10.6 |nompi_h7c3c948_1111 3.1 MB conda-forge heapdict-1.0.1 | py_0 7 KB conda-forge icu-58.2 | hf484d3e_1000 22.6 MB conda-forge idna-2.10 | pyh9f0ad1d_0 52 KB conda-forge imagesize-1.2.0 | py_0 8 KB conda-forge importlib-metadata-3.4.0 | py37h89c1867_0 20 KB conda-forge importlib_metadata-3.4.0 | hd8ed1ab_0 3 KB conda-forge intel-openmp-2020.2 | 254 786 KB ipykernel-5.4.3 | py37h888b3d9_0 165 KB conda-forge ipython-5.8.0 | py37_1 1.1 MB conda-forge ipython_genutils-0.2.0 | py_1 21 KB conda-forge ipywidgets-7.6.3 | pyhd3deb0d_0 101 KB conda-forge jinja2-2.11.3 | pyh44b312d_0 93 KB conda-forge jpeg-9d | h36c2ea0_0 264 KB conda-forge jsonschema-3.2.0 | py_2 45 KB conda-forge jupyter-1.0.0 | py37h89c1867_6 6 KB conda-forge jupyter_client-6.1.11 | pyhd8ed1ab_1 76 KB conda-forge jupyter_console-5.2.0 | py37_1 34 KB conda-forge jupyter_core-4.7.1 | py37h89c1867_0 72 KB conda-forge jupyterlab_pygments-0.1.2 | pyh9f0ad1d_0 8 KB conda-forge jupyterlab_widgets-1.0.0 | pyhd8ed1ab_1 130 KB conda-forge kiwisolver-1.3.1 | py37hc928c03_0 86 KB conda-forge krb5-1.17.2 | h926e7f8_0 1.4 MB conda-forge latexcodec-2.0.1 | pyh9f0ad1d_0 18 KB conda-forge libcurl-7.71.1 | hcdd3856_8 312 KB conda-forge libev-4.33 | h516909a_1 104 KB conda-forge libgcc-7.2.0 | h69d50b8_2 304 KB conda-forge libgfortran-ng-7.5.0 | h14aa051_18 22 KB conda-forge libgfortran4-7.5.0 | h14aa051_18 1.3 MB conda-forge libllvm10-10.0.1 | he513fc3_3 26.4 MB conda-forge libnetcdf-4.7.4 |nompi_h56d31a8_107 1.3 MB conda-forge libnghttp2-1.41.0 | h8cfc5f6_2 774 KB conda-forge libpng-1.6.37 | h21135ba_2 306 KB conda-forge libsodium-1.0.18 | h36c2ea0_1 366 KB conda-forge libssh2-1.9.0 | hab1572f_5 225 KB conda-forge libtiff-4.0.10 | hc3755c2_1005 602 KB conda-forge libuuid-2.32.1 | h14c3975_1000 26 KB conda-forge libxcb-1.13 | h14c3975_1002 396 KB conda-forge libxml2-2.9.10 | hb55368b_3 1.2 MB libxslt-1.1.34 | hc22bd24_0 432 KB llvmlite-0.34.0 | py37h269e1b5_4 297 KB locket-0.2.0 | py_2 6 KB conda-forge lxml-4.6.2 | py37h9120a33_0 1.2 MB lz4-c-1.9.2 | he1b5a44_3 203 KB conda-forge lzo-2.10 | h516909a_1000 314 KB conda-forge markupsafe-1.1.1 | py37hb5d75c8_2 27 KB conda-forge matplotlib-3.3.2 | h06a4308_0 24 KB matplotlib-base-3.3.2 | py37h817c723_0 5.1 MB mdtraj-1.9.5 | py37h113d463_0 1.7 MB conda-forge mistune-0.8.4 |py37h4abf009_1002 54 KB conda-forge mkl-2020.2 | 256 138.3 MB mkl-service-2.3.0 | py37h8f50634_2 54 KB conda-forge mkl_fft-1.2.0 | py37h161383b_1 164 KB conda-forge mkl_random-1.2.0 | py37h9fdb41a_1 313 KB conda-forge mock-4.0.3 | py37h89c1867_1 51 KB conda-forge mpich-3.2 | 4 5.3 MB omnia mpiplus-v0.0.1 |py37hc8dfbb8_1002 23 KB conda-forge msgpack-python-1.0.1 | py37hff7bd54_0 82 KB nbclient-0.5.1 | py_0 60 KB conda-forge nbconvert-6.0.7 | py37h89c1867_3 535 KB conda-forge nbformat-5.1.2 | pyhd8ed1ab_1 66 KB conda-forge nest-asyncio-1.4.3 | pyhd8ed1ab_0 9 KB conda-forge netcdf4-1.5.3 |nompi_py37hdc49583_105 547 KB conda-forge networkx-2.5 | py_0 1.2 MB conda-forge notebook-6.2.0 | py37h89c1867_0 6.2 MB conda-forge numba-0.51.2 | py37h9fdb41a_0 3.6 MB conda-forge numexpr-2.7.2 | py37hb2eb853_0 184 KB numpy-1.19.2 | py37h54aff64_0 22 KB numpy-base-1.19.2 | py37hfa32c7d_0 4.1 MB numpydoc-1.1.0 | py_1 42 KB conda-forge olefile-0.46 | pyh9f0ad1d_1 32 KB conda-forge openmm-7.4.2 |py37_cuda101_rc_1 11.9 MB omnia openmmtools-0.20.0 | py37_0 9.0 MB omnia openmoltools-0.8.5 | py37_0 4.3 MB omnia packaging-20.8 | pyhd3deb0d_0 34 KB conda-forge pandas-1.1.4 | py37h10a2094_0 10.5 MB conda-forge pandoc-2.11.4 | h7f98852_0 17.9 MB conda-forge pandocfilters-1.4.2 | py_1 9 KB conda-forge parmed-3.2.0 | py37_0 1.1 MB omnia partd-1.1.0 | py_0 17 KB conda-forge patsy-0.5.1 | py_0 187 KB conda-forge pcre-8.44 | he1b5a44_0 261 KB conda-forge pdbfixer-1.6 | py_1 167 KB omnia perses-0.5.0 | py37_0 1.6 MB omnia pexpect-4.8.0 | pyh9f0ad1d_2 47 KB conda-forge pickleshare-0.7.5 | py_1003 9 KB conda-forge pillow-6.2.1 | py37h6b7be26_0 637 KB conda-forge progressbar2-3.53.1 | pyh9f0ad1d_0 25 KB conda-forge prometheus_client-0.9.0 | pyhd3deb0d_0 44 KB conda-forge prompt_toolkit-1.0.15 | py_1 154 KB conda-forge psutil-5.4.8 |py37h14c3975_1000 312 KB conda-forge pthread-stubs-0.4 | h36c2ea0_1001 5 KB conda-forge ptyprocess-0.7.0 | pyhd3deb0d_0 16 KB conda-forge pybtex-0.24.0 | py37h89c1867_0 305 KB conda-forge pybtex-docutils-1.0.0 | py37h89c1867_0 10 KB conda-forge pycparser-2.20 | pyh9f0ad1d_2 94 KB conda-forge pygments-2.7.4 | pyhd8ed1ab_0 708 KB conda-forge pymbar-3.0.5 | py37ha21ca33_1 119 KB conda-forge pyopenssl-20.0.1 | pyhd8ed1ab_0 48 KB conda-forge pyparsing-2.4.7 | pyh9f0ad1d_0 60 KB conda-forge pyqt-5.9.2 | py37hcca6a23_4 5.7 MB conda-forge pyrsistent-0.17.3 | py37h4abf009_1 89 KB conda-forge pysocks-1.7.1 | py37h89c1867_3 27 KB conda-forge pytables-3.6.1 | py37he17a9a8_3 1.4 MB conda-forge python-dateutil-2.8.1 | py_0 220 KB conda-forge python-utils-2.5.5 | pyh44b312d_0 15 KB conda-forge python_abi-3.7 | 1_cp37m 4 KB conda-forge pytz-2021.1 | pyhd8ed1ab_0 239 KB conda-forge pyyaml-5.3.1 | py37hb5d75c8_1 185 KB conda-forge pyzmq-20.0.0 | py37h5a562af_1 507 KB conda-forge qt-5.9.7 | h5867ecd_1 68.5 MB qtconsole-5.0.2 | pyhd8ed1ab_0 89 KB conda-forge qtpy-1.9.0 | py_0 34 KB conda-forge requests-2.25.1 | pyhd3deb0d_0 51 KB conda-forge scipy-1.5.2 | py37h0b6359f_0 14.3 MB seaborn-0.11.1 | hd8ed1ab_1 4 KB conda-forge seaborn-base-0.11.1 | pyhd8ed1ab_1 217 KB conda-forge send2trash-1.5.0 | py_0 12 KB conda-forge simplegeneric-0.8.1 | py_1 7 KB conda-forge sip-4.19.8 | py37hf484d3e_0 274 KB six-1.15.0 | pyh9f0ad1d_0 14 KB conda-forge snappy-1.1.8 | he1b5a44_3 32 KB conda-forge snowballstemmer-2.1.0 | pyhd8ed1ab_0 57 KB conda-forge sortedcontainers-2.3.0 | pyhd8ed1ab_0 26 KB conda-forge sphinx-3.4.3 | pyhd8ed1ab_0 1.5 MB conda-forge sphinxcontrib-applehelp-1.0.2| py_0 28 KB conda-forge sphinxcontrib-bibtex-2.1.4 | pyhd8ed1ab_0 17 KB conda-forge sphinxcontrib-devhelp-1.0.2| py_0 22 KB conda-forge sphinxcontrib-htmlhelp-1.0.3| py_0 27 KB conda-forge sphinxcontrib-jsmath-1.0.1 | py_0 7 KB conda-forge sphinxcontrib-qthelp-1.0.3 | py_0 25 KB conda-forge sphinxcontrib-serializinghtml-1.1.4| py_0 24 KB conda-forge statsmodels-0.12.1 | py37ha21ca33_1 11.0 MB conda-forge tblib-1.6.0 | py_0 14 KB conda-forge terminado-0.9.2 | py37h89c1867_0 26 KB conda-forge testpath-0.4.4 | py_0 85 KB conda-forge toolz-0.11.1 | py_0 46 KB conda-forge tornado-6.1 | py37h4abf009_0 645 KB conda-forge traitlets-5.0.5 | py_0 81 KB conda-forge typing_extensions-3.7.4.3 | py_0 25 KB conda-forge urllib3-1.26.3 | pyhd8ed1ab_0 99 KB conda-forge wcwidth-0.2.5 | pyh9f0ad1d_2 33 KB conda-forge webencodings-0.5.1 | py_1 12 KB conda-forge widgetsnbextension-3.5.1 | py37h89c1867_4 1.8 MB conda-forge xorg-libxau-1.0.9 | h14c3975_0 13 KB conda-forge xorg-libxdmcp-1.1.3 | h516909a_0 18 KB conda-forge yaml-0.2.5 | h516909a_0 82 KB conda-forge yank-0.25.2 | py37_0 2.3 MB omnia zeromq-4.3.3 | h58526e2_3 288 KB conda-forge zict-2.0.0 | py_0 10 KB conda-forge zipp-3.4.0 | py_0 11 KB conda-forge zstd-1.4.5 | h9ceee32_0 619 KB ------------------------------------------------------------ Total: 561.6 MB The following NEW packages will be INSTALLED: alabaster conda-forge/noarch::alabaster-0.7.12-py_0 ambertools omnia/linux-64::ambertools-18.0-0 argon2-cffi conda-forge/linux-64::argon2-cffi-20.1.0-py37h4abf009_2 astunparse conda-forge/noarch::astunparse-1.6.3-pyhd8ed1ab_0 async_generator conda-forge/noarch::async_generator-1.10-py_0 attrs conda-forge/noarch::attrs-20.3.0-pyhd3deb0d_0 babel conda-forge/noarch::babel-2.9.0-pyhd3deb0d_0 backports conda-forge/noarch::backports-1.0-py_2 backports.functoo~ conda-forge/noarch::backports.functools_lru_cache-1.6.1-py_0 blas pkgs/main/linux-64::blas-1.0-mkl bleach conda-forge/noarch::bleach-3.2.3-pyh44b312d_0 blosc pkgs/main/linux-64::blosc-1.20.1-hd408876_0 bokeh conda-forge/linux-64::bokeh-1.4.0-py37hc8dfbb8_1 brotlipy conda-forge/linux-64::brotlipy-0.7.0-py37hb5d75c8_1001 bzip2 conda-forge/linux-64::bzip2-1.0.8-h516909a_3 c-ares conda-forge/linux-64::c-ares-1.17.1-h36c2ea0_0 cerberus omnia/noarch::cerberus-1.1-py_1 cffi pkgs/main/linux-64::cffi-1.14.4-py37h261ae71_0 cftime conda-forge/linux-64::cftime-1.2.1-py37h161383b_1 chardet conda-forge/linux-64::chardet-4.0.0-py37h89c1867_1 click conda-forge/noarch::click-7.1.2-pyh9f0ad1d_0 cloudpickle conda-forge/noarch::cloudpickle-1.6.0-py_0 clusterutils omnia/linux-64::clusterutils-0.3.1-py37_1 colorama conda-forge/noarch::colorama-0.4.4-pyh9f0ad1d_0 cryptography pkgs/main/linux-64::cryptography-3.3.1-py37h3c74f83_0 curl conda-forge/linux-64::curl-7.71.1-he644dc0_8 cycler conda-forge/noarch::cycler-0.10.0-py_2 cython conda-forge/linux-64::cython-0.29.21-py37hb892b2f_1 cytoolz conda-forge/linux-64::cytoolz-0.11.0-py37h4abf009_1 dask conda-forge/noarch::dask-2021.1.1-pyhd8ed1ab_0 dask-core conda-forge/noarch::dask-core-2021.1.1-pyhd8ed1ab_0 dbus conda-forge/linux-64::dbus-1.13.6-he372182_0 decorator conda-forge/noarch::decorator-4.4.2-py_0 defusedxml conda-forge/noarch::defusedxml-0.6.0-py_0 distributed conda-forge/linux-64::distributed-2021.1.1-py37h89c1867_0 docopt conda-forge/noarch::docopt-0.6.2-py_1 docutils conda-forge/linux-64::docutils-0.16-py37h89c1867_3 entrypoints conda-forge/noarch::entrypoints-0.3-pyhd8ed1ab_1003 expat pkgs/main/linux-64::expat-2.2.10-he6710b0_2 fftw3f omnia/linux-64::fftw3f-3.3.4-2 fontconfig conda-forge/linux-64::fontconfig-2.13.1-he4413a7_1000 freetype conda-forge/linux-64::freetype-2.10.4-h7ca028e_0 fsspec conda-forge/noarch::fsspec-0.8.5-pyhd8ed1ab_0 glib pkgs/main/linux-64::glib-2.66.1-h92f7085_0 gst-plugins-base pkgs/main/linux-64::gst-plugins-base-1.14.0-hbbd80ab_1 gstreamer pkgs/main/linux-64::gstreamer-1.14.0-h28cd5cc_2 hdf4 conda-forge/linux-64::hdf4-4.2.13-hf30be14_1003 hdf5 conda-forge/linux-64::hdf5-1.10.6-nompi_h7c3c948_1111 heapdict conda-forge/noarch::heapdict-1.0.1-py_0 icu conda-forge/linux-64::icu-58.2-hf484d3e_1000 idna conda-forge/noarch::idna-2.10-pyh9f0ad1d_0 imagesize conda-forge/noarch::imagesize-1.2.0-py_0 importlib-metadata conda-forge/linux-64::importlib-metadata-3.4.0-py37h89c1867_0 importlib_metadata conda-forge/noarch::importlib_metadata-3.4.0-hd8ed1ab_0 intel-openmp pkgs/main/linux-64::intel-openmp-2020.2-254 ipykernel conda-forge/linux-64::ipykernel-5.4.3-py37h888b3d9_0 ipython conda-forge/linux-64::ipython-5.8.0-py37_1 ipython_genutils conda-forge/noarch::ipython_genutils-0.2.0-py_1 ipywidgets conda-forge/noarch::ipywidgets-7.6.3-pyhd3deb0d_0 jinja2 conda-forge/noarch::jinja2-2.11.3-pyh44b312d_0 jpeg conda-forge/linux-64::jpeg-9d-h36c2ea0_0 jsonschema conda-forge/noarch::jsonschema-3.2.0-py_2 jupyter conda-forge/linux-64::jupyter-1.0.0-py37h89c1867_6 jupyter_client conda-forge/noarch::jupyter_client-6.1.11-pyhd8ed1ab_1 jupyter_console conda-forge/linux-64::jupyter_console-5.2.0-py37_1 jupyter_core conda-forge/linux-64::jupyter_core-4.7.1-py37h89c1867_0 jupyterlab_pygmen~ conda-forge/noarch::jupyterlab_pygments-0.1.2-pyh9f0ad1d_0 jupyterlab_widgets conda-forge/noarch::jupyterlab_widgets-1.0.0-pyhd8ed1ab_1 kiwisolver conda-forge/linux-64::kiwisolver-1.3.1-py37hc928c03_0 krb5 conda-forge/linux-64::krb5-1.17.2-h926e7f8_0 latexcodec conda-forge/noarch::latexcodec-2.0.1-pyh9f0ad1d_0 libcurl conda-forge/linux-64::libcurl-7.71.1-hcdd3856_8 libev conda-forge/linux-64::libev-4.33-h516909a_1 libgcc conda-forge/linux-64::libgcc-7.2.0-h69d50b8_2 libgfortran-ng conda-forge/linux-64::libgfortran-ng-7.5.0-h14aa051_18 libgfortran4 conda-forge/linux-64::libgfortran4-7.5.0-h14aa051_18 libllvm10 conda-forge/linux-64::libllvm10-10.0.1-he513fc3_3 libnetcdf conda-forge/linux-64::libnetcdf-4.7.4-nompi_h56d31a8_107 libnghttp2 conda-forge/linux-64::libnghttp2-1.41.0-h8cfc5f6_2 libpng conda-forge/linux-64::libpng-1.6.37-h21135ba_2 libsodium conda-forge/linux-64::libsodium-1.0.18-h36c2ea0_1 libssh2 conda-forge/linux-64::libssh2-1.9.0-hab1572f_5 libtiff conda-forge/linux-64::libtiff-4.0.10-hc3755c2_1005 libuuid conda-forge/linux-64::libuuid-2.32.1-h14c3975_1000 libxcb conda-forge/linux-64::libxcb-1.13-h14c3975_1002 libxml2 pkgs/main/linux-64::libxml2-2.9.10-hb55368b_3 libxslt pkgs/main/linux-64::libxslt-1.1.34-hc22bd24_0 llvmlite pkgs/main/linux-64::llvmlite-0.34.0-py37h269e1b5_4 locket conda-forge/noarch::locket-0.2.0-py_2 lxml pkgs/main/linux-64::lxml-4.6.2-py37h9120a33_0 lz4-c conda-forge/linux-64::lz4-c-1.9.2-he1b5a44_3 lzo conda-forge/linux-64::lzo-2.10-h516909a_1000 markupsafe conda-forge/linux-64::markupsafe-1.1.1-py37hb5d75c8_2 matplotlib pkgs/main/linux-64::matplotlib-3.3.2-h06a4308_0 matplotlib-base pkgs/main/linux-64::matplotlib-base-3.3.2-py37h817c723_0 mdtraj conda-forge/linux-64::mdtraj-1.9.5-py37h113d463_0 mistune conda-forge/linux-64::mistune-0.8.4-py37h4abf009_1002 mkl pkgs/main/linux-64::mkl-2020.2-256 mkl-service conda-forge/linux-64::mkl-service-2.3.0-py37h8f50634_2 mkl_fft conda-forge/linux-64::mkl_fft-1.2.0-py37h161383b_1 mkl_random conda-forge/linux-64::mkl_random-1.2.0-py37h9fdb41a_1 mock conda-forge/linux-64::mock-4.0.3-py37h89c1867_1 mpich omnia/linux-64::mpich-3.2-4 mpiplus conda-forge/linux-64::mpiplus-v0.0.1-py37hc8dfbb8_1002 msgpack-python pkgs/main/linux-64::msgpack-python-1.0.1-py37hff7bd54_0 nbclient conda-forge/noarch::nbclient-0.5.1-py_0 nbconvert conda-forge/linux-64::nbconvert-6.0.7-py37h89c1867_3 nbformat conda-forge/noarch::nbformat-5.1.2-pyhd8ed1ab_1 nest-asyncio conda-forge/noarch::nest-asyncio-1.4.3-pyhd8ed1ab_0 netcdf4 conda-forge/linux-64::netcdf4-1.5.3-nompi_py37hdc49583_105 networkx conda-forge/noarch::networkx-2.5-py_0 notebook conda-forge/linux-64::notebook-6.2.0-py37h89c1867_0 numba conda-forge/linux-64::numba-0.51.2-py37h9fdb41a_0 numexpr pkgs/main/linux-64::numexpr-2.7.2-py37hb2eb853_0 numpy pkgs/main/linux-64::numpy-1.19.2-py37h54aff64_0 numpy-base pkgs/main/linux-64::numpy-base-1.19.2-py37hfa32c7d_0 numpydoc conda-forge/noarch::numpydoc-1.1.0-py_1 olefile conda-forge/noarch::olefile-0.46-pyh9f0ad1d_1 openmm omnia/linux-64::openmm-7.4.2-py37_cuda101_rc_1 openmmtools omnia/linux-64::openmmtools-0.20.0-py37_0 openmoltools omnia/linux-64::openmoltools-0.8.5-py37_0 packaging conda-forge/noarch::packaging-20.8-pyhd3deb0d_0 pandas conda-forge/linux-64::pandas-1.1.4-py37h10a2094_0 pandoc conda-forge/linux-64::pandoc-2.11.4-h7f98852_0 pandocfilters conda-forge/noarch::pandocfilters-1.4.2-py_1 parmed omnia/linux-64::parmed-3.2.0-py37_0 partd conda-forge/noarch::partd-1.1.0-py_0 patsy conda-forge/noarch::patsy-0.5.1-py_0 pcre conda-forge/linux-64::pcre-8.44-he1b5a44_0 pdbfixer omnia/noarch::pdbfixer-1.6-py_1 perses omnia/linux-64::perses-0.5.0-py37_0 pexpect conda-forge/noarch::pexpect-4.8.0-pyh9f0ad1d_2 pickleshare conda-forge/noarch::pickleshare-0.7.5-py_1003 pillow conda-forge/linux-64::pillow-6.2.1-py37h6b7be26_0 progressbar2 conda-forge/noarch::progressbar2-3.53.1-pyh9f0ad1d_0 prometheus_client conda-forge/noarch::prometheus_client-0.9.0-pyhd3deb0d_0 prompt_toolkit conda-forge/noarch::prompt_toolkit-1.0.15-py_1 psutil conda-forge/linux-64::psutil-5.4.8-py37h14c3975_1000 pthread-stubs conda-forge/linux-64::pthread-stubs-0.4-h36c2ea0_1001 ptyprocess conda-forge/noarch::ptyprocess-0.7.0-pyhd3deb0d_0 pybtex conda-forge/linux-64::pybtex-0.24.0-py37h89c1867_0 pybtex-docutils conda-forge/linux-64::pybtex-docutils-1.0.0-py37h89c1867_0 pycparser conda-forge/noarch::pycparser-2.20-pyh9f0ad1d_2 pygments conda-forge/noarch::pygments-2.7.4-pyhd8ed1ab_0 pymbar conda-forge/linux-64::pymbar-3.0.5-py37ha21ca33_1 pyopenssl conda-forge/noarch::pyopenssl-20.0.1-pyhd8ed1ab_0 pyparsing conda-forge/noarch::pyparsing-2.4.7-pyh9f0ad1d_0 pyqt conda-forge/linux-64::pyqt-5.9.2-py37hcca6a23_4 pyrsistent conda-forge/linux-64::pyrsistent-0.17.3-py37h4abf009_1 pysocks conda-forge/linux-64::pysocks-1.7.1-py37h89c1867_3 pytables conda-forge/linux-64::pytables-3.6.1-py37he17a9a8_3 python-dateutil conda-forge/noarch::python-dateutil-2.8.1-py_0 python-utils conda-forge/noarch::python-utils-2.5.5-pyh44b312d_0 python_abi conda-forge/linux-64::python_abi-3.7-1_cp37m pytz conda-forge/noarch::pytz-2021.1-pyhd8ed1ab_0 pyyaml conda-forge/linux-64::pyyaml-5.3.1-py37hb5d75c8_1 pyzmq conda-forge/linux-64::pyzmq-20.0.0-py37h5a562af_1 qt pkgs/main/linux-64::qt-5.9.7-h5867ecd_1 qtconsole conda-forge/noarch::qtconsole-5.0.2-pyhd8ed1ab_0 qtpy conda-forge/noarch::qtpy-1.9.0-py_0 requests conda-forge/noarch::requests-2.25.1-pyhd3deb0d_0 scipy pkgs/main/linux-64::scipy-1.5.2-py37h0b6359f_0 seaborn conda-forge/noarch::seaborn-0.11.1-hd8ed1ab_1 seaborn-base conda-forge/noarch::seaborn-base-0.11.1-pyhd8ed1ab_1 send2trash conda-forge/noarch::send2trash-1.5.0-py_0 simplegeneric conda-forge/noarch::simplegeneric-0.8.1-py_1 sip pkgs/main/linux-64::sip-4.19.8-py37hf484d3e_0 six conda-forge/noarch::six-1.15.0-pyh9f0ad1d_0 snappy conda-forge/linux-64::snappy-1.1.8-he1b5a44_3 snowballstemmer conda-forge/noarch::snowballstemmer-2.1.0-pyhd8ed1ab_0 sortedcontainers conda-forge/noarch::sortedcontainers-2.3.0-pyhd8ed1ab_0 sphinx conda-forge/noarch::sphinx-3.4.3-pyhd8ed1ab_0 sphinxcontrib-app~ conda-forge/noarch::sphinxcontrib-applehelp-1.0.2-py_0 sphinxcontrib-bib~ conda-forge/noarch::sphinxcontrib-bibtex-2.1.4-pyhd8ed1ab_0 sphinxcontrib-dev~ conda-forge/noarch::sphinxcontrib-devhelp-1.0.2-py_0 sphinxcontrib-htm~ conda-forge/noarch::sphinxcontrib-htmlhelp-1.0.3-py_0 sphinxcontrib-jsm~ conda-forge/noarch::sphinxcontrib-jsmath-1.0.1-py_0 sphinxcontrib-qth~ conda-forge/noarch::sphinxcontrib-qthelp-1.0.3-py_0 sphinxcontrib-ser~ conda-forge/noarch::sphinxcontrib-serializinghtml-1.1.4-py_0 statsmodels conda-forge/linux-64::statsmodels-0.12.1-py37ha21ca33_1 tblib conda-forge/noarch::tblib-1.6.0-py_0 terminado conda-forge/linux-64::terminado-0.9.2-py37h89c1867_0 testpath conda-forge/noarch::testpath-0.4.4-py_0 toolz conda-forge/noarch::toolz-0.11.1-py_0 tornado conda-forge/linux-64::tornado-6.1-py37h4abf009_0 traitlets conda-forge/noarch::traitlets-5.0.5-py_0 typing_extensions conda-forge/noarch::typing_extensions-3.7.4.3-py_0 urllib3 conda-forge/noarch::urllib3-1.26.3-pyhd8ed1ab_0 wcwidth conda-forge/noarch::wcwidth-0.2.5-pyh9f0ad1d_2 webencodings conda-forge/noarch::webencodings-0.5.1-py_1 widgetsnbextension conda-forge/linux-64::widgetsnbextension-3.5.1-py37h89c1867_4 xorg-libxau conda-forge/linux-64::xorg-libxau-1.0.9-h14c3975_0 xorg-libxdmcp conda-forge/linux-64::xorg-libxdmcp-1.1.3-h516909a_0 yaml conda-forge/linux-64::yaml-0.2.5-h516909a_0 yank omnia/linux-64::yank-0.25.2-py37_0 zeromq conda-forge/linux-64::zeromq-4.3.3-h58526e2_3 zict conda-forge/noarch::zict-2.0.0-py_0 zipp conda-forge/noarch::zipp-3.4.0-py_0 zstd pkgs/main/linux-64::zstd-1.4.5-h9ceee32_0 The following packages will be UPDATED: certifi pkgs/main::certifi-2020.12.5-py37h06a~ --> conda-forge::certifi-2020.12.5-py37h89c1867_1 The following packages will be SUPERSEDED by a higher-priority channel: ca-certificates pkgs/main::ca-certificates-2021.1.19-~ --> conda-forge::ca-certificates-2020.12.5-ha878542_0 Proceed ([y]/n)? ``` </details> I'll use this issue to document other changes that might need to happen as I get things setup locally
non_test
update install instructions it looks like the conda forge channel is necessary for installation without it we get this really confusing error from conda ➜ conda install c omnia perses collecting package metadata current repodata json done solving environment failed with initial frozen solve retrying with flexible solve solving environment failed with repodata from current repodata json will retry with next repodata source collecting package metadata repodata json done solving environment failed with initial frozen solve retrying with flexible solve solving environment found conflicts looking for incompatible packages this can take several minutes press ctrl c to abort failed unsatisfiableerror the following specifications were found to be incompatible with each other output in format requested package available versions i ve never seen this error fail so hard that it doesn t even say which package it can t find when i add c conda forge it works click to expand ➜ conda install c omnia c conda forge perses collecting package metadata current repodata json done solving environment done package plan environment location home mmh envs perses added updated specs perses the following packages will be downloaded package build alabaster py kb conda forge ambertools mb omnia cffi kb conda forge astunparse kb conda forge async generator py kb conda forge attrs kb conda forge babel mb conda forge backports py kb conda forge backports functools lru cache py kb conda forge blas mkl kb bleach kb conda forge blosc kb bokeh mb conda forge brotlipy kb conda forge kb conda forge c ares kb conda forge ca certificates kb conda forge cerberus py kb omnia certifi kb conda forge cffi kb cftime kb conda forge chardet kb conda forge click kb conda forge cloudpickle py kb conda forge clusterutils kb omnia colorama kb conda forge cryptography kb curl kb conda forge cycler py kb conda forge cython mb conda forge cytoolz kb conda forge dask kb conda forge dask core kb conda forge dbus kb conda forge decorator py kb conda forge defusedxml py kb conda forge distributed mb conda forge docopt py kb conda forge docutils kb conda forge entrypoints kb conda forge expat kb mb omnia fontconfig kb conda forge freetype kb conda forge fsspec kb conda forge glib mb gst plugins base mb gstreamer mb kb conda forge nompi mb conda forge heapdict py kb conda forge icu mb conda forge idna kb conda forge imagesize py kb conda forge importlib metadata kb conda forge importlib metadata kb conda forge intel openmp kb ipykernel kb conda forge ipython mb conda forge ipython genutils py kb conda forge ipywidgets kb conda forge kb conda forge jpeg kb conda forge jsonschema py kb conda forge jupyter kb conda forge jupyter client kb conda forge jupyter console kb conda forge jupyter core kb conda forge jupyterlab pygments kb conda forge jupyterlab widgets kb conda forge kiwisolver kb conda forge mb conda forge latexcodec kb conda forge libcurl kb conda forge libev kb conda forge libgcc kb conda forge libgfortran ng kb conda forge mb conda forge mb conda forge libnetcdf nompi mb conda forge kb conda forge libpng kb conda forge libsodium kb conda forge kb conda forge libtiff kb conda forge libuuid kb conda forge libxcb kb conda forge mb libxslt kb llvmlite kb locket py kb conda forge lxml mb c kb conda forge lzo kb conda forge markupsafe kb conda forge matplotlib kb matplotlib base mb mdtraj mb conda forge mistune kb conda forge mkl mb mkl service kb conda forge mkl fft kb conda forge mkl random kb conda forge mock kb conda forge mpich mb omnia mpiplus kb conda forge msgpack python kb nbclient py kb conda forge nbconvert kb conda forge nbformat kb conda forge nest asyncio kb conda forge nompi kb conda forge networkx py mb conda forge notebook mb conda forge numba mb conda forge numexpr kb numpy kb numpy base mb numpydoc py kb conda forge olefile kb conda forge openmm rc mb omnia openmmtools mb omnia openmoltools mb omnia packaging kb conda forge pandas mb conda forge pandoc mb conda forge pandocfilters py kb conda forge parmed mb omnia partd py kb conda forge patsy py kb conda forge pcre kb conda forge pdbfixer py kb omnia perses mb omnia pexpect kb conda forge pickleshare py kb conda forge pillow kb conda forge kb conda forge prometheus client kb conda forge prompt toolkit py kb conda forge psutil kb conda forge pthread stubs kb conda forge ptyprocess kb conda forge pybtex kb conda forge pybtex docutils kb conda forge pycparser kb conda forge pygments kb conda forge pymbar kb conda forge pyopenssl kb conda forge pyparsing kb conda forge pyqt mb conda forge pyrsistent kb conda forge pysocks kb conda forge pytables mb conda forge python dateutil py kb conda forge python utils kb conda forge python abi kb conda forge pytz kb conda forge pyyaml kb conda forge pyzmq kb conda forge qt mb qtconsole kb conda forge qtpy py kb conda forge requests kb conda forge scipy mb seaborn kb conda forge seaborn base kb conda forge py kb conda forge simplegeneric py kb conda forge sip kb six kb conda forge snappy kb conda forge snowballstemmer kb conda forge sortedcontainers kb conda forge sphinx mb conda forge sphinxcontrib applehelp py kb conda forge sphinxcontrib bibtex kb conda forge sphinxcontrib devhelp py kb conda forge sphinxcontrib htmlhelp py kb conda forge sphinxcontrib jsmath py kb conda forge sphinxcontrib qthelp py kb conda forge sphinxcontrib serializinghtml py kb conda forge statsmodels mb conda forge tblib py kb conda forge terminado kb conda forge testpath py kb conda forge toolz py kb conda forge tornado kb conda forge traitlets py kb conda forge typing extensions py kb conda forge kb conda forge wcwidth kb conda forge webencodings py kb conda forge widgetsnbextension mb conda forge xorg libxau kb conda forge xorg libxdmcp kb conda forge yaml kb conda forge yank mb omnia zeromq kb conda forge zict py kb conda forge zipp py kb conda forge zstd kb total mb the following new packages will be installed alabaster conda forge noarch alabaster py ambertools omnia linux ambertools cffi conda forge linux cffi astunparse conda forge noarch astunparse async generator conda forge noarch async generator py attrs conda forge noarch attrs babel conda forge noarch babel backports conda forge noarch backports py backports functoo conda forge noarch backports functools lru cache py blas pkgs main linux blas mkl bleach conda forge noarch bleach blosc pkgs main linux blosc bokeh conda forge linux bokeh brotlipy conda forge linux brotlipy conda forge linux c ares conda forge linux c ares cerberus omnia noarch cerberus py cffi pkgs main linux cffi cftime conda forge linux cftime chardet conda forge linux chardet click conda forge noarch click cloudpickle conda forge noarch cloudpickle py clusterutils omnia linux clusterutils colorama conda forge noarch colorama cryptography pkgs main linux cryptography curl conda forge linux curl cycler conda forge noarch cycler py cython conda forge linux cython cytoolz conda forge linux cytoolz dask conda forge noarch dask dask core conda forge noarch dask core dbus conda forge linux dbus decorator conda forge noarch decorator py defusedxml conda forge noarch defusedxml py distributed conda forge linux distributed docopt conda forge noarch docopt py docutils conda forge linux docutils entrypoints conda forge noarch entrypoints expat pkgs main linux expat omnia linux fontconfig conda forge linux fontconfig freetype conda forge linux freetype fsspec conda forge noarch fsspec glib pkgs main linux glib gst plugins base pkgs main linux gst plugins base gstreamer pkgs main linux gstreamer conda forge linux conda forge linux nompi heapdict conda forge noarch heapdict py icu conda forge linux icu idna conda forge noarch idna imagesize conda forge noarch imagesize py importlib metadata conda forge linux importlib metadata importlib metadata conda forge noarch importlib metadata intel openmp pkgs main linux intel openmp ipykernel conda forge linux ipykernel ipython conda forge linux ipython ipython genutils conda forge noarch ipython genutils py ipywidgets conda forge noarch ipywidgets conda forge noarch jpeg conda forge linux jpeg jsonschema conda forge noarch jsonschema py jupyter conda forge linux jupyter jupyter client conda forge noarch jupyter client jupyter console conda forge linux jupyter console jupyter core conda forge linux jupyter core jupyterlab pygmen conda forge noarch jupyterlab pygments jupyterlab widgets conda forge noarch jupyterlab widgets kiwisolver conda forge linux kiwisolver conda forge linux latexcodec conda forge noarch latexcodec libcurl conda forge linux libcurl libev conda forge linux libev libgcc conda forge linux libgcc libgfortran ng conda forge linux libgfortran ng conda forge linux conda forge linux libnetcdf conda forge linux libnetcdf nompi conda forge linux libpng conda forge linux libpng libsodium conda forge linux libsodium conda forge linux libtiff conda forge linux libtiff libuuid conda forge linux libuuid libxcb conda forge linux libxcb pkgs main linux libxslt pkgs main linux libxslt llvmlite pkgs main linux llvmlite locket conda forge noarch locket py lxml pkgs main linux lxml c conda forge linux c lzo conda forge linux lzo markupsafe conda forge linux markupsafe matplotlib pkgs main linux matplotlib matplotlib base pkgs main linux matplotlib base mdtraj conda forge linux mdtraj mistune conda forge linux mistune mkl pkgs main linux mkl mkl service conda forge linux mkl service mkl fft conda forge linux mkl fft mkl random conda forge linux mkl random mock conda forge linux mock mpich omnia linux mpich mpiplus conda forge linux mpiplus msgpack python pkgs main linux msgpack python nbclient conda forge noarch nbclient py nbconvert conda forge linux nbconvert nbformat conda forge noarch nbformat nest asyncio conda forge noarch nest asyncio conda forge linux nompi networkx conda forge noarch networkx py notebook conda forge linux notebook numba conda forge linux numba numexpr pkgs main linux numexpr numpy pkgs main linux numpy numpy base pkgs main linux numpy base numpydoc conda forge noarch numpydoc py olefile conda forge noarch olefile openmm omnia linux openmm rc openmmtools omnia linux openmmtools openmoltools omnia linux openmoltools packaging conda forge noarch packaging pandas conda forge linux pandas pandoc conda forge linux pandoc pandocfilters conda forge noarch pandocfilters py parmed omnia linux parmed partd conda forge noarch partd py patsy conda forge noarch patsy py pcre conda forge linux pcre pdbfixer omnia noarch pdbfixer py perses omnia linux perses pexpect conda forge noarch pexpect pickleshare conda forge noarch pickleshare py pillow conda forge linux pillow conda forge noarch prometheus client conda forge noarch prometheus client prompt toolkit conda forge noarch prompt toolkit py psutil conda forge linux psutil pthread stubs conda forge linux pthread stubs ptyprocess conda forge noarch ptyprocess pybtex conda forge linux pybtex pybtex docutils conda forge linux pybtex docutils pycparser conda forge noarch pycparser pygments conda forge noarch pygments pymbar conda forge linux pymbar pyopenssl conda forge noarch pyopenssl pyparsing conda forge noarch pyparsing pyqt conda forge linux pyqt pyrsistent conda forge linux pyrsistent pysocks conda forge linux pysocks pytables conda forge linux pytables python dateutil conda forge noarch python dateutil py python utils conda forge noarch python utils python abi conda forge linux python abi pytz conda forge noarch pytz pyyaml conda forge linux pyyaml pyzmq conda forge linux pyzmq qt pkgs main linux qt qtconsole conda forge noarch qtconsole qtpy conda forge noarch qtpy py requests conda forge noarch requests scipy pkgs main linux scipy seaborn conda forge noarch seaborn seaborn base conda forge noarch seaborn base conda forge noarch py simplegeneric conda forge noarch simplegeneric py sip pkgs main linux sip six conda forge noarch six snappy conda forge linux snappy snowballstemmer conda forge noarch snowballstemmer sortedcontainers conda forge noarch sortedcontainers sphinx conda forge noarch sphinx sphinxcontrib app conda forge noarch sphinxcontrib applehelp py sphinxcontrib bib conda forge noarch sphinxcontrib bibtex sphinxcontrib dev conda forge noarch sphinxcontrib devhelp py sphinxcontrib htm conda forge noarch sphinxcontrib htmlhelp py sphinxcontrib jsm conda forge noarch sphinxcontrib jsmath py sphinxcontrib qth conda forge noarch sphinxcontrib qthelp py sphinxcontrib ser conda forge noarch sphinxcontrib serializinghtml py statsmodels conda forge linux statsmodels tblib conda forge noarch tblib py terminado conda forge linux terminado testpath conda forge noarch testpath py toolz conda forge noarch toolz py tornado conda forge linux tornado traitlets conda forge noarch traitlets py typing extensions conda forge noarch typing extensions py conda forge noarch wcwidth conda forge noarch wcwidth webencodings conda forge noarch webencodings py widgetsnbextension conda forge linux widgetsnbextension xorg libxau conda forge linux xorg libxau xorg libxdmcp conda forge linux xorg libxdmcp yaml conda forge linux yaml yank omnia linux yank zeromq conda forge linux zeromq zict conda forge noarch zict py zipp conda forge noarch zipp py zstd pkgs main linux zstd the following packages will be updated certifi pkgs main certifi conda forge certifi the following packages will be superseded by a higher priority channel ca certificates pkgs main ca certificates conda forge ca certificates proceed n i ll use this issue to document other changes that might need to happen as i get things setup locally
0
627,423
19,904,648,741
IssuesEvent
2022-01-25 11:29:18
SAP/xsk
https://api.github.com/repos/SAP/xsk
closed
[IDE] Console view not working locally in Safari
bug priority-low IDE shadow
**Describe the bug** Console view in XSK run locally in Docker returns an error `[ERROR] Connection problem! Check security roles assignments.` > What version of the XSK are you using? 0.9.2 **To Reproduce** Steps to reproduce the behavior: 1. Run xsk:0.9.2 locally 2. Open Safari 3. Open XSK 4. See error **Expected behavior** Logs are streamed and the error is not shown **Desktop (please complete the following information):** - OS: MacOS 11.6 - Browser: Safari - Version: 15.0
1.0
[IDE] Console view not working locally in Safari - **Describe the bug** Console view in XSK run locally in Docker returns an error `[ERROR] Connection problem! Check security roles assignments.` > What version of the XSK are you using? 0.9.2 **To Reproduce** Steps to reproduce the behavior: 1. Run xsk:0.9.2 locally 2. Open Safari 3. Open XSK 4. See error **Expected behavior** Logs are streamed and the error is not shown **Desktop (please complete the following information):** - OS: MacOS 11.6 - Browser: Safari - Version: 15.0
non_test
console view not working locally in safari describe the bug console view in xsk run locally in docker returns an error connection problem check security roles assignments what version of the xsk are you using to reproduce steps to reproduce the behavior run xsk locally open safari open xsk see error expected behavior logs are streamed and the error is not shown desktop please complete the following information os macos browser safari version
0
91,317
15,856,395,758
IssuesEvent
2021-04-08 02:14:39
n-devs/testTungTonScript
https://api.github.com/repos/n-devs/testTungTonScript
opened
CVE-2020-7789 (Medium) detected in node-notifier-5.4.0.tgz
security vulnerability
## CVE-2020-7789 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-notifier-5.4.0.tgz</b></p></summary> <p>A Node.js module for sending notifications on native Mac, Windows (post and pre 8) and Linux (or Growl as fallback)</p> <p>Library home page: <a href="https://registry.npmjs.org/node-notifier/-/node-notifier-5.4.0.tgz">https://registry.npmjs.org/node-notifier/-/node-notifier-5.4.0.tgz</a></p> <p>Path to dependency file: /testTungTonScript/package.json</p> <p>Path to vulnerable library: testTungTonScript/node_modules/node-notifier/package.json</p> <p> Dependency Hierarchy: - react-native-0.54.0.tgz (Root Library) - :x: **node-notifier-5.4.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package node-notifier before 9.0.0. It allows an attacker to run arbitrary commands on Linux machines due to the options params not being sanitised when being passed an array. <p>Publish Date: 2020-12-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7789>CVE-2020-7789</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7789">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7789</a></p> <p>Release Date: 2020-12-11</p> <p>Fix Resolution: 9.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7789 (Medium) detected in node-notifier-5.4.0.tgz - ## CVE-2020-7789 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-notifier-5.4.0.tgz</b></p></summary> <p>A Node.js module for sending notifications on native Mac, Windows (post and pre 8) and Linux (or Growl as fallback)</p> <p>Library home page: <a href="https://registry.npmjs.org/node-notifier/-/node-notifier-5.4.0.tgz">https://registry.npmjs.org/node-notifier/-/node-notifier-5.4.0.tgz</a></p> <p>Path to dependency file: /testTungTonScript/package.json</p> <p>Path to vulnerable library: testTungTonScript/node_modules/node-notifier/package.json</p> <p> Dependency Hierarchy: - react-native-0.54.0.tgz (Root Library) - :x: **node-notifier-5.4.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package node-notifier before 9.0.0. It allows an attacker to run arbitrary commands on Linux machines due to the options params not being sanitised when being passed an array. <p>Publish Date: 2020-12-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7789>CVE-2020-7789</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7789">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7789</a></p> <p>Release Date: 2020-12-11</p> <p>Fix Resolution: 9.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in node notifier tgz cve medium severity vulnerability vulnerable library node notifier tgz a node js module for sending notifications on native mac windows post and pre and linux or growl as fallback library home page a href path to dependency file testtungtonscript package json path to vulnerable library testtungtonscript node modules node notifier package json dependency hierarchy react native tgz root library x node notifier tgz vulnerable library vulnerability details this affects the package node notifier before it allows an attacker to run arbitrary commands on linux machines due to the options params not being sanitised when being passed an array publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
500,317
14,495,977,066
IssuesEvent
2020-12-11 12:03:01
enso-org/enso
https://api.github.com/repos/enso-org/enso
opened
Substitute Loggers Warning in Language Server Tests
Category: Tooling Difficulty: Intermediate Priority: Low Type: Bug
<!-- Please ensure that you are running the latest version of Enso before reporting the bug! It may have been fixed since. --> ### General Summary <!-- - Please include a high-level description of your bug here. --> When running `language-server` tests, the SLF4J API issues a warning about substitute loggers access. Ideally all logging should be initialized in such a way that substitute loggers are not accessed. Weirdly, the access seems to not be to our logging service logger, but something Akka specific: `akka.event.slf4j.Slf4jLogger`; oddly enough even though this logger seems to be accessed it does not incur a conflict (as usually having multiple SLF4J backends on the classpath would result in a warning). ### Steps to Reproduce Run `sbt language-server/test`; inspect the output. ### Expected Result <!-- - A description of the results you expected from the reproduction steps. --> The tests run without warnings. ### Actual Result <!-- - A description of what actually happens when you run these steps. - Please include any error output if relevant. --> The tests do run, but the first few lines include the following warning: ``` SLF4J: The following set of substitute loggers may have been accessed SLF4J: during the initialization phase. Logging calls during this SLF4J: phase were not honored. However, subsequent logging calls to these SLF4J: loggers will work as normally expected. SLF4J: See also http://www.slf4j.org/codes.html#substituteLogger SLF4J: akka.event.slf4j.Slf4jLogger ``` ### Enso Version <!-- - Please include the output of `enso --version`. For example: ``` Enso Compiler and Runtime Version: 0.1.0 Built with: scala-2.13.3 for GraalVM 20.2.0 Built from: main* @ 919ffbdfacc44cc35a1b38f1bad5b573acdbe358 Running on: OpenJDK 64-Bit Server VM, GraalVM Community, JDK 11.0.8+10-jvmci-20.2-b03 Linux 4.15.0-112-generic (amd64) ``` --> Reproducible at 75e3136221a0c8b8cc8aaf7396292ab5867b3248 but it has been happening for a longer time.
1.0
Substitute Loggers Warning in Language Server Tests - <!-- Please ensure that you are running the latest version of Enso before reporting the bug! It may have been fixed since. --> ### General Summary <!-- - Please include a high-level description of your bug here. --> When running `language-server` tests, the SLF4J API issues a warning about substitute loggers access. Ideally all logging should be initialized in such a way that substitute loggers are not accessed. Weirdly, the access seems to not be to our logging service logger, but something Akka specific: `akka.event.slf4j.Slf4jLogger`; oddly enough even though this logger seems to be accessed it does not incur a conflict (as usually having multiple SLF4J backends on the classpath would result in a warning). ### Steps to Reproduce Run `sbt language-server/test`; inspect the output. ### Expected Result <!-- - A description of the results you expected from the reproduction steps. --> The tests run without warnings. ### Actual Result <!-- - A description of what actually happens when you run these steps. - Please include any error output if relevant. --> The tests do run, but the first few lines include the following warning: ``` SLF4J: The following set of substitute loggers may have been accessed SLF4J: during the initialization phase. Logging calls during this SLF4J: phase were not honored. However, subsequent logging calls to these SLF4J: loggers will work as normally expected. SLF4J: See also http://www.slf4j.org/codes.html#substituteLogger SLF4J: akka.event.slf4j.Slf4jLogger ``` ### Enso Version <!-- - Please include the output of `enso --version`. For example: ``` Enso Compiler and Runtime Version: 0.1.0 Built with: scala-2.13.3 for GraalVM 20.2.0 Built from: main* @ 919ffbdfacc44cc35a1b38f1bad5b573acdbe358 Running on: OpenJDK 64-Bit Server VM, GraalVM Community, JDK 11.0.8+10-jvmci-20.2-b03 Linux 4.15.0-112-generic (amd64) ``` --> Reproducible at 75e3136221a0c8b8cc8aaf7396292ab5867b3248 but it has been happening for a longer time.
non_test
substitute loggers warning in language server tests please ensure that you are running the latest version of enso before reporting the bug it may have been fixed since general summary please include a high level description of your bug here when running language server tests the api issues a warning about substitute loggers access ideally all logging should be initialized in such a way that substitute loggers are not accessed weirdly the access seems to not be to our logging service logger but something akka specific akka event oddly enough even though this logger seems to be accessed it does not incur a conflict as usually having multiple backends on the classpath would result in a warning steps to reproduce run sbt language server test inspect the output expected result a description of the results you expected from the reproduction steps the tests run without warnings actual result a description of what actually happens when you run these steps please include any error output if relevant the tests do run but the first few lines include the following warning the following set of substitute loggers may have been accessed during the initialization phase logging calls during this phase were not honored however subsequent logging calls to these loggers will work as normally expected see also akka event enso version please include the output of enso version for example enso compiler and runtime version built with scala for graalvm built from main running on openjdk bit server vm graalvm community jdk jvmci linux generic reproducible at but it has been happening for a longer time
0
12,075
7,777,673,641
IssuesEvent
2018-06-05 12:02:16
MoonchildProductions/UXP
https://api.github.com/repos/MoonchildProductions/UXP
opened
Disable CSS Errors logging by default to improve performance
Component: DevTools Performance
By default CSS parser sends hundreds of errors to the Error Console for almost every page. This info is completely useless for the regular browsing and only wastes CPU time and memory. Such load can be quite significant, just open Error Console for evidences. Setting layout.css.report_errors default to false will disable CSS Errors logging. The devs should turn it on only for actual purposes. Extension devs that rely on this should take steps to enable this feature if and when they need it in their extension.
True
Disable CSS Errors logging by default to improve performance - By default CSS parser sends hundreds of errors to the Error Console for almost every page. This info is completely useless for the regular browsing and only wastes CPU time and memory. Such load can be quite significant, just open Error Console for evidences. Setting layout.css.report_errors default to false will disable CSS Errors logging. The devs should turn it on only for actual purposes. Extension devs that rely on this should take steps to enable this feature if and when they need it in their extension.
non_test
disable css errors logging by default to improve performance by default css parser sends hundreds of errors to the error console for almost every page this info is completely useless for the regular browsing and only wastes cpu time and memory such load can be quite significant just open error console for evidences setting layout css report errors default to false will disable css errors logging the devs should turn it on only for actual purposes extension devs that rely on this should take steps to enable this feature if and when they need it in their extension
0
138,797
11,217,655,207
IssuesEvent
2020-01-07 09:44:29
FreeRDP/FreeRDP
https://api.github.com/repos/FreeRDP/FreeRDP
closed
Usb device redirecting not working
channel:usb fixed-waiting-test
Currently pulled. In associate with issue #2078. New log messages not showing fatal handler: signum=11, but it's still crashing. command line: ``` user@ubuntu-test:~/FreeRDP-master$ sudo xfreerdp /u:"user" /p:"password" /v:XXX.XXX.XXX.XXX /f /rfx /usb:id,dev:04b8:014a ``` log is: ``` [10:06:54:033] [6578:6579] [INFO][com.freerdp.client.common.cmdline] - loading channel drdynvc [10:06:54:086] [6578:6579] [ERROR][com.freerdp.locale] - Unable to get current timezone rule [10:06:55:215] [6578:6580] [INFO][com.freerdp.channels.dvcman.client] - Loading Dynamic Virtual Channel urbdrc [10:06:55:224] [6578:6580] [WARN][com.freerdp.channels.urbdrc.client] - bus:0 dev:0 not exist in udevman [10:06:55:224] [6578:6580] [ERROR][com.freerdp.channels.urbdrc.client] - VID: 0x04B8 PID: 0x014A [10:06:55:229] [6578:6580] [WARN][com.freerdp.channels.urbdrc.client] - bus:8 dev:3 not exist in udevman ``` It gave some new information about the problem that udevman is guilty
1.0
Usb device redirecting not working - Currently pulled. In associate with issue #2078. New log messages not showing fatal handler: signum=11, but it's still crashing. command line: ``` user@ubuntu-test:~/FreeRDP-master$ sudo xfreerdp /u:"user" /p:"password" /v:XXX.XXX.XXX.XXX /f /rfx /usb:id,dev:04b8:014a ``` log is: ``` [10:06:54:033] [6578:6579] [INFO][com.freerdp.client.common.cmdline] - loading channel drdynvc [10:06:54:086] [6578:6579] [ERROR][com.freerdp.locale] - Unable to get current timezone rule [10:06:55:215] [6578:6580] [INFO][com.freerdp.channels.dvcman.client] - Loading Dynamic Virtual Channel urbdrc [10:06:55:224] [6578:6580] [WARN][com.freerdp.channels.urbdrc.client] - bus:0 dev:0 not exist in udevman [10:06:55:224] [6578:6580] [ERROR][com.freerdp.channels.urbdrc.client] - VID: 0x04B8 PID: 0x014A [10:06:55:229] [6578:6580] [WARN][com.freerdp.channels.urbdrc.client] - bus:8 dev:3 not exist in udevman ``` It gave some new information about the problem that udevman is guilty
test
usb device redirecting not working currently pulled in associate with issue new log messages not showing fatal handler signum but it s still crashing command line user ubuntu test freerdp master sudo xfreerdp u user p password v xxx xxx xxx xxx f rfx usb id dev log is loading channel drdynvc unable to get current timezone rule loading dynamic virtual channel urbdrc bus dev not exist in udevman vid pid bus dev not exist in udevman it gave some new information about the problem that udevman is guilty
1
824,237
31,146,204,697
IssuesEvent
2023-08-16 06:39:51
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.xfinity.com - see bug description
browser-firefox priority-important engine-gecko
<!-- @browser: Firefox 116.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:109.0) Gecko/20100101 Firefox/116.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/125793 --> **URL**: https://www.xfinity.com **Browser / Version**: Firefox 116.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Edge **Problem type**: Something else **Description**: won't connect. Says SECURE CONNECTION FAILED. PEER'S CERTIFICATE HAS BEEN REVOKED. NOT A PROBLEM ON MICROSOFT EDGE BROWSER **Steps to Reproduce**: WORKS ON MICROSOFT EDGE. PROBLEM OCCURRED 2 DAYS AGO ON FIREFOX. NOTING CHANGED ON FIREFOX EXCEPT AUTOMATIC UPDATE. ALL OTHER WEBSITES WORK <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.xfinity.com - see bug description - <!-- @browser: Firefox 116.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:109.0) Gecko/20100101 Firefox/116.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/125793 --> **URL**: https://www.xfinity.com **Browser / Version**: Firefox 116.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Edge **Problem type**: Something else **Description**: won't connect. Says SECURE CONNECTION FAILED. PEER'S CERTIFICATE HAS BEEN REVOKED. NOT A PROBLEM ON MICROSOFT EDGE BROWSER **Steps to Reproduce**: WORKS ON MICROSOFT EDGE. PROBLEM OCCURRED 2 DAYS AGO ON FIREFOX. NOTING CHANGED ON FIREFOX EXCEPT AUTOMATIC UPDATE. ALL OTHER WEBSITES WORK <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
see bug description url browser version firefox operating system windows tested another browser yes edge problem type something else description won t connect says secure connection failed peer s certificate has been revoked not a problem on microsoft edge browser steps to reproduce works on microsoft edge problem occurred days ago on firefox noting changed on firefox except automatic update all other websites work browser configuration none from with ❤️
0
168,631
13,097,826,332
IssuesEvent
2020-08-03 18:13:14
pkgjs/meet
https://api.github.com/repos/pkgjs/meet
closed
Test Meeting 2020-08-04
meeting test
## Date/Time | Timezone | Date/Time | |----------|-----------| | America/Los_Angeles | Tue 04-Aug-2020 10:00 (10:00 AM) | | America/Denver | Tue 04-Aug-2020 11:00 (11:00 AM) | | America/Chicago | Tue 04-Aug-2020 12:00 (12:00 PM) | | America/New_York | Tue 04-Aug-2020 13:00 (01:00 PM) | | Europe/London | Tue 04-Aug-2020 18:00 (06:00 PM) | | Europe/Amsterdam | Tue 04-Aug-2020 19:00 (07:00 PM) | | Europe/Moscow | Tue 04-Aug-2020 20:00 (08:00 PM) | | Asia/Kolkata | Tue 04-Aug-2020 22:30 (10:30 PM) | | Asia/Shanghai | Wed 05-Aug-2020 01:00 (01:00 AM) | | Asia/Tokyo | Wed 05-Aug-2020 02:00 (02:00 AM) | | Australia/Sydney | Wed 05-Aug-2020 03:00 (03:00 AM) | Or in your local time: * https://www.timeanddate.com/worldclock/?iso=2020-08-04T17:00:00 ## Agenda Extracted from **meeting-agenda-test** labelled issues and pull requests from **pkgjs/meet** prior to the meeting. ## Links * Minutes: ### Joining the meeting * link for participants: * For those who just want to watch:
1.0
Test Meeting 2020-08-04 - ## Date/Time | Timezone | Date/Time | |----------|-----------| | America/Los_Angeles | Tue 04-Aug-2020 10:00 (10:00 AM) | | America/Denver | Tue 04-Aug-2020 11:00 (11:00 AM) | | America/Chicago | Tue 04-Aug-2020 12:00 (12:00 PM) | | America/New_York | Tue 04-Aug-2020 13:00 (01:00 PM) | | Europe/London | Tue 04-Aug-2020 18:00 (06:00 PM) | | Europe/Amsterdam | Tue 04-Aug-2020 19:00 (07:00 PM) | | Europe/Moscow | Tue 04-Aug-2020 20:00 (08:00 PM) | | Asia/Kolkata | Tue 04-Aug-2020 22:30 (10:30 PM) | | Asia/Shanghai | Wed 05-Aug-2020 01:00 (01:00 AM) | | Asia/Tokyo | Wed 05-Aug-2020 02:00 (02:00 AM) | | Australia/Sydney | Wed 05-Aug-2020 03:00 (03:00 AM) | Or in your local time: * https://www.timeanddate.com/worldclock/?iso=2020-08-04T17:00:00 ## Agenda Extracted from **meeting-agenda-test** labelled issues and pull requests from **pkgjs/meet** prior to the meeting. ## Links * Minutes: ### Joining the meeting * link for participants: * For those who just want to watch:
test
test meeting date time timezone date time america los angeles tue aug am america denver tue aug am america chicago tue aug pm america new york tue aug pm europe london tue aug pm europe amsterdam tue aug pm europe moscow tue aug pm asia kolkata tue aug pm asia shanghai wed aug am asia tokyo wed aug am australia sydney wed aug am or in your local time agenda extracted from meeting agenda test labelled issues and pull requests from pkgjs meet prior to the meeting links minutes joining the meeting link for participants for those who just want to watch
1
152,387
12,103,425,119
IssuesEvent
2020-04-20 18:22:32
BryanSWeber/CUNYAIModule
https://api.github.com/repos/BryanSWeber/CUNYAIModule
closed
Bot has an IsPlaceableCUNY check inaccurately found:
Needs Testing
http://www.openbw.com/replay-viewer/?rep=https://basilicum.bytekeeper.org/bots/Bryan%20Weber/Bryan%20Weber%20vs%20Dragon%20Python%20CTR_E9DCAED0.rep So what all needs to be checked for buildings?
1.0
Bot has an IsPlaceableCUNY check inaccurately found: - http://www.openbw.com/replay-viewer/?rep=https://basilicum.bytekeeper.org/bots/Bryan%20Weber/Bryan%20Weber%20vs%20Dragon%20Python%20CTR_E9DCAED0.rep So what all needs to be checked for buildings?
test
bot has an isplaceablecuny check inaccurately found so what all needs to be checked for buildings
1
278,123
8,636,957,422
IssuesEvent
2018-11-23 09:38:56
christophe-calmejane/Hive
https://api.github.com/repos/christophe-calmejane/Hive
closed
Focus (and modification) lost when a new entity is discovered
UI bug good first issue help wanted high priority
When a new entity is discovered, a reselection is made in the entity list, cascading to the entity inspector. If the user is editing an EditBox or a ComboBox, the focus is lost as all pending modifications.
1.0
Focus (and modification) lost when a new entity is discovered - When a new entity is discovered, a reselection is made in the entity list, cascading to the entity inspector. If the user is editing an EditBox or a ComboBox, the focus is lost as all pending modifications.
non_test
focus and modification lost when a new entity is discovered when a new entity is discovered a reselection is made in the entity list cascading to the entity inspector if the user is editing an editbox or a combobox the focus is lost as all pending modifications
0
302,027
26,118,537,515
IssuesEvent
2022-12-28 09:40:24
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: roachmart/partition=true failed
C-test-failure O-robot X-duplicate O-roachtest branch-master release-blocker T-kv
roachtest.roachmart/partition=true [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8046095?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8046095?buildTab=artifacts#/roachmart/partition=true) on master @ [10266a323f94c3cf397d5de590e512d987e63e22](https://github.com/cockroachdb/cockroach/commits/10266a323f94c3cf397d5de590e512d987e63e22): ``` test artifacts and logs in: /artifacts/roachmart/partition=true/run_1 (test_impl.go:291).Fatal: output in run_154656.438493202_n1_workload_init_roachmart: ./workload init roachmart --local-zone=us-central1-b --local-percent=90 --users=10 --orders=100 --partition=true returned: COMMAND_PROBLEM: exit status 1 ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_fs=ext4</code> , <code>ROACHTEST_localSSD=true</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*roachmart/partition=true.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-22673
2.0
roachtest: roachmart/partition=true failed - roachtest.roachmart/partition=true [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8046095?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8046095?buildTab=artifacts#/roachmart/partition=true) on master @ [10266a323f94c3cf397d5de590e512d987e63e22](https://github.com/cockroachdb/cockroach/commits/10266a323f94c3cf397d5de590e512d987e63e22): ``` test artifacts and logs in: /artifacts/roachmart/partition=true/run_1 (test_impl.go:291).Fatal: output in run_154656.438493202_n1_workload_init_roachmart: ./workload init roachmart --local-zone=us-central1-b --local-percent=90 --users=10 --orders=100 --partition=true returned: COMMAND_PROBLEM: exit status 1 ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_fs=ext4</code> , <code>ROACHTEST_localSSD=true</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*roachmart/partition=true.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-22673
test
roachtest roachmart partition true failed roachtest roachmart partition true with on master test artifacts and logs in artifacts roachmart partition true run test impl go fatal output in run workload init roachmart workload init roachmart local zone us b local percent users orders partition true returned command problem exit status parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest fs roachtest localssd true roachtest ssd help see see cc cockroachdb kv triage jira issue crdb
1
249,829
21,193,952,133
IssuesEvent
2022-04-08 21:00:29
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
closed
There is a background color for the last property value on 'Properties' panel after switching to the next item by pressing 'Tab'
:heavy_check_mark: merged :wheelchair: accessibility 🧪 testing
**Storage Explorer Version**: 1.20.0-dev **Build Number**: 20210519.5 **Branch**: main **Platform/OS**: Windows 10/ Linux Ubuntu 18.04/ MacOS Big Sur 11.3 **Architecture**: ia32/x64 **How Found**: Exploratory testing **Regression From**: Not a regression ## Steps to Reproduce ## 1. Select one item which has properties on its Properties panel (Like one queue). 2. Press 'Tab' to switch among the current items. 3. After switch from the last property value on the Properties panel to the next item. 4. Check there is no background color for the last property value. ## Expected Experience ## There is no background color for the property value. ## Actual Experience ## There is a background color for the property value. ![image](https://user-images.githubusercontent.com/41351993/118955317-29f12c80-b991-11eb-8398-0a3548f80ca6.png)
1.0
There is a background color for the last property value on 'Properties' panel after switching to the next item by pressing 'Tab' - **Storage Explorer Version**: 1.20.0-dev **Build Number**: 20210519.5 **Branch**: main **Platform/OS**: Windows 10/ Linux Ubuntu 18.04/ MacOS Big Sur 11.3 **Architecture**: ia32/x64 **How Found**: Exploratory testing **Regression From**: Not a regression ## Steps to Reproduce ## 1. Select one item which has properties on its Properties panel (Like one queue). 2. Press 'Tab' to switch among the current items. 3. After switch from the last property value on the Properties panel to the next item. 4. Check there is no background color for the last property value. ## Expected Experience ## There is no background color for the property value. ## Actual Experience ## There is a background color for the property value. ![image](https://user-images.githubusercontent.com/41351993/118955317-29f12c80-b991-11eb-8398-0a3548f80ca6.png)
test
there is a background color for the last property value on properties panel after switching to the next item by pressing tab storage explorer version dev build number branch main platform os windows linux ubuntu macos big sur architecture how found exploratory testing regression from not a regression steps to reproduce select one item which has properties on its properties panel like one queue press tab to switch among the current items after switch from the last property value on the properties panel to the next item check there is no background color for the last property value expected experience there is no background color for the property value actual experience there is a background color for the property value
1
120,066
10,092,043,989
IssuesEvent
2019-07-26 15:38:40
ampproject/amp-wp
https://api.github.com/repos/ampproject/amp-wp
opened
E2E Tests: Page Block
AMP Stories Needs Tests [Type] Enhancement
* [ ] Verify background color functionality * [ ] Verify gradients work as expected * [ ] Verify opacity setting works as expected Related: #2902
1.0
E2E Tests: Page Block - * [ ] Verify background color functionality * [ ] Verify gradients work as expected * [ ] Verify opacity setting works as expected Related: #2902
test
tests page block verify background color functionality verify gradients work as expected verify opacity setting works as expected related
1
97,504
8,657,983,272
IssuesEvent
2018-11-27 23:01:54
envoyproxy/envoy
https://api.github.com/repos/envoyproxy/envoy
closed
TSAN error in SignalAction::sigHandler
help wanted test flakes
*Description*: The [CircleCI TSAN run](https://circleci.com/gh/envoyproxy/envoy/24955?utm_campaign=vcs-integration-link&utm_medium=referral&utm_source=github-build-link) for one of my pull requests caught what appears to be an existing problem: a signal-unsafe call to `new` due to string processing within a signal handler. *Repro steps*: This happened at least once by running the integration tests with TSAN. I'm not sure whether it happens deterministically. *Call Stack*: ``` WARNING: ThreadSanitizer: signal-unsafe call inside of a signal (pid=12258) #0 operator new(unsigned long) ??:? (ratelimit_integration_test+0x5c1e43) #1 void std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >::_M_construct<char const*>(char const*, char const*, std::forward_iterator_tag) /usr/bin/../lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/basic_string.tcc:223 (ratelimit_integration_test+0xc7a0ec) #2 std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >::basic_string(char const*, unsigned long, std::allocator<char> const&) ??:? (libstdc++.so.6+0x121c04) #3 Envoy::Logger::LockingStderrOrFileSink::log(spdlog::details::log_msg const&) /proc/self/cwd/external/envoy/source/common/common/logger.cc:46 (ratelimit_integration_test+0x19df8d2) #4 spdlog::logger::_sink_it(spdlog::details::log_msg&) /proc/self/cwd/external/com_github_gabime_spdlog/include/spdlog/details/logger_impl.h:321 (ratelimit_integration_test+0x5d63cc) #5 void spdlog::logger::log<char const*, void const*>(spdlog::level::level_enum, char const*, char const* const&, void const* const&) /proc/self/cwd/external/com_github_gabime_spdlog/include/spdlog/details/logger_impl.h:74 (ratelimit_integration_test+0x19d74f9) #6 void spdlog::logger::critical<char const*, void const*>(char const*, char const* const&, void const* const&) /proc/self/cwd/external/com_github_gabime_spdlog/include/spdlog/details/logger_impl.h:163 (ratelimit_integration_test+0x19d737d) #7 Envoy::BackwardsTrace::logFault(char const*, void const*) /proc/self/cwd/bazel-out/k8-dbg/bin/external/envoy/source/server/_virtual_includes/backtrace_lib/server/backtrace.h:101 (ratelimit_integration_test+0x19d670f) #8 Envoy::SignalAction::sigHandler(int, siginfo_t*, void*) /proc/self/cwd/external/envoy/source/exe/signal_action.cc:30 (ratelimit_integration_test+0x19d528b) #9 __tsan::CallUserSignalHandler(__tsan::ThreadState*, bool, bool, bool, int, my_siginfo_t*, void*) crtstuff.c:? (ratelimit_integration_test+0x530c49) #10 non-virtual thunk to Envoy::(anonymous namespace)::RatelimitIntegrationTest_Ok_Test::TestBody() external/envoy/test/integration/ratelimit_integration_test.cc:? (ratelimit_integration_test+0x5c88af) #11 void testing::internal::HandleSehExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2401 (ratelimit_integration_test+0x1f08206) #12 void testing::internal::HandleExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2437 (ratelimit_integration_test+0x1eee961) #13 testing::Test::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2473 (ratelimit_integration_test+0x1ed723e) #14 testing::TestInfo::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2655 (ratelimit_integration_test+0x1ed82f1) #15 testing::TestCase::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2773 (ratelimit_integration_test+0x1ed8d37) #16 testing::internal::UnitTestImpl::RunAllTests() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:4673 (ratelimit_integration_test+0x1ee455b) #17 bool testing::internal::HandleSehExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2401 (ratelimit_integration_test+0x1f0d916) #18 bool testing::internal::HandleExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2437 (ratelimit_integration_test+0x1ef24d7) #19 testing::UnitTest::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:4281 (ratelimit_integration_test+0x1ee3f6b) #20 RUN_ALL_TESTS() /proc/self/cwd/external/com_google_googletest/googletest/include/gtest/gtest.h:2237 (ratelimit_integration_test+0x15fbf27) #21 Envoy::TestRunner::RunTests(int, char**) /proc/self/cwd/external/envoy/test/test_runner.h:33 (ratelimit_integration_test+0x15fbde0) #22 main /proc/self/cwd/external/envoy/test/main.cc:35 (ratelimit_integration_test+0x15fae52) ```
1.0
TSAN error in SignalAction::sigHandler - *Description*: The [CircleCI TSAN run](https://circleci.com/gh/envoyproxy/envoy/24955?utm_campaign=vcs-integration-link&utm_medium=referral&utm_source=github-build-link) for one of my pull requests caught what appears to be an existing problem: a signal-unsafe call to `new` due to string processing within a signal handler. *Repro steps*: This happened at least once by running the integration tests with TSAN. I'm not sure whether it happens deterministically. *Call Stack*: ``` WARNING: ThreadSanitizer: signal-unsafe call inside of a signal (pid=12258) #0 operator new(unsigned long) ??:? (ratelimit_integration_test+0x5c1e43) #1 void std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >::_M_construct<char const*>(char const*, char const*, std::forward_iterator_tag) /usr/bin/../lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/basic_string.tcc:223 (ratelimit_integration_test+0xc7a0ec) #2 std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >::basic_string(char const*, unsigned long, std::allocator<char> const&) ??:? (libstdc++.so.6+0x121c04) #3 Envoy::Logger::LockingStderrOrFileSink::log(spdlog::details::log_msg const&) /proc/self/cwd/external/envoy/source/common/common/logger.cc:46 (ratelimit_integration_test+0x19df8d2) #4 spdlog::logger::_sink_it(spdlog::details::log_msg&) /proc/self/cwd/external/com_github_gabime_spdlog/include/spdlog/details/logger_impl.h:321 (ratelimit_integration_test+0x5d63cc) #5 void spdlog::logger::log<char const*, void const*>(spdlog::level::level_enum, char const*, char const* const&, void const* const&) /proc/self/cwd/external/com_github_gabime_spdlog/include/spdlog/details/logger_impl.h:74 (ratelimit_integration_test+0x19d74f9) #6 void spdlog::logger::critical<char const*, void const*>(char const*, char const* const&, void const* const&) /proc/self/cwd/external/com_github_gabime_spdlog/include/spdlog/details/logger_impl.h:163 (ratelimit_integration_test+0x19d737d) #7 Envoy::BackwardsTrace::logFault(char const*, void const*) /proc/self/cwd/bazel-out/k8-dbg/bin/external/envoy/source/server/_virtual_includes/backtrace_lib/server/backtrace.h:101 (ratelimit_integration_test+0x19d670f) #8 Envoy::SignalAction::sigHandler(int, siginfo_t*, void*) /proc/self/cwd/external/envoy/source/exe/signal_action.cc:30 (ratelimit_integration_test+0x19d528b) #9 __tsan::CallUserSignalHandler(__tsan::ThreadState*, bool, bool, bool, int, my_siginfo_t*, void*) crtstuff.c:? (ratelimit_integration_test+0x530c49) #10 non-virtual thunk to Envoy::(anonymous namespace)::RatelimitIntegrationTest_Ok_Test::TestBody() external/envoy/test/integration/ratelimit_integration_test.cc:? (ratelimit_integration_test+0x5c88af) #11 void testing::internal::HandleSehExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2401 (ratelimit_integration_test+0x1f08206) #12 void testing::internal::HandleExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2437 (ratelimit_integration_test+0x1eee961) #13 testing::Test::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2473 (ratelimit_integration_test+0x1ed723e) #14 testing::TestInfo::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2655 (ratelimit_integration_test+0x1ed82f1) #15 testing::TestCase::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2773 (ratelimit_integration_test+0x1ed8d37) #16 testing::internal::UnitTestImpl::RunAllTests() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:4673 (ratelimit_integration_test+0x1ee455b) #17 bool testing::internal::HandleSehExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2401 (ratelimit_integration_test+0x1f0d916) #18 bool testing::internal::HandleExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:2437 (ratelimit_integration_test+0x1ef24d7) #19 testing::UnitTest::Run() /proc/self/cwd/external/com_google_googletest/googletest/src/gtest.cc:4281 (ratelimit_integration_test+0x1ee3f6b) #20 RUN_ALL_TESTS() /proc/self/cwd/external/com_google_googletest/googletest/include/gtest/gtest.h:2237 (ratelimit_integration_test+0x15fbf27) #21 Envoy::TestRunner::RunTests(int, char**) /proc/self/cwd/external/envoy/test/test_runner.h:33 (ratelimit_integration_test+0x15fbde0) #22 main /proc/self/cwd/external/envoy/test/main.cc:35 (ratelimit_integration_test+0x15fae52) ```
test
tsan error in signalaction sighandler description the for one of my pull requests caught what appears to be an existing problem a signal unsafe call to new due to string processing within a signal handler repro steps this happened at least once by running the integration tests with tsan i m not sure whether it happens deterministically call stack warning threadsanitizer signal unsafe call inside of a signal pid operator new unsigned long ratelimit integration test void std basic string std allocator m construct char const char const std forward iterator tag usr bin lib gcc linux gnu include c bits basic string tcc ratelimit integration test std basic string std allocator basic string char const unsigned long std allocator const libstdc so envoy logger lockingstderrorfilesink log spdlog details log msg const proc self cwd external envoy source common common logger cc ratelimit integration test spdlog logger sink it spdlog details log msg proc self cwd external com github gabime spdlog include spdlog details logger impl h ratelimit integration test void spdlog logger log spdlog level level enum char const char const const void const const proc self cwd external com github gabime spdlog include spdlog details logger impl h ratelimit integration test void spdlog logger critical char const char const const void const const proc self cwd external com github gabime spdlog include spdlog details logger impl h ratelimit integration test envoy backwardstrace logfault char const void const proc self cwd bazel out dbg bin external envoy source server virtual includes backtrace lib server backtrace h ratelimit integration test envoy signalaction sighandler int siginfo t void proc self cwd external envoy source exe signal action cc ratelimit integration test tsan callusersignalhandler tsan threadstate bool bool bool int my siginfo t void crtstuff c ratelimit integration test non virtual thunk to envoy anonymous namespace ratelimitintegrationtest ok test testbody external envoy test integration ratelimit integration test cc ratelimit integration test void testing internal handlesehexceptionsinmethodifsupported testing test void testing test char const proc self cwd external com google googletest googletest src gtest cc ratelimit integration test void testing internal handleexceptionsinmethodifsupported testing test void testing test char const proc self cwd external com google googletest googletest src gtest cc ratelimit integration test testing test run proc self cwd external com google googletest googletest src gtest cc ratelimit integration test testing testinfo run proc self cwd external com google googletest googletest src gtest cc ratelimit integration test testing testcase run proc self cwd external com google googletest googletest src gtest cc ratelimit integration test testing internal unittestimpl runalltests proc self cwd external com google googletest googletest src gtest cc ratelimit integration test bool testing internal handlesehexceptionsinmethodifsupported testing internal unittestimpl bool testing internal unittestimpl char const proc self cwd external com google googletest googletest src gtest cc ratelimit integration test bool testing internal handleexceptionsinmethodifsupported testing internal unittestimpl bool testing internal unittestimpl char const proc self cwd external com google googletest googletest src gtest cc ratelimit integration test testing unittest run proc self cwd external com google googletest googletest src gtest cc ratelimit integration test run all tests proc self cwd external com google googletest googletest include gtest gtest h ratelimit integration test envoy testrunner runtests int char proc self cwd external envoy test test runner h ratelimit integration test main proc self cwd external envoy test main cc ratelimit integration test
1
69,203
7,126,966,131
IssuesEvent
2018-01-20 16:30:25
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
teamcity: failed tests on master: test/TestBackupRestoreInterleaved, test/TestBackupRestoreInterleaved/all_tables_in_interleave_hierarchy, testrace/TestBackupRestoreInterleaved, testrace/TestBackupRestoreInterleaved/all_tables_in_interleave_hierarchy
Robot test-failure
The following tests appear to have failed: [#474455](https://teamcity.cockroachdb.com/viewLog.html?buildId=474455): ``` --- FAIL: test/TestBackupRestoreInterleaved (0.000s) Test ended in panic. ------- Stdout: ------- W180110 18:02:59.680952 118774 server/status/runtime.go:109 Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180110 18:02:59.682776 118774 server/config.go:518 [n?] 1 storage engine initialized I180110 18:02:59.682793 118774 server/config.go:521 [n?] RocksDB cache size: 128 MiB I180110 18:02:59.682798 118774 server/config.go:521 [n?] store 0: in-memory, size 0 B I180110 18:02:59.684988 118774 server/node.go:361 [n?] **** cluster 12dcf5c1-311b-4753-a474-73554bd7f94f has been created I180110 18:02:59.685014 118774 server/server.go:934 [n?] **** add additional nodes by specifying --join=127.0.0.1:35537 I180110 18:02:59.685616 118774 storage/store.go:1312 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available I180110 18:02:59.685656 118774 server/node.go:491 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=3.2 KiB), ranges=1, leases=0, writes=0.00, bytesPerReplica={p10=3322.00 p25=3322.00 p50=3322.00 p75=3322.00 p90=3322.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00} I180110 18:02:59.685688 118774 server/node.go:339 [n1] node ID 1 initialized I180110 18:02:59.685757 118774 gossip/gossip.go:332 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:35537" > attrs:<> locality:<> ServerVersion:<major_val:1 minor_val:1 patch:0 unstable:9 > I180110 18:02:59.685924 118774 storage/stores.go:331 [n1] read 0 node addresses from persistent storage I180110 18:02:59.686060 118774 server/node.go:632 [n1] connecting to gossip network to verify cluster ID... I180110 18:02:59.686094 118774 server/node.go:657 [n1] node connected via gossip and verified as part of cluster "12dcf5c1-311b-4753-a474-73554bd7f94f" I180110 18:02:59.686121 118774 server/node.go:433 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes [] I180110 18:02:59.686252 118774 sql/distsql_physical_planner.go:122 [n1] creating DistSQLPlanner with address {tcp 127.0.0.1:35537} I180110 18:02:59.687294 118745 storage/replica_command.go:819 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2] I180110 18:02:59.693580 118774 server/server.go:1161 [n1] starting https server at 127.0.0.1:40697 I180110 18:02:59.693633 118774 server/server.go:1162 [n1] starting grpc/postgres server at 127.0.0.1:35537 I180110 18:02:59.693657 118774 server/server.go:1163 [n1] advertising CockroachDB node at 127.0.0.1:35537 I180110 18:02:59.698407 118745 storage/replica_command.go:819 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3] I180110 18:02:59.703687 118745 storage/replica_command.go:819 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4] W180110 18:02:59.704283 119042 storage/intent_resolver.go:324 [n1,s1] failed to push during intent resolution: failed to push "sql txn implicit" id=96eef510 key=/Table/SystemConfigSpan/Start rw=true pri=0.03373129 iso=SERIALIZABLE stat=PENDING epo=0 ts=1515607379.699252305,0 orig=1515607379.699252305,0 max=1515607379.699252305,0 wto=false rop=false seq=7 I180110 18:02:59.705378 118774 sql/event_log.go:115 [n1] Event: "alter_table", target: 12, info: {TableName:eventlog Statement:ALTER TABLE system.eventlog ALTER COLUMN "uniqueID" SET DEFAULT uuid_v4() User:node MutationID:0 CascadeDroppedViews:[]} I180110 18:02:59.709553 118745 storage/replica_command.go:819 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5] I180110 18:02:59.711655 118774 sql/lease.go:348 [n1] publish: descID=12 (eventlog) version=2 mtime=2018-01-10 18:02:59.711397351 +0000 UTC I180110 18:02:59.714515 118745 storage/replica_command.go:819 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6] I180110 18:02:59.722104 118745 storage/replica_command.go:819 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7] I180110 18:02:59.728036 118745 storage/replica_command.go:819 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8] W180110 18:02:59.728476 118812 storage/intent_resolver.go:324 [n1,s1] failed to push during intent resolution: failed to push "sql txn implicit" id=f8a38c98 key=/Table/SystemConfigSpan/Start rw=true pri=0.00799806 iso=SERIALIZABLE stat=PENDING epo=0 ts=1515607379.723877453,0 orig=1515607379.723877453,0 max=1515607379.723877453,0 wto=false rop=false seq=7 I180110 18:02:59.729842 118774 sql/event_log.go:115 [n1] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:node} I180110 18:02:59.732898 118745 storage/replica_command.go:819 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9] I180110 18:02:59.738942 118745 storage/replica_command.go:819 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10] I180110 18:02:59.741515 118774 sql/event_log.go:115 [n1] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:$1 User:node} I180110 18:02:59.743707 118745 storage/replica_command.go:819 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11] I180110 18:02:59.745378 118774 sql/event_log.go:115 [n1] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:node} I180110 18:02:59.748179 118745 storage/replica_command.go:819 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12] I180110 18:02:59.754047 118745 storage/replica_command.go:819 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13] I180110 18:02:59.756675 118774 sql/event_log.go:115 [n1] Event: "alter_table", target: 4, info: {TableName:users Statement:ALTER TABLE system.users ADD COLUMN IF NOT EXISTS "isRole" BOOL NOT NULL DEFAULT false User:node MutationID:1 CascadeDroppedViews:[]} I180110 18:02:59.757649 118745 storage/replica_command.go:819 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14] I180110 18:02:59.761927 118774 sql/lease.go:348 [n1] publish: descID=4 (users) version=2 mtime=2018-01-10 18:02:59.761827399 +0000 UTC I180110 18:02:59.765394 118745 storage/replica_command.go:819 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15] I180110 18:02:59.766836 118774 sql/lease.go:348 [n1] publish: descID=4 (users) version=3 mtime=2018-01-10 18:02:59.766683769 +0000 UTC I180110 18:02:59.769003 118745 storage/replica_command.go:819 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16] I180110 18:02:59.771316 118774 sql/backfill.go:133 [n1] Running backfill for "users", v=3, m=1 I180110 18:02:59.771931 118745 storage/replica_command.go:819 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17] I180110 18:02:59.774793 118745 storage/replica_command.go:819 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18] I180110 18:02:59.778046 118745 storage/replica_command.go:819 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19] I180110 18:02:59.781235 118774 sql/lease.go:348 [n1] publish: descID=4 (users) version=4 mtime=2018-01-10 18:02:59.781034178 +0000 UTC I180110 18:02:59.782374 118745 storage/replica_command.go:819 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20] I180110 18:02:59.784585 118774 sql/event_log.go:115 [n1] Event: "finish_schema_change", target: 4, info: {MutationID:1} I180110 18:02:59.785559 118774 sql/lease.go:274 publish (count leases): descID=4 name=users version=3 count=1 I180110 18:02:59.818441 118774 server/server.go:1232 [n1] done ensuring all necessary migrations have run I180110 18:02:59.818491 118774 server/server.go:1235 [n1] serving sql connections I180110 18:02:59.820529 119248 sql/event_log.go:115 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:35537} Attrs: Locality: ServerVersion:1.1-9} ClusterID:12dcf5c1-311b-4753-a474-73554bd7f94f StartedAt:1515607379686099339 LastUp:1515607379686099339} I180110 18:02:59.835610 119285 sql/event_log.go:115 [client=127.0.0.1:59480,user=root,n1] Event: "create_database", target: 50, info: {DatabaseName:data Statement:CREATE DATABASE data User:root} I180110 18:02:59.836073 118745 storage/replica_command.go:819 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21] I180110 18:02:59.839557 119285 sql/event_log.go:115 [client=127.0.0.1:59480,user=root,n1] Event: "create_table", target: 51, info: {TableName:bank Statement:CREATE TABLE bank (id INT PRIMARY KEY, balance INT, payload STRING, FAMILY (id, balance, payload)) User:root} I180110 18:02:59.840359 118745 storage/replica_command.go:819 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22] I180110 18:02:59.842925 119285 storage/replica_command.go:819 [n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51/1/1 [r23] I180110 18:02:59.894721 119285 storage/replica_command.go:819 [n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/51/1/1 [r24] I180110 18:02:59.897970 119285 storage/replica_command.go:819 [n1,s1,r24/1:/{Table/51/1/1-Max}] initiating a split of this range at key /Table/51/1/2 [r25] I180110 18:02:59.902866 119285 storage/replica_command.go:819 [n1,s1,r25/1:/{Table/51/1/2-Max}] initiating a split of this range at key /Table/51/1/3 [r26] I180110 18:02:59.906764 119285 storage/replica_command.go:819 [n1,s1,r26/1:/{Table/51/1/3-Max}] initiating a split of this range at key /Table/51/1/4 [r27] I180110 18:02:59.911998 119285 storage/replica_command.go:819 [n1,s1,r27/1:/{Table/51/1/4-Max}] initiating a split of this range at key /Table/51/1/5 [r28] I180110 18:02:59.915310 119285 storage/replica_command.go:819 [n1,s1,r28/1:/{Table/51/1/5-Max}] initiating a split of this range at key /Table/51/1/6 [r29] I180110 18:02:59.918539 119285 storage/replica_command.go:819 [n1,s1,r29/1:/{Table/51/1/6-Max}] initiating a split of this range at key /Table/51/1/7 [r30] I180110 18:02:59.922686 119285 storage/replica_command.go:819 [n1,s1,r30/1:/{Table/51/1/7-Max}] initiating a split of this range at key /Table/51/1/8 [r31] I180110 18:02:59.926944 119285 storage/replica_command.go:819 [n1,s1,r31/1:/{Table/51/1/8-Max}] initiating a split of this range at key /Table/51/1/9 [r32] I180110 18:02:59.934885 119285 sql/event_log.go:115 [client=127.0.0.1:59480,user=root,n1] Event: "create_table", target: 52, info: {TableName:i0 Statement:CREATE TABLE i0 (a INT, b INT, PRIMARY KEY (a, b)) INTERLEAVE IN PARENT bank (a) User:root} I180110 18:02:59.936802 118745 storage/replica_command.go:819 [split,n1,s1,r32/1:/{Table/51/1/9-Max}] initiating a split of this range at key /Table/52 [r33] I180110 18:02:59.936960 119285 sql/lease.go:348 [client=127.0.0.1:59480,user=root,n1] publish: descID=52 (i0) version=2 mtime=2018-01-10 18:02:59.936866919 +0000 UTC I180110 18:02:59.941868 119285 sql/lease.go:348 [client=127.0.0.1:59480,user=root,n1] publish: descID=51 (bank) version=2 mtime=2018-01-10 18:02:59.941755855 +0000 UTC I180110 18:02:59.943463 119285 sql/lease.go:274 publish (count leases): descID=51 name=bank version=1 count=1 I180110 18:02:59.979049 119285 sql/event_log.go:115 [client=127.0.0.1:59480,user=root,n1] Event: "create_table", target: 53, info: {TableName:i0_0 Statement:CREATE TABLE i0_0 (a INT, b INT, c INT, PRIMARY KEY (a, b, c)) INTERLEAVE IN PARENT i0 (a, b) User:root} I180110 18:02:59.979712 118745 storage/replica_command.go:819 [split,n1,s1,r33/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r34] I180110 18:02:59.983045 119285 sql/lease.go:348 [client=127.0.0.1:59480,user=root,n1] publish: descID=53 (i0_0) version=2 mtime=2018-01-10 18:02:59.982717895 +0000 UTC I180110 18:02:59.987749 119285 sql/lease.go:348 [client=127.0.0.1:59480,user=root,n1] publish: descID=52 (i0) version=3 mtime=2018-01-10 18:02:59.987656881 +0000 UTC I180110 18:02:59.996752 119285 sql/event_log.go:115 [client=127.0.0.1:59480,user=root,n1] Event: "create_table", target: 54, info: {TableName:i1 Statement:CREATE TABLE i1 (a INT, b INT, PRIMARY KEY (a, b)) INTERLEAVE IN PARENT bank (a) User:root} I180110 18:02:59.997271 118745 storage/replica_command.go:819 [split,n1,s1,r34/1:/{Table/53-Max}] initiating a split of this range at key /Table/54 [r35] I180110 18:02:59.998973 119285 sql/lease.go:348 [client=127.0.0.1:59480,user=root,n1] publish: descID=54 (i1) version=2 mtime=2018-01-10 18:02:59.998869144 +0000 UTC I180110 18:03:00.005665 119285 sql/lease.go:348 [client=127.0.0.1:59480,user=root,n1] publish: descID=51 (bank) version=3 mtime=2018-01-10 18:03:00.00552806 +0000 UTC I180110 18:03:00.006862 119285 sql/lease.go:274 publish (count leases): descID=51 name=bank version=2 count=1 I180110 18:03:00.070675 119285 storage/replica_command.go:819 [n1,s1,r22/1:/Table/51{-/1/1}] initiating a split of this range at key /Table/51/1/0/#/52/1/1 [r36] I180110 18:03:00.076517 119285 storage/replica_command.go:819 [n1,s1,r36/1:/Table/51/1/{0/#/52…-1}] initiating a split of this range at key /Table/51/1/0/#/52/1/2 [r37] I180110 18:03:00.080343 119285 storage/replica_command.go:819 [n1,s1,r24/1:/Table/51/1/{1-2}] initiating a split of this range at key /Table/51/1/1/#/52/1/1 [r38] I180110 18:03:00.084681 119285 storage/replica_command.go:819 [n1,s1,r38/1:/Table/51/1/{1/#/52…-2}] initiating a split of this range at key /Table/51/1/1/#/52/1/2 [r39] I180110 18:03:00.090001 119285 storage/replica_command.go:819 [n1,s1,r25/1:/Table/51/1/{2-3}] initiating a split of this range at key /Table/51/1/2/#/52/1/1 [r40] I180110 18:03:00.093756 119285 storage/replica_command.go:819 [n1,s1,r40/1:/Table/51/1/{2/#/52…-3}] initiating a split of this range at key /Table/51/1/2/#/52/1/2 [r41] I180110 18:03:00.096434 119285 storage/replica_command.go:819 [n1,s1,r26/1:/Table/51/1/{3-4}] initiating a split of this range at key /Table/51/1/3/#/52/1/1 [r42] I180110 18:03:00.099404 119285 storage/replica_command.go:819 [n1,s1,r42/1:/Table/51/1/{3/#/52…-4}] initiating a split of this range at key /Table/51/1/3/#/52/1/2 [r43] I180110 18:03:00.103815 119285 storage/replica_command.go:819 [n1,s1,r27/1:/Table/51/1/{4-5}] initiating a split of this range at key /Table/51/1/4/#/52/1/1 [r44] I180110 18:03:00.107627 119285 storage/replica_command.go:819 [n1,s1,r44/1:/Table/51/1/{4/#/52…-5}] initiating a split of this range at key /Table/51/1/4/#/52/1/2 [r45] I180110 18:03:00.114279 119285 storage/replica_command.go:819 [n1,s1,r36/1:/Table/51/1/0/#/52/1/{1-2}] initiating a split of this range at key /Table/51/1/0/#/52/1/1/#/53/1 [r46] I180110 18:03:00.121526 119285 storage/replica_command.go:819 [n1,s1,r37/1:/Table/51/1/{0/#/52…-1}] initiating a split of this range at key /Table/51/1/0/#/52/1/2/#/53/1 [r47] I180110 18:03:00.127570 119285 storage/replica_command.go:819 [n1,s1,r38/1:/Table/51/1/1/#/52/1/{1-2}] initiating a split of this range at key /Table/51/1/1/#/52/1/1/#/53/1 [r48] I180110 18:03:00.131410 119285 storage/replica_command.go:819 [n1,s1,r39/1:/Table/51/1/{1/#/52…-2}] initiating a split of this range at key /Table/51/1/1/#/52/1/2/#/53/1 [r49] I180110 18:03:00.134666 119285 storage/replica_command.go:819 [n1,s1,r40/1:/Table/51/1/2/#/52/1/{1-2}] initiating a split of this range at key /Table/51/1/2/#/52/1/1/#/53/1 [r50] I180110 18:03:00.138930 119285 storage/replica_command.go:819 [n1,s1,r41/1:/Table/51/1/{2/#/52…-3}] initiating a split of this range at key /Table/51/1/2/#/52/1/2/#/53/1 [r51] I180110 18:03:00.142102 119285 storage/replica_command.go:819 [n1,s1,r42/1:/Table/51/1/3/#/52/1/{1-2}] initiating a split of this range at key /Table/51/1/3/#/52/1/1/#/53/1 [r52] I180110 18:03:00.146773 119285 storage/replica_command.go:819 [n1,s1,r43/1:/Table/51/1/{3/#/52…-4}] initiating a split of this range at key /Table/51/1/3/#/52/1/2/#/53/1 [r53] I180110 18:03:00.153089 119285 storage/replica_command.go:819 [n1,s1,r44/1:/Table/51/1/4/#/52/1/{1-2}] initiating a split of this range at key /Table/51/1/4/#/52/1/1/#/53/1 [r54] I180110 18:03:00.156921 119285 storage/replica_command.go:819 [n1,s1,r45/1:/Table/51/1/{4/#/52…-5}] initiating a split of this range at key /Table/51/1/4/#/52/1/2/#/53/1 [r55] I180110 18:03:00.164474 119285 storage/replica_command.go:819 [n1,s1,r47/1:/Table/51/1/{0/#/52…-1}] initiating a split of this range at key /Table/51/1/0/#/54/1/1 [r56] I180110 18:03:00.173135 119285 storage/replica_command.go:819 [n1,s1,r56/1:/Table/51/1/{0/#/54…-1}] initiating a split of this range at key /Table/51/1/0/#/54/1/2 [r57] I180110 18:03:00.178226 119285 storage/replica_command.go:819 [n1,s1,r49/1:/Table/51/1/{1/#/52…-2}] initiating a split of this range at key /Table/51/1/1/#/54/1/1 [r58] I180110 18:03:00.182743 119285 storage/replica_command.go:819 [n1,s1,r58/1:/Table/51/1/{1/#/54…-2}] initiating a split of this range at key /Table/51/1/1/#/54/1/2 [r59] I180110 18:03:00.187677 119285 storage/replica_command.go:819 [n1,s1,r51/1:/Table/51/1/{2/#/52…-3}] initiating a split of this range at key /Table/51/1/2/#/54/1/1 [r60] I180110 18:03:00.191390 119285 storage/replica_command.go:819 [n1,s1,r60/1:/Table/51/1/{2/#/54…-3}] initiating a split of this range at key /Table/51/1/2/#/54/1/2 [r61] I180110 18:03:00.196240 119285 storage/replica_command.go:819 [n1,s1,r53/1:/Table/51/1/{3/#/52…-4}] initiating a split of this range at key /Table/51/1/3/#/54/1/1 [r62] I180110 18:03:00.199671 119285 storage/replica_command.go:819 [n1,s1,r62/1:/Table/51/1/{3/#/54…-4}] initiating a split of this range at key /Table/51/1/3/#/54/1/2 [r63] I180110 18:03:00.202414 119285 storage/replica_command.go:819 [n1,s1,r55/1:/Table/51/1/{4/#/52…-5}] initiating a split of this range at key /Table/51/1/4/#/54/1/1 [r64] I180110 18:03:00.206213 119285 storage/replica_command.go:819 [n1,s1,r64/1:/Table/51/1/{4/#/54…-5}] initiating a split of this range at key /Table/51/1/4/#/54/1/2 [r65] I180110 18:03:00.221491 119727 ccl/storageccl/export.go:124 [n1,s1,r47/1:/Table/51/1/0/#/5{2/1/…-4/1/1}] export [/Table/51/1/0/#/52/1/2/#/53/1,/Table/51/1/0/#/54/1/1) I180110 18:03:00.221634 119726 ccl/storageccl/export.go:124 [n1,s1,r37/1:/Table/51/1/0/#/52/1/2{-/…}] export [/Table/51/1/0/#/52/1/2,/Table/51/1/0/#/52/1/2/#/53/1) I180110 18:03:00.221726 119724 ccl/storageccl/export.go:124 [n1,s1,r36/1:/Table/51/1/0/#/52/1/1{-/…}] export [/Table/51/1/0/#/52/1/1,/Table/51/1/0/#/52/1/1/#/53/1) I180110 18:03:00.221634 119723 ccl/storageccl/export.go:124 [n1,s1,r22/1:/Table/51{-/1/0/#/5…}] export [/Table/51/1,/Table/51/1/0/#/52/1/1) I180110 18:03:00.221657 119725 ccl/storageccl/export.go:124 [n1,s1,r46/1:/Table/51/1/0/#/52/1/{1/…-2}] export [/Table/51/1/0/#/52/1/1/#/53/1,/Table/51/1/0/#/52/1/2) I180110 18:03:00.224829 119762 ccl/storageccl/export.go:124 [n1,s1,r57/1:/Table/51/1/{0/#/54…-1}] export [/Table/51/1/0/#/54/1/2,/Table/51/1/1) I180110 18:03:00.224909 119729 ccl/storageccl/export.go:124 [n1,s1,r56/1:/Table/51/1/0/#/54/1/{1-2}] export [/Table/51/1/0/#/54/1/1,/Table/51/1/0/#/54/1/2) I180110 18:03:00.225032 119752 ccl/storageccl/export.go:124 [n1,s1,r38/1:/Table/51/1/1/#/52/1/1{-/…}] export [/Table/51/1/1/#/52/1/1,/Table/51/1/1/#/52/1/1/#/53/1) I180110 18:03:00.225148 119751 ccl/storageccl/export.go:124 [n1,s1,r24/1:/Table/51/1/1{-/#/52/…}] export [/Table/51/1/1,/Table/51/1/1/#/52/1/1) I180110 18:03:00.227625 119637 ccl/storageccl/export.go:124 [n1,s1,r48/1:/Table/51/1/1/#/52/1/{1/…-2}] export [/Table/51/1/1/#/52/1/1/#/53/1,/Table/51/1/1/#/52/1/2) I180110 18:03:00.227662 119763 ccl/storageccl/export.go:124 [n1,s1,r39/1:/Table/51/1/1/#/52/1/2{-/…}] export [/Table/51/1/1/#/52/1/2,/Table/51/1/1/#/52/1/2/#/53/1) I180110 18:03:00.227783 119695 ccl/storageccl/export.go:124 [n1,s1,r49/1:/Table/51/1/1/#/5{2/1/…-4/1/1}] export [/Table/51/1/1/#/52/1/2/#/53/1,/Table/51/1/1/#/54/1/1) I180110 18:03:00.228066 119778 ccl/storageccl/export.go:124 [n1,s1,r59/1:/Table/51/1/{1/#/54…-2}] export [/Table/51/1/1/#/54/1/2,/Table/51/1/2) I180110 18:03:00.227784 119753 ccl/storageccl/export.go:124 [n1,s1,r58/1:/Table/51/1/1/#/54/1/{1-2}] export [/Table/51/1/1/#/54/1/1,/Table/51/1/1/#/54/1/2) I180110 18:03:00.231000 119639 ccl/storageccl/export.go:124 [n1,s1,r40/1:/Table/51/1/2/#/52/1/1{-/…}] export [/Table/51/1/2/#/52/1/1,/Table/51/1/2/#/52/1/1/#/53/1) I180110 18:03:00.231028 119638 ccl/storageccl/export.go:124 [n1,s1,r25/1:/Table/51/1/2{-/#/52/…}] export [/Table/51/1/2,/Table/51/1/2/#/52/1/1) I180110 18:03:00.231226 119764 ccl/storageccl/export.go:124 [n1,s1,r41/1:/Table/51/1/2/#/52/1/2{-/…}] export [/Table/51/1/2/#/52/1/2,/Table/51/1/2/#/52/1/2/#/53/1) I180110 18:03:00.231284 119754 ccl/storageccl/export.go:124 [n1,s1,r50/1:/Table/51/1/2/#/52/1/{1/…-2}] export [/Table/51/1/2/#/52/1/1/#/53/1,/Table/51/1/2/#/52/1/2) I180110 18:03:00.231284 119765 ccl/storageccl/export.go:124 [n1,s1,r51/1:/Table/51/1/2/#/5{2/1/…-4/1/1}] export [/Table/51/1/2/#/52/1/2/#/53/1,/Table/51/1/2/#/54/1/1) I180110 18:03:00.234051 119608 ccl/storageccl/export.go:124 [n1,s1,r26/1:/Table/51/1/3{-/#/52/…}] export [/Table/51/1/3,/Table/51/1/3/#/52/1/1) I180110 18:03:00.234057 119641 ccl/storageccl/export.go:124 [n1,s1,r61/1:/Table/51/1/{2/#/54…-3}] export [/Table/51/1/2/#/54/1/2,/Table/51/1/3) I180110 18:03:00.234232 119640 ccl/storageccl/export.go:124 [n1,s1,r60/1:/Table/51/1/2/#/54/1/{1-2}] export [/Table/51/1/2/#/54/1/1,/Table/51/1/2/#/54/1/2) I180110 18:03:00.234519 119564 ccl/storageccl/export.go:124 [n1,s1,r42/1:/Table/51/1/3/#/52/1/1{-/…}] export [/Table/51/1/3/#/52/1/1,/Table/51/1/3/#/52/1/1/#/53/1) I180110 18:03:00.234570 119780 ccl/storageccl/export.go:124 [n1,s1,r52/1:/Table/51/1/3/#/52/1/{1/…-2}] export [/Table/51/1/3/#/52/1/1/#/53/1,/Table/51/1/3/#/52/1/2) I180110 18:03:00.236747 119610 ccl/storageccl/export.go:124 [n1,s1,r53/1:/Table/51/1/3/#/5{2/1/…-4/1/1}] export [/Table/51/1/3/#/52/1/2/#/53/1,/Table/51/1/3/#/54/1/1) I180110 18:03:00.236764 119609 ccl/storageccl/export.go:124 [n1,s1,r43/1:/Table/51/1/3/#/52/1/2{-/…}] export [/Table/51/1/3/#/52/1/2,/Table/51/1/3/#/52/1/2/#/53/1) I180110 18:03:00.236833 119644 ccl/storageccl/export.go:124 [n1,s1,r27/1:/Table/51/1/4{-/#/52/…}] export [/Table/51/1/4,/Table/51/1/4/#/52/1/1) I180110 18:03:00.236857 119643 ccl/storageccl/export.go:124 [n1,s1,r63/1:/Table/51/1/{3/#/54…-4}] export [/Table/51/1/3/#/54/1/2,/Table/51/1/4) I180110 18:03:00.236842 119642 ccl/storageccl/export.go:124 [n1,s1,r62/1:/Table/51/1/3/#/54/1/{1-2}] export [/Table/51/1/3/#/54/1/1,/Table/51/1/3/#/54/1/2) I180110 18:03:00.239370 119696 ccl/storageccl/export.go:124 [n1,s1,r54/1:/Table/51/1/4/#/52/1/{1/…-2}] export [/Table/51/1/4/#/52/1/1/#/53/1,/Table/51/1/4/#/52/1/2) I180110 18:03:00.239471 119782 ccl/storageccl/export.go:124 [n1,s1,r64/1:/Table/51/1/4/#/54/1/{1-2}] export [/Table/51/1/4/#/54/1/1,/Table/51/1/4/#/54/1/2) I180110 18:03:00.239371 119611 ccl/storageccl/export.go:124 [n1,s1,r44/1:/Table/51/1/4/#/52/1/1{-/…}] export [/Table/51/1/4/#/52/1/1,/Table/51/1/4/#/52/1/1/#/53/1) I180110 18:03:00.239882 119645 ccl/storageccl/export.go:124 [n1,s1,r45/1:/Table/51/1/4/#/52/1/2{-/…}] export [/Table/51/1/4/#/52/1/2,/Table/51/1/4/#/52/1/2/#/53/1) I180110 18:03:00.239915 119646 ccl/storageccl/export.go:124 [n1,s1,r55/1:/Table/51/1/4/#/5{2/1/…-4/1/1}] export [/Table/51/1/4/#/52/1/2/#/53/1,/Table/51/1/4/#/54/1/1) I180110 18:03:00.242495 119794 ccl/storageccl/export.go:124 [n1,s1,r28/1:/Table/51/1/{5-6}] export [/Table/51/1/5,/Table/51/1/6) I180110 18:03:00.242578 119567 ccl/storageccl/export.go:124 [n1,s1,r30/1:/Table/51/1/{7-8}] export [/Table/51/1/7,/Table/51/1/8) I180110 18:03:00.242631 119783 ccl/storageccl/export.go:124 [n1,s1,r31/1:/Table/51/1/{8-9}] export [/Table/51/1/8,/Table/51/1/9) I180110 18:03:00.242578 119648 ccl/storageccl/export.go:124 [n1,s1,r29/1:/Table/51/1/{6-7}] export [/Table/51/1/6,/Table/51/1/7) I180110 18:03:00.242583 119697 ccl/storageccl/export.go:124 [n1,s1,r65/1:/Table/51/1/{4/#/54…-5}] export [/Table/51/1/4/#/54/1/2,/Table/51/1/5) I180110 18:03:00.245395 119568 ccl/storageccl/export.go:124 [n1,s1,r32/1:/Table/5{1/1/9-2}] export [/Table/51/1/9,/Table/51/2) --- FAIL: test/TestBackupRestoreInterleaved/all_tables_in_interleave_hierarchy (0.000s) Test ended in panic. ------- Stdout: ------- W180110 18:03:00.275286 119831 server/status/runtime.go:109 Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180110 18:03:00.281610 119831 server/config.go:518 [n?] 1 storage engine initialized I180110 18:03:00.281656 119831 server/config.go:521 [n?] RocksDB cache size: 128 MiB I180110 18:03:00.281667 119831 server/config.go:521 [n?] store 0: in-memory, size 0 B I180110 18:03:00.285388 119831 server/node.go:361 [n?] **** cluster e5fa619e-84b4-4546-b83e-035c0afbcb10 has been created I180110 18:03:00.285454 119831 server/server.go:934 [n?] **** add additional nodes by specifying --join=127.0.0.1:35605 I180110 18:03:00.286184 119831 storage/store.go:1312 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available I180110 18:03:00.286249 119831 server/node.go:491 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=3.2 KiB), ranges=1, leases=0, writes=0.00, bytesPerReplica={p10=3322.00 p25=3322.00 p50=3322.00 p75=3322.00 p90=3322.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00} I180110 18:03:00.286281 119831 server/node.go:339 [n1] node ID 1 initialized I180110 18:03:00.286323 119831 gossip/gossip.go:332 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:35605" > attrs:<> locality:<> ServerVersion:<major_val:1 minor_val:1 patch:0 unstable:9 > I180110 18:03:00.286446 119831 storage/stores.go:331 [n1] read 0 node addresses from persistent storage I180110 18:03:00.286553 119831 server/node.go:632 [n1] connecting to gossip network to verify cluster ID... I180110 18:03:00.286585 119831 server/node.go:657 [n1] node connected via gossip and verified as part of cluster "e5fa619e-84b4-4546-b83e-035c0afbcb10" I180110 18:03:00.286610 119831 server/node.go:433 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes [] I180110 18:03:00.286667 119831 sql/distsql_physical_planner.go:122 [n1] creating DistSQLPlanner with address {tcp 127.0.0.1:35605} I180110 18:03:00.287975 119768 storage/replica_command.go:819 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2] I180110 18:03:00.290150 119831 server/server.go:1161 [n1] starting https server at 127.0.0.1:46291 I180110 18:03:00.290180 119831 server/server.go:1162 [n1] starting grpc/postgres server at 127.0.0.1:35605 I180110 18:03:00.290190 119831 server/server.go:1163 [n1] advertising CockroachDB node at 127.0.0.1:35605 I180110 18:03:00.303117 119768 storage/replica_command.go:819 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3] I180110 18:03:00.309002 119768 storage/replica_command.go:819 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4] W180110 18:03:00.310214 120037 storage/intent_resolver.go:324 [n1,s1] failed to push during intent resolution: failed to push "sql txn implicit" id=dca6c167 key=/Table/SystemConfigSpan/Start rw=true pri=0.03337446 iso=SERIALIZABLE stat=PENDING epo=0 ts=1515607380.304345637,0 orig=1515607380.304345637,0 max=1515607380.304345637,0 wto=false rop=false seq=7 I180110 18:03:00.310846 119831 sql/event_log.go:115 [n1] Event: "alter_table", target: 12, info: {TableName:eventlog Statement:ALTER TABLE system.eventlog ALTER COLUMN "uniqueID" SET DEFAULT uuid_v4() User:node MutationID:0 CascadeDroppedViews:[]} W180110 18:03:00.313544 120067 storage/intent_resolver.go:324 [n1,s1] failed to push during intent resolution: failed to push "split" id=56ea7810 key=/Local/Range/System/NodeLiveness/RangeDescriptor rw=true pri=0.00888184 iso=SERIALIZABLE stat=PENDING epo=0 ts=1515607380.309028972,0 orig=1515607380.309028972,0 max=1515607380.309028972,0 wto=false rop=false seq=3 I180110 18:03:00.314936 119768 storage/replica_command.go:819 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5] I180110 18:03:00.315957 119831 sql/lease.go:348 [n1] publish: descID=12 (eventlog) version=2 mtime=2018-01-10 18:03:00.315829678 +0000 UTC I180110 18:03:00.319063 119768 storage/replica_command.go:819 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6] I180110 18:03:00.325466 119768 storage/replica_command.go:819 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7] I180110 18:03:00.328366 119831 sql/event_log.go:115 [n1] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:node} I180110 18:03:00.329842 119768 storage/replica_command.go:819 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8] I180110 18:03:00.335208 119768 storage/replica_command.go:819 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9] I180110 18:03:00.339886 119831 sql/event_log.go:115 [n1] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:$1 User:node} I180110 18:03:00.340273 119768 storage/replica_command.go:819 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10] I180110 18:03:00.343763 119768 storage/replica_command.go:819 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11] I180110 18:03:00.345078 119831 sql/event_log.go:115 [n1] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:node} I180110 18:03:00.348228 119768 storage/replica_command.go:819 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12] I180110 18:03:00.351319 119768 storage/replica_command.go:819 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13] I180110 18:03:00.357426 119768 storage/replica_command.go:819 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14] I180110 18:03:00.359871 119831 sql/event_log.go:115 [n1] Event: "alter_table", target: 4, info: {TableName:users Statement:ALTER TABLE system.users ADD COLUMN IF NOT EXISTS "isRole" BOOL NOT NULL DEFAULT false User:node MutationID:1 CascadeDroppedViews:[]} I180110 18:03:00.361795 119768 storage/replica_command.go:819 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15] I180110 18:03:00.362485 119831 sql/lease.go:348 [n1] publish: descID=4 (users) version=2 mtime=2018-01-10 18:03:00.362378 +0000 UTC I180110 18:03:00.365402 119768 storage/replica_command.go:819 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16] I180110 18:03:00.367126 119831 sql/lease.go:348 [n1] publish: descID=4 (users) version=3 mtime=2018-01-10 18:03:00.366998973 +0000 UTC I180110 18:03:00.369005 119768 storage/replica_command.go:819 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17] I180110 18:03:00.371107 119831 sql/backfill.go:133 [n1] Running backfill for "users", v=3, m=1 I180110 18:03:00.372228 119768 storage/replica_command.go:819 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18] I180110 18:03:00.375920 119768 storage/replica_command.go:819 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19] I180110 18:03:00.380211 119768 storage/replica_command.go:819 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20] I180110 18:03:00.380849 119831 sql/lease.go:348 [n1] publish: descID=4 (users) version=4 mtime=2018-01-10 18:03:00.380596668 +0000 UTC I180110 18:03:00.385248 119831 sql/event_log.go:115 [n1] Event: "finish_schema_change", target: 4, info: {MutationID:1} I180110 18:03:00.386245 119831 sql/lease.go:274 publish (count leases): descID=4 name=users version=3 count=1 I180110 18:03:00.418753 119831 server/server.go:1232 [n1] done ensuring all necessary migrations have run I180110 18:03:00.418794 119831 server/server.go:1235 [n1] serving sql connections I180110 18:03:00.421106 120299 sql/event_log.go:115 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:35605} Attrs: Locality: ServerVersion:1.1-9} ClusterID:e5fa619e-84b4-4546-b83e-035c0afbcb10 StartedAt:1515607380286590075 LastUp:1515607380286590075} I180110 18:03:00.433218 120174 sql/event_log.go:115 [client=127.0.0.1:50924,user=root,n1] Event: "create_database", target: 50, info: {DatabaseName:ignored Statement:CREATE DATABASE ignored User:root} I180110 18:03:00.433578 119768 storage/replica_command.go:819 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21] I180110 18:03:00.435962 120174 sql/event_log.go:115 [client=127.0.0.1:50924,user=root,n1] Event: "create_database", target: 51, info: {DatabaseName:data Statement:CREATE DATABASE data User:root} I180110 18:03:00.439496 119768 storage/replica_command.go:819 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22] I180110 18:03:00.445885 120339 storage/replica_command.go:819 [n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23] I180110 18:03:00.453388 120339 storage/replica_command.go:819 [n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/52/1/0/#/55/1/2 [r24] I180110 18:03:00.454107 120341 storage/replica_command.go:819 [n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/52/1/0/#/53/1/1 [r25] I180110 18:03:00.462172 120339 storage/replica_command.go:819 [n1,s1,r24/1:/{Table/52/1/0…-Max}] initiating a split of this range at key /Table/52/1/1/#/55/1/1 [r26] I180110 18:03:00.463059 120340 storage/replica_command.go:819 [n1,s1,r24/1:/{Table/52/1/0…-Max}] initiating a split of this range at key /Table/52/1/1 [r27] W180110 18:03:00.468477 120310 storage/replica.go:2174 [n1,s1,r24/1:/Table/52/1/{0/#/55…-1/#/55…}] context canceled before command queue: 1 QueryTxn I180110 18:03:00.511020 120341 storage/replica_command.go:819 [n1,s1,r23/1:/Table/52{-/1/0/#/5…}] initiating a split of this range at key /Table/52/1/0/#/53/1/1 [r28] I180110 18:03:00.515293 120340 storage/replica_command.go:819 [n1,s1,r24/1:/Table/52/1/{0/#/55…-1/#/55…}] initiating a split of this range at key /Table/52/1/1 [r29] I180110 18:03:00.517083 120341 storage/replica_command.go:819 [n1,s1,r28/1:/Table/52/1/0/#/5{3/1/1-5/1/2}] initiating a split of this range at key /Table/52/1/0/#/53/1/1/#/54/1 [r30] I180110 18:03:00.521277 120340 storage/replica_command.go:819 [n1,s1,r29/1:/Table/52/1/1{-/#/55/…}] initiating a split of this range at key /Table/52/1/1/#/53/1/1 [r31] I180110 18:03:00.522960 120341 storage/replica_command.go:819 [n1,s1,r30/1:/Table/52/1/0/#/5{3/1/…-5/1/2}] initiating a split of this range at key /Table/52/1/0/#/53/1/2 [r32] I180110 18:03:00.526073 120341 storage/replica_command.go:819 [n1,s1,r32/1:/Table/52/1/0/#/5{3/1/2-5/1/2}] initiating a split of this range at key /Table/52/1/0/#/53/1/2/#/54/1 [r33] I180110 18:03:00.527339 120340 storage/replica_command.go:819 [n1,s1,r31/1:/Table/52/1/1/#/5{3/1/1-5/1/1}] initiating a split of this range at key /Table/52/1/1/#/53/1/1/#/54/1 [r34] I180110 18:03:00.530427 120340 storage/replica_command.go:819 [n1,s1,r34/1:/Table/52/1/1/#/5{3/1/…-5/1/1}] initiating a split of this range at key /Table/52/1/1/#/53/1/2 [r35] I180110 18:03:00.531312 120341 storage/replica_command.go:819 [n1,s1,r33/1:/Table/52/1/0/#/5{3/1/…-5/1/2}] initiating a split of this range at key /Table/52/1/0/#/55/1/1 [r36] I180110 18:03:00.534390 120340 storage/replica_command.go:819 [n1,s1,r35/1:/Table/52/1/1/#/5{3/1/2-5/1/1}] initiating a split of this range at key /Table/52/1/1/#/53/1/2/#/54/1 [r37] I180110 18:03:00.536031 120341 storage/replica_command.go:819 [n1,s1,r26/1:/{Table/52/1/1…-Max}] initiating a split of this range at key /Table/52/1/1/#/55/1/2 [r38] I180110 18:03:00.536192 120339 storage/replica_command.go:819 [n1,s1,r26/1:/{Table/52/1/1…-Max}] initiating a split of this range at key /Table/52/1/2/#/53/1/2/#/54/1 [r39] W180110 18:03:00.539966 120490 storage/replica.go:2174 [n1,s1,r26/1:/Table/52/1/1/#/55/1/{1-2}] context canceled before command queue: 1 QueryTxn I180110 18:03:00.540236 120341 storage/replica_command.go:819 [n1,s1,r38/1:/{Table/52/1/1…-Max}] initiating a split of this range at key /Table/52/1/2 [r40] I180110 18:03:00.545769 120341 storage/replica_command.go:819 [n1,s1,r40/1:/{Table/52/1/2-Max}] initiating a split of this range at key /Table/52/1/2/#/53/1/1 [r41] I180110 18:03:00.550971 120341 storage/replica_command.go:819 [n1,s1,r41/1:/{Table/52/1/2…-Max}] initiating a split of this range at key /Table/52/1/2/#/53/1/1/#/54/1 [r42] I180110 18:03:00.558917 120341 storage/replica_command.go:819 [n1,s1,r42/1:/{Table/52/1/2…-Max}] initiating a split of this range at key /Table/52/1/2/#/53/1/2 [r43] I180110 18:03:00.594970 120339 storage/replica_command.go:819 [n1,s1,r43/1:/{Table/52/1/2…-Max}] initiating a split of this range at key /Table/52/1/2/#/53/1/2/#/54/1 [r44] I180110 18:03:00.599792 120339 storage/replica_command.go:819 [n1,s1,r44/1:/{Table/52/1/2…-Max}] initiating a split of this range at key /Table/52/1/3/#/53/1/2 [r45] I180110 18:03:00.600503 120340 storage/replica_command.go:819 [n1,s1,r44/1:/{Table/52/1/2…-Max}] initiating a split of this range at key /Table/52/1/2/#/55/1/1 [r46] W180110 18:03:00.607605 120619 storage/replica.go:2174 [n1,s1,r44/1:/Table/52/1/{2/#/53…-3/#/53…}] context canceled before command queue: 1 QueryTxn I180110 18:03:00.607847 120339 storage/replica_command.go:819 [n1,s1,r45/1:/{Table/52/1/3…-Max}] initiating a split of this range at key /Table/52/1/4/#/53/1/1/#/54/1 [r47] I180110 18:03:00.608619 120341 storage/replica_command.go:819 [n1,s1,r45/1:/{Table/52/1/3…-Max}] initiating a split of this range at key /Table/52/1/3/#/53/1/2/#/54/1 [r48] W180110 18:03:00.614903 120572 storage/replica.go:2174 [n1,s1,r45/1:/Table/52/1/{3/#/53…-4/#/53…}] context canceled before command queue: 1 QueryTxn I180110 18:03:00.654583 120340 storage/replica_command.go:819 [n1,s1,r44/1:/Table/52/1/{2/#/53…-3/#/53…}] initiating a split of this range at key /Table/52/1/2/#/55/1/1 [r49] I180110 18:03:00.660522 120340 storage/replica_command.go:819 [n1,s1,r49/1:/Table/52/1/{2/#/55…-3/#/53…}] initiating a split of this range at key /Table/52/1/2/#/55/1/2 [r50] I180110 18:03:00.664488 120341 storage/replica_command.go:819 [n1,s1,r45/1:/Table/52/1/{3/#/53…-4/#/53…}] initiating a split of this range at key /Table/52/1/3/#/53/1/2/#/54/1 [r51] I180110 18:03:00.667950 120340 storage/replica_command.go:819 [n1,s1,r50/1:/Table/52/1/{2/#/55…-3/#/53…}] initiating a split of this range at key /Table/52/1/3 [r52] I180110 18:03:00.669715 120341 storage/replica_command.go:819 [n1,s1,r51/1:/Table/52/1/{3/#/53…-4/#/53…}] initiating a split of this range at key /Table/52/1/3/#/55/1/1 [r53] I180110 18:03:00.673659 120340 storage/replica_command.go:819 [n1,s1,r52/1:/Table/52/1/3{-/#/53/…}] initiating a split of this range at key /Table/52/1/3/#/53/1/1 [r54] I180110 18:03:00.673762 120341 storage/replica_command.go:819 [n1,s1,r53/1:/Table/52/1/{3/#/55…-4/#/53…}] initiating a split of this range at key /Table/52/1/3/#/55/1/2 [r55] I180110 18:03:00.677249 120340 storage/replica_command.go:819 [n1,s1,r54/1:/Table/52/1/3/#/53/1/{1-2}] initiating a split of this range at key /Table/52/1/3/#/53/1/1/#/54/1 [r56] I180110 18:03:00.677963 120341 storage/replica_command.go:819 [n1,s1,r55/1:/Table/52/1/{3/#/55…-4/#/53…}] initiating a split of this range at key /Table/52/1/4 [r57] I180110 18:03:00.682706 120339 storage/replica_command.go:819 [n1,s1,r47/1:/{Table/52/1/4…-Max}] initiating a split of this range at key /Table/52/1/6 [r58] I180110 18:03:00.683136 120341 storage/replica_command.go:819 [n1,s1,r57/1:/Table/52/1/4{-/#/53/…}] initiating a split of this range at key /Table/52/1/4/#/53/1/1 [r59] I180110 18:03:00.683623 120340 storage/replica_command.go:819 [n1,s1,r47/1:/{Table/52/1/4…-Max}] initiating a split of this range at key /Table/52/1/4/#/53/1/2 [r60] W180110 18:03:00.686715 120739 storage/replica.go:2174 [n1,s1,r47/1:/Table/52/1/{4/#/53…-6}] context canceled before command queue: 1 QueryTxn I180110 18:03:00.687267 120341 storage/replica_command.go:819 [n1,s1,r58/1:/{Table/52/1/6-Max}] initiating a split of this range at key /Table/52/1/7 [r61] I180110 18:03:00.692013 120341 storage/replica_command.go:819 [n1,s1,r61/1:/{Table/52/1/7-Max}] initiating a split of this range at key /Table/52/1/8 [r62] I180110 18:03:00.700994 120341 storage/replica_command.go:819 [n1,s1,r62/1:/{Table/52/1/8-Max}] initiating a split of this range at key /Table/52/1/9 [r63] panic: test timed out after 4m0s goroutine 120721 [running]: testing.startAlarm.func1() /usr/local/go/src/testing/testing.go:1145 +0xf9 created by time.goFunc /usr/local/go/src/time/sleep.go:170 +0x44 goroutine 1 [chan receive]: testing.(*T).Run(0xc4200c2690, 0x233f0da, 0x1c, 0x23ff290, 0x8d4f01) /usr/local/go/src/testing/testing.go:790 +0x2fc testing.runTests.func1(0xc4200c2690) /usr/local/go/src/testing/testing.go:1004 +0x64 testing.tRunner(0xc4200c2690, 0xc42067bdb8) /usr/local/go/src/testing/testing.go:746 +0xd0 testing.runTests(0xc4201ddc00, 0x3410660, 0x38, 0x38, 0x0) /usr/local/go/src/testing/testing.go:1002 +0x2d8 testing.(*M).Run(0xc420a27f18, 0xc420086500) /usr/local/go/src/testing/testing.go:921 +0x111 github.com/cockroachdb/cockroach/pkg/ccl/sqlccl.TestMain(0xc42067bf18) /go/src/github.com/cockroachdb/cockroach/pkg/ccl/sqlccl/main_test.go:31 +0xda main.main() github.com/cockroachdb/cockroach/pkg/ccl/sqlccl/_test/_testmain.go:172 +0xdb goroutine 24 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).flushDaemon(0x365efa0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1043 +0x81 created by github.com/cockroachdb/cockroach/pkg/util/log.init.0 /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:581 +0xbf goroutine 50 [syscall, 4 minutes]: os/signal.signal_recv(0x0) /usr/local/go/src/runtime/sigqueue.go:131 +0xa6 os/signal.loop() /usr/local/go/src/os/signal/signal_unix.go:22 +0x22 created by os/signal.init.0 /usr/local/go/src/os/signal/signal_unix.go:28 +0x41 goroutine 118774 [chan receive]: testing.(*T).Run(0xc4215f64b0, 0x234c57a, 0x22, 0xc4202f3080, 0x1a) /usr/local/go/src/testing/testing.go:790 +0x2fc github.com/cockroachdb/cockroach/pkg/ccl/sqlccl_test.TestBackupRestoreInterleaved(0xc4215f64b0) /go/src/github.com/cockroachdb/cockroach/pkg/ccl/sqlccl/backup_test.go:1065 +0x937 testing.tRunner(0xc4215f64b0, 0x23ff290) /usr/local/go/src/testing/testing.go:746 +0xd0 created by testing.(*T).Run /usr/local/go/src/testing/testing.go:789 +0x2de goroutine 5 [select, locked to thread]: runtime.gopark(0x2403108, 0x0, 0x2319caf, 0x6, 0x18, 0x1) /usr/local/go/src/runtime/proc.go:277 +0x12c runtime.selectgo(0xc420041f50, 0xc42048ed80) /usr/local/go/src/runtime/select.go:395 +0x1138 runtime.ensureSigM.func1() /usr/local/go/src/runtime/signal_unix.go:511 +0x220 runtime.goexit() /usr/local/go/src/runtime/asm_amd64.s:2337 +0x1 goroutine 118948 [semacquire]: sync.runtime_notifyListWait(0xc423412890, 0xc4000003ae) /usr/local/go/src/runtime/sema.go:507 +0x110 sync.(*Cond).Wait(0xc423412880) /usr/local/go/src/sync/cond.go:56 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420b5d440, 0x343ba20, 0xc423a6ecf0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:197 +0x7c github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2(0x343ba20, 0xc423a6ecf0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:166 +0x3e github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420344f10, 0xc4210b58c0, 0xc420344ee0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 120716 [semacquire]: sync.runtime_Semacquire(0xc42008f6d4) /usr/local/go/src/runtime/sema.go:56 +0x39 sync.(*WaitGroup).Wait(0xc42008f6c8) /usr/local/go/src/sync/waitgroup.go:131 +0x72 github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup.(*Group).Wait(0xc42008f6c0, 0xc4216028a0, 0x2345377) /go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup/errgroup.go:41 +0x31 github.com/cockroachdb/cockroach/pkg/ccl/storageccl.evalImport(0x343ba20, 0xc4216028a0, 0x344ec60, 0xc423526000, 0x150884e1f9f193d6, 0x0, 0x100000001, 0x1, 0x3f, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/ccl/storageccl/import.go:311 +0x2084 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).executeAdminBatch(0xc423526000, 0x343ba20, 0xc4216028a0, 0x150884e1f9f193d6, 0x0, 0x100000001, 0x1, 0x3f, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2485 +0xacd github.com/cockroachdb/cockroach/pkg/storage.(*Replica).Send(0xc423526000, 0x343ba20, 0xc4216028a0, 0x150884e1f9f193d6, 0x0, 0x100000001, 0x1, 0x3f, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:1919 +0x60f github.com/cockroachdb/cockroach/pkg/storage.(*Store).Send(0xc42354c400, 0x343ba20, 0xc421602870, 0x150884e1f9f193d6, 0x0, 0x100000001, 0x1, 0x3f, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:2688 +0x688 github.com/cockroachdb/cockroach/pkg/storage.(*Stores).Send(0xc423198bb0, 0x343ba20, 0xc421602810, 0x0, 0x0, 0x100000001, 0x1, 0x3f, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/stores.go:214 +0x214 github.com/cockroachdb/cockroach/pkg/server.(*Node).batchInternal.func1(0x343ba20, 0xc421602810, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/server/node.go:865 +0x192 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunTaskWithErr(0xc4253dbb00, 0x343ba20, 0xc421602810, 0x232938b, 0x10, 0xc422157cf8, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:245 +0xf7 github.com/cockroachdb/cockroach/pkg/server.(*Node).batchInternal(0xc424f8d200, 0x343ba20, 0xc421602810, 0xc423e212a8, 0xc421602810, 0x14ecafe, 0x343ba20) /go/src/github.com/cockroachdb/cockroach/pkg/server/node.go:852 +0x152 github.com/cockroachdb/cockroach/pkg/server.(*Node).Batch(0xc424f8d200, 0x343ba20, 0xc421602810, 0xc423e212a8, 0x1f, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/server/node.go:893 +0x99 github.com/cockroachdb/cockroach/pkg/kv.(*grpcTransport).send.func1(0xc4218db490, 0xc423e21290, 0x343ba20, 0xc421602750, 0x100000000, 0xc42256e1d8, 0x8580c8) /go/src/github.com/cockroachdb/cockroach/pkg/kv/transport.go:243 +0x6b0 github.com/cockroachdb/cockroach/pkg/kv.(*grpcTransport).send(0xc4218db490, 0x343ba20, 0xc421602750, 0xc421fd9ce0, 0xf, 0xc42275ce60, 0x0, 0x0, 0x100000001, 0x1, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/transport.go:271 +0x9a github.com/cockroachdb/cockroach/pkg/kv.(*grpcTransport).SendNext(0xc4218db490, 0x343ba20, 0xc421602750, 0xc4234e5b60) /go/src/github.com/cockroachdb/cockroach/pkg/kv/transport.go:208 +0x31b github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendToReplicas(0xc42474d1d0, 0x343ba20, 0xc421602750, 0xc42474d220, 0x3f, 0xc42356b3c0, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1246 +0x3a4 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendRPC(0xc42474d1d0, 0x343ba20, 0xc421602750, 0x3f, 0xc42356b3c0, 0x1, 0x1, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:382 +0x27b github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendSingleRange(0xc42474d1d0, 0x343ba20, 0xc421602750, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:446 +0x16e github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendPartialBatch(0xc42474d1d0, 0x343ba20, 0xc421602750, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1024 +0x347 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges(0xc42474d1d0, 0x343ba20, 0xc421602750, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:700 +0x16f8 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send(0xc42474d1d0, 0x343ba20, 0xc421602750, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:621 +0x3f5 github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send(0xc4228b22a0, 0x343ba20, 0xc421602750, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:444 +0x1ce github.com/cockroachdb/cockroach/pkg/internal/client.SendWrappedWith(0x343b960, 0xc420878840, 0x341b060, 0xc4228b22a0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/sender.go:62 +0x143 github.com/cockroachdb/cockroach/pkg/internal/client.SendWrapped(0x343b960, 0xc420878840, 0x341b060, 0xc4228b22a0, 0x34496e0, 0xc4253e6be0, 0xc422397880, 0xc424252f78, 0x16d2656) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/sender.go:79 +0xaf github.com/cockroachdb/cockroach/pkg/ccl/sqlccl.restore.func4(0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/ccl/sqlccl/restore.go:967 +0x13e github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup.(*Group).Go.func1(0xc420878880, 0xc4234e5b00) /go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup/errgroup.go:58 +0x57 created by github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup.(*Group).Go /go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup/errgroup.go:55 +0x66 goroutine 118949 [semacquire]: sync.runtime_notifyListWait(0xc423412890, 0xc4000003a9) /usr/local/go/src/runtime/sema.go:507 +0x110 sync.(*Cond).Wait(0xc423412880) /usr/local/go/src/sync/cond.go:56 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420b5d440, 0x343ba20, 0xc425475e90) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:197 +0x7c github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2(0x343ba20, 0xc425475e90) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:166 +0x3e github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420344f40, 0xc4210b58c0, 0xc420344f20) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 119790 [select]: github.com/cockroachdb/cockroach/pkg/server.(*Server).startSampleEnvironment.func1(0x343ba20, 0xc420846120) /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:1416 +0x16c github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc424060280, 0xc4253dbb00, 0xc4201dde00) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 118988 [select]: github.com/cockroachdb/cockroach/pkg/sql.(*DistSQLPlanner).initRunners.func1(0x343ba20, 0xc421227050) /go/src/github.com/cockroachdb/cockroach/pkg/sql/distsql_running.go:86 +0x167 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc4204538d0, 0xc4210b58c0, 0xc4204538c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 118965 [semacquire]: sync.runtime_Semacquire(0xc420b5d4c4) /usr/local/go/src/runtime/sema.go:56 +0x39 sync.(*WaitGroup).Wait(0xc420b5d4b8) /usr/local/go/src/sync/waitgroup.go:131 +0x72 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Wait(0xc420b5d440, 0x343ba20, 0xc425475fb0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:172 +0x31 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Wait-fm(0x343ba20, 0xc425475fb0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3834 +0x3e github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420345270, 0xc4210b58c0, 0xc420345260) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 118820 [chan receive]: github.com/cockroachdb/cockroach/pkg/gossip.(*server).start.func3(0x343ba20, 0xc42303c3f0) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/server.go:371 +0x5a github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc4203763e0, 0xc4210b58c0, 0xc422f01b00) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 118818 [chan receive]: github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux.muxListener.Accept(...) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux/cmux.go:184 github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux.(*muxListener).Accept(0xc422f016a0, 0x24024e8, 0xc4207cba40, 0x34478c0, 0xc422a1b220) <autogenerated>:1 +0x65 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).Serve(0xc4207cba40, 0x343aba0, 0xc422f016a0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:463 +0x196 github.com/cockroachdb/cockroach/pkg/server.(*Server).Start.func7(0x343ba20, 0xc42303c3c0) /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:795 +0x43 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc4203761c0, 0xc4210b58c0, 0xc422f01980) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 119938 [semacquire]: sync.runtime_notifyListWait(0xc4208d8310, 0xc400000325) /usr/local/go/src/runtime/sema.go:507 +0x110 sync.(*Cond).Wait(0xc4208d8300) /usr/local/go/src/sync/cond.go:56 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc4226ac7e0, 0x343ba20, 0xc420afda10) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:197 +0x7c github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2(0x343ba20, 0xc420afda10) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:166 +0x3e github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420452b80, 0xc4253dbb00, 0xc420452b70) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 119973 [semacquire]: sync.runtime_notifyListWait(0xc4208d8310, 0xc40000033c) /usr/local/go/src/runtime/sema.go:507 +0x110 sync.(*Cond).Wait(0xc4208d8300) /usr/local/go/src/sync/cond.go:56 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc4226ac7e0, 0x343ba20, 0xc420d947e0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:197 +0x7c github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2(0x343ba20, 0xc420d947e0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:166 +0x3e github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc4204531f0, 0xc4253dbb00, 0xc4204531e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 118906 [semacquire]: sync.runtime_notifyListWait(0xc423412890, 0xc4000003b5) /usr/local/go/src/runtime/sema.go:507 +0x110 sync.(*Cond).Wait(0xc423412880) /usr/local/go/src/sync/cond.go:56 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420b5d440, 0x343ba20, 0xc42303c870) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:197 +0x7c github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2(0x343ba20, 0xc42303c870) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:166 +0x3e github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420344680, 0xc4210b58c0, 0xc420344670) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 118911 [semacquire]: sync.runtime_notifyListWait(0xc423412890, 0xc4000003b8) /usr/local/go/src/runtime/sema.go:507 +0x110 sync.(*Cond).Wait(0xc423412880) /usr/local/go/src/sync/cond.go:56 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420b5d440, 0x343ba20, 0xc423a6eab0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:197 +0x7c github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2(0x343ba20, 0xc423a6eab0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:166 +0x3e github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420344770, 0xc4210b58c0, 0xc420344760) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 118928 [semacquire]: sync.runtime_notifyListWait(0xc423412890, 0xc4000003c2) /usr/local/go/src/runtime/sema.go:507 +0x110 sync.(*Cond).Wait(0xc423412880) /usr/local/go/src/sync/cond.go:56 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420b5d440, 0x343ba20, 0xc42303cb70) /go/src/github.com/co ``` Please assign, take a look and update the issue accordingly.
1.0
teamcity: failed tests on master: test/TestBackupRestoreInterleaved, test/TestBackupRestoreInterleaved/all_tables_in_interleave_hierarchy, testrace/TestBackupRestoreInterleaved, testrace/TestBackupRestoreInterleaved/all_tables_in_interleave_hierarchy - The following tests appear to have failed: [#474455](https://teamcity.cockroachdb.com/viewLog.html?buildId=474455): ``` --- FAIL: test/TestBackupRestoreInterleaved (0.000s) Test ended in panic. ------- Stdout: ------- W180110 18:02:59.680952 118774 server/status/runtime.go:109 Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180110 18:02:59.682776 118774 server/config.go:518 [n?] 1 storage engine initialized I180110 18:02:59.682793 118774 server/config.go:521 [n?] RocksDB cache size: 128 MiB I180110 18:02:59.682798 118774 server/config.go:521 [n?] store 0: in-memory, size 0 B I180110 18:02:59.684988 118774 server/node.go:361 [n?] **** cluster 12dcf5c1-311b-4753-a474-73554bd7f94f has been created I180110 18:02:59.685014 118774 server/server.go:934 [n?] **** add additional nodes by specifying --join=127.0.0.1:35537 I180110 18:02:59.685616 118774 storage/store.go:1312 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available I180110 18:02:59.685656 118774 server/node.go:491 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=3.2 KiB), ranges=1, leases=0, writes=0.00, bytesPerReplica={p10=3322.00 p25=3322.00 p50=3322.00 p75=3322.00 p90=3322.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00} I180110 18:02:59.685688 118774 server/node.go:339 [n1] node ID 1 initialized I180110 18:02:59.685757 118774 gossip/gossip.go:332 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:35537" > attrs:<> locality:<> ServerVersion:<major_val:1 minor_val:1 patch:0 unstable:9 > I180110 18:02:59.685924 118774 storage/stores.go:331 [n1] read 0 node addresses from persistent storage I180110 18:02:59.686060 118774 server/node.go:632 [n1] connecting to gossip network to verify cluster ID... I180110 18:02:59.686094 118774 server/node.go:657 [n1] node connected via gossip and verified as part of cluster "12dcf5c1-311b-4753-a474-73554bd7f94f" I180110 18:02:59.686121 118774 server/node.go:433 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes [] I180110 18:02:59.686252 118774 sql/distsql_physical_planner.go:122 [n1] creating DistSQLPlanner with address {tcp 127.0.0.1:35537} I180110 18:02:59.687294 118745 storage/replica_command.go:819 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2] I180110 18:02:59.693580 118774 server/server.go:1161 [n1] starting https server at 127.0.0.1:40697 I180110 18:02:59.693633 118774 server/server.go:1162 [n1] starting grpc/postgres server at 127.0.0.1:35537 I180110 18:02:59.693657 118774 server/server.go:1163 [n1] advertising CockroachDB node at 127.0.0.1:35537 I180110 18:02:59.698407 118745 storage/replica_command.go:819 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3] I180110 18:02:59.703687 118745 storage/replica_command.go:819 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4] W180110 18:02:59.704283 119042 storage/intent_resolver.go:324 [n1,s1] failed to push during intent resolution: failed to push "sql txn implicit" id=96eef510 key=/Table/SystemConfigSpan/Start rw=true pri=0.03373129 iso=SERIALIZABLE stat=PENDING epo=0 ts=1515607379.699252305,0 orig=1515607379.699252305,0 max=1515607379.699252305,0 wto=false rop=false seq=7 I180110 18:02:59.705378 118774 sql/event_log.go:115 [n1] Event: "alter_table", target: 12, info: {TableName:eventlog Statement:ALTER TABLE system.eventlog ALTER COLUMN "uniqueID" SET DEFAULT uuid_v4() User:node MutationID:0 CascadeDroppedViews:[]} I180110 18:02:59.709553 118745 storage/replica_command.go:819 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5] I180110 18:02:59.711655 118774 sql/lease.go:348 [n1] publish: descID=12 (eventlog) version=2 mtime=2018-01-10 18:02:59.711397351 +0000 UTC I180110 18:02:59.714515 118745 storage/replica_command.go:819 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6] I180110 18:02:59.722104 118745 storage/replica_command.go:819 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7] I180110 18:02:59.728036 118745 storage/replica_command.go:819 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8] W180110 18:02:59.728476 118812 storage/intent_resolver.go:324 [n1,s1] failed to push during intent resolution: failed to push "sql txn implicit" id=f8a38c98 key=/Table/SystemConfigSpan/Start rw=true pri=0.00799806 iso=SERIALIZABLE stat=PENDING epo=0 ts=1515607379.723877453,0 orig=1515607379.723877453,0 max=1515607379.723877453,0 wto=false rop=false seq=7 I180110 18:02:59.729842 118774 sql/event_log.go:115 [n1] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:node} I180110 18:02:59.732898 118745 storage/replica_command.go:819 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9] I180110 18:02:59.738942 118745 storage/replica_command.go:819 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10] I180110 18:02:59.741515 118774 sql/event_log.go:115 [n1] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:$1 User:node} I180110 18:02:59.743707 118745 storage/replica_command.go:819 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11] I180110 18:02:59.745378 118774 sql/event_log.go:115 [n1] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:node} I180110 18:02:59.748179 118745 storage/replica_command.go:819 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12] I180110 18:02:59.754047 118745 storage/replica_command.go:819 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13] I180110 18:02:59.756675 118774 sql/event_log.go:115 [n1] Event: "alter_table", target: 4, info: {TableName:users Statement:ALTER TABLE system.users ADD COLUMN IF NOT EXISTS "isRole" BOOL NOT NULL DEFAULT false User:node MutationID:1 CascadeDroppedViews:[]} I180110 18:02:59.757649 118745 storage/replica_command.go:819 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14] I180110 18:02:59.761927 118774 sql/lease.go:348 [n1] publish: descID=4 (users) version=2 mtime=2018-01-10 18:02:59.761827399 +0000 UTC I180110 18:02:59.765394 118745 storage/replica_command.go:819 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15] I180110 18:02:59.766836 118774 sql/lease.go:348 [n1] publish: descID=4 (users) version=3 mtime=2018-01-10 18:02:59.766683769 +0000 UTC I180110 18:02:59.769003 118745 storage/replica_command.go:819 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16] I180110 18:02:59.771316 118774 sql/backfill.go:133 [n1] Running backfill for "users", v=3, m=1 I180110 18:02:59.771931 118745 storage/replica_command.go:819 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17] I180110 18:02:59.774793 118745 storage/replica_command.go:819 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18] I180110 18:02:59.778046 118745 storage/replica_command.go:819 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19] I180110 18:02:59.781235 118774 sql/lease.go:348 [n1] publish: descID=4 (users) version=4 mtime=2018-01-10 18:02:59.781034178 +0000 UTC I180110 18:02:59.782374 118745 storage/replica_command.go:819 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20] I180110 18:02:59.784585 118774 sql/event_log.go:115 [n1] Event: "finish_schema_change", target: 4, info: {MutationID:1} I180110 18:02:59.785559 118774 sql/lease.go:274 publish (count leases): descID=4 name=users version=3 count=1 I180110 18:02:59.818441 118774 server/server.go:1232 [n1] done ensuring all necessary migrations have run I180110 18:02:59.818491 118774 server/server.go:1235 [n1] serving sql connections I180110 18:02:59.820529 119248 sql/event_log.go:115 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:35537} Attrs: Locality: ServerVersion:1.1-9} ClusterID:12dcf5c1-311b-4753-a474-73554bd7f94f StartedAt:1515607379686099339 LastUp:1515607379686099339} I180110 18:02:59.835610 119285 sql/event_log.go:115 [client=127.0.0.1:59480,user=root,n1] Event: "create_database", target: 50, info: {DatabaseName:data Statement:CREATE DATABASE data User:root} I180110 18:02:59.836073 118745 storage/replica_command.go:819 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21] I180110 18:02:59.839557 119285 sql/event_log.go:115 [client=127.0.0.1:59480,user=root,n1] Event: "create_table", target: 51, info: {TableName:bank Statement:CREATE TABLE bank (id INT PRIMARY KEY, balance INT, payload STRING, FAMILY (id, balance, payload)) User:root} I180110 18:02:59.840359 118745 storage/replica_command.go:819 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22] I180110 18:02:59.842925 119285 storage/replica_command.go:819 [n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51/1/1 [r23] I180110 18:02:59.894721 119285 storage/replica_command.go:819 [n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/51/1/1 [r24] I180110 18:02:59.897970 119285 storage/replica_command.go:819 [n1,s1,r24/1:/{Table/51/1/1-Max}] initiating a split of this range at key /Table/51/1/2 [r25] I180110 18:02:59.902866 119285 storage/replica_command.go:819 [n1,s1,r25/1:/{Table/51/1/2-Max}] initiating a split of this range at key /Table/51/1/3 [r26] I180110 18:02:59.906764 119285 storage/replica_command.go:819 [n1,s1,r26/1:/{Table/51/1/3-Max}] initiating a split of this range at key /Table/51/1/4 [r27] I180110 18:02:59.911998 119285 storage/replica_command.go:819 [n1,s1,r27/1:/{Table/51/1/4-Max}] initiating a split of this range at key /Table/51/1/5 [r28] I180110 18:02:59.915310 119285 storage/replica_command.go:819 [n1,s1,r28/1:/{Table/51/1/5-Max}] initiating a split of this range at key /Table/51/1/6 [r29] I180110 18:02:59.918539 119285 storage/replica_command.go:819 [n1,s1,r29/1:/{Table/51/1/6-Max}] initiating a split of this range at key /Table/51/1/7 [r30] I180110 18:02:59.922686 119285 storage/replica_command.go:819 [n1,s1,r30/1:/{Table/51/1/7-Max}] initiating a split of this range at key /Table/51/1/8 [r31] I180110 18:02:59.926944 119285 storage/replica_command.go:819 [n1,s1,r31/1:/{Table/51/1/8-Max}] initiating a split of this range at key /Table/51/1/9 [r32] I180110 18:02:59.934885 119285 sql/event_log.go:115 [client=127.0.0.1:59480,user=root,n1] Event: "create_table", target: 52, info: {TableName:i0 Statement:CREATE TABLE i0 (a INT, b INT, PRIMARY KEY (a, b)) INTERLEAVE IN PARENT bank (a) User:root} I180110 18:02:59.936802 118745 storage/replica_command.go:819 [split,n1,s1,r32/1:/{Table/51/1/9-Max}] initiating a split of this range at key /Table/52 [r33] I180110 18:02:59.936960 119285 sql/lease.go:348 [client=127.0.0.1:59480,user=root,n1] publish: descID=52 (i0) version=2 mtime=2018-01-10 18:02:59.936866919 +0000 UTC I180110 18:02:59.941868 119285 sql/lease.go:348 [client=127.0.0.1:59480,user=root,n1] publish: descID=51 (bank) version=2 mtime=2018-01-10 18:02:59.941755855 +0000 UTC I180110 18:02:59.943463 119285 sql/lease.go:274 publish (count leases): descID=51 name=bank version=1 count=1 I180110 18:02:59.979049 119285 sql/event_log.go:115 [client=127.0.0.1:59480,user=root,n1] Event: "create_table", target: 53, info: {TableName:i0_0 Statement:CREATE TABLE i0_0 (a INT, b INT, c INT, PRIMARY KEY (a, b, c)) INTERLEAVE IN PARENT i0 (a, b) User:root} I180110 18:02:59.979712 118745 storage/replica_command.go:819 [split,n1,s1,r33/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r34] I180110 18:02:59.983045 119285 sql/lease.go:348 [client=127.0.0.1:59480,user=root,n1] publish: descID=53 (i0_0) version=2 mtime=2018-01-10 18:02:59.982717895 +0000 UTC I180110 18:02:59.987749 119285 sql/lease.go:348 [client=127.0.0.1:59480,user=root,n1] publish: descID=52 (i0) version=3 mtime=2018-01-10 18:02:59.987656881 +0000 UTC I180110 18:02:59.996752 119285 sql/event_log.go:115 [client=127.0.0.1:59480,user=root,n1] Event: "create_table", target: 54, info: {TableName:i1 Statement:CREATE TABLE i1 (a INT, b INT, PRIMARY KEY (a, b)) INTERLEAVE IN PARENT bank (a) User:root} I180110 18:02:59.997271 118745 storage/replica_command.go:819 [split,n1,s1,r34/1:/{Table/53-Max}] initiating a split of this range at key /Table/54 [r35] I180110 18:02:59.998973 119285 sql/lease.go:348 [client=127.0.0.1:59480,user=root,n1] publish: descID=54 (i1) version=2 mtime=2018-01-10 18:02:59.998869144 +0000 UTC I180110 18:03:00.005665 119285 sql/lease.go:348 [client=127.0.0.1:59480,user=root,n1] publish: descID=51 (bank) version=3 mtime=2018-01-10 18:03:00.00552806 +0000 UTC I180110 18:03:00.006862 119285 sql/lease.go:274 publish (count leases): descID=51 name=bank version=2 count=1 I180110 18:03:00.070675 119285 storage/replica_command.go:819 [n1,s1,r22/1:/Table/51{-/1/1}] initiating a split of this range at key /Table/51/1/0/#/52/1/1 [r36] I180110 18:03:00.076517 119285 storage/replica_command.go:819 [n1,s1,r36/1:/Table/51/1/{0/#/52…-1}] initiating a split of this range at key /Table/51/1/0/#/52/1/2 [r37] I180110 18:03:00.080343 119285 storage/replica_command.go:819 [n1,s1,r24/1:/Table/51/1/{1-2}] initiating a split of this range at key /Table/51/1/1/#/52/1/1 [r38] I180110 18:03:00.084681 119285 storage/replica_command.go:819 [n1,s1,r38/1:/Table/51/1/{1/#/52…-2}] initiating a split of this range at key /Table/51/1/1/#/52/1/2 [r39] I180110 18:03:00.090001 119285 storage/replica_command.go:819 [n1,s1,r25/1:/Table/51/1/{2-3}] initiating a split of this range at key /Table/51/1/2/#/52/1/1 [r40] I180110 18:03:00.093756 119285 storage/replica_command.go:819 [n1,s1,r40/1:/Table/51/1/{2/#/52…-3}] initiating a split of this range at key /Table/51/1/2/#/52/1/2 [r41] I180110 18:03:00.096434 119285 storage/replica_command.go:819 [n1,s1,r26/1:/Table/51/1/{3-4}] initiating a split of this range at key /Table/51/1/3/#/52/1/1 [r42] I180110 18:03:00.099404 119285 storage/replica_command.go:819 [n1,s1,r42/1:/Table/51/1/{3/#/52…-4}] initiating a split of this range at key /Table/51/1/3/#/52/1/2 [r43] I180110 18:03:00.103815 119285 storage/replica_command.go:819 [n1,s1,r27/1:/Table/51/1/{4-5}] initiating a split of this range at key /Table/51/1/4/#/52/1/1 [r44] I180110 18:03:00.107627 119285 storage/replica_command.go:819 [n1,s1,r44/1:/Table/51/1/{4/#/52…-5}] initiating a split of this range at key /Table/51/1/4/#/52/1/2 [r45] I180110 18:03:00.114279 119285 storage/replica_command.go:819 [n1,s1,r36/1:/Table/51/1/0/#/52/1/{1-2}] initiating a split of this range at key /Table/51/1/0/#/52/1/1/#/53/1 [r46] I180110 18:03:00.121526 119285 storage/replica_command.go:819 [n1,s1,r37/1:/Table/51/1/{0/#/52…-1}] initiating a split of this range at key /Table/51/1/0/#/52/1/2/#/53/1 [r47] I180110 18:03:00.127570 119285 storage/replica_command.go:819 [n1,s1,r38/1:/Table/51/1/1/#/52/1/{1-2}] initiating a split of this range at key /Table/51/1/1/#/52/1/1/#/53/1 [r48] I180110 18:03:00.131410 119285 storage/replica_command.go:819 [n1,s1,r39/1:/Table/51/1/{1/#/52…-2}] initiating a split of this range at key /Table/51/1/1/#/52/1/2/#/53/1 [r49] I180110 18:03:00.134666 119285 storage/replica_command.go:819 [n1,s1,r40/1:/Table/51/1/2/#/52/1/{1-2}] initiating a split of this range at key /Table/51/1/2/#/52/1/1/#/53/1 [r50] I180110 18:03:00.138930 119285 storage/replica_command.go:819 [n1,s1,r41/1:/Table/51/1/{2/#/52…-3}] initiating a split of this range at key /Table/51/1/2/#/52/1/2/#/53/1 [r51] I180110 18:03:00.142102 119285 storage/replica_command.go:819 [n1,s1,r42/1:/Table/51/1/3/#/52/1/{1-2}] initiating a split of this range at key /Table/51/1/3/#/52/1/1/#/53/1 [r52] I180110 18:03:00.146773 119285 storage/replica_command.go:819 [n1,s1,r43/1:/Table/51/1/{3/#/52…-4}] initiating a split of this range at key /Table/51/1/3/#/52/1/2/#/53/1 [r53] I180110 18:03:00.153089 119285 storage/replica_command.go:819 [n1,s1,r44/1:/Table/51/1/4/#/52/1/{1-2}] initiating a split of this range at key /Table/51/1/4/#/52/1/1/#/53/1 [r54] I180110 18:03:00.156921 119285 storage/replica_command.go:819 [n1,s1,r45/1:/Table/51/1/{4/#/52…-5}] initiating a split of this range at key /Table/51/1/4/#/52/1/2/#/53/1 [r55] I180110 18:03:00.164474 119285 storage/replica_command.go:819 [n1,s1,r47/1:/Table/51/1/{0/#/52…-1}] initiating a split of this range at key /Table/51/1/0/#/54/1/1 [r56] I180110 18:03:00.173135 119285 storage/replica_command.go:819 [n1,s1,r56/1:/Table/51/1/{0/#/54…-1}] initiating a split of this range at key /Table/51/1/0/#/54/1/2 [r57] I180110 18:03:00.178226 119285 storage/replica_command.go:819 [n1,s1,r49/1:/Table/51/1/{1/#/52…-2}] initiating a split of this range at key /Table/51/1/1/#/54/1/1 [r58] I180110 18:03:00.182743 119285 storage/replica_command.go:819 [n1,s1,r58/1:/Table/51/1/{1/#/54…-2}] initiating a split of this range at key /Table/51/1/1/#/54/1/2 [r59] I180110 18:03:00.187677 119285 storage/replica_command.go:819 [n1,s1,r51/1:/Table/51/1/{2/#/52…-3}] initiating a split of this range at key /Table/51/1/2/#/54/1/1 [r60] I180110 18:03:00.191390 119285 storage/replica_command.go:819 [n1,s1,r60/1:/Table/51/1/{2/#/54…-3}] initiating a split of this range at key /Table/51/1/2/#/54/1/2 [r61] I180110 18:03:00.196240 119285 storage/replica_command.go:819 [n1,s1,r53/1:/Table/51/1/{3/#/52…-4}] initiating a split of this range at key /Table/51/1/3/#/54/1/1 [r62] I180110 18:03:00.199671 119285 storage/replica_command.go:819 [n1,s1,r62/1:/Table/51/1/{3/#/54…-4}] initiating a split of this range at key /Table/51/1/3/#/54/1/2 [r63] I180110 18:03:00.202414 119285 storage/replica_command.go:819 [n1,s1,r55/1:/Table/51/1/{4/#/52…-5}] initiating a split of this range at key /Table/51/1/4/#/54/1/1 [r64] I180110 18:03:00.206213 119285 storage/replica_command.go:819 [n1,s1,r64/1:/Table/51/1/{4/#/54…-5}] initiating a split of this range at key /Table/51/1/4/#/54/1/2 [r65] I180110 18:03:00.221491 119727 ccl/storageccl/export.go:124 [n1,s1,r47/1:/Table/51/1/0/#/5{2/1/…-4/1/1}] export [/Table/51/1/0/#/52/1/2/#/53/1,/Table/51/1/0/#/54/1/1) I180110 18:03:00.221634 119726 ccl/storageccl/export.go:124 [n1,s1,r37/1:/Table/51/1/0/#/52/1/2{-/…}] export [/Table/51/1/0/#/52/1/2,/Table/51/1/0/#/52/1/2/#/53/1) I180110 18:03:00.221726 119724 ccl/storageccl/export.go:124 [n1,s1,r36/1:/Table/51/1/0/#/52/1/1{-/…}] export [/Table/51/1/0/#/52/1/1,/Table/51/1/0/#/52/1/1/#/53/1) I180110 18:03:00.221634 119723 ccl/storageccl/export.go:124 [n1,s1,r22/1:/Table/51{-/1/0/#/5…}] export [/Table/51/1,/Table/51/1/0/#/52/1/1) I180110 18:03:00.221657 119725 ccl/storageccl/export.go:124 [n1,s1,r46/1:/Table/51/1/0/#/52/1/{1/…-2}] export [/Table/51/1/0/#/52/1/1/#/53/1,/Table/51/1/0/#/52/1/2) I180110 18:03:00.224829 119762 ccl/storageccl/export.go:124 [n1,s1,r57/1:/Table/51/1/{0/#/54…-1}] export [/Table/51/1/0/#/54/1/2,/Table/51/1/1) I180110 18:03:00.224909 119729 ccl/storageccl/export.go:124 [n1,s1,r56/1:/Table/51/1/0/#/54/1/{1-2}] export [/Table/51/1/0/#/54/1/1,/Table/51/1/0/#/54/1/2) I180110 18:03:00.225032 119752 ccl/storageccl/export.go:124 [n1,s1,r38/1:/Table/51/1/1/#/52/1/1{-/…}] export [/Table/51/1/1/#/52/1/1,/Table/51/1/1/#/52/1/1/#/53/1) I180110 18:03:00.225148 119751 ccl/storageccl/export.go:124 [n1,s1,r24/1:/Table/51/1/1{-/#/52/…}] export [/Table/51/1/1,/Table/51/1/1/#/52/1/1) I180110 18:03:00.227625 119637 ccl/storageccl/export.go:124 [n1,s1,r48/1:/Table/51/1/1/#/52/1/{1/…-2}] export [/Table/51/1/1/#/52/1/1/#/53/1,/Table/51/1/1/#/52/1/2) I180110 18:03:00.227662 119763 ccl/storageccl/export.go:124 [n1,s1,r39/1:/Table/51/1/1/#/52/1/2{-/…}] export [/Table/51/1/1/#/52/1/2,/Table/51/1/1/#/52/1/2/#/53/1) I180110 18:03:00.227783 119695 ccl/storageccl/export.go:124 [n1,s1,r49/1:/Table/51/1/1/#/5{2/1/…-4/1/1}] export [/Table/51/1/1/#/52/1/2/#/53/1,/Table/51/1/1/#/54/1/1) I180110 18:03:00.228066 119778 ccl/storageccl/export.go:124 [n1,s1,r59/1:/Table/51/1/{1/#/54…-2}] export [/Table/51/1/1/#/54/1/2,/Table/51/1/2) I180110 18:03:00.227784 119753 ccl/storageccl/export.go:124 [n1,s1,r58/1:/Table/51/1/1/#/54/1/{1-2}] export [/Table/51/1/1/#/54/1/1,/Table/51/1/1/#/54/1/2) I180110 18:03:00.231000 119639 ccl/storageccl/export.go:124 [n1,s1,r40/1:/Table/51/1/2/#/52/1/1{-/…}] export [/Table/51/1/2/#/52/1/1,/Table/51/1/2/#/52/1/1/#/53/1) I180110 18:03:00.231028 119638 ccl/storageccl/export.go:124 [n1,s1,r25/1:/Table/51/1/2{-/#/52/…}] export [/Table/51/1/2,/Table/51/1/2/#/52/1/1) I180110 18:03:00.231226 119764 ccl/storageccl/export.go:124 [n1,s1,r41/1:/Table/51/1/2/#/52/1/2{-/…}] export [/Table/51/1/2/#/52/1/2,/Table/51/1/2/#/52/1/2/#/53/1) I180110 18:03:00.231284 119754 ccl/storageccl/export.go:124 [n1,s1,r50/1:/Table/51/1/2/#/52/1/{1/…-2}] export [/Table/51/1/2/#/52/1/1/#/53/1,/Table/51/1/2/#/52/1/2) I180110 18:03:00.231284 119765 ccl/storageccl/export.go:124 [n1,s1,r51/1:/Table/51/1/2/#/5{2/1/…-4/1/1}] export [/Table/51/1/2/#/52/1/2/#/53/1,/Table/51/1/2/#/54/1/1) I180110 18:03:00.234051 119608 ccl/storageccl/export.go:124 [n1,s1,r26/1:/Table/51/1/3{-/#/52/…}] export [/Table/51/1/3,/Table/51/1/3/#/52/1/1) I180110 18:03:00.234057 119641 ccl/storageccl/export.go:124 [n1,s1,r61/1:/Table/51/1/{2/#/54…-3}] export [/Table/51/1/2/#/54/1/2,/Table/51/1/3) I180110 18:03:00.234232 119640 ccl/storageccl/export.go:124 [n1,s1,r60/1:/Table/51/1/2/#/54/1/{1-2}] export [/Table/51/1/2/#/54/1/1,/Table/51/1/2/#/54/1/2) I180110 18:03:00.234519 119564 ccl/storageccl/export.go:124 [n1,s1,r42/1:/Table/51/1/3/#/52/1/1{-/…}] export [/Table/51/1/3/#/52/1/1,/Table/51/1/3/#/52/1/1/#/53/1) I180110 18:03:00.234570 119780 ccl/storageccl/export.go:124 [n1,s1,r52/1:/Table/51/1/3/#/52/1/{1/…-2}] export [/Table/51/1/3/#/52/1/1/#/53/1,/Table/51/1/3/#/52/1/2) I180110 18:03:00.236747 119610 ccl/storageccl/export.go:124 [n1,s1,r53/1:/Table/51/1/3/#/5{2/1/…-4/1/1}] export [/Table/51/1/3/#/52/1/2/#/53/1,/Table/51/1/3/#/54/1/1) I180110 18:03:00.236764 119609 ccl/storageccl/export.go:124 [n1,s1,r43/1:/Table/51/1/3/#/52/1/2{-/…}] export [/Table/51/1/3/#/52/1/2,/Table/51/1/3/#/52/1/2/#/53/1) I180110 18:03:00.236833 119644 ccl/storageccl/export.go:124 [n1,s1,r27/1:/Table/51/1/4{-/#/52/…}] export [/Table/51/1/4,/Table/51/1/4/#/52/1/1) I180110 18:03:00.236857 119643 ccl/storageccl/export.go:124 [n1,s1,r63/1:/Table/51/1/{3/#/54…-4}] export [/Table/51/1/3/#/54/1/2,/Table/51/1/4) I180110 18:03:00.236842 119642 ccl/storageccl/export.go:124 [n1,s1,r62/1:/Table/51/1/3/#/54/1/{1-2}] export [/Table/51/1/3/#/54/1/1,/Table/51/1/3/#/54/1/2) I180110 18:03:00.239370 119696 ccl/storageccl/export.go:124 [n1,s1,r54/1:/Table/51/1/4/#/52/1/{1/…-2}] export [/Table/51/1/4/#/52/1/1/#/53/1,/Table/51/1/4/#/52/1/2) I180110 18:03:00.239471 119782 ccl/storageccl/export.go:124 [n1,s1,r64/1:/Table/51/1/4/#/54/1/{1-2}] export [/Table/51/1/4/#/54/1/1,/Table/51/1/4/#/54/1/2) I180110 18:03:00.239371 119611 ccl/storageccl/export.go:124 [n1,s1,r44/1:/Table/51/1/4/#/52/1/1{-/…}] export [/Table/51/1/4/#/52/1/1,/Table/51/1/4/#/52/1/1/#/53/1) I180110 18:03:00.239882 119645 ccl/storageccl/export.go:124 [n1,s1,r45/1:/Table/51/1/4/#/52/1/2{-/…}] export [/Table/51/1/4/#/52/1/2,/Table/51/1/4/#/52/1/2/#/53/1) I180110 18:03:00.239915 119646 ccl/storageccl/export.go:124 [n1,s1,r55/1:/Table/51/1/4/#/5{2/1/…-4/1/1}] export [/Table/51/1/4/#/52/1/2/#/53/1,/Table/51/1/4/#/54/1/1) I180110 18:03:00.242495 119794 ccl/storageccl/export.go:124 [n1,s1,r28/1:/Table/51/1/{5-6}] export [/Table/51/1/5,/Table/51/1/6) I180110 18:03:00.242578 119567 ccl/storageccl/export.go:124 [n1,s1,r30/1:/Table/51/1/{7-8}] export [/Table/51/1/7,/Table/51/1/8) I180110 18:03:00.242631 119783 ccl/storageccl/export.go:124 [n1,s1,r31/1:/Table/51/1/{8-9}] export [/Table/51/1/8,/Table/51/1/9) I180110 18:03:00.242578 119648 ccl/storageccl/export.go:124 [n1,s1,r29/1:/Table/51/1/{6-7}] export [/Table/51/1/6,/Table/51/1/7) I180110 18:03:00.242583 119697 ccl/storageccl/export.go:124 [n1,s1,r65/1:/Table/51/1/{4/#/54…-5}] export [/Table/51/1/4/#/54/1/2,/Table/51/1/5) I180110 18:03:00.245395 119568 ccl/storageccl/export.go:124 [n1,s1,r32/1:/Table/5{1/1/9-2}] export [/Table/51/1/9,/Table/51/2) --- FAIL: test/TestBackupRestoreInterleaved/all_tables_in_interleave_hierarchy (0.000s) Test ended in panic. ------- Stdout: ------- W180110 18:03:00.275286 119831 server/status/runtime.go:109 Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180110 18:03:00.281610 119831 server/config.go:518 [n?] 1 storage engine initialized I180110 18:03:00.281656 119831 server/config.go:521 [n?] RocksDB cache size: 128 MiB I180110 18:03:00.281667 119831 server/config.go:521 [n?] store 0: in-memory, size 0 B I180110 18:03:00.285388 119831 server/node.go:361 [n?] **** cluster e5fa619e-84b4-4546-b83e-035c0afbcb10 has been created I180110 18:03:00.285454 119831 server/server.go:934 [n?] **** add additional nodes by specifying --join=127.0.0.1:35605 I180110 18:03:00.286184 119831 storage/store.go:1312 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available I180110 18:03:00.286249 119831 server/node.go:491 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=3.2 KiB), ranges=1, leases=0, writes=0.00, bytesPerReplica={p10=3322.00 p25=3322.00 p50=3322.00 p75=3322.00 p90=3322.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00} I180110 18:03:00.286281 119831 server/node.go:339 [n1] node ID 1 initialized I180110 18:03:00.286323 119831 gossip/gossip.go:332 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:35605" > attrs:<> locality:<> ServerVersion:<major_val:1 minor_val:1 patch:0 unstable:9 > I180110 18:03:00.286446 119831 storage/stores.go:331 [n1] read 0 node addresses from persistent storage I180110 18:03:00.286553 119831 server/node.go:632 [n1] connecting to gossip network to verify cluster ID... I180110 18:03:00.286585 119831 server/node.go:657 [n1] node connected via gossip and verified as part of cluster "e5fa619e-84b4-4546-b83e-035c0afbcb10" I180110 18:03:00.286610 119831 server/node.go:433 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes [] I180110 18:03:00.286667 119831 sql/distsql_physical_planner.go:122 [n1] creating DistSQLPlanner with address {tcp 127.0.0.1:35605} I180110 18:03:00.287975 119768 storage/replica_command.go:819 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2] I180110 18:03:00.290150 119831 server/server.go:1161 [n1] starting https server at 127.0.0.1:46291 I180110 18:03:00.290180 119831 server/server.go:1162 [n1] starting grpc/postgres server at 127.0.0.1:35605 I180110 18:03:00.290190 119831 server/server.go:1163 [n1] advertising CockroachDB node at 127.0.0.1:35605 I180110 18:03:00.303117 119768 storage/replica_command.go:819 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3] I180110 18:03:00.309002 119768 storage/replica_command.go:819 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4] W180110 18:03:00.310214 120037 storage/intent_resolver.go:324 [n1,s1] failed to push during intent resolution: failed to push "sql txn implicit" id=dca6c167 key=/Table/SystemConfigSpan/Start rw=true pri=0.03337446 iso=SERIALIZABLE stat=PENDING epo=0 ts=1515607380.304345637,0 orig=1515607380.304345637,0 max=1515607380.304345637,0 wto=false rop=false seq=7 I180110 18:03:00.310846 119831 sql/event_log.go:115 [n1] Event: "alter_table", target: 12, info: {TableName:eventlog Statement:ALTER TABLE system.eventlog ALTER COLUMN "uniqueID" SET DEFAULT uuid_v4() User:node MutationID:0 CascadeDroppedViews:[]} W180110 18:03:00.313544 120067 storage/intent_resolver.go:324 [n1,s1] failed to push during intent resolution: failed to push "split" id=56ea7810 key=/Local/Range/System/NodeLiveness/RangeDescriptor rw=true pri=0.00888184 iso=SERIALIZABLE stat=PENDING epo=0 ts=1515607380.309028972,0 orig=1515607380.309028972,0 max=1515607380.309028972,0 wto=false rop=false seq=3 I180110 18:03:00.314936 119768 storage/replica_command.go:819 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5] I180110 18:03:00.315957 119831 sql/lease.go:348 [n1] publish: descID=12 (eventlog) version=2 mtime=2018-01-10 18:03:00.315829678 +0000 UTC I180110 18:03:00.319063 119768 storage/replica_command.go:819 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6] I180110 18:03:00.325466 119768 storage/replica_command.go:819 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7] I180110 18:03:00.328366 119831 sql/event_log.go:115 [n1] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:node} I180110 18:03:00.329842 119768 storage/replica_command.go:819 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8] I180110 18:03:00.335208 119768 storage/replica_command.go:819 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9] I180110 18:03:00.339886 119831 sql/event_log.go:115 [n1] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:$1 User:node} I180110 18:03:00.340273 119768 storage/replica_command.go:819 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10] I180110 18:03:00.343763 119768 storage/replica_command.go:819 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11] I180110 18:03:00.345078 119831 sql/event_log.go:115 [n1] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:node} I180110 18:03:00.348228 119768 storage/replica_command.go:819 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12] I180110 18:03:00.351319 119768 storage/replica_command.go:819 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13] I180110 18:03:00.357426 119768 storage/replica_command.go:819 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14] I180110 18:03:00.359871 119831 sql/event_log.go:115 [n1] Event: "alter_table", target: 4, info: {TableName:users Statement:ALTER TABLE system.users ADD COLUMN IF NOT EXISTS "isRole" BOOL NOT NULL DEFAULT false User:node MutationID:1 CascadeDroppedViews:[]} I180110 18:03:00.361795 119768 storage/replica_command.go:819 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15] I180110 18:03:00.362485 119831 sql/lease.go:348 [n1] publish: descID=4 (users) version=2 mtime=2018-01-10 18:03:00.362378 +0000 UTC I180110 18:03:00.365402 119768 storage/replica_command.go:819 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16] I180110 18:03:00.367126 119831 sql/lease.go:348 [n1] publish: descID=4 (users) version=3 mtime=2018-01-10 18:03:00.366998973 +0000 UTC I180110 18:03:00.369005 119768 storage/replica_command.go:819 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17] I180110 18:03:00.371107 119831 sql/backfill.go:133 [n1] Running backfill for "users", v=3, m=1 I180110 18:03:00.372228 119768 storage/replica_command.go:819 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18] I180110 18:03:00.375920 119768 storage/replica_command.go:819 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19] I180110 18:03:00.380211 119768 storage/replica_command.go:819 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20] I180110 18:03:00.380849 119831 sql/lease.go:348 [n1] publish: descID=4 (users) version=4 mtime=2018-01-10 18:03:00.380596668 +0000 UTC I180110 18:03:00.385248 119831 sql/event_log.go:115 [n1] Event: "finish_schema_change", target: 4, info: {MutationID:1} I180110 18:03:00.386245 119831 sql/lease.go:274 publish (count leases): descID=4 name=users version=3 count=1 I180110 18:03:00.418753 119831 server/server.go:1232 [n1] done ensuring all necessary migrations have run I180110 18:03:00.418794 119831 server/server.go:1235 [n1] serving sql connections I180110 18:03:00.421106 120299 sql/event_log.go:115 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:35605} Attrs: Locality: ServerVersion:1.1-9} ClusterID:e5fa619e-84b4-4546-b83e-035c0afbcb10 StartedAt:1515607380286590075 LastUp:1515607380286590075} I180110 18:03:00.433218 120174 sql/event_log.go:115 [client=127.0.0.1:50924,user=root,n1] Event: "create_database", target: 50, info: {DatabaseName:ignored Statement:CREATE DATABASE ignored User:root} I180110 18:03:00.433578 119768 storage/replica_command.go:819 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21] I180110 18:03:00.435962 120174 sql/event_log.go:115 [client=127.0.0.1:50924,user=root,n1] Event: "create_database", target: 51, info: {DatabaseName:data Statement:CREATE DATABASE data User:root} I180110 18:03:00.439496 119768 storage/replica_command.go:819 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22] I180110 18:03:00.445885 120339 storage/replica_command.go:819 [n1,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23] I180110 18:03:00.453388 120339 storage/replica_command.go:819 [n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/52/1/0/#/55/1/2 [r24] I180110 18:03:00.454107 120341 storage/replica_command.go:819 [n1,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/52/1/0/#/53/1/1 [r25] I180110 18:03:00.462172 120339 storage/replica_command.go:819 [n1,s1,r24/1:/{Table/52/1/0…-Max}] initiating a split of this range at key /Table/52/1/1/#/55/1/1 [r26] I180110 18:03:00.463059 120340 storage/replica_command.go:819 [n1,s1,r24/1:/{Table/52/1/0…-Max}] initiating a split of this range at key /Table/52/1/1 [r27] W180110 18:03:00.468477 120310 storage/replica.go:2174 [n1,s1,r24/1:/Table/52/1/{0/#/55…-1/#/55…}] context canceled before command queue: 1 QueryTxn I180110 18:03:00.511020 120341 storage/replica_command.go:819 [n1,s1,r23/1:/Table/52{-/1/0/#/5…}] initiating a split of this range at key /Table/52/1/0/#/53/1/1 [r28] I180110 18:03:00.515293 120340 storage/replica_command.go:819 [n1,s1,r24/1:/Table/52/1/{0/#/55…-1/#/55…}] initiating a split of this range at key /Table/52/1/1 [r29] I180110 18:03:00.517083 120341 storage/replica_command.go:819 [n1,s1,r28/1:/Table/52/1/0/#/5{3/1/1-5/1/2}] initiating a split of this range at key /Table/52/1/0/#/53/1/1/#/54/1 [r30] I180110 18:03:00.521277 120340 storage/replica_command.go:819 [n1,s1,r29/1:/Table/52/1/1{-/#/55/…}] initiating a split of this range at key /Table/52/1/1/#/53/1/1 [r31] I180110 18:03:00.522960 120341 storage/replica_command.go:819 [n1,s1,r30/1:/Table/52/1/0/#/5{3/1/…-5/1/2}] initiating a split of this range at key /Table/52/1/0/#/53/1/2 [r32] I180110 18:03:00.526073 120341 storage/replica_command.go:819 [n1,s1,r32/1:/Table/52/1/0/#/5{3/1/2-5/1/2}] initiating a split of this range at key /Table/52/1/0/#/53/1/2/#/54/1 [r33] I180110 18:03:00.527339 120340 storage/replica_command.go:819 [n1,s1,r31/1:/Table/52/1/1/#/5{3/1/1-5/1/1}] initiating a split of this range at key /Table/52/1/1/#/53/1/1/#/54/1 [r34] I180110 18:03:00.530427 120340 storage/replica_command.go:819 [n1,s1,r34/1:/Table/52/1/1/#/5{3/1/…-5/1/1}] initiating a split of this range at key /Table/52/1/1/#/53/1/2 [r35] I180110 18:03:00.531312 120341 storage/replica_command.go:819 [n1,s1,r33/1:/Table/52/1/0/#/5{3/1/…-5/1/2}] initiating a split of this range at key /Table/52/1/0/#/55/1/1 [r36] I180110 18:03:00.534390 120340 storage/replica_command.go:819 [n1,s1,r35/1:/Table/52/1/1/#/5{3/1/2-5/1/1}] initiating a split of this range at key /Table/52/1/1/#/53/1/2/#/54/1 [r37] I180110 18:03:00.536031 120341 storage/replica_command.go:819 [n1,s1,r26/1:/{Table/52/1/1…-Max}] initiating a split of this range at key /Table/52/1/1/#/55/1/2 [r38] I180110 18:03:00.536192 120339 storage/replica_command.go:819 [n1,s1,r26/1:/{Table/52/1/1…-Max}] initiating a split of this range at key /Table/52/1/2/#/53/1/2/#/54/1 [r39] W180110 18:03:00.539966 120490 storage/replica.go:2174 [n1,s1,r26/1:/Table/52/1/1/#/55/1/{1-2}] context canceled before command queue: 1 QueryTxn I180110 18:03:00.540236 120341 storage/replica_command.go:819 [n1,s1,r38/1:/{Table/52/1/1…-Max}] initiating a split of this range at key /Table/52/1/2 [r40] I180110 18:03:00.545769 120341 storage/replica_command.go:819 [n1,s1,r40/1:/{Table/52/1/2-Max}] initiating a split of this range at key /Table/52/1/2/#/53/1/1 [r41] I180110 18:03:00.550971 120341 storage/replica_command.go:819 [n1,s1,r41/1:/{Table/52/1/2…-Max}] initiating a split of this range at key /Table/52/1/2/#/53/1/1/#/54/1 [r42] I180110 18:03:00.558917 120341 storage/replica_command.go:819 [n1,s1,r42/1:/{Table/52/1/2…-Max}] initiating a split of this range at key /Table/52/1/2/#/53/1/2 [r43] I180110 18:03:00.594970 120339 storage/replica_command.go:819 [n1,s1,r43/1:/{Table/52/1/2…-Max}] initiating a split of this range at key /Table/52/1/2/#/53/1/2/#/54/1 [r44] I180110 18:03:00.599792 120339 storage/replica_command.go:819 [n1,s1,r44/1:/{Table/52/1/2…-Max}] initiating a split of this range at key /Table/52/1/3/#/53/1/2 [r45] I180110 18:03:00.600503 120340 storage/replica_command.go:819 [n1,s1,r44/1:/{Table/52/1/2…-Max}] initiating a split of this range at key /Table/52/1/2/#/55/1/1 [r46] W180110 18:03:00.607605 120619 storage/replica.go:2174 [n1,s1,r44/1:/Table/52/1/{2/#/53…-3/#/53…}] context canceled before command queue: 1 QueryTxn I180110 18:03:00.607847 120339 storage/replica_command.go:819 [n1,s1,r45/1:/{Table/52/1/3…-Max}] initiating a split of this range at key /Table/52/1/4/#/53/1/1/#/54/1 [r47] I180110 18:03:00.608619 120341 storage/replica_command.go:819 [n1,s1,r45/1:/{Table/52/1/3…-Max}] initiating a split of this range at key /Table/52/1/3/#/53/1/2/#/54/1 [r48] W180110 18:03:00.614903 120572 storage/replica.go:2174 [n1,s1,r45/1:/Table/52/1/{3/#/53…-4/#/53…}] context canceled before command queue: 1 QueryTxn I180110 18:03:00.654583 120340 storage/replica_command.go:819 [n1,s1,r44/1:/Table/52/1/{2/#/53…-3/#/53…}] initiating a split of this range at key /Table/52/1/2/#/55/1/1 [r49] I180110 18:03:00.660522 120340 storage/replica_command.go:819 [n1,s1,r49/1:/Table/52/1/{2/#/55…-3/#/53…}] initiating a split of this range at key /Table/52/1/2/#/55/1/2 [r50] I180110 18:03:00.664488 120341 storage/replica_command.go:819 [n1,s1,r45/1:/Table/52/1/{3/#/53…-4/#/53…}] initiating a split of this range at key /Table/52/1/3/#/53/1/2/#/54/1 [r51] I180110 18:03:00.667950 120340 storage/replica_command.go:819 [n1,s1,r50/1:/Table/52/1/{2/#/55…-3/#/53…}] initiating a split of this range at key /Table/52/1/3 [r52] I180110 18:03:00.669715 120341 storage/replica_command.go:819 [n1,s1,r51/1:/Table/52/1/{3/#/53…-4/#/53…}] initiating a split of this range at key /Table/52/1/3/#/55/1/1 [r53] I180110 18:03:00.673659 120340 storage/replica_command.go:819 [n1,s1,r52/1:/Table/52/1/3{-/#/53/…}] initiating a split of this range at key /Table/52/1/3/#/53/1/1 [r54] I180110 18:03:00.673762 120341 storage/replica_command.go:819 [n1,s1,r53/1:/Table/52/1/{3/#/55…-4/#/53…}] initiating a split of this range at key /Table/52/1/3/#/55/1/2 [r55] I180110 18:03:00.677249 120340 storage/replica_command.go:819 [n1,s1,r54/1:/Table/52/1/3/#/53/1/{1-2}] initiating a split of this range at key /Table/52/1/3/#/53/1/1/#/54/1 [r56] I180110 18:03:00.677963 120341 storage/replica_command.go:819 [n1,s1,r55/1:/Table/52/1/{3/#/55…-4/#/53…}] initiating a split of this range at key /Table/52/1/4 [r57] I180110 18:03:00.682706 120339 storage/replica_command.go:819 [n1,s1,r47/1:/{Table/52/1/4…-Max}] initiating a split of this range at key /Table/52/1/6 [r58] I180110 18:03:00.683136 120341 storage/replica_command.go:819 [n1,s1,r57/1:/Table/52/1/4{-/#/53/…}] initiating a split of this range at key /Table/52/1/4/#/53/1/1 [r59] I180110 18:03:00.683623 120340 storage/replica_command.go:819 [n1,s1,r47/1:/{Table/52/1/4…-Max}] initiating a split of this range at key /Table/52/1/4/#/53/1/2 [r60] W180110 18:03:00.686715 120739 storage/replica.go:2174 [n1,s1,r47/1:/Table/52/1/{4/#/53…-6}] context canceled before command queue: 1 QueryTxn I180110 18:03:00.687267 120341 storage/replica_command.go:819 [n1,s1,r58/1:/{Table/52/1/6-Max}] initiating a split of this range at key /Table/52/1/7 [r61] I180110 18:03:00.692013 120341 storage/replica_command.go:819 [n1,s1,r61/1:/{Table/52/1/7-Max}] initiating a split of this range at key /Table/52/1/8 [r62] I180110 18:03:00.700994 120341 storage/replica_command.go:819 [n1,s1,r62/1:/{Table/52/1/8-Max}] initiating a split of this range at key /Table/52/1/9 [r63] panic: test timed out after 4m0s goroutine 120721 [running]: testing.startAlarm.func1() /usr/local/go/src/testing/testing.go:1145 +0xf9 created by time.goFunc /usr/local/go/src/time/sleep.go:170 +0x44 goroutine 1 [chan receive]: testing.(*T).Run(0xc4200c2690, 0x233f0da, 0x1c, 0x23ff290, 0x8d4f01) /usr/local/go/src/testing/testing.go:790 +0x2fc testing.runTests.func1(0xc4200c2690) /usr/local/go/src/testing/testing.go:1004 +0x64 testing.tRunner(0xc4200c2690, 0xc42067bdb8) /usr/local/go/src/testing/testing.go:746 +0xd0 testing.runTests(0xc4201ddc00, 0x3410660, 0x38, 0x38, 0x0) /usr/local/go/src/testing/testing.go:1002 +0x2d8 testing.(*M).Run(0xc420a27f18, 0xc420086500) /usr/local/go/src/testing/testing.go:921 +0x111 github.com/cockroachdb/cockroach/pkg/ccl/sqlccl.TestMain(0xc42067bf18) /go/src/github.com/cockroachdb/cockroach/pkg/ccl/sqlccl/main_test.go:31 +0xda main.main() github.com/cockroachdb/cockroach/pkg/ccl/sqlccl/_test/_testmain.go:172 +0xdb goroutine 24 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).flushDaemon(0x365efa0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1043 +0x81 created by github.com/cockroachdb/cockroach/pkg/util/log.init.0 /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:581 +0xbf goroutine 50 [syscall, 4 minutes]: os/signal.signal_recv(0x0) /usr/local/go/src/runtime/sigqueue.go:131 +0xa6 os/signal.loop() /usr/local/go/src/os/signal/signal_unix.go:22 +0x22 created by os/signal.init.0 /usr/local/go/src/os/signal/signal_unix.go:28 +0x41 goroutine 118774 [chan receive]: testing.(*T).Run(0xc4215f64b0, 0x234c57a, 0x22, 0xc4202f3080, 0x1a) /usr/local/go/src/testing/testing.go:790 +0x2fc github.com/cockroachdb/cockroach/pkg/ccl/sqlccl_test.TestBackupRestoreInterleaved(0xc4215f64b0) /go/src/github.com/cockroachdb/cockroach/pkg/ccl/sqlccl/backup_test.go:1065 +0x937 testing.tRunner(0xc4215f64b0, 0x23ff290) /usr/local/go/src/testing/testing.go:746 +0xd0 created by testing.(*T).Run /usr/local/go/src/testing/testing.go:789 +0x2de goroutine 5 [select, locked to thread]: runtime.gopark(0x2403108, 0x0, 0x2319caf, 0x6, 0x18, 0x1) /usr/local/go/src/runtime/proc.go:277 +0x12c runtime.selectgo(0xc420041f50, 0xc42048ed80) /usr/local/go/src/runtime/select.go:395 +0x1138 runtime.ensureSigM.func1() /usr/local/go/src/runtime/signal_unix.go:511 +0x220 runtime.goexit() /usr/local/go/src/runtime/asm_amd64.s:2337 +0x1 goroutine 118948 [semacquire]: sync.runtime_notifyListWait(0xc423412890, 0xc4000003ae) /usr/local/go/src/runtime/sema.go:507 +0x110 sync.(*Cond).Wait(0xc423412880) /usr/local/go/src/sync/cond.go:56 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420b5d440, 0x343ba20, 0xc423a6ecf0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:197 +0x7c github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2(0x343ba20, 0xc423a6ecf0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:166 +0x3e github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420344f10, 0xc4210b58c0, 0xc420344ee0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 120716 [semacquire]: sync.runtime_Semacquire(0xc42008f6d4) /usr/local/go/src/runtime/sema.go:56 +0x39 sync.(*WaitGroup).Wait(0xc42008f6c8) /usr/local/go/src/sync/waitgroup.go:131 +0x72 github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup.(*Group).Wait(0xc42008f6c0, 0xc4216028a0, 0x2345377) /go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup/errgroup.go:41 +0x31 github.com/cockroachdb/cockroach/pkg/ccl/storageccl.evalImport(0x343ba20, 0xc4216028a0, 0x344ec60, 0xc423526000, 0x150884e1f9f193d6, 0x0, 0x100000001, 0x1, 0x3f, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/ccl/storageccl/import.go:311 +0x2084 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).executeAdminBatch(0xc423526000, 0x343ba20, 0xc4216028a0, 0x150884e1f9f193d6, 0x0, 0x100000001, 0x1, 0x3f, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2485 +0xacd github.com/cockroachdb/cockroach/pkg/storage.(*Replica).Send(0xc423526000, 0x343ba20, 0xc4216028a0, 0x150884e1f9f193d6, 0x0, 0x100000001, 0x1, 0x3f, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:1919 +0x60f github.com/cockroachdb/cockroach/pkg/storage.(*Store).Send(0xc42354c400, 0x343ba20, 0xc421602870, 0x150884e1f9f193d6, 0x0, 0x100000001, 0x1, 0x3f, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:2688 +0x688 github.com/cockroachdb/cockroach/pkg/storage.(*Stores).Send(0xc423198bb0, 0x343ba20, 0xc421602810, 0x0, 0x0, 0x100000001, 0x1, 0x3f, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/stores.go:214 +0x214 github.com/cockroachdb/cockroach/pkg/server.(*Node).batchInternal.func1(0x343ba20, 0xc421602810, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/server/node.go:865 +0x192 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunTaskWithErr(0xc4253dbb00, 0x343ba20, 0xc421602810, 0x232938b, 0x10, 0xc422157cf8, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:245 +0xf7 github.com/cockroachdb/cockroach/pkg/server.(*Node).batchInternal(0xc424f8d200, 0x343ba20, 0xc421602810, 0xc423e212a8, 0xc421602810, 0x14ecafe, 0x343ba20) /go/src/github.com/cockroachdb/cockroach/pkg/server/node.go:852 +0x152 github.com/cockroachdb/cockroach/pkg/server.(*Node).Batch(0xc424f8d200, 0x343ba20, 0xc421602810, 0xc423e212a8, 0x1f, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/server/node.go:893 +0x99 github.com/cockroachdb/cockroach/pkg/kv.(*grpcTransport).send.func1(0xc4218db490, 0xc423e21290, 0x343ba20, 0xc421602750, 0x100000000, 0xc42256e1d8, 0x8580c8) /go/src/github.com/cockroachdb/cockroach/pkg/kv/transport.go:243 +0x6b0 github.com/cockroachdb/cockroach/pkg/kv.(*grpcTransport).send(0xc4218db490, 0x343ba20, 0xc421602750, 0xc421fd9ce0, 0xf, 0xc42275ce60, 0x0, 0x0, 0x100000001, 0x1, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/transport.go:271 +0x9a github.com/cockroachdb/cockroach/pkg/kv.(*grpcTransport).SendNext(0xc4218db490, 0x343ba20, 0xc421602750, 0xc4234e5b60) /go/src/github.com/cockroachdb/cockroach/pkg/kv/transport.go:208 +0x31b github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendToReplicas(0xc42474d1d0, 0x343ba20, 0xc421602750, 0xc42474d220, 0x3f, 0xc42356b3c0, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1246 +0x3a4 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendRPC(0xc42474d1d0, 0x343ba20, 0xc421602750, 0x3f, 0xc42356b3c0, 0x1, 0x1, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:382 +0x27b github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendSingleRange(0xc42474d1d0, 0x343ba20, 0xc421602750, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:446 +0x16e github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendPartialBatch(0xc42474d1d0, 0x343ba20, 0xc421602750, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1024 +0x347 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges(0xc42474d1d0, 0x343ba20, 0xc421602750, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:700 +0x16f8 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send(0xc42474d1d0, 0x343ba20, 0xc421602750, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:621 +0x3f5 github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send(0xc4228b22a0, 0x343ba20, 0xc421602750, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:444 +0x1ce github.com/cockroachdb/cockroach/pkg/internal/client.SendWrappedWith(0x343b960, 0xc420878840, 0x341b060, 0xc4228b22a0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/sender.go:62 +0x143 github.com/cockroachdb/cockroach/pkg/internal/client.SendWrapped(0x343b960, 0xc420878840, 0x341b060, 0xc4228b22a0, 0x34496e0, 0xc4253e6be0, 0xc422397880, 0xc424252f78, 0x16d2656) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/sender.go:79 +0xaf github.com/cockroachdb/cockroach/pkg/ccl/sqlccl.restore.func4(0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/ccl/sqlccl/restore.go:967 +0x13e github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup.(*Group).Go.func1(0xc420878880, 0xc4234e5b00) /go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup/errgroup.go:58 +0x57 created by github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup.(*Group).Go /go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup/errgroup.go:55 +0x66 goroutine 118949 [semacquire]: sync.runtime_notifyListWait(0xc423412890, 0xc4000003a9) /usr/local/go/src/runtime/sema.go:507 +0x110 sync.(*Cond).Wait(0xc423412880) /usr/local/go/src/sync/cond.go:56 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420b5d440, 0x343ba20, 0xc425475e90) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:197 +0x7c github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2(0x343ba20, 0xc425475e90) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:166 +0x3e github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420344f40, 0xc4210b58c0, 0xc420344f20) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 119790 [select]: github.com/cockroachdb/cockroach/pkg/server.(*Server).startSampleEnvironment.func1(0x343ba20, 0xc420846120) /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:1416 +0x16c github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc424060280, 0xc4253dbb00, 0xc4201dde00) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 118988 [select]: github.com/cockroachdb/cockroach/pkg/sql.(*DistSQLPlanner).initRunners.func1(0x343ba20, 0xc421227050) /go/src/github.com/cockroachdb/cockroach/pkg/sql/distsql_running.go:86 +0x167 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc4204538d0, 0xc4210b58c0, 0xc4204538c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 118965 [semacquire]: sync.runtime_Semacquire(0xc420b5d4c4) /usr/local/go/src/runtime/sema.go:56 +0x39 sync.(*WaitGroup).Wait(0xc420b5d4b8) /usr/local/go/src/sync/waitgroup.go:131 +0x72 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Wait(0xc420b5d440, 0x343ba20, 0xc425475fb0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:172 +0x31 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Wait-fm(0x343ba20, 0xc425475fb0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3834 +0x3e github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420345270, 0xc4210b58c0, 0xc420345260) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 118820 [chan receive]: github.com/cockroachdb/cockroach/pkg/gossip.(*server).start.func3(0x343ba20, 0xc42303c3f0) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/server.go:371 +0x5a github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc4203763e0, 0xc4210b58c0, 0xc422f01b00) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 118818 [chan receive]: github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux.muxListener.Accept(...) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux/cmux.go:184 github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux.(*muxListener).Accept(0xc422f016a0, 0x24024e8, 0xc4207cba40, 0x34478c0, 0xc422a1b220) <autogenerated>:1 +0x65 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).Serve(0xc4207cba40, 0x343aba0, 0xc422f016a0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:463 +0x196 github.com/cockroachdb/cockroach/pkg/server.(*Server).Start.func7(0x343ba20, 0xc42303c3c0) /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:795 +0x43 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc4203761c0, 0xc4210b58c0, 0xc422f01980) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 119938 [semacquire]: sync.runtime_notifyListWait(0xc4208d8310, 0xc400000325) /usr/local/go/src/runtime/sema.go:507 +0x110 sync.(*Cond).Wait(0xc4208d8300) /usr/local/go/src/sync/cond.go:56 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc4226ac7e0, 0x343ba20, 0xc420afda10) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:197 +0x7c github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2(0x343ba20, 0xc420afda10) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:166 +0x3e github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420452b80, 0xc4253dbb00, 0xc420452b70) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 119973 [semacquire]: sync.runtime_notifyListWait(0xc4208d8310, 0xc40000033c) /usr/local/go/src/runtime/sema.go:507 +0x110 sync.(*Cond).Wait(0xc4208d8300) /usr/local/go/src/sync/cond.go:56 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc4226ac7e0, 0x343ba20, 0xc420d947e0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:197 +0x7c github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2(0x343ba20, 0xc420d947e0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:166 +0x3e github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc4204531f0, 0xc4253dbb00, 0xc4204531e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 118906 [semacquire]: sync.runtime_notifyListWait(0xc423412890, 0xc4000003b5) /usr/local/go/src/runtime/sema.go:507 +0x110 sync.(*Cond).Wait(0xc423412880) /usr/local/go/src/sync/cond.go:56 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420b5d440, 0x343ba20, 0xc42303c870) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:197 +0x7c github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2(0x343ba20, 0xc42303c870) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:166 +0x3e github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420344680, 0xc4210b58c0, 0xc420344670) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 118911 [semacquire]: sync.runtime_notifyListWait(0xc423412890, 0xc4000003b8) /usr/local/go/src/runtime/sema.go:507 +0x110 sync.(*Cond).Wait(0xc423412880) /usr/local/go/src/sync/cond.go:56 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420b5d440, 0x343ba20, 0xc423a6eab0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:197 +0x7c github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2(0x343ba20, 0xc423a6eab0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:166 +0x3e github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420344770, 0xc4210b58c0, 0xc420344760) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:195 +0xf3 created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:188 +0xad goroutine 118928 [semacquire]: sync.runtime_notifyListWait(0xc423412890, 0xc4000003c2) /usr/local/go/src/runtime/sema.go:507 +0x110 sync.(*Cond).Wait(0xc423412880) /usr/local/go/src/sync/cond.go:56 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420b5d440, 0x343ba20, 0xc42303cb70) /go/src/github.com/co ``` Please assign, take a look and update the issue accordingly.
test
teamcity failed tests on master test testbackuprestoreinterleaved test testbackuprestoreinterleaved all tables in interleave hierarchy testrace testbackuprestoreinterleaved testrace testbackuprestoreinterleaved all tables in interleave hierarchy the following tests appear to have failed fail test testbackuprestoreinterleaved test ended in panic stdout server status runtime go could not parse build timestamp parsing time as cannot parse as server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b server node go cluster has been created server server go add additional nodes by specifying join storage store go failed initial metrics computation system config not yet available server node go initialized store disk capacity mib available mib used b logicalbytes kib ranges leases writes bytesperreplica writesperreplica server node go node id initialized gossip gossip go nodedescriptor set to node id address attrs locality serverversion storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster server node go node started with engine s and attributes sql distsql physical planner go creating distsqlplanner with address tcp storage replica command go initiating a split of this range at key system server server go starting https server at server server go starting grpc postgres server at server server go advertising cockroachdb node at storage replica command go initiating a split of this range at key system nodeliveness storage replica command go initiating a split of this range at key system nodelivenessmax storage intent resolver go failed to push during intent resolution failed to push sql txn implicit id key table systemconfigspan start rw true pri iso serializable stat pending epo ts orig max wto false rop false seq sql event log go event alter table target info tablename eventlog statement alter table system eventlog alter column uniqueid set default uuid user node mutationid cascadedroppedviews storage replica command go initiating a split of this range at key system tsd sql lease go publish descid eventlog version mtime utc storage replica command go initiating a split of this range at key system tse storage replica command go initiating a split of this range at key table systemconfigspan start storage replica command go initiating a split of this range at key table storage intent resolver go failed to push during intent resolution failed to push sql txn implicit id key table systemconfigspan start rw true pri iso serializable stat pending epo ts orig max wto false rop false seq sql event log go event set cluster setting target info settingname diagnostics reporting enabled value true user node storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname version value user node storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname trace debug enable value false user node storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event alter table target info tablename users statement alter table system users add column if not exists isrole bool not null default false user node mutationid cascadedroppedviews storage replica command go initiating a split of this range at key table sql lease go publish descid users version mtime utc storage replica command go initiating a split of this range at key table sql lease go publish descid users version mtime utc storage replica command go initiating a split of this range at key table sql backfill go running backfill for users v m storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql lease go publish descid users version mtime utc storage replica command go initiating a split of this range at key table sql event log go event finish schema change target info mutationid sql lease go publish count leases descid name users version count server server go done ensuring all necessary migrations have run server server go serving sql connections sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion clusterid startedat lastup sql event log go event create database target info databasename data statement create database data user root storage replica command go initiating a split of this range at key table sql event log go event create table target info tablename bank statement create table bank id int primary key balance int payload string family id balance payload user root storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event create table target info tablename statement create table a int b int primary key a b interleave in parent bank a user root storage replica command go initiating a split of this range at key table sql lease go publish descid version mtime utc sql lease go publish descid bank version mtime utc sql lease go publish count leases descid name bank version count sql event log go event create table target info tablename statement create table a int b int c int primary key a b c interleave in parent a b user root storage replica command go initiating a split of this range at key table sql lease go publish descid version mtime utc sql lease go publish descid version mtime utc sql event log go event create table target info tablename statement create table a int b int primary key a b interleave in parent bank a user root storage replica command go initiating a split of this range at key table sql lease go publish descid version mtime utc sql lease go publish descid bank version mtime utc sql lease go publish count leases descid name bank version count storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table ccl storageccl export go export table table fail test testbackuprestoreinterleaved all tables in interleave hierarchy test ended in panic stdout server status runtime go could not parse build timestamp parsing time as cannot parse as server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b server node go cluster has been created server server go add additional nodes by specifying join storage store go failed initial metrics computation system config not yet available server node go initialized store disk capacity mib available mib used b logicalbytes kib ranges leases writes bytesperreplica writesperreplica server node go node id initialized gossip gossip go nodedescriptor set to node id address attrs locality serverversion storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster server node go node started with engine s and attributes sql distsql physical planner go creating distsqlplanner with address tcp storage replica command go initiating a split of this range at key system server server go starting https server at server server go starting grpc postgres server at server server go advertising cockroachdb node at storage replica command go initiating a split of this range at key system nodeliveness storage replica command go initiating a split of this range at key system nodelivenessmax storage intent resolver go failed to push during intent resolution failed to push sql txn implicit id key table systemconfigspan start rw true pri iso serializable stat pending epo ts orig max wto false rop false seq sql event log go event alter table target info tablename eventlog statement alter table system eventlog alter column uniqueid set default uuid user node mutationid cascadedroppedviews storage intent resolver go failed to push during intent resolution failed to push split id key local range system nodeliveness rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key system tsd sql lease go publish descid eventlog version mtime utc storage replica command go initiating a split of this range at key system tse storage replica command go initiating a split of this range at key table systemconfigspan start sql event log go event set cluster setting target info settingname diagnostics reporting enabled value true user node storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname version value user node storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname trace debug enable value false user node storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event alter table target info tablename users statement alter table system users add column if not exists isrole bool not null default false user node mutationid cascadedroppedviews storage replica command go initiating a split of this range at key table sql lease go publish descid users version mtime utc storage replica command go initiating a split of this range at key table sql lease go publish descid users version mtime utc storage replica command go initiating a split of this range at key table sql backfill go running backfill for users v m storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql lease go publish descid users version mtime utc sql event log go event finish schema change target info mutationid sql lease go publish count leases descid name users version count server server go done ensuring all necessary migrations have run server server go serving sql connections sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion clusterid startedat lastup sql event log go event create database target info databasename ignored statement create database ignored user root storage replica command go initiating a split of this range at key table sql event log go event create database target info databasename data statement create database data user root storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica go context canceled before command queue querytxn storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica go context canceled before command queue querytxn storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica go context canceled before command queue querytxn storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica go context canceled before command queue querytxn storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica go context canceled before command queue querytxn storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table panic test timed out after goroutine testing startalarm usr local go src testing testing go created by time gofunc usr local go src time sleep go goroutine testing t run usr local go src testing testing go testing runtests usr local go src testing testing go testing trunner usr local go src testing testing go testing runtests usr local go src testing testing go testing m run usr local go src testing testing go github com cockroachdb cockroach pkg ccl sqlccl testmain go src github com cockroachdb cockroach pkg ccl sqlccl main test go main main github com cockroachdb cockroach pkg ccl sqlccl test testmain go goroutine github com cockroachdb cockroach pkg util log loggingt flushdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log init go src github com cockroachdb cockroach pkg util log clog go goroutine os signal signal recv usr local go src runtime sigqueue go os signal loop usr local go src os signal signal unix go created by os signal init usr local go src os signal signal unix go goroutine testing t run usr local go src testing testing go github com cockroachdb cockroach pkg ccl sqlccl test testbackuprestoreinterleaved go src github com cockroachdb cockroach pkg ccl sqlccl backup test go testing trunner usr local go src testing testing go created by testing t run usr local go src testing testing go goroutine runtime gopark usr local go src runtime proc go runtime selectgo usr local go src runtime select go runtime ensuresigm usr local go src runtime signal unix go runtime goexit usr local go src runtime asm s goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime semacquire usr local go src runtime sema go sync waitgroup wait usr local go src sync waitgroup go github com cockroachdb cockroach vendor golang org x sync errgroup group wait go src github com cockroachdb cockroach vendor golang org x sync errgroup errgroup go github com cockroachdb cockroach pkg ccl storageccl evalimport go src github com cockroachdb cockroach pkg ccl storageccl import go github com cockroachdb cockroach pkg storage replica executeadminbatch go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage replica send go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage store send go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg storage stores send go src github com cockroachdb cockroach pkg storage stores go github com cockroachdb cockroach pkg server node batchinternal go src github com cockroachdb cockroach pkg server node go github com cockroachdb cockroach pkg util stop stopper runtaskwitherr go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg server node batchinternal go src github com cockroachdb cockroach pkg server node go github com cockroachdb cockroach pkg server node batch go src github com cockroachdb cockroach pkg server node go github com cockroachdb cockroach pkg kv grpctransport send go src github com cockroachdb cockroach pkg kv transport go github com cockroachdb cockroach pkg kv grpctransport send go src github com cockroachdb cockroach pkg kv transport go github com cockroachdb cockroach pkg kv grpctransport sendnext go src github com cockroachdb cockroach pkg kv transport go github com cockroachdb cockroach pkg kv distsender sendtoreplicas go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendrpc go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendsinglerange go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendpartialbatch go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender divideandsendbatchtoranges go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender send go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv txncoordsender send go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg internal client sendwrappedwith go src github com cockroachdb cockroach pkg internal client sender go github com cockroachdb cockroach pkg internal client sendwrapped go src github com cockroachdb cockroach pkg internal client sender go github com cockroachdb cockroach pkg ccl sqlccl restore go src github com cockroachdb cockroach pkg ccl sqlccl restore go github com cockroachdb cockroach vendor golang org x sync errgroup group go go src github com cockroachdb cockroach vendor golang org x sync errgroup errgroup go created by github com cockroachdb cockroach vendor golang org x sync errgroup group go go src github com cockroachdb cockroach vendor golang org x sync errgroup errgroup go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg server server startsampleenvironment go src github com cockroachdb cockroach pkg server server go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg sql distsqlplanner initrunners go src github com cockroachdb cockroach pkg sql distsql running go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime semacquire usr local go src runtime sema go sync waitgroup wait usr local go src sync waitgroup go github com cockroachdb cockroach pkg storage raftscheduler wait go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler wait fm go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg gossip server start go src github com cockroachdb cockroach pkg gossip server go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach vendor github com cockroachdb cmux muxlistener accept go src github com cockroachdb cockroach vendor github com cockroachdb cmux cmux go github com cockroachdb cockroach vendor github com cockroachdb cmux muxlistener accept github com cockroachdb cockroach vendor google golang org grpc server serve go src github com cockroachdb cockroach vendor google golang org grpc server go github com cockroachdb cockroach pkg server server start go src github com cockroachdb cockroach pkg server server go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com co please assign take a look and update the issue accordingly
1
60,392
25,112,192,331
IssuesEvent
2022-11-08 21:35:07
BCDevOps/developer-experience
https://api.github.com/repos/BCDevOps/developer-experience
closed
Just-ask move image and repo for feature development
team/bcgov-platform-services app-development
**Describe the issue** Just-ask code and image are in Patrick's personal github account and dockerhub, as we are fully adapt this tool, it is safer to have this code and image under bcgov repo. **Additional context** Add any other context, attachments or screenshots **How does this benefit the users of our platform?** **Definition of done** rebuild the pipeline with new image source
1.0
Just-ask move image and repo for feature development - **Describe the issue** Just-ask code and image are in Patrick's personal github account and dockerhub, as we are fully adapt this tool, it is safer to have this code and image under bcgov repo. **Additional context** Add any other context, attachments or screenshots **How does this benefit the users of our platform?** **Definition of done** rebuild the pipeline with new image source
non_test
just ask move image and repo for feature development describe the issue just ask code and image are in patrick s personal github account and dockerhub as we are fully adapt this tool it is safer to have this code and image under bcgov repo additional context add any other context attachments or screenshots how does this benefit the users of our platform definition of done rebuild the pipeline with new image source
0
27,258
4,289,717,459
IssuesEvent
2016-07-18 05:35:07
NativeScript/nativescript-cli
https://api.github.com/repos/NativeScript/nativescript-cli
closed
`tns livesync android` not updating application consistently
2 - Ready For Test question
I have seen other closed issues reporting similar issues; however, I appear to still be having an issue with `tns livesync android`. Whenever I run this command it sometimes updates the application and sometimes doesn't. I have run `tns livesync android --emulator --watch` and had the application refresh with the newly updated code as expected; however, it will randomly stop updating. I will `Ctrl-C` then re-run `tns livesync android --emulator --watch` and it won't pick up any changes. I receive no errors, and the application refreshes - just without any changes. The only fix currently seems to be restarting my computer. I am able to get the expected functionality after this. Help is appreciated, thank you.
1.0
`tns livesync android` not updating application consistently - I have seen other closed issues reporting similar issues; however, I appear to still be having an issue with `tns livesync android`. Whenever I run this command it sometimes updates the application and sometimes doesn't. I have run `tns livesync android --emulator --watch` and had the application refresh with the newly updated code as expected; however, it will randomly stop updating. I will `Ctrl-C` then re-run `tns livesync android --emulator --watch` and it won't pick up any changes. I receive no errors, and the application refreshes - just without any changes. The only fix currently seems to be restarting my computer. I am able to get the expected functionality after this. Help is appreciated, thank you.
test
tns livesync android not updating application consistently i have seen other closed issues reporting similar issues however i appear to still be having an issue with tns livesync android whenever i run this command it sometimes updates the application and sometimes doesn t i have run tns livesync android emulator watch and had the application refresh with the newly updated code as expected however it will randomly stop updating i will ctrl c then re run tns livesync android emulator watch and it won t pick up any changes i receive no errors and the application refreshes just without any changes the only fix currently seems to be restarting my computer i am able to get the expected functionality after this help is appreciated thank you
1
492,262
14,199,320,613
IssuesEvent
2020-11-16 02:00:10
nhn/tui.grid
https://api.github.com/repos/nhn/tui.grid
closed
Memory leak or endless cycle Tree UI
Bug Priority: High
**Describe the bug** If call `grid.resetOriginData();` and `grid.restore()` browser freezes. **To Reproduce** Steps to reproduce the behavior: 1. Go to https://cdpn.io/b1tc0re/debug/WZzdWr/XBAGRqLDpnRr 2. Click on button resetOriginData 3. Click on button restore 4. See error **Screenshots** ![2](https://user-images.githubusercontent.com/8367655/93948974-758f0000-fd48-11ea-9f63-a37ef50561c6.gif) **Desktop (please complete the following information):** - OS: Windows 10 - Browser chrome , safari - Version Chrome 85.0.4183.102
1.0
Memory leak or endless cycle Tree UI - **Describe the bug** If call `grid.resetOriginData();` and `grid.restore()` browser freezes. **To Reproduce** Steps to reproduce the behavior: 1. Go to https://cdpn.io/b1tc0re/debug/WZzdWr/XBAGRqLDpnRr 2. Click on button resetOriginData 3. Click on button restore 4. See error **Screenshots** ![2](https://user-images.githubusercontent.com/8367655/93948974-758f0000-fd48-11ea-9f63-a37ef50561c6.gif) **Desktop (please complete the following information):** - OS: Windows 10 - Browser chrome , safari - Version Chrome 85.0.4183.102
non_test
memory leak or endless cycle tree ui describe the bug if call grid resetorigindata and grid restore browser freezes to reproduce steps to reproduce the behavior go to click on button resetorigindata click on button restore see error screenshots desktop please complete the following information os windows browser chrome safari version chrome
0
120,435
17,644,195,664
IssuesEvent
2021-08-20 01:55:39
logbie/HyperGAN
https://api.github.com/repos/logbie/HyperGAN
opened
CVE-2021-29540 (High) detected in tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl
security vulnerability
## CVE-2021-29540 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/0a/93/c7bca39b23aae45cd2e85ad3871c81eccc63b9c5276e926511e2e5b0879d/tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/0a/93/c7bca39b23aae45cd2e85ad3871c81eccc63b9c5276e926511e2e5b0879d/tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: HyperGAN/requirements.txt</p> <p>Path to vulnerable library: HyperGAN/requirements.txt</p> <p> Dependency Hierarchy: - :x: **tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an end-to-end open source platform for machine learning. An attacker can cause a heap buffer overflow to occur in `Conv2DBackpropFilter`. This is because the implementation(https://github.com/tensorflow/tensorflow/blob/1b0296c3b8dd9bd948f924aa8cd62f87dbb7c3da/tensorflow/core/kernels/conv_grad_filter_ops.cc#L495-L497) computes the size of the filter tensor but does not validate that it matches the number of elements in `filter_sizes`. Later, when reading/writing to this buffer, code uses the value computed here, instead of the number of elements in the tensor. The fix will be included in TensorFlow 2.5.0. We will also cherrypick this commit on TensorFlow 2.4.2, TensorFlow 2.3.3, TensorFlow 2.2.3 and TensorFlow 2.1.4, as these are also affected and still in supported range. <p>Publish Date: 2021-05-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29540>CVE-2021-29540</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-xgc3-m89p-vr3x">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-xgc3-m89p-vr3x</a></p> <p>Release Date: 2021-05-14</p> <p>Fix Resolution: tensorflow - 2.5.0, tensorflow-cpu - 2.5.0, tensorflow-gpu - 2.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-29540 (High) detected in tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl - ## CVE-2021-29540 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/0a/93/c7bca39b23aae45cd2e85ad3871c81eccc63b9c5276e926511e2e5b0879d/tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/0a/93/c7bca39b23aae45cd2e85ad3871c81eccc63b9c5276e926511e2e5b0879d/tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: HyperGAN/requirements.txt</p> <p>Path to vulnerable library: HyperGAN/requirements.txt</p> <p> Dependency Hierarchy: - :x: **tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an end-to-end open source platform for machine learning. An attacker can cause a heap buffer overflow to occur in `Conv2DBackpropFilter`. This is because the implementation(https://github.com/tensorflow/tensorflow/blob/1b0296c3b8dd9bd948f924aa8cd62f87dbb7c3da/tensorflow/core/kernels/conv_grad_filter_ops.cc#L495-L497) computes the size of the filter tensor but does not validate that it matches the number of elements in `filter_sizes`. Later, when reading/writing to this buffer, code uses the value computed here, instead of the number of elements in the tensor. The fix will be included in TensorFlow 2.5.0. We will also cherrypick this commit on TensorFlow 2.4.2, TensorFlow 2.3.3, TensorFlow 2.2.3 and TensorFlow 2.1.4, as these are also affected and still in supported range. <p>Publish Date: 2021-05-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29540>CVE-2021-29540</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-xgc3-m89p-vr3x">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-xgc3-m89p-vr3x</a></p> <p>Release Date: 2021-05-14</p> <p>Fix Resolution: tensorflow - 2.5.0, tensorflow-cpu - 2.5.0, tensorflow-gpu - 2.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in tensorflow gpu whl cve high severity vulnerability vulnerable library tensorflow gpu whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file hypergan requirements txt path to vulnerable library hypergan requirements txt dependency hierarchy x tensorflow gpu whl vulnerable library vulnerability details tensorflow is an end to end open source platform for machine learning an attacker can cause a heap buffer overflow to occur in this is because the implementation computes the size of the filter tensor but does not validate that it matches the number of elements in filter sizes later when reading writing to this buffer code uses the value computed here instead of the number of elements in the tensor the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with whitesource
0
127,061
10,450,565,011
IssuesEvent
2019-09-19 10:50:43
govim/govim
https://api.github.com/repos/govim/govim
opened
testing: move away from :call feedkeys to :exe "normal ..."
enhancement testing
Per Bram's advice in https://groups.google.com/d/msg/vim_dev/op_PKiE9iog/V7H3U2ZfAQAJ For example the following `feedkeys` call: ``` call feedkeys("A\<C-x>\<C-o>\<C-n>", "xt") ``` is replaced by the equivalent: ``` exe "normal A\<C-x>\<C-o>\<C-n>" ``` FYI @leitzler
1.0
testing: move away from :call feedkeys to :exe "normal ..." - Per Bram's advice in https://groups.google.com/d/msg/vim_dev/op_PKiE9iog/V7H3U2ZfAQAJ For example the following `feedkeys` call: ``` call feedkeys("A\<C-x>\<C-o>\<C-n>", "xt") ``` is replaced by the equivalent: ``` exe "normal A\<C-x>\<C-o>\<C-n>" ``` FYI @leitzler
test
testing move away from call feedkeys to exe normal per bram s advice in for example the following feedkeys call call feedkeys a xt is replaced by the equivalent exe normal a fyi leitzler
1
11,034
7,035,217,768
IssuesEvent
2017-12-27 21:59:35
christhekeele/mnemonix
https://api.github.com/repos/christhekeele/mnemonix
closed
Re-enable the Expiry feature
Kind: Chore Status: Planned Type: Usability
This had to be globally commented out during a refactor, but needs to come back online before the next release. It's the prime opportunity to implement some of the other planned expiry enhancements as well.
True
Re-enable the Expiry feature - This had to be globally commented out during a refactor, but needs to come back online before the next release. It's the prime opportunity to implement some of the other planned expiry enhancements as well.
non_test
re enable the expiry feature this had to be globally commented out during a refactor but needs to come back online before the next release it s the prime opportunity to implement some of the other planned expiry enhancements as well
0
718,879
24,734,854,028
IssuesEvent
2022-10-20 20:53:19
yugabyte/yugabyte-db
https://api.github.com/repos/yugabyte/yugabyte-db
closed
[DocDB] The statistic threads_running_thread_pool for YCQL shows incorrect value
kind/bug area/docdb priority/medium
Jira Link: [DB-3154](https://yugabyte.atlassian.net/browse/DB-3154) ### Description When querying the metrics for the YCQL (12000) endpoint, I get the following values: - metrics endpoint: 18446744073709551609 - prometheus-metrics endpoint: -7 These do not seem like realistic numbers to me, and therefore give the impression this statistic is broken. I now find this value with version 2.15.2.0-b83, but witnessed this value for quite some time.
1.0
[DocDB] The statistic threads_running_thread_pool for YCQL shows incorrect value - Jira Link: [DB-3154](https://yugabyte.atlassian.net/browse/DB-3154) ### Description When querying the metrics for the YCQL (12000) endpoint, I get the following values: - metrics endpoint: 18446744073709551609 - prometheus-metrics endpoint: -7 These do not seem like realistic numbers to me, and therefore give the impression this statistic is broken. I now find this value with version 2.15.2.0-b83, but witnessed this value for quite some time.
non_test
the statistic threads running thread pool for ycql shows incorrect value jira link description when querying the metrics for the ycql endpoint i get the following values metrics endpoint prometheus metrics endpoint these do not seem like realistic numbers to me and therefore give the impression this statistic is broken i now find this value with version but witnessed this value for quite some time
0
79,732
23,031,200,204
IssuesEvent
2022-07-22 14:04:54
NixOS/nixpkgs
https://api.github.com/repos/NixOS/nixpkgs
closed
netcoredbg fails to build
0.kind: build failure
### Steps To Reproduce Steps to reproduce the behavior: 1. build *netcoredbg* ### Build log https://gist.github.com/mdarocha/2625e779c17587bfe6010f8bfa7d410d ### Notify maintainers <!-- Please @ people who are in the `meta.maintainers` list of the offending package or module. If in doubt, check `git blame` for whoever last touched something. --> @leo60228 ### Metadata Please run `nix-shell -p nix-info --run "nix-info -m"` and paste the result. ```console [user@system:~]$ nix-shell -p nix-info --run "nix-info -m" - system: `"x86_64-linux"` - host os: `Linux 5.15.53, NixOS, 22.11 (Raccoon), 22.11.20220716.8f48571` - multi-user?: `yes` - sandbox: `yes` - version: `nix-env (Nix) 2.10.3` - nixpkgs: `/nix/store/a885zpv9ys2p2x7qnzqvxlsy321mclip-source` ```
1.0
netcoredbg fails to build - ### Steps To Reproduce Steps to reproduce the behavior: 1. build *netcoredbg* ### Build log https://gist.github.com/mdarocha/2625e779c17587bfe6010f8bfa7d410d ### Notify maintainers <!-- Please @ people who are in the `meta.maintainers` list of the offending package or module. If in doubt, check `git blame` for whoever last touched something. --> @leo60228 ### Metadata Please run `nix-shell -p nix-info --run "nix-info -m"` and paste the result. ```console [user@system:~]$ nix-shell -p nix-info --run "nix-info -m" - system: `"x86_64-linux"` - host os: `Linux 5.15.53, NixOS, 22.11 (Raccoon), 22.11.20220716.8f48571` - multi-user?: `yes` - sandbox: `yes` - version: `nix-env (Nix) 2.10.3` - nixpkgs: `/nix/store/a885zpv9ys2p2x7qnzqvxlsy321mclip-source` ```
non_test
netcoredbg fails to build steps to reproduce steps to reproduce the behavior build netcoredbg build log notify maintainers please people who are in the meta maintainers list of the offending package or module if in doubt check git blame for whoever last touched something metadata please run nix shell p nix info run nix info m and paste the result console nix shell p nix info run nix info m system linux host os linux nixos raccoon multi user yes sandbox yes version nix env nix nixpkgs nix store source
0
370,307
25,898,658,425
IssuesEvent
2022-12-15 02:10:26
apache/daffodil-vscode
https://api.github.com/repos/apache/daffodil-vscode
closed
Change extension name to "Apache Daffodil™ Extension for Visual Studio Code"
documentation
According to this guide (https://code.visualstudio.com/brand), "Apache Daffodil™ Extension for Visual Studio Code" would be the proper way to name the extension rather than "Apache Daffodil VS Code Extension" used currently.
1.0
Change extension name to "Apache Daffodil™ Extension for Visual Studio Code" - According to this guide (https://code.visualstudio.com/brand), "Apache Daffodil™ Extension for Visual Studio Code" would be the proper way to name the extension rather than "Apache Daffodil VS Code Extension" used currently.
non_test
change extension name to apache daffodil™ extension for visual studio code according to this guide apache daffodil™ extension for visual studio code would be the proper way to name the extension rather than apache daffodil vs code extension used currently
0
26,306
5,243,497,419
IssuesEvent
2017-01-31 20:52:51
palantir/blueprint
https://api.github.com/repos/palantir/blueprint
closed
Feature Request: Add Button "type" prop
Domain: documentation Type: question
It'd be quite nice to have a `type` `prop` on the `Button` component. Currently, it renders as `<button type="button" .../>`. In a `form`, if there is a `<button type="submit"`, then the `onSubmit` event is fired if return key is pressed from an input field. This is a handy default behaviour which is worth keeping IMO. I can make a PR adding a `type` prop to `Button` component and `ButtonType` (?) enum to implement this behaviour. Thanks!
1.0
Feature Request: Add Button "type" prop - It'd be quite nice to have a `type` `prop` on the `Button` component. Currently, it renders as `<button type="button" .../>`. In a `form`, if there is a `<button type="submit"`, then the `onSubmit` event is fired if return key is pressed from an input field. This is a handy default behaviour which is worth keeping IMO. I can make a PR adding a `type` prop to `Button` component and `ButtonType` (?) enum to implement this behaviour. Thanks!
non_test
feature request add button type prop it d be quite nice to have a type prop on the button component currently it renders as in a form if there is a button type submit then the onsubmit event is fired if return key is pressed from an input field this is a handy default behaviour which is worth keeping imo i can make a pr adding a type prop to button component and buttontype enum to implement this behaviour thanks
0
1,788
3,371,286,808
IssuesEvent
2015-11-23 18:28:52
NeuronRobotics/nrjavaserial
https://api.github.com/repos/NeuronRobotics/nrjavaserial
opened
GEt Gradle to build with Java 1.5 compatibility
infrastructure
#21 asked us to compile for Java 1.5, and we've maintained that between then and now (I think). However, Gradle doesn't run on Java 1.5 – it has a minimum of 1.6. Before we build a new JAR for the 3.12.0 release, we need to figure out the configuration required to get Gradle to target 1.5 from a 1.6 compiler, or to get Gradle to invoke a 1.5 `javac` when running under 1.6.
1.0
GEt Gradle to build with Java 1.5 compatibility - #21 asked us to compile for Java 1.5, and we've maintained that between then and now (I think). However, Gradle doesn't run on Java 1.5 – it has a minimum of 1.6. Before we build a new JAR for the 3.12.0 release, we need to figure out the configuration required to get Gradle to target 1.5 from a 1.6 compiler, or to get Gradle to invoke a 1.5 `javac` when running under 1.6.
non_test
get gradle to build with java compatibility asked us to compile for java and we ve maintained that between then and now i think however gradle doesn t run on java – it has a minimum of before we build a new jar for the release we need to figure out the configuration required to get gradle to target from a compiler or to get gradle to invoke a javac when running under
0
282,476
21,315,491,895
IssuesEvent
2022-04-16 07:39:31
vishandi/pe
https://api.github.com/repos/vishandi/pe
opened
Unclear References to Woofie
type.DocumentationBug severity.Low
![image.png](https://raw.githubusercontent.com/vishandi/pe/master/files/a3040fe6-8fe9-4894-b75f-b9393f499a01.png) <br> (Notice the example, who is Woofie?) ![image.png](https://raw.githubusercontent.com/vishandi/pe/master/files/43ec8c59-a442-4fa3-9f5e-9a54be325817.png) <br> (Same thing here) <!--session: 1650093091321-c72dc0c1-7596-48c2-93ca-dcd0db112d5f--> <!--Version: Web v3.4.2-->
1.0
Unclear References to Woofie - ![image.png](https://raw.githubusercontent.com/vishandi/pe/master/files/a3040fe6-8fe9-4894-b75f-b9393f499a01.png) <br> (Notice the example, who is Woofie?) ![image.png](https://raw.githubusercontent.com/vishandi/pe/master/files/43ec8c59-a442-4fa3-9f5e-9a54be325817.png) <br> (Same thing here) <!--session: 1650093091321-c72dc0c1-7596-48c2-93ca-dcd0db112d5f--> <!--Version: Web v3.4.2-->
non_test
unclear references to woofie notice the example who is woofie same thing here
0
4,889
4,703,276,487
IssuesEvent
2016-10-13 07:18:54
mapbox/mapbox-gl-js
https://api.github.com/repos/mapbox/mapbox-gl-js
closed
Ensure removing releases tile resources
performance
If you use `Map#removeSource`, the source is removed but `unloadTile` isn't called for any currently loaded tiles, so their WebGL resources may continue to consume memory for an indefinite amount of time, including until the WebGL context is destroyed.
True
Ensure removing releases tile resources - If you use `Map#removeSource`, the source is removed but `unloadTile` isn't called for any currently loaded tiles, so their WebGL resources may continue to consume memory for an indefinite amount of time, including until the WebGL context is destroyed.
non_test
ensure removing releases tile resources if you use map removesource the source is removed but unloadtile isn t called for any currently loaded tiles so their webgl resources may continue to consume memory for an indefinite amount of time including until the webgl context is destroyed
0
100,504
30,721,721,981
IssuesEvent
2023-07-27 16:23:58
dotnet/dnceng
https://api.github.com/repos/dotnet/dnceng
closed
Build failed: dotnet-arcade-validation-official/main #20230711.1
Build Failed
Build [#20230711.1](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=2219115) partiallySucceeded ## :warning: : internal / dotnet-arcade-validation-official partiallySucceeded ### Summary **Finished** - Tue, 11 Jul 2023 22:30:07 GMT **Duration** - 98 minutes **Requested for** - DotNet Bot **Reason** - batchedCI ### Details #### Build - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/111) - Component Governance detected 3 security related alerts at or above '"High"' severity. Microsoft’s Open Source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components. Vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency. - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/112) - Component Governance detected 3 security related alerts at or above '"High"' severity. Microsoft’s Open Source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components. Vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency. - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/20) - Component Governance detected 2 security related alerts at or above '"High"' severity. Microsoft’s Open Source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components. Vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency. - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/21) - Component Governance detected 2 security related alerts at or above '"High"' severity. Microsoft’s Open Source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components. Vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency. - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/38) - Component Governance detected 2 security related alerts at or above '"High"' severity. Microsoft’s Open Source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components. Vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency. - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/40) - Component Governance detected 2 security related alerts at or above '"High"' severity. Microsoft’s Open Source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components. Vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency. - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/108) - Component Governance detected 2 security related alerts at or above '"High"' severity. Microsoft’s Open Source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components. Vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency. - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/102) - Component Governance detected 3 security related alerts at or above '"High"' severity. Microsoft’s Open Source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components. Vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency. - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/122) - Error: The signing api didn't complete in time (331 seconds), failed to sign SBOM. - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/181) - Component Governance detected 3 security related alerts at or above '"High"' severity. Microsoft’s Open Source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components. Vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency. #### Create BAR ID Tag - :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/324) - Artifact Create_BAR_ID_Tag_Create_BAR_ID_Tag_SBOM already exists for build 2219115. ### Changes - [dd2ef2a0](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/dd2ef2a05b13a984a3c7b750244a24d0b367899b) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20230711.1 (#3872)
1.0
Build failed: dotnet-arcade-validation-official/main #20230711.1 - Build [#20230711.1](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=2219115) partiallySucceeded ## :warning: : internal / dotnet-arcade-validation-official partiallySucceeded ### Summary **Finished** - Tue, 11 Jul 2023 22:30:07 GMT **Duration** - 98 minutes **Requested for** - DotNet Bot **Reason** - batchedCI ### Details #### Build - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/111) - Component Governance detected 3 security related alerts at or above '"High"' severity. Microsoft’s Open Source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components. Vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency. - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/112) - Component Governance detected 3 security related alerts at or above '"High"' severity. Microsoft’s Open Source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components. Vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency. - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/20) - Component Governance detected 2 security related alerts at or above '"High"' severity. Microsoft’s Open Source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components. Vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency. - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/21) - Component Governance detected 2 security related alerts at or above '"High"' severity. Microsoft’s Open Source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components. Vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency. - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/38) - Component Governance detected 2 security related alerts at or above '"High"' severity. Microsoft’s Open Source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components. Vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency. - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/40) - Component Governance detected 2 security related alerts at or above '"High"' severity. Microsoft’s Open Source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components. Vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency. - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/108) - Component Governance detected 2 security related alerts at or above '"High"' severity. Microsoft’s Open Source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components. Vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency. - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/102) - Component Governance detected 3 security related alerts at or above '"High"' severity. Microsoft’s Open Source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components. Vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency. - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/122) - Error: The signing api didn't complete in time (331 seconds), failed to sign SBOM. - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/181) - Component Governance detected 3 security related alerts at or above '"High"' severity. Microsoft’s Open Source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components. Vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency. #### Create BAR ID Tag - :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2219115/logs/324) - Artifact Create_BAR_ID_Tag_Create_BAR_ID_Tag_SBOM already exists for build 2219115. ### Changes - [dd2ef2a0](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/dd2ef2a05b13a984a3c7b750244a24d0b367899b) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20230711.1 (#3872)
non_test
build failed dotnet arcade validation official main build partiallysucceeded warning internal dotnet arcade validation official partiallysucceeded summary finished tue jul gmt duration minutes requested for dotnet bot reason batchedci details build warning component governance detected security related alerts at or above high severity microsoft’s open source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency warning component governance detected security related alerts at or above high severity microsoft’s open source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency warning component governance detected security related alerts at or above high severity microsoft’s open source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency warning component governance detected security related alerts at or above high severity microsoft’s open source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency warning component governance detected security related alerts at or above high severity microsoft’s open source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency warning component governance detected security related alerts at or above high severity microsoft’s open source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency warning component governance detected security related alerts at or above high severity microsoft’s open source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency warning component governance detected security related alerts at or above high severity microsoft’s open source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency warning error the signing api didn t complete in time seconds failed to sign sbom warning component governance detected security related alerts at or above high severity microsoft’s open source policy requires that all high and critical security vulnerabilities found by this task be addressed by upgrading vulnerable components vulnerabilities in indirect dependencies should be addressed by upgrading the root dependency create bar id tag x artifact create bar id tag create bar id tag sbom already exists for build changes dotnet maestro update dependencies from build
0
344,717
30,755,160,678
IssuesEvent
2023-07-29 01:20:02
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Intermittent failure in CriticalFinalizer test
disabled-test area-VM-coreclr
Affected tests: - `baseservices/finalization/CriticalFinalizer/CriticalFinalizer.sh` - `baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd` Runfo failures in [last 30 days - .sh](https://runfo.azurewebsites.net/search/tests/?q=started%3A%7E30+definition%3Aruntime+name%3A%22baseservices%2Ffinalization%2Fcriticalfinalizer%2Fcriticalfinalizer.sh%22) and [last 30 days - .cmd](https://runfo.azurewebsites.net/search/tests/?q=started%3A%7E30+definition%3Aruntime+name%3A%22baseservices%5C%5Cfinalization%5C%5Ccriticalfinalizer%5C%5Ccriticalfinalizer.cmd%22): - 9/23 - 188 failures - First failure in Rolling run https://runfo.azurewebsites.net/view/build/?number=24556 on 9/20 ... more than 50x failures per day CoreCLR: https://helixre107v0xdeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-pull-75573-merge-b0042e942c0145f38e/baseservices.finalization/1/console.e135272f.log?helixlogtype=result ``` BEGIN EXECUTION "C:\h\w\A68508F4\p\corerun.exe" -p "System.Reflection.Metadata.MetadataUpdater.IsSupported=false" CriticalFinalizer.dll Finalized 0 Normal and 0 Critical objects. The first Critical object was finalized after 0 Normal objects. Expected: 100 Actual: 101 END EXECUTION - FAILED ``` Mono: https://helixre107v0xdeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-pull-74586-merge-662a1339f83d41b786/baseservices.finalization/1/console.0af36b40.log?helixlogtype=result ``` BEGIN EXECUTION /tmp/helix/working/A1850911/p/corerun -p System.Reflection.Metadata.MetadataUpdater.IsSupported=false CriticalFinalizer.dll '' Finalized 100 Normal and 99 Critical objects. The first Critical object was finalized after 100 Normal objects. Expected: 100 Actual: 101 END EXECUTION - FAILED ``` ``` { "ErrorMessage":"CriticalFinalizer" } ``` <!--Known issue error report start --> ### Report |Build|Definition|Step Name|Console log| |---|---|---|---| |[28655](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28655)|dotnet/runtime|LLVM AOT compile CoreCLR tests|[Log](https://dev.azure.com/dnceng-public/cbb18261-c48f-4abb-8651-8cdcb5474649/_apis/build/builds/28655/logs/1777) |Build|Definition|Test| |---|---|---| |[29097](https://dev.azure.com/dnceng-public/public/_build/results?buildId=29097)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=29097&view=ms.vss-test-web.build-test-results-tab&runId=586054&resultId=103265)| |[28963](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28963)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28963&view=ms.vss-test-web.build-test-results-tab&runId=585362&resultId=100006)| |[28964](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28964)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28964&view=ms.vss-test-web.build-test-results-tab&runId=585370&resultId=110809)| |[29051](https://dev.azure.com/dnceng-public/public/_build/results?buildId=29051)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=29051&view=ms.vss-test-web.build-test-results-tab&runId=585598&resultId=103004)| |[29016](https://dev.azure.com/dnceng-public/public/_build/results?buildId=29016)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=29016&view=ms.vss-test-web.build-test-results-tab&runId=585520&resultId=103041)| |[29012](https://dev.azure.com/dnceng-public/public/_build/results?buildId=29012)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=29012&view=ms.vss-test-web.build-test-results-tab&runId=585474&resultId=103041)| |[29010](https://dev.azure.com/dnceng-public/public/_build/results?buildId=29010)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=29010&view=ms.vss-test-web.build-test-results-tab&runId=585434&resultId=100069)| |[28977](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28977)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28977&view=ms.vss-test-web.build-test-results-tab&runId=585086&resultId=103262)| |[28942](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28942)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28942&view=ms.vss-test-web.build-test-results-tab&runId=584954&resultId=100005)| |[28861](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28861)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28861&view=ms.vss-test-web.build-test-results-tab&runId=584842&resultId=100142)| |[28621](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28621)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28621&view=ms.vss-test-web.build-test-results-tab&runId=583814&resultId=110809)| |[28899](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28899)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28899&view=ms.vss-test-web.build-test-results-tab&runId=583614&resultId=101098)| |[28892](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28892)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28892&view=ms.vss-test-web.build-test-results-tab&runId=583586&resultId=101314)| |[28655](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28655)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28655&view=ms.vss-test-web.build-test-results-tab&runId=582966&resultId=100089)| |[28841](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28841)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28841&view=ms.vss-test-web.build-test-results-tab&runId=582902&resultId=100188)| |[28819](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28819)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28819&view=ms.vss-test-web.build-test-results-tab&runId=582510&resultId=103262)| |[28749](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28749)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28749&view=ms.vss-test-web.build-test-results-tab&runId=581790&resultId=110005)| |[28766](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28766)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28766&view=ms.vss-test-web.build-test-results-tab&runId=581716&resultId=100122)| |[28692](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28692)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28692&view=ms.vss-test-web.build-test-results-tab&runId=581136&resultId=100204)| |[28682](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28682)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28682&view=ms.vss-test-web.build-test-results-tab&runId=581120&resultId=100224)| |[28649](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28649)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28649&view=ms.vss-test-web.build-test-results-tab&runId=578836&resultId=103262)| |[28606](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28606)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28606&view=ms.vss-test-web.build-test-results-tab&runId=577392&resultId=110809)| |[28615](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28615)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28615&view=ms.vss-test-web.build-test-results-tab&runId=577274&resultId=100989)| |[28597](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28597)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28597&view=ms.vss-test-web.build-test-results-tab&runId=576808&resultId=100304)| |[28592](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28592)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28592&view=ms.vss-test-web.build-test-results-tab&runId=576724&resultId=100269)| |[28591](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28591)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28591&view=ms.vss-test-web.build-test-results-tab&runId=576700&resultId=100192)| |[27258](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27258)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27258&view=ms.vss-test-web.build-test-results-tab&runId=575262&resultId=100071)| |[28513](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28513)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28513&view=ms.vss-test-web.build-test-results-tab&runId=574886&resultId=110809)| |[28494](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28494)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28494&view=ms.vss-test-web.build-test-results-tab&runId=574254&resultId=103262)| |[28476](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28476)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28476&view=ms.vss-test-web.build-test-results-tab&runId=573856&resultId=103262)| |[28414](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28414)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28414&view=ms.vss-test-web.build-test-results-tab&runId=573554&resultId=110809)| |[27149](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27149)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27149&view=ms.vss-test-web.build-test-results-tab&runId=572978&resultId=103260)| |[28383](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28383)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28383&view=ms.vss-test-web.build-test-results-tab&runId=572118&resultId=100254)| |[28313](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28313)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28313&view=ms.vss-test-web.build-test-results-tab&runId=571294&resultId=103262)| |[28293](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28293)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28293&view=ms.vss-test-web.build-test-results-tab&runId=570840&resultId=103262)| |[28278](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28278)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28278&view=ms.vss-test-web.build-test-results-tab&runId=571006&resultId=101258)| |[28251](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28251)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28251&view=ms.vss-test-web.build-test-results-tab&runId=570948&resultId=101314)| |[28306](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28306)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28306&view=ms.vss-test-web.build-test-results-tab&runId=570932&resultId=101059)| |[28149](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28149)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28149&view=ms.vss-test-web.build-test-results-tab&runId=568764&resultId=100166)| |[27779](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27779)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27779&view=ms.vss-test-web.build-test-results-tab&runId=567762&resultId=103261)| |[27987](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27987)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27987&view=ms.vss-test-web.build-test-results-tab&runId=565762&resultId=102210)| |[27102](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27102)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27102&view=ms.vss-test-web.build-test-results-tab&runId=566954&resultId=100022)| |[28056](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28056)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28056&view=ms.vss-test-web.build-test-results-tab&runId=566628&resultId=100144)| |[28047](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28047)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28047&view=ms.vss-test-web.build-test-results-tab&runId=565850&resultId=103003)| |[28023](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28023)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28023&view=ms.vss-test-web.build-test-results-tab&runId=566096&resultId=101314)| |[27880](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27880)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27880&view=ms.vss-test-web.build-test-results-tab&runId=565226&resultId=100076)| |[27960](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27960)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27960&view=ms.vss-test-web.build-test-results-tab&runId=564516&resultId=103261)| |[27940](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27940)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27940&view=ms.vss-test-web.build-test-results-tab&runId=564364&resultId=100391)| |[27866](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27866)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27866&view=ms.vss-test-web.build-test-results-tab&runId=562906&resultId=100005)| |[26219](https://dev.azure.com/dnceng-public/public/_build/results?buildId=26219)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=26219&view=ms.vss-test-web.build-test-results-tab&runId=562654&resultId=100085)| |[27820](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27820)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27820&view=ms.vss-test-web.build-test-results-tab&runId=561308&resultId=103261)| |[27804](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27804)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27804&view=ms.vss-test-web.build-test-results-tab&runId=561162&resultId=103261)| |[27855](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27855)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27855&view=ms.vss-test-web.build-test-results-tab&runId=562074&resultId=101059)| |[27831](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27831)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27831&view=ms.vss-test-web.build-test-results-tab&runId=561664&resultId=100259)| |[27789](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27789)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27789&view=ms.vss-test-web.build-test-results-tab&runId=561164&resultId=100804)| |[27742](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27742)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27742&view=ms.vss-test-web.build-test-results-tab&runId=560558&resultId=103260)| |[27754](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27754)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27754&view=ms.vss-test-web.build-test-results-tab&runId=560328&resultId=103002)| |[27703](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27703)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27703&view=ms.vss-test-web.build-test-results-tab&runId=559784&resultId=100032)| |[26661](https://dev.azure.com/dnceng-public/public/_build/results?buildId=26661)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=26661&view=ms.vss-test-web.build-test-results-tab&runId=558390&resultId=103259)| |[27657](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27657)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27657&view=ms.vss-test-web.build-test-results-tab&runId=558654&resultId=100278)| |[27643](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27643)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27643&view=ms.vss-test-web.build-test-results-tab&runId=558066&resultId=100130)| |[27588](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27588)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27588&view=ms.vss-test-web.build-test-results-tab&runId=557792&resultId=100158)| |[27630](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27630)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27630&view=ms.vss-test-web.build-test-results-tab&runId=557690&resultId=103002)| |[27557](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27557)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27557&view=ms.vss-test-web.build-test-results-tab&runId=556856&resultId=100002)| #### Summary |24-Hour Hit Count|7-Day Hit Count|1-Month Count| |---|---|---| |33|65|65| <!--Known issue error report end -->
1.0
Intermittent failure in CriticalFinalizer test - Affected tests: - `baseservices/finalization/CriticalFinalizer/CriticalFinalizer.sh` - `baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd` Runfo failures in [last 30 days - .sh](https://runfo.azurewebsites.net/search/tests/?q=started%3A%7E30+definition%3Aruntime+name%3A%22baseservices%2Ffinalization%2Fcriticalfinalizer%2Fcriticalfinalizer.sh%22) and [last 30 days - .cmd](https://runfo.azurewebsites.net/search/tests/?q=started%3A%7E30+definition%3Aruntime+name%3A%22baseservices%5C%5Cfinalization%5C%5Ccriticalfinalizer%5C%5Ccriticalfinalizer.cmd%22): - 9/23 - 188 failures - First failure in Rolling run https://runfo.azurewebsites.net/view/build/?number=24556 on 9/20 ... more than 50x failures per day CoreCLR: https://helixre107v0xdeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-pull-75573-merge-b0042e942c0145f38e/baseservices.finalization/1/console.e135272f.log?helixlogtype=result ``` BEGIN EXECUTION "C:\h\w\A68508F4\p\corerun.exe" -p "System.Reflection.Metadata.MetadataUpdater.IsSupported=false" CriticalFinalizer.dll Finalized 0 Normal and 0 Critical objects. The first Critical object was finalized after 0 Normal objects. Expected: 100 Actual: 101 END EXECUTION - FAILED ``` Mono: https://helixre107v0xdeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-pull-74586-merge-662a1339f83d41b786/baseservices.finalization/1/console.0af36b40.log?helixlogtype=result ``` BEGIN EXECUTION /tmp/helix/working/A1850911/p/corerun -p System.Reflection.Metadata.MetadataUpdater.IsSupported=false CriticalFinalizer.dll '' Finalized 100 Normal and 99 Critical objects. The first Critical object was finalized after 100 Normal objects. Expected: 100 Actual: 101 END EXECUTION - FAILED ``` ``` { "ErrorMessage":"CriticalFinalizer" } ``` <!--Known issue error report start --> ### Report |Build|Definition|Step Name|Console log| |---|---|---|---| |[28655](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28655)|dotnet/runtime|LLVM AOT compile CoreCLR tests|[Log](https://dev.azure.com/dnceng-public/cbb18261-c48f-4abb-8651-8cdcb5474649/_apis/build/builds/28655/logs/1777) |Build|Definition|Test| |---|---|---| |[29097](https://dev.azure.com/dnceng-public/public/_build/results?buildId=29097)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=29097&view=ms.vss-test-web.build-test-results-tab&runId=586054&resultId=103265)| |[28963](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28963)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28963&view=ms.vss-test-web.build-test-results-tab&runId=585362&resultId=100006)| |[28964](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28964)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28964&view=ms.vss-test-web.build-test-results-tab&runId=585370&resultId=110809)| |[29051](https://dev.azure.com/dnceng-public/public/_build/results?buildId=29051)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=29051&view=ms.vss-test-web.build-test-results-tab&runId=585598&resultId=103004)| |[29016](https://dev.azure.com/dnceng-public/public/_build/results?buildId=29016)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=29016&view=ms.vss-test-web.build-test-results-tab&runId=585520&resultId=103041)| |[29012](https://dev.azure.com/dnceng-public/public/_build/results?buildId=29012)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=29012&view=ms.vss-test-web.build-test-results-tab&runId=585474&resultId=103041)| |[29010](https://dev.azure.com/dnceng-public/public/_build/results?buildId=29010)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=29010&view=ms.vss-test-web.build-test-results-tab&runId=585434&resultId=100069)| |[28977](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28977)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28977&view=ms.vss-test-web.build-test-results-tab&runId=585086&resultId=103262)| |[28942](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28942)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28942&view=ms.vss-test-web.build-test-results-tab&runId=584954&resultId=100005)| |[28861](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28861)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28861&view=ms.vss-test-web.build-test-results-tab&runId=584842&resultId=100142)| |[28621](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28621)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28621&view=ms.vss-test-web.build-test-results-tab&runId=583814&resultId=110809)| |[28899](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28899)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28899&view=ms.vss-test-web.build-test-results-tab&runId=583614&resultId=101098)| |[28892](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28892)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28892&view=ms.vss-test-web.build-test-results-tab&runId=583586&resultId=101314)| |[28655](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28655)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28655&view=ms.vss-test-web.build-test-results-tab&runId=582966&resultId=100089)| |[28841](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28841)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28841&view=ms.vss-test-web.build-test-results-tab&runId=582902&resultId=100188)| |[28819](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28819)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28819&view=ms.vss-test-web.build-test-results-tab&runId=582510&resultId=103262)| |[28749](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28749)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28749&view=ms.vss-test-web.build-test-results-tab&runId=581790&resultId=110005)| |[28766](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28766)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28766&view=ms.vss-test-web.build-test-results-tab&runId=581716&resultId=100122)| |[28692](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28692)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28692&view=ms.vss-test-web.build-test-results-tab&runId=581136&resultId=100204)| |[28682](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28682)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28682&view=ms.vss-test-web.build-test-results-tab&runId=581120&resultId=100224)| |[28649](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28649)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28649&view=ms.vss-test-web.build-test-results-tab&runId=578836&resultId=103262)| |[28606](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28606)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28606&view=ms.vss-test-web.build-test-results-tab&runId=577392&resultId=110809)| |[28615](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28615)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28615&view=ms.vss-test-web.build-test-results-tab&runId=577274&resultId=100989)| |[28597](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28597)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28597&view=ms.vss-test-web.build-test-results-tab&runId=576808&resultId=100304)| |[28592](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28592)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28592&view=ms.vss-test-web.build-test-results-tab&runId=576724&resultId=100269)| |[28591](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28591)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28591&view=ms.vss-test-web.build-test-results-tab&runId=576700&resultId=100192)| |[27258](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27258)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27258&view=ms.vss-test-web.build-test-results-tab&runId=575262&resultId=100071)| |[28513](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28513)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28513&view=ms.vss-test-web.build-test-results-tab&runId=574886&resultId=110809)| |[28494](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28494)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28494&view=ms.vss-test-web.build-test-results-tab&runId=574254&resultId=103262)| |[28476](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28476)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28476&view=ms.vss-test-web.build-test-results-tab&runId=573856&resultId=103262)| |[28414](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28414)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28414&view=ms.vss-test-web.build-test-results-tab&runId=573554&resultId=110809)| |[27149](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27149)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27149&view=ms.vss-test-web.build-test-results-tab&runId=572978&resultId=103260)| |[28383](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28383)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28383&view=ms.vss-test-web.build-test-results-tab&runId=572118&resultId=100254)| |[28313](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28313)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28313&view=ms.vss-test-web.build-test-results-tab&runId=571294&resultId=103262)| |[28293](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28293)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28293&view=ms.vss-test-web.build-test-results-tab&runId=570840&resultId=103262)| |[28278](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28278)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28278&view=ms.vss-test-web.build-test-results-tab&runId=571006&resultId=101258)| |[28251](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28251)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28251&view=ms.vss-test-web.build-test-results-tab&runId=570948&resultId=101314)| |[28306](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28306)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28306&view=ms.vss-test-web.build-test-results-tab&runId=570932&resultId=101059)| |[28149](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28149)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28149&view=ms.vss-test-web.build-test-results-tab&runId=568764&resultId=100166)| |[27779](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27779)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27779&view=ms.vss-test-web.build-test-results-tab&runId=567762&resultId=103261)| |[27987](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27987)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27987&view=ms.vss-test-web.build-test-results-tab&runId=565762&resultId=102210)| |[27102](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27102)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27102&view=ms.vss-test-web.build-test-results-tab&runId=566954&resultId=100022)| |[28056](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28056)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28056&view=ms.vss-test-web.build-test-results-tab&runId=566628&resultId=100144)| |[28047](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28047)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28047&view=ms.vss-test-web.build-test-results-tab&runId=565850&resultId=103003)| |[28023](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28023)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=28023&view=ms.vss-test-web.build-test-results-tab&runId=566096&resultId=101314)| |[27880](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27880)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27880&view=ms.vss-test-web.build-test-results-tab&runId=565226&resultId=100076)| |[27960](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27960)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27960&view=ms.vss-test-web.build-test-results-tab&runId=564516&resultId=103261)| |[27940](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27940)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27940&view=ms.vss-test-web.build-test-results-tab&runId=564364&resultId=100391)| |[27866](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27866)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27866&view=ms.vss-test-web.build-test-results-tab&runId=562906&resultId=100005)| |[26219](https://dev.azure.com/dnceng-public/public/_build/results?buildId=26219)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=26219&view=ms.vss-test-web.build-test-results-tab&runId=562654&resultId=100085)| |[27820](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27820)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27820&view=ms.vss-test-web.build-test-results-tab&runId=561308&resultId=103261)| |[27804](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27804)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27804&view=ms.vss-test-web.build-test-results-tab&runId=561162&resultId=103261)| |[27855](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27855)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27855&view=ms.vss-test-web.build-test-results-tab&runId=562074&resultId=101059)| |[27831](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27831)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27831&view=ms.vss-test-web.build-test-results-tab&runId=561664&resultId=100259)| |[27789](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27789)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27789&view=ms.vss-test-web.build-test-results-tab&runId=561164&resultId=100804)| |[27742](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27742)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27742&view=ms.vss-test-web.build-test-results-tab&runId=560558&resultId=103260)| |[27754](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27754)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27754&view=ms.vss-test-web.build-test-results-tab&runId=560328&resultId=103002)| |[27703](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27703)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27703&view=ms.vss-test-web.build-test-results-tab&runId=559784&resultId=100032)| |[26661](https://dev.azure.com/dnceng-public/public/_build/results?buildId=26661)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=26661&view=ms.vss-test-web.build-test-results-tab&runId=558390&resultId=103259)| |[27657](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27657)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27657&view=ms.vss-test-web.build-test-results-tab&runId=558654&resultId=100278)| |[27643](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27643)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27643&view=ms.vss-test-web.build-test-results-tab&runId=558066&resultId=100130)| |[27588](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27588)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27588&view=ms.vss-test-web.build-test-results-tab&runId=557792&resultId=100158)| |[27630](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27630)|dotnet/runtime|[baseservices.finalization.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27630&view=ms.vss-test-web.build-test-results-tab&runId=557690&resultId=103002)| |[27557](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27557)|dotnet/runtime|[baseservices\\finalization\\CriticalFinalizer\\CriticalFinalizer.cmd](https://dev.azure.com/dnceng-public/public/_build/results?buildId=27557&view=ms.vss-test-web.build-test-results-tab&runId=556856&resultId=100002)| #### Summary |24-Hour Hit Count|7-Day Hit Count|1-Month Count| |---|---|---| |33|65|65| <!--Known issue error report end -->
test
intermittent failure in criticalfinalizer test affected tests baseservices finalization criticalfinalizer criticalfinalizer sh baseservices finalization criticalfinalizer criticalfinalizer cmd runfo failures in and failures first failure in rolling run on more than failures per day coreclr begin execution c h w p corerun exe p system reflection metadata metadataupdater issupported false criticalfinalizer dll finalized normal and critical objects the first critical object was finalized after normal objects expected actual end execution failed mono begin execution tmp helix working p corerun p system reflection metadata metadataupdater issupported false criticalfinalizer dll finalized normal and critical objects the first critical object was finalized after normal objects expected actual end execution failed errormessage criticalfinalizer report build definition step name console log aot compile coreclr tests build definition test summary hour hit count day hit count month count
1
160,960
12,527,837,439
IssuesEvent
2020-06-04 08:35:57
MPDL/KEEPER
https://api.github.com/repos/MPDL/KEEPER
closed
E-mail about Bloxberg transaction not available
test
Testserver: qa keeper Browser: firefox quantum 67.0.4 (64-Bit) Version: Seafile 6.3.12 Actions: select file -> file menue bar -> "certify data via Bloxberg" -> Observation: -> info "Transacrion succeded" is displayed, but no e-mail sended Update: notification just taked longer as expected, so just e-mail missing Expected Result: user gets informed and e-mail
1.0
E-mail about Bloxberg transaction not available - Testserver: qa keeper Browser: firefox quantum 67.0.4 (64-Bit) Version: Seafile 6.3.12 Actions: select file -> file menue bar -> "certify data via Bloxberg" -> Observation: -> info "Transacrion succeded" is displayed, but no e-mail sended Update: notification just taked longer as expected, so just e-mail missing Expected Result: user gets informed and e-mail
test
e mail about bloxberg transaction not available testserver qa keeper browser firefox quantum bit version seafile actions select file file menue bar certify data via bloxberg observation info transacrion succeded is displayed but no e mail sended update notification just taked longer as expected so just e mail missing expected result user gets informed and e mail
1
439,110
12,677,554,981
IssuesEvent
2020-06-19 08:00:13
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.chip.de - design is broken
browser-firefox-mobile engine-gecko priority-important
<!-- @browser: Firefox Mobile 79.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Mobile; rv:79.0) Gecko/79.0 Firefox/79.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://www.chip.de/news/Vier-Labor-Updates-veroeffentlicht-Diese-Fritzboxen-bekommen-den-neuen-Patch_165971368.html **Browser / Version**: Firefox Mobile 79.0 **Operating System**: Android 8.0.0 **Tested Another Browser**: Yes Chrome **Problem type**: Design is broken **Description**: Items not fully visible **Steps to Reproduce**: Die Webseite läd nur im einfachen Designund nicht im "normalen" Design. <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200617093637</li><li>channel: nightly</li><li>hasTouchScreen: true</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.chip.de - design is broken - <!-- @browser: Firefox Mobile 79.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Mobile; rv:79.0) Gecko/79.0 Firefox/79.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://www.chip.de/news/Vier-Labor-Updates-veroeffentlicht-Diese-Fritzboxen-bekommen-den-neuen-Patch_165971368.html **Browser / Version**: Firefox Mobile 79.0 **Operating System**: Android 8.0.0 **Tested Another Browser**: Yes Chrome **Problem type**: Design is broken **Description**: Items not fully visible **Steps to Reproduce**: Die Webseite läd nur im einfachen Designund nicht im "normalen" Design. <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200617093637</li><li>channel: nightly</li><li>hasTouchScreen: true</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
design is broken url browser version firefox mobile operating system android tested another browser yes chrome problem type design is broken description items not fully visible steps to reproduce die webseite läd nur im einfachen designund nicht im normalen design browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true from with ❤️
0
349,895
31,839,561,080
IssuesEvent
2023-09-14 15:25:51
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Manual test run on Android Tab for 1.58.x - Release
tests QA/Yes release-notes/exclude OS/Android ARM
## Installer - [x] Check that installer is close to the size of the last release - [x] Check the Brave version in About and make sure it is EXACTLY as expected ## Startup - [x] Verify that Brave is only contacting `*.brave.com` endpoints on first launch using either `Charles Proxy`, `Fiddler` or `Wireshark` (or a similar application) - [x] Verify that opening a NTP doesn't trigger any outbound connections related to widgets without user interaction ## Visual look - [x] Make sure thereafter every merge - [x] No Chrome/Chromium words appear on `brave://version` - [x] No Chrome/Chromium words appear on normal or private tabs - [x] No Chrome/Chromium words appear in site settings for `Location` / `Camera` / `Microphone` / `Augmented Reality` - [x] No Chrome/Chromium icons are shown in normal or private tabs ## Custom tabs - [x] Make sure Brave handles links from Gmail, Slack - [x] Ensure custom tabs work even with sync enabled/disabled ## Tab Groups - [x] Ensure tab-group is enabled by default - [x] Verify toggling tab-group setting triggers a relaunch request - [x] Verify tab-group is not reverted back to default setting after browser restart ## Developer Tools - [x] Verify you can inspect sub-links via dev tools ## Clear Data - [x] Verify Clear Data on exit works as intended ## Settings and Bottom bar - [x] Verify changing default settings are retained and don't cause the browser to crash - [x] Verify bottom bar buttons (Home/Bookmark/Search/Tabs) work as expected - [x] Verify items in hamburger menu/share menu to ensure nothing unexpected has been added ## IPFS, ENS & Unstoppable Domain - [x] Verify settings for IPFS, ENS & Unstoppable Domain show up under Brave Shields & Privacy settings ## Downloads - [x] Verify downloading a file works and that all actions on the download item work. - [x] Verify that PDF is downloaded over HTTPS at `https://basicattentiontoken.org/BasicAttentionTokenWhitePaper-4.pdf` - [x] Verify that PDF is downloaded over HTTP at `http://www.pdf995.com/samples/pdf.pdf` ## Bravery settings - [x] Check that HTTPS Everywhere works by loading `http://https-everywhere.badssl.com/` - [x] Turning HTTPS Everywhere off and shields off both disable the redirect to `https://https-everywhere.badssl.com/` - [x] Check that toggling to blocking and allow ads works as expected - [x] Verify that clicking through a cert error in `https://badssl.com/` works - [x] Visit `https://twitter.com/` and then turn on script blocking, nothing should load. Allow it from the script blocking UI in the URL bar and it should work. - [x] Verify that default Bravery settings take effect on pages with no site settings - [x] Verify that 3rd party storage results are blank at `https://jsfiddle.net/7ke9r14a/7/` when 3rd party cookies are blocked - [x] Choose a DNS provider from the list in Settings | Privacy | Use Secure DNS, load `https://browserleaks.com/dns`, and verify your ISP's DNS resolvers aren't detected and shown; only your chosen DoH provider should appear. - [x] Run all six test configurations on `https://dev-pages.brave.software/storage/ephemeral-storage.html` and confirm results are as expected per each test listed. ### Fingerprint Tests - [x] Visit `https://browserleaks.com/webrtc`, ensure 2 blocked items are listed in shields - [x] Test that `https://diafygi.github.io/webrtc-ips/` doesn't leak IP address for each option under `Settings -> Privacy and Security -> WebRTC IP handling policy` ## Content Tests - [x] Go to `https://brianbondy.com/` and click on the twitter icon on the top right. Verify that context menus work in the new twitter tab - [x] Go to `https://feedly.com` and make sure that the password can be saved. Make sure the saved password is auto-populated when you visit the site again - [x] Verify that `https://mixed-script.badssl.com/` shows up as grey not red (no mixed content scripts are run) ## Brave Rewards - [x] Verify you are able to create a new Rewards profile and are in the unverified state by default - [x] Verify when you visit a creator in this state the panel shows a prompt to connect a custodian and no BAT information (earnings, balance, etc) is displayed - [x] Verify when you visit brave://rewards in this state there are no Auto Contribution, Tipping, or summary panels on this page - [x] Verify you can toggle ads off/on from this page and the panel responds accordingly - [x] Verify you can toggle rewards off/on from the panel and the brave://rewards page responds accordingly - [x] Verify you are able to connect a custodian - [x] Verify Rewards balance shows correct BAT and USD value on brave://rewards and panel after you connect - [x] Verify when you click on the BR panel while on a site, the panel displays site-specific information (site favicon, domain, attention %) - [x] Verify BR panel and tip banner shows message about an unverified creator - [x] Verify you are able to tip a creator who has the same custodian as you - [x] Verify you are able to perform an auto contribution - [x] Verify if you disable auto-contribute you are still able to tip creators - [x] Verify if auto-contribute is disabled AC does not occur - [x] Verify that there is no "Disconnect" option once connected to a custodian - [x] Verify that you are able to reset rewards ## Brave Ads - [x] Verify ads is auto-enabled when rewards is enabled for the supported region - [x] Verify ads are only shown when the app is being used - [x] Verify ad notification are shown based on ads per hour setting - [x] Verify ad notifications stack up in notification tray - [x] Verify swipe left/right dismisses the ad notification when shown and is not stored in the notification tray - [x] Verify clicking on an ad notification shows the landing page - [x] Verify `view`,`clicked` and `landed` and `dismiss` states are logged based on the action ## Sync - [x] Verify you are able to join sync chain by scanning the QR code - [x] Verify you are able to join sync chain using code words - [x] Verify you are able to create a sync chain on the device and add other devices to the chain via QR code/Code words - [x] Verify once sync chain is created, `Categories` option is shown in devices list - [x] Verify only `Bookmarks` is enabled in `Categories` by default - [x] Verify enabling `Sync everything` enables all other switches - [x] Verify existing bookmarks before joining sync chain also gets sync'd to all devices on the sync chain - [x] Verify `Autofill` data is sync'd to all devices when enabled - [x] Verify `History` is sync'd to all devices - [x] Verify `Open tab` form other devices shows up under history below the device name - [x] Verify `Password` is sync'd to all devices - [x] Verify `Settings` sync's site settings on all devices - [x] Verify sync works on an upgrade profile and new bookmarks added post-upgrade sync's across devices on the chain - [x] Verify adding a bookmark on custom tab gets sync'd across all devices in the chain - [x] Verify you are able to create a standalone sync chain with one device - [x] Verify `Remove this device` shows a confirmation alert before removing the sync chain on the device ## Top sites view - [x] Long-press on top sites to get to deletion mode, and delete a top site (note this will stop that site from showing up again on top sites, so you may not want to do this a site you want to keep there) ## Session storage - [x] Verify that tabs restore when closed, including active tab ## Upgrade from previous version Examples of pre-requisites before upgrading from previous version to the build being tested: * visit several websites so `Top Tiles` under New Tab Page get updated/populated * add several websites to the Android home screen via `Add to Home screen` (from the hamburger/setting menu) * add several bookmarks include folders * change/update several settings (Example: default search engine, adding a custom home page, changing site settings, changing several privacy settings etc..) * create/enable a Brave Wallet (or restore if individual who's going through the upgrade has one available) * enable Brave News and add/remove several sources * change the default shield settings on varios websites (Example: enable script blocking/change FP/Ad blocking to strict, disable shields) * Login into several websites and leave them opened * Login into several websites and close the tabs so they're not opened when upgrading * Login into a website and leave it as the active tab while upgrading * enable/create a sync chain with several devices **`Upgrade Cases`** - [x] Ensure that `brave://version` displays both the correct Brave version and expected Chromium version - [x] Ensure that bookmarks and folders from the previous release have been retained/can be loaded without issues - [x] Ensure that previously opened tabs are retained and can be lazy loaded with out issues - [x] Ensure that websites that have been logged into and were opened while upgrading are still logged in (ensure cookies are not cleared/lost) - [x] Ensure that websites that have been logged into but weren't open when upgrading are still logged in when loading in new tab (ensure cookies are not cleared/lost) - [x] Ensure that the correct website loads as the active tab (as per the pre-requisites, this tab should be logged into a website) - [x] Ensure that the shields settings from the previous versions are retained on each website - [x] Ensure that saved passwords are being displayed under `Settings` -> `Passwords` and can be autofilled without issues - [x] Ensure that the sync chain is preserved and syncing still works across the devices that have been added - [x] Ensure that sharing a links/tabs using `Send to your device` works as expected - [x] Ensure that Brave News is still enabled and the sources changes from the previous release are retained - [x] Ensure that the various setting changes via `Settings` are retained (Example: enable script blocking/change FP/Ad blocking to strict, disable shields) - [x] Ensure that IPFS, ENS and Unstoppable Domain settings are retained from the previous version - [x] Ensure that `History` is retained from the previous version - [x] Ensure that the website shortcuts added via `Add to Home screen` are still visible on the Android home screen and tapping on icons loads the appropriate webpage - [x] Ensure that the `Tab Group` setting doesn't reset/change after upgrading - [x] Ensure that `Clear data on exit` is retained from the previous version - [x] Rewards - [x] BAT balance is retained - [x] Auto-contribute list is retained - [x] Both Tips and Monthly Contributions are retained - [x] Summary panel transactions list is retained - [x] Changes to rewards settings are retained - [x] Ads - [x] Both `Estimated pending rewards` & `Ad notifications received this month` are retained - [x] Changes to ads settings are retained - [x] Ensure that ads are not being enabled when upgrading to a new version if they were disabled - [x] Ensure that ads are not disabled when upgrading to a new version if they were enabled
1.0
Manual test run on Android Tab for 1.58.x - Release - ## Installer - [x] Check that installer is close to the size of the last release - [x] Check the Brave version in About and make sure it is EXACTLY as expected ## Startup - [x] Verify that Brave is only contacting `*.brave.com` endpoints on first launch using either `Charles Proxy`, `Fiddler` or `Wireshark` (or a similar application) - [x] Verify that opening a NTP doesn't trigger any outbound connections related to widgets without user interaction ## Visual look - [x] Make sure thereafter every merge - [x] No Chrome/Chromium words appear on `brave://version` - [x] No Chrome/Chromium words appear on normal or private tabs - [x] No Chrome/Chromium words appear in site settings for `Location` / `Camera` / `Microphone` / `Augmented Reality` - [x] No Chrome/Chromium icons are shown in normal or private tabs ## Custom tabs - [x] Make sure Brave handles links from Gmail, Slack - [x] Ensure custom tabs work even with sync enabled/disabled ## Tab Groups - [x] Ensure tab-group is enabled by default - [x] Verify toggling tab-group setting triggers a relaunch request - [x] Verify tab-group is not reverted back to default setting after browser restart ## Developer Tools - [x] Verify you can inspect sub-links via dev tools ## Clear Data - [x] Verify Clear Data on exit works as intended ## Settings and Bottom bar - [x] Verify changing default settings are retained and don't cause the browser to crash - [x] Verify bottom bar buttons (Home/Bookmark/Search/Tabs) work as expected - [x] Verify items in hamburger menu/share menu to ensure nothing unexpected has been added ## IPFS, ENS & Unstoppable Domain - [x] Verify settings for IPFS, ENS & Unstoppable Domain show up under Brave Shields & Privacy settings ## Downloads - [x] Verify downloading a file works and that all actions on the download item work. - [x] Verify that PDF is downloaded over HTTPS at `https://basicattentiontoken.org/BasicAttentionTokenWhitePaper-4.pdf` - [x] Verify that PDF is downloaded over HTTP at `http://www.pdf995.com/samples/pdf.pdf` ## Bravery settings - [x] Check that HTTPS Everywhere works by loading `http://https-everywhere.badssl.com/` - [x] Turning HTTPS Everywhere off and shields off both disable the redirect to `https://https-everywhere.badssl.com/` - [x] Check that toggling to blocking and allow ads works as expected - [x] Verify that clicking through a cert error in `https://badssl.com/` works - [x] Visit `https://twitter.com/` and then turn on script blocking, nothing should load. Allow it from the script blocking UI in the URL bar and it should work. - [x] Verify that default Bravery settings take effect on pages with no site settings - [x] Verify that 3rd party storage results are blank at `https://jsfiddle.net/7ke9r14a/7/` when 3rd party cookies are blocked - [x] Choose a DNS provider from the list in Settings | Privacy | Use Secure DNS, load `https://browserleaks.com/dns`, and verify your ISP's DNS resolvers aren't detected and shown; only your chosen DoH provider should appear. - [x] Run all six test configurations on `https://dev-pages.brave.software/storage/ephemeral-storage.html` and confirm results are as expected per each test listed. ### Fingerprint Tests - [x] Visit `https://browserleaks.com/webrtc`, ensure 2 blocked items are listed in shields - [x] Test that `https://diafygi.github.io/webrtc-ips/` doesn't leak IP address for each option under `Settings -> Privacy and Security -> WebRTC IP handling policy` ## Content Tests - [x] Go to `https://brianbondy.com/` and click on the twitter icon on the top right. Verify that context menus work in the new twitter tab - [x] Go to `https://feedly.com` and make sure that the password can be saved. Make sure the saved password is auto-populated when you visit the site again - [x] Verify that `https://mixed-script.badssl.com/` shows up as grey not red (no mixed content scripts are run) ## Brave Rewards - [x] Verify you are able to create a new Rewards profile and are in the unverified state by default - [x] Verify when you visit a creator in this state the panel shows a prompt to connect a custodian and no BAT information (earnings, balance, etc) is displayed - [x] Verify when you visit brave://rewards in this state there are no Auto Contribution, Tipping, or summary panels on this page - [x] Verify you can toggle ads off/on from this page and the panel responds accordingly - [x] Verify you can toggle rewards off/on from the panel and the brave://rewards page responds accordingly - [x] Verify you are able to connect a custodian - [x] Verify Rewards balance shows correct BAT and USD value on brave://rewards and panel after you connect - [x] Verify when you click on the BR panel while on a site, the panel displays site-specific information (site favicon, domain, attention %) - [x] Verify BR panel and tip banner shows message about an unverified creator - [x] Verify you are able to tip a creator who has the same custodian as you - [x] Verify you are able to perform an auto contribution - [x] Verify if you disable auto-contribute you are still able to tip creators - [x] Verify if auto-contribute is disabled AC does not occur - [x] Verify that there is no "Disconnect" option once connected to a custodian - [x] Verify that you are able to reset rewards ## Brave Ads - [x] Verify ads is auto-enabled when rewards is enabled for the supported region - [x] Verify ads are only shown when the app is being used - [x] Verify ad notification are shown based on ads per hour setting - [x] Verify ad notifications stack up in notification tray - [x] Verify swipe left/right dismisses the ad notification when shown and is not stored in the notification tray - [x] Verify clicking on an ad notification shows the landing page - [x] Verify `view`,`clicked` and `landed` and `dismiss` states are logged based on the action ## Sync - [x] Verify you are able to join sync chain by scanning the QR code - [x] Verify you are able to join sync chain using code words - [x] Verify you are able to create a sync chain on the device and add other devices to the chain via QR code/Code words - [x] Verify once sync chain is created, `Categories` option is shown in devices list - [x] Verify only `Bookmarks` is enabled in `Categories` by default - [x] Verify enabling `Sync everything` enables all other switches - [x] Verify existing bookmarks before joining sync chain also gets sync'd to all devices on the sync chain - [x] Verify `Autofill` data is sync'd to all devices when enabled - [x] Verify `History` is sync'd to all devices - [x] Verify `Open tab` form other devices shows up under history below the device name - [x] Verify `Password` is sync'd to all devices - [x] Verify `Settings` sync's site settings on all devices - [x] Verify sync works on an upgrade profile and new bookmarks added post-upgrade sync's across devices on the chain - [x] Verify adding a bookmark on custom tab gets sync'd across all devices in the chain - [x] Verify you are able to create a standalone sync chain with one device - [x] Verify `Remove this device` shows a confirmation alert before removing the sync chain on the device ## Top sites view - [x] Long-press on top sites to get to deletion mode, and delete a top site (note this will stop that site from showing up again on top sites, so you may not want to do this a site you want to keep there) ## Session storage - [x] Verify that tabs restore when closed, including active tab ## Upgrade from previous version Examples of pre-requisites before upgrading from previous version to the build being tested: * visit several websites so `Top Tiles` under New Tab Page get updated/populated * add several websites to the Android home screen via `Add to Home screen` (from the hamburger/setting menu) * add several bookmarks include folders * change/update several settings (Example: default search engine, adding a custom home page, changing site settings, changing several privacy settings etc..) * create/enable a Brave Wallet (or restore if individual who's going through the upgrade has one available) * enable Brave News and add/remove several sources * change the default shield settings on varios websites (Example: enable script blocking/change FP/Ad blocking to strict, disable shields) * Login into several websites and leave them opened * Login into several websites and close the tabs so they're not opened when upgrading * Login into a website and leave it as the active tab while upgrading * enable/create a sync chain with several devices **`Upgrade Cases`** - [x] Ensure that `brave://version` displays both the correct Brave version and expected Chromium version - [x] Ensure that bookmarks and folders from the previous release have been retained/can be loaded without issues - [x] Ensure that previously opened tabs are retained and can be lazy loaded with out issues - [x] Ensure that websites that have been logged into and were opened while upgrading are still logged in (ensure cookies are not cleared/lost) - [x] Ensure that websites that have been logged into but weren't open when upgrading are still logged in when loading in new tab (ensure cookies are not cleared/lost) - [x] Ensure that the correct website loads as the active tab (as per the pre-requisites, this tab should be logged into a website) - [x] Ensure that the shields settings from the previous versions are retained on each website - [x] Ensure that saved passwords are being displayed under `Settings` -> `Passwords` and can be autofilled without issues - [x] Ensure that the sync chain is preserved and syncing still works across the devices that have been added - [x] Ensure that sharing a links/tabs using `Send to your device` works as expected - [x] Ensure that Brave News is still enabled and the sources changes from the previous release are retained - [x] Ensure that the various setting changes via `Settings` are retained (Example: enable script blocking/change FP/Ad blocking to strict, disable shields) - [x] Ensure that IPFS, ENS and Unstoppable Domain settings are retained from the previous version - [x] Ensure that `History` is retained from the previous version - [x] Ensure that the website shortcuts added via `Add to Home screen` are still visible on the Android home screen and tapping on icons loads the appropriate webpage - [x] Ensure that the `Tab Group` setting doesn't reset/change after upgrading - [x] Ensure that `Clear data on exit` is retained from the previous version - [x] Rewards - [x] BAT balance is retained - [x] Auto-contribute list is retained - [x] Both Tips and Monthly Contributions are retained - [x] Summary panel transactions list is retained - [x] Changes to rewards settings are retained - [x] Ads - [x] Both `Estimated pending rewards` & `Ad notifications received this month` are retained - [x] Changes to ads settings are retained - [x] Ensure that ads are not being enabled when upgrading to a new version if they were disabled - [x] Ensure that ads are not disabled when upgrading to a new version if they were enabled
test
manual test run on android tab for x release installer check that installer is close to the size of the last release check the brave version in about and make sure it is exactly as expected startup verify that brave is only contacting brave com endpoints on first launch using either charles proxy fiddler or wireshark or a similar application verify that opening a ntp doesn t trigger any outbound connections related to widgets without user interaction visual look make sure thereafter every merge no chrome chromium words appear on brave version no chrome chromium words appear on normal or private tabs no chrome chromium words appear in site settings for location camera microphone augmented reality no chrome chromium icons are shown in normal or private tabs custom tabs make sure brave handles links from gmail slack ensure custom tabs work even with sync enabled disabled tab groups ensure tab group is enabled by default verify toggling tab group setting triggers a relaunch request verify tab group is not reverted back to default setting after browser restart developer tools verify you can inspect sub links via dev tools clear data verify clear data on exit works as intended settings and bottom bar verify changing default settings are retained and don t cause the browser to crash verify bottom bar buttons home bookmark search tabs work as expected verify items in hamburger menu share menu to ensure nothing unexpected has been added ipfs ens unstoppable domain verify settings for ipfs ens unstoppable domain show up under brave shields privacy settings downloads verify downloading a file works and that all actions on the download item work verify that pdf is downloaded over https at verify that pdf is downloaded over http at bravery settings check that https everywhere works by loading turning https everywhere off and shields off both disable the redirect to check that toggling to blocking and allow ads works as expected verify that clicking through a cert error in works visit and then turn on script blocking nothing should load allow it from the script blocking ui in the url bar and it should work verify that default bravery settings take effect on pages with no site settings verify that party storage results are blank at when party cookies are blocked choose a dns provider from the list in settings privacy use secure dns load and verify your isp s dns resolvers aren t detected and shown only your chosen doh provider should appear run all six test configurations on and confirm results are as expected per each test listed fingerprint tests visit ensure blocked items are listed in shields test that doesn t leak ip address for each option under settings privacy and security webrtc ip handling policy content tests go to and click on the twitter icon on the top right verify that context menus work in the new twitter tab go to and make sure that the password can be saved make sure the saved password is auto populated when you visit the site again verify that shows up as grey not red no mixed content scripts are run brave rewards verify you are able to create a new rewards profile and are in the unverified state by default verify when you visit a creator in this state the panel shows a prompt to connect a custodian and no bat information earnings balance etc is displayed verify when you visit brave rewards in this state there are no auto contribution tipping or summary panels on this page verify you can toggle ads off on from this page and the panel responds accordingly verify you can toggle rewards off on from the panel and the brave rewards page responds accordingly verify you are able to connect a custodian verify rewards balance shows correct bat and usd value on brave rewards and panel after you connect verify when you click on the br panel while on a site the panel displays site specific information site favicon domain attention verify br panel and tip banner shows message about an unverified creator verify you are able to tip a creator who has the same custodian as you verify you are able to perform an auto contribution verify if you disable auto contribute you are still able to tip creators verify if auto contribute is disabled ac does not occur verify that there is no disconnect option once connected to a custodian verify that you are able to reset rewards brave ads verify ads is auto enabled when rewards is enabled for the supported region verify ads are only shown when the app is being used verify ad notification are shown based on ads per hour setting verify ad notifications stack up in notification tray verify swipe left right dismisses the ad notification when shown and is not stored in the notification tray verify clicking on an ad notification shows the landing page verify view clicked and landed and dismiss states are logged based on the action sync verify you are able to join sync chain by scanning the qr code verify you are able to join sync chain using code words verify you are able to create a sync chain on the device and add other devices to the chain via qr code code words verify once sync chain is created categories option is shown in devices list verify only bookmarks is enabled in categories by default verify enabling sync everything enables all other switches verify existing bookmarks before joining sync chain also gets sync d to all devices on the sync chain verify autofill data is sync d to all devices when enabled verify history is sync d to all devices verify open tab form other devices shows up under history below the device name verify password is sync d to all devices verify settings sync s site settings on all devices verify sync works on an upgrade profile and new bookmarks added post upgrade sync s across devices on the chain verify adding a bookmark on custom tab gets sync d across all devices in the chain verify you are able to create a standalone sync chain with one device verify remove this device shows a confirmation alert before removing the sync chain on the device top sites view long press on top sites to get to deletion mode and delete a top site note this will stop that site from showing up again on top sites so you may not want to do this a site you want to keep there session storage verify that tabs restore when closed including active tab upgrade from previous version examples of pre requisites before upgrading from previous version to the build being tested visit several websites so top tiles under new tab page get updated populated add several websites to the android home screen via add to home screen from the hamburger setting menu add several bookmarks include folders change update several settings example default search engine adding a custom home page changing site settings changing several privacy settings etc create enable a brave wallet or restore if individual who s going through the upgrade has one available enable brave news and add remove several sources change the default shield settings on varios websites example enable script blocking change fp ad blocking to strict disable shields login into several websites and leave them opened login into several websites and close the tabs so they re not opened when upgrading login into a website and leave it as the active tab while upgrading enable create a sync chain with several devices upgrade cases ensure that brave version displays both the correct brave version and expected chromium version ensure that bookmarks and folders from the previous release have been retained can be loaded without issues ensure that previously opened tabs are retained and can be lazy loaded with out issues ensure that websites that have been logged into and were opened while upgrading are still logged in ensure cookies are not cleared lost ensure that websites that have been logged into but weren t open when upgrading are still logged in when loading in new tab ensure cookies are not cleared lost ensure that the correct website loads as the active tab as per the pre requisites this tab should be logged into a website ensure that the shields settings from the previous versions are retained on each website ensure that saved passwords are being displayed under settings passwords and can be autofilled without issues ensure that the sync chain is preserved and syncing still works across the devices that have been added ensure that sharing a links tabs using send to your device works as expected ensure that brave news is still enabled and the sources changes from the previous release are retained ensure that the various setting changes via settings are retained example enable script blocking change fp ad blocking to strict disable shields ensure that ipfs ens and unstoppable domain settings are retained from the previous version ensure that history is retained from the previous version ensure that the website shortcuts added via add to home screen are still visible on the android home screen and tapping on icons loads the appropriate webpage ensure that the tab group setting doesn t reset change after upgrading ensure that clear data on exit is retained from the previous version rewards bat balance is retained auto contribute list is retained both tips and monthly contributions are retained summary panel transactions list is retained changes to rewards settings are retained ads both estimated pending rewards ad notifications received this month are retained changes to ads settings are retained ensure that ads are not being enabled when upgrading to a new version if they were disabled ensure that ads are not disabled when upgrading to a new version if they were enabled
1
371,567
25,956,755,277
IssuesEvent
2022-12-18 10:36:38
44r0n/miorganizador
https://api.github.com/repos/44r0n/miorganizador
opened
Control de finanzas. Kakebo
documentation enhancement
**¿Tu característica está relacionada con un problema? Por favor, descríbela./Is your feature request related to a problem? Please describe.** Control de finanzas estilo [Kakebo](https://www.google.com/search?hl=es&q=kakebo) **Detalla la solución que te gustaría./Describe the solution you'd like** En la apertura de la aplicación debería preguntar si se han hecho movimientos hoy. El formulario de entrada de gastos debe ser simple, con candidad, concepto y categoría debería bastar. Al inicio de cada mes, sugerir automáticamente al iniciar la aplicación, junto a los movimientos, el cierre del mes anterior, mostrando el resumen del mes y preguntando si se han cumplido los objetivos. Después del cierre se deberá realizar la introducción de una apertura de mes. Usando como plantilla la del mes pasado. El usuario podrá cambiar los datos. Un mes cerrado no se podrá modificar. Las categorías serán las predefinidas por [Kakebo](https://www.google.com/search?hl=es&q=kakebo) de momento, - **Gastos de primera necesidad**: Incluye gastos indispensables para la supervivencia, tales como comida, farmacia, transportes y combustible. - **Opcionales**: Incluye las compras no indispensables, tables como tiendas, regalos o caprichos. - **Cultura y tiempo libre**: Incluye los gastos relacionados con cultura y entretenimiento, tales como libros, música, cine y restauración. - **Extras e imprevistos**: Incluye los gastos que no caben en las categorías anteriores y son gastos excepcionales, tales como viajes, imprevistos en casa o reparaciones. **Alternativas que has considerado./Describe alternatives you've considered** Hay formas de controlar esto en un excel, pero estaría bien tenerlo automatizado.
1.0
Control de finanzas. Kakebo - **¿Tu característica está relacionada con un problema? Por favor, descríbela./Is your feature request related to a problem? Please describe.** Control de finanzas estilo [Kakebo](https://www.google.com/search?hl=es&q=kakebo) **Detalla la solución que te gustaría./Describe the solution you'd like** En la apertura de la aplicación debería preguntar si se han hecho movimientos hoy. El formulario de entrada de gastos debe ser simple, con candidad, concepto y categoría debería bastar. Al inicio de cada mes, sugerir automáticamente al iniciar la aplicación, junto a los movimientos, el cierre del mes anterior, mostrando el resumen del mes y preguntando si se han cumplido los objetivos. Después del cierre se deberá realizar la introducción de una apertura de mes. Usando como plantilla la del mes pasado. El usuario podrá cambiar los datos. Un mes cerrado no se podrá modificar. Las categorías serán las predefinidas por [Kakebo](https://www.google.com/search?hl=es&q=kakebo) de momento, - **Gastos de primera necesidad**: Incluye gastos indispensables para la supervivencia, tales como comida, farmacia, transportes y combustible. - **Opcionales**: Incluye las compras no indispensables, tables como tiendas, regalos o caprichos. - **Cultura y tiempo libre**: Incluye los gastos relacionados con cultura y entretenimiento, tales como libros, música, cine y restauración. - **Extras e imprevistos**: Incluye los gastos que no caben en las categorías anteriores y son gastos excepcionales, tales como viajes, imprevistos en casa o reparaciones. **Alternativas que has considerado./Describe alternatives you've considered** Hay formas de controlar esto en un excel, pero estaría bien tenerlo automatizado.
non_test
control de finanzas kakebo ¿tu característica está relacionada con un problema por favor descríbela is your feature request related to a problem please describe control de finanzas estilo detalla la solución que te gustaría describe the solution you d like en la apertura de la aplicación debería preguntar si se han hecho movimientos hoy el formulario de entrada de gastos debe ser simple con candidad concepto y categoría debería bastar al inicio de cada mes sugerir automáticamente al iniciar la aplicación junto a los movimientos el cierre del mes anterior mostrando el resumen del mes y preguntando si se han cumplido los objetivos después del cierre se deberá realizar la introducción de una apertura de mes usando como plantilla la del mes pasado el usuario podrá cambiar los datos un mes cerrado no se podrá modificar las categorías serán las predefinidas por de momento gastos de primera necesidad incluye gastos indispensables para la supervivencia tales como comida farmacia transportes y combustible opcionales incluye las compras no indispensables tables como tiendas regalos o caprichos cultura y tiempo libre incluye los gastos relacionados con cultura y entretenimiento tales como libros música cine y restauración extras e imprevistos incluye los gastos que no caben en las categorías anteriores y son gastos excepcionales tales como viajes imprevistos en casa o reparaciones alternativas que has considerado describe alternatives you ve considered hay formas de controlar esto en un excel pero estaría bien tenerlo automatizado
0
35,608
4,997,680,767
IssuesEvent
2016-12-09 17:26:40
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
github.com/cockroachdb/cockroach/pkg/storage: TestStoreRangeSystemSplits failed under stress
Robot test-failure
SHA: https://github.com/cockroachdb/cockroach/commits/d879e9a153024cd4c4d6082a9aa78cc90994d276 Parameters: ``` COCKROACH_PROPOSER_EVALUATED_KV=false TAGS=deadlock GOFLAGS= ``` Stress build found a failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=85800&tab=buildLog ``` github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0x8, 0xc43970bf00) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204cc0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2350998 [semacquire]: sync.runtime_notifyListWait(0xc426e02390, 0xc400000187) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc426e02380) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204ca0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351011 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc4386ce3c0, 0x18a0520, 0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/metric.(*Histogram).RecordValue(0xc426689080, 0x2968) /go/src/github.com/cockroachdb/cockroach/pkg/util/metric/metric.go:217 +0x3a github.com/cockroachdb/cockroach/pkg/storage.newReplica.func4(0x2968) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:603 +0x46 github.com/cockroachdb/cockroach/pkg/util/syncutil.ThresholdLogger.func1(0x2968) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:53 +0x6c github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Unlock(0xc42c07e0f0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:97 +0xb6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tickRaftMuLocked(0xc42c07e000, 0x1c35d00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2720 +0x1d9 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tick(0xc42c07e000, 0xc429202c00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2673 +0x85 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0x16, 0xc4391adef0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204e60) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2392633 [semacquire]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc4386ce280, 0xc430e73860) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2351008 [running]: goroutine running on other thread; stack unavailable created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2350740 [chan receive]: github.com/cockroachdb/cockroach/pkg/storage/engine.(*RocksDB).open.func1(0xc432777e00) /go/src/github.com/cockroachdb/cockroach/pkg/storage/engine/rocksdb.go:373 +0x44 created by github.com/cockroachdb/cockroach/pkg/storage/engine.(*RocksDB).open /go/src/github.com/cockroachdb/cockroach/pkg/storage/engine/rocksdb.go:374 +0x702 goroutine 2351040 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).maybeAddToPurgatory.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:637 +0xaef github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42fde5800) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2392603 [semacquire]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc4386ce1c0, 0xc430e72d20) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2392645 [semacquire]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc4386ce320, 0xc4311702a0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2392658 [semacquire]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc4386ce3d0, 0xc431170e40) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2392577 [runnable]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc4386ce130, 0xc430e72540) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2350781 [select]: github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).printStatsLoop(0xc42c62e820, 0x2b67bf4ed6f8, 0xc4200943f8) /go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:209 +0xbd3 github.com/cockroachdb/cockroach/pkg/kv.NewTxnCoordSender.func1() /go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:193 +0x98 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc426a54d70) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351005 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc4386ce130, 0x18a0520, 0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/metric.(*Histogram).RecordValue(0xc426689080, 0x14820) /go/src/github.com/cockroachdb/cockroach/pkg/util/metric/metric.go:217 +0x3a github.com/cockroachdb/cockroach/pkg/storage.newReplica.func4(0x14820) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:603 +0x46 github.com/cockroachdb/cockroach/pkg/util/syncutil.ThresholdLogger.func1(0x14820) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:53 +0x6c github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Unlock(0xc422ae0ff0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:97 +0xb6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tickRaftMuLocked(0xc422ae0f00, 0x1c35d00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2720 +0x1d9 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tick(0xc422ae0f00, 0xc429202c00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2673 +0x85 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0xd, 0xc43972bf00) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204da0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2392627 [semacquire]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc4386ce220, 0xc430e73380) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2350995 [chan receive]: github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:171 +0x55 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204c40) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2350742 [chan receive]: github.com/cockroachdb/cockroach/pkg/rpc.NewContext.func1() /go/src/github.com/cockroachdb/cockroach/pkg/rpc/context.go:137 +0x76 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42328eb80) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2392644 [semacquire]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc4386ce310, 0xc4311701e0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2351098 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tryAddWriteCmd(0xc431ee0a00, 0x2b67bf5f1878, 0xc43073b530, 0x7b, 0x749, 0x100000001, 0x1, 0x1a, 0x0, 0xc42c9e2c00, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2021 +0xb17 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).addWriteCmd(0xc431ee0a00, 0x2b67bf5f1878, 0xc43073b530, 0x7b, 0x749, 0x100000001, 0x1, 0x1a, 0x0, 0xc42c9e2c00, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:1847 +0x90 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).Send(0xc431ee0a00, 0x2b67bf5f1878, 0xc43073b530, 0x7b, 0x749, 0x100000001, 0x1, 0x1a, 0x0, 0xc42c9e2c00, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:1299 +0x222 github.com/cockroachdb/cockroach/pkg/storage.(*Store).Send(0xc431d86700, 0x2b67bf5f1878, 0xc43073b440, 0x7b, 0x749, 0x100000001, 0x1, 0x1a, 0x0, 0xc42c9e2c00, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:2471 +0x6c5 github.com/cockroachdb/cockroach/pkg/storage.(*Stores).Send(0xc4266b2a20, 0x2b67bf5f1878, 0xc43073b2c0, 0x0, 0x0, 0x100000001, 0x1, 0x1a, 0x0, 0xc42c9e2c00, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/stores.go:187 +0x202 github.com/cockroachdb/cockroach/pkg/kv.(*senderTransport).SendNext(0xc433296aa0, 0xc431af3740) /go/src/github.com/cockroachdb/cockroach/pkg/kv/transport.go:309 +0x1f8 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendToReplicas(0xc4319bdc20, 0x2b67bf5f1800, 0xc43869aa80, 0x1dcd6500, 0xc42caebf80, 0xc4319bdc58, 0x4, 0xc431583840, 0x1, 0x1, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1142 +0x35d github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendRPC(0xc4319bdc20, 0x2b67bf5f1800, 0xc43869aa80, 0x4, 0xc431583840, 0x1, 0x1, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:410 +0x372 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendSingleRange(0xc4319bdc20, 0x2b67bf515760, 0xc42c7ccb40, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc42c9e2b40, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:479 +0x16e github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendPartialBatch(0xc4319bdc20, 0x2b67bf515760, 0xc42c7ccb40, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc42c9e2b40, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:931 +0x2c7 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges(0xc4319bdc20, 0x2b67bf515760, 0xc42c7ccb40, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc42c9e2b40, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:810 +0x41f github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send(0xc4319bdc20, 0x2b67bf515760, 0xc42c7ccb40, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc42c9e2a80, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:625 +0x2d8 github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send(0xc42c62e820, 0x2b67bf515760, 0xc42c7ccb40, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc42c9e29c0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:418 +0x7e9 github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).sendInternal(0xc42a7bde40, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc42a7bde88, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:587 +0xfd github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).send(0xc42a7bde40, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:711 +0x498 github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).(github.com/cockroachdb/cockroach/pkg/internal/client.send)-fm(0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:313 +0x58 github.com/cockroachdb/cockroach/pkg/internal/client.sendAndFill(0xc4326591b8, 0xc421d60600, 0x0, 0xc43883df50) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:418 +0x168 github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Run(0xc42a7bde40, 0xc421d60600, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:313 +0xc6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).adminSplitWithDescriptor.func1(0xc42a7bde40, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:2407 +0x27b github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn.func1(0xc42a7bde40, 0xc43883ded0, 0x4000000000000000, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:468 +0x2f github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Exec(0xc42a7bde40, 0xc438830101, 0x0, 0xc43883dec0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:520 +0x215 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn(0xc432dc26e0, 0x2b67bf515760, 0xc42c7ccb40, 0xc438670640, 0xc4326598b0, 0x2) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:469 +0x1fa github.com/cockroachdb/cockroach/pkg/storage.(*Replica).adminSplitWithDescriptor(0xc431ee0a00, 0x2b67bf515760, 0xc42c7ccb40, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc438766928, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:2449 +0xab4 github.com/cockroachdb/cockroach/pkg/storage.(*splitQueue).process(0xc4269e87a0, 0x2b67bf515760, 0xc42c7ccb40, 0xc4224373e0, 0xc431ee0a00, 0xc430826000, 0x1d, 0x20, 0x0, 0x2) /go/src/github.com/cockroachdb/cockroach/pkg/storage/split_queue.go:111 +0x290 github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processReplica(0xc4319bde00, 0x2b67bf5f1878, 0xc422437320, 0xc431ee0a00, 0xc42a27ff80, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:579 +0x3e6 github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1.2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:499 +0xf2 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunTask(0xc42dc082d0, 0xc432659e40, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:224 +0x100 github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:506 +0x34d github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42936dae0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2350732 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Replica).beginCmds(0xc42e294f00, 0x2b67bf5f1878, 0xc422437680, 0xc4323487e0, 0x7b, 0x74d, 0xc4325b2370) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:1513 +0x10da github.com/cockroachdb/cockroach/pkg/storage.(*Replica).addReadOnlyCmd(0xc42e294f00, 0x2b67bf5f1878, 0xc422437680, 0x7b, 0x6e8, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:1770 +0x6ee github.com/cockroachdb/cockroach/pkg/storage.(*Replica).Send(0xc42e294f00, 0x2b67bf5f1878, 0xc422437680, 0x7b, 0x6e8, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:1302 +0x48f github.com/cockroachdb/cockroach/pkg/storage.(*Store).Send(0xc431d86700, 0x2b67bf5f1878, 0xc4324c2840, 0x7b, 0x6e8, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:2471 +0x6c5 github.com/cockroachdb/cockroach/pkg/storage.(*Stores).Send(0xc4266b2a20, 0x2b67bf5f1878, 0xc4324c27b0, 0x0, 0x0, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/stores.go:187 +0x202 github.com/cockroachdb/cockroach/pkg/kv.(*senderTransport).SendNext(0xc432c435e0, 0xc4331d7560) /go/src/github.com/cockroachdb/cockroach/pkg/kv/transport.go:309 +0x1f8 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendToReplicas(0xc4319bdc20, 0x2b67bf5f1800, 0xc438a9ce40, 0x1dcd6500, 0xc42caebf80, 0xc4319bdc58, 0x1, 0xc42d7228e0, 0x1, 0x1, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1142 +0x35d github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendRPC(0xc4319bdc20, 0x2b67bf5f1800, 0xc438a9ce40, 0x1, 0xc42d7228e0, 0x1, 0x1, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:410 +0x372 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendSingleRange(0xc4319bdc20, 0x2b67bf5f1878, 0xc4324c23f0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:479 +0x16e github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendPartialBatch(0xc4319bdc20, 0x2b67bf5f1878, 0xc4324c23f0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:931 +0x2c7 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges(0xc4319bdc20, 0x2b67bf5f1878, 0xc4324c23f0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:810 +0x41f github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send(0xc4319bdc20, 0x2b67bf5f1878, 0xc4324c23f0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:625 +0x2d8 github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send(0xc42c62e820, 0x2b67bf5f1878, 0xc4324c23f0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:418 +0x7e9 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).send(0xc432dc26e0, 0x2b67bf4ed6f8, 0xc420094410, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:513 +0xbf github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Run.func1(0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:447 +0x6e github.com/cockroachdb/cockroach/pkg/internal/client.sendAndFill(0xc42d722780, 0xc432f7ac00, 0x0, 0xc42d722760) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:418 +0x168 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Run(0xc432dc26e0, 0x2b67bf4ed6f8, 0xc420094410, 0xc432f7ac00, 0xc42d722760, 0xc42d92ae80) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:449 +0xac github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).scan(0xc432dc26e0, 0x2b67bf4ed6f8, 0xc420094410, 0x195bee0, 0xc42d722740, 0x195bee0, 0xc42d722760, 0x0, 0x2d722700, 0xc4345e1968, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:307 +0x105 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Scan(0xc432dc26e0, 0x2b67bf4ed6f8, 0xc420094410, 0x195bee0, 0xc42d722740, 0x195bee0, 0xc42d722760, 0x0, 0x1c2eadd, 0xc41f3a8225, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:318 +0x9a github.com/cockroachdb/cockroach/pkg/storage_test.TestStoreRangeSystemSplits.func2.1(0xecfdc7d23, 0xc43ac4d307) /go/src/github.com/cockroachdb/cockroach/pkg/storage/client_split_test.go:867 +0x15e github.com/cockroachdb/cockroach/pkg/util.RetryForDuration(0xa7a358200, 0xc4345e1c80, 0xc4345e1bb8, 0x270) /go/src/github.com/cockroachdb/cockroach/pkg/util/retry.go:31 +0xcd github.com/cockroachdb/cockroach/pkg/testutils.SucceedsSoonDepth(0x1, 0x26294a0, 0xc4311069c0, 0xc4345e1c80) /go/src/github.com/cockroachdb/cockroach/pkg/testutils/soon.go:41 +0x41 github.com/cockroachdb/cockroach/pkg/storage_test.TestStoreRangeSystemSplits.func2(0x16) /go/src/github.com/cockroachdb/cockroach/pkg/storage/client_split_test.go:879 +0x67d github.com/cockroachdb/cockroach/pkg/storage_test.TestStoreRangeSystemSplits(0xc4311069c0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/client_split_test.go:897 +0x448 testing.tRunner(0xc4311069c0, 0x1c35998) /usr/local/go/src/testing/testing.go:610 +0x81 created by testing.(*T).Run /usr/local/go/src/testing/testing.go:646 +0x2ec goroutine 2392148 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tryAddWriteCmd(0xc42e294f00, 0x2b67bf5f1878, 0xc43073ad50, 0x7b, 0x745, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2021 +0xb17 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).addWriteCmd(0xc42e294f00, 0x2b67bf5f1878, 0xc43073ad50, 0x7b, 0x745, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:1847 +0x90 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).Send(0xc42e294f00, 0x2b67bf5f1878, 0xc43073ad50, 0x7b, 0x745, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:1299 +0x222 github.com/cockroachdb/cockroach/pkg/storage.(*Store).Send(0xc431d86700, 0x2b67bf5f1878, 0xc43073acc0, 0x7b, 0x745, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:2471 +0x6c5 github.com/cockroachdb/cockroach/pkg/storage.(*Stores).Send(0xc4266b2a20, 0x2b67bf5f1878, 0xc43073ac00, 0x0, 0x0, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/stores.go:187 +0x202 github.com/cockroachdb/cockroach/pkg/kv.(*senderTransport).SendNext(0xc4332966e0, 0xc43074e780) /go/src/github.com/cockroachdb/cockroach/pkg/kv/transport.go:309 +0x1f8 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendToReplicas(0xc4319bdc20, 0x2b67bf5f1800, 0xc43889b780, 0x1dcd6500, 0xc42caebf80, 0xc4319bdc58, 0x1, 0xc43053f3e0, 0x1, 0x1, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1142 +0x35d github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendRPC(0xc4319bdc20, 0x2b67bf5f1800, 0xc43889b780, 0x1, 0xc43053f3e0, 0x1, 0x1, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:410 +0x372 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendSingleRange(0xc4319bdc20, 0x2b67bf5f1878, 0xc4392719e0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:479 +0x16e github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendPartialBatch(0xc4319bdc20, 0x2b67bf5f1878, 0xc4392719e0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:931 +0x2c7 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges(0xc4319bdc20, 0x2b67bf5f1878, 0xc4392719e0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:810 +0x41f github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send(0xc4319bdc20, 0x2b67bf5f1878, 0xc4392719e0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:625 +0x2d8 github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send(0xc42c62e820, 0x2b67bf5f1878, 0xc4392719e0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:418 +0x7e9 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).send(0xc432dc26e0, 0x2b67bf5f1878, 0xc4392719e0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:513 +0xbf github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Run.func1(0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:447 +0x6e github.com/cockroachdb/cockroach/pkg/internal/client.sendAndFill(0xc43053f2a0, 0xc431fd2600, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:418 +0x168 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Run(0xc432dc26e0, 0x2b67bf5f1878, 0xc4392719e0, 0xc431fd2600, 0x2625a01, 0xc439271a70) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:449 +0xac github.com/cockroachdb/cockroach/pkg/storage.(*intentResolver).resolveIntents.func1(0x19516a0, 0xc439271a70) /go/src/github.com/cockroachdb/cockroach/pkg/storage/intent_resolver.go:450 +0x4c github.com/cockroachdb/cockroach/pkg/storage.(*intentResolver).resolveIntents(0xc423226c40, 0x2b67bf5f1878, 0xc4392719e0, 0xc42a2dc7e0, 0x2, 0x2, 0x18a0001, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/intent_resolver.go:461 +0x5de github.com/cockroachdb/cockroach/pkg/storage.(*intentResolver).processIntents(0xc423226c40, 0x2b67bf4ed6f8, 0xc420094410, 0xc42c07ea00, 0x2625560, 0xc427c9d0e0, 0xc42a2dc7e0, 0x2, 0x2, 0x7b, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/intent_resolver.go:345 +0x449 github.com/cockroachdb/cockroach/pkg/storage.(*intentResolver).processIntentsAsync.func1(0x2b67bf4ed6f8, 0xc420094410) /go/src/github.com/cockroachdb/cockroach/pkg/storage/intent_resolver.go:277 +0x7f github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunLimitedAsyncTask.func1(0xc42dc082d0, 0x230c20d, 0x1a, 0x116, 0xc425098c00, 0x0, 0x0, 0xc43073a8a0, 0x2b67bf4ed6f8, 0xc420094410) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:326 +0x10c created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunLimitedAsyncTask /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:327 +0x3f0 goroutine 2351051 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Store).startGossip.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:1287 +0x39b github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42ac14690) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2392610 [semacquire]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc438767660, 0xc422481c80) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2350996 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc4386ce3d0, 0x18a0520, 0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/metric.(*Histogram).RecordValue(0xc426689080, 0x5999a) /go/src/github.com/cockroachdb/cockroach/pkg/util/metric/metric.go:217 +0x3a github.com/cockroachdb/cockroach/pkg/storage.newReplica.func4(0x5999a) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:603 +0x46 github.com/cockroachdb/cockroach/pkg/util/syncutil.ThresholdLogger.func1(0x5999a) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:53 +0x6c github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Unlock(0xc4232919f0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:97 +0xb6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tickRaftMuLocked(0xc423291900, 0x1c35d00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2720 +0x1d9 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tick(0xc423291900, 0xc429202c00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2673 +0x85 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0x12, 0xc42f981ef0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204c60) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351049 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Store).Start.func3() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:1201 +0x165 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42936d8a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2350997 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc4386ce280, 0x18a0520, 0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/metric.(*Histogram).RecordValue(0xc426689080, 0x2bc0) /go/src/github.com/cockroachdb/cockroach/pkg/util/metric/metric.go:217 +0x3a github.com/cockroachdb/cockroach/pkg/storage.newReplica.func4(0x2bc0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:603 +0x46 github.com/cockroachdb/cockroach/pkg/util/syncutil.ThresholdLogger.func1(0x2bc0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:53 +0x6c github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Unlock(0xc42fe30ff0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:97 +0xb6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tickRaftMuLocked(0xc42fe30f00, 0x1c35d00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2720 +0x1d9 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tick(0xc42fe30f00, 0xc429202c00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2673 +0x85 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0xf, 0xc43951bef0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204c80) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351012 [semacquire]: sync.runtime_Semacquire(0xc421bd0db4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*WaitGroup).Wait(0xc421bd0da8) /usr/local/go/src/sync/waitgroup.go:131 +0x97 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Wait(0xc421bd0d10) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:187 +0x34 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processRaft.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3357 +0x3a github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc4269a40d0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351007 [runnable]: sync.runtime_Semacquire(0xc4201522a4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4201522a0) /usr/local/go/src/sync/mutex.go:85 +0xd0 github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*lockOrder).PreLock(0xc4201522a0, 0x4, 0x18a0520, 0xc4269e8640) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:243 +0xc4 github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.PreLock(0x4, 0x18a0520, 0xc4269e8640) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:129 +0x4b github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc4386ce0e0, 0x18a0520, 0xc4269e8640) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:145 +0x93 github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4269e8640) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Lock(0xc431d869d8) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:86 +0x34 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0x9, 0xc431bdff00) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3332 +0x76 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204de0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351000 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc4386ce320, 0x18a0520, 0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/metric.(*Histogram).RecordValue(0xc426689080, 0x4b191) /go/src/github.com/cockroachdb/cockroach/pkg/util/metric/metric.go:217 +0x3a github.com/cockroachdb/cockroach/pkg/storage.newReplica.func4(0x4b191) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:603 +0x46 github.com/cockroachdb/cockroach/pkg/util/syncutil.ThresholdLogger.func1(0x4b191) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:53 +0x6c github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Unlock(0xc4321714f0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:97 +0xb6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tickRaftMuLocked(0xc432171400, 0x1c35d00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2720 +0x1d9 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tick(0xc432171400, 0xc429202c00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2673 +0x85 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0x2, 0xc439745f00) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204ce0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351050 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Store).startGossip.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:1287 +0x39b github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42ac14660) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351004 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc438766db0, 0x18a0520, 0xc4202c8560) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4202c8560) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/storage.TrackRaftProtos.func1(0x2619c00, 0xc434ba2280) /go/src/github.com/cockroachdb/cockroach/pkg/storage/track_raft_protos.go:52 +0xdb github.com/cockroachdb/cockroach/pkg/roachpb.(*Value).SetProto(0xc4392133f8, 0x2619c00, 0xc434ba2280, 0x10, 0x20) /go/src/github.com/cockroachdb/cockroach/pkg/roachpb/data.go:369 +0x118 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).append(0xc431ee0a00, 0x2b67bf5f1878, 0xc422437830, 0x2626560, 0xc43021a478, 0xb, 0x2a, 0xc434ba2280, 0x1, 0x1, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_raftstorage.go:504 +0x144 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).handleRaftReadyRaftMuLocked(0xc431ee0a00, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2561 +0x1028 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).handleRaftReady(0xc431ee0a00, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2453 +0xc5 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processReady(0xc431d86700, 0x1a) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3296 +0x137 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:230 +0x28c github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204d80) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351097 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:475 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42936dac0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351100 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:475 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42936db20) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2392651 [semacquire]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc4386ce370, 0xc4311707e0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2351006 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc4386ce2e0, 0x18a0520, 0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/metric.(*Histogram).RecordValue(0xc426689080, 0x4fb0) /go/src/github.com/cockroachdb/cockroach/pkg/util/metric/metric.go:217 +0x3a github.com/cockroachdb/cockroach/pkg/storage.newReplica.func4(0x4fb0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:603 +0x46 github.com/cockroachdb/cockroach/pkg/util/syncutil.ThresholdLogger.func1(0x4fb0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:53 +0x6c github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Unlock(0xc42c6680f0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:97 +0xb6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tickRaftMuLocked(0xc42c668000, 0x1c35d00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2723 +0x221 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tick(0xc42c668000, 0xc429202c00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2673 +0x85 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0x18, 0xc42edcbef0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204dc0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351099 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:475 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42936db00) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351424 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*idAllocator).start.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/id_alloc.go:133 +0x7c8 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc4268d2fc0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351101 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:475 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42936db40) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351001 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc4386ce310, 0x18a0520, 0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/metric.(*Histogram).RecordValue(0xc426689080, 0x2774) /go/src/github.com/cockroachdb/cockroach/pkg/util/metric/metric.go:217 +0x3a github.com/cockroachdb/cockroach/pkg/storage.newReplica.func4(0x2774) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:603 +0x46 github.com/cockroachdb/cockroach/pkg/util/syncutil.ThresholdLogger.func1(0x2774) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:53 +0x6c github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Unlock(0xc42fe30af0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:97 +0xb6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tickRaftMuLocked(0xc42fe30a00, 0x1c35d00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2720 +0x1d9 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tick(0xc42fe30a00, 0xc429202c00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2673 +0x85 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0xe, 0xc4397c9f00) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204d20) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2392441 [semacquire]: sync.runtime_Semacquire(0xc4202c8564) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4202c8560) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc438766db0, 0xc430616e40) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2392657 [semacquire]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc4386ce3c0, 0xc431170d80) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2351102 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:475 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42936db60) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351010 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc438767660, 0x18a0520, 0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/metric.(*Histogram).RecordValue(0xc426689080, 0x2bc0) /go/src/github.com/cockroachdb/cockroach/pkg/util/metric/metric.go:217 +0x3a github.com/cockroachdb/cockroach/pkg/storage.newReplica.func4(0x2bc0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:603 +0x46 github.com/cockroachdb/cockroach/pkg/util/syncutil.ThresholdLogger.func1(0x2bc0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:53 +0x6c github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Unlock(0xc42ffbaaf0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:97 +0xb6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tickRaftMuLocked(0xc42ffbaa00, 0x1c35d00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2720 +0x1d9 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tick(0xc42ffbaa00, 0xc429202c00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2673 +0x85 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0x5, 0xc430ffff00) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204e40) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2392640 [semacquire]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc4386ce2e0, 0xc430e73e00) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2351013 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Store).raftTickLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3383 +0x334 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc4269a4140) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351014 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Store).startCoalescedHeartbeatsLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3414 +0x167 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc4269a4150) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351003 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc4386ce220, 0x18a0520, 0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/metric.(*Histogram).RecordValue(0xc426689080, 0x283c) /go/src/github.com/cockroachdb/cockroach/pkg/util/metric/metric.go:217 +0x3a github.com/cockroachdb/cockroach/pkg/storage.newReplica.func4(0x283c) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:603 +0x46 github.com/cockroachdb/cockroach/pkg/util/syncutil.ThresholdLogger.func1(0x283c) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:53 +0x6c github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Unlock(0xc427e76af0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:97 +0xb6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tickRaftMuLocked(0xc427e76a00, 0x1c35d00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2720 +0x1d9 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tick(0xc427e76a00, 0xc429202c00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2673 +0x85 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0xa, 0xc432f9ff00) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204d60) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351009 [semacquire]: sync.runtime_notifyListWait(0xc426e02390, 0xc400000186) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc426e02380) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204e20) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351002 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc4386ce370, 0x18a0520, 0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/metric.(*Histogram).RecordValue(0xc426689080, 0x28a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/metric/metric.go:217 +0x3a github.com/cockroachdb/cockroach/pkg/storage.newReplica.func4(0x28a0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:603 +0x46 github.com/cockroachdb/cockroach/pkg/util/syncutil.ThresholdLogger.func1(0x28a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:53 +0x6c github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Unlock(0xc436e3aaf0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:97 +0xb6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tickRaftMuLocked(0xc436e3aa00, 0x1c35d00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2720 +0x1d9 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tick(0xc436e3aa00, 0xc429202c00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2673 +0x85 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0x11, 0xc4322a5f00) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204d40) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 ERROR: exit status 2 make: *** [stress] Error 1 1 runs completed, 1 failures, over 1m1s Makefile:138: recipe for target 'stress' failed ```
1.0
github.com/cockroachdb/cockroach/pkg/storage: TestStoreRangeSystemSplits failed under stress - SHA: https://github.com/cockroachdb/cockroach/commits/d879e9a153024cd4c4d6082a9aa78cc90994d276 Parameters: ``` COCKROACH_PROPOSER_EVALUATED_KV=false TAGS=deadlock GOFLAGS= ``` Stress build found a failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=85800&tab=buildLog ``` github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0x8, 0xc43970bf00) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204cc0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2350998 [semacquire]: sync.runtime_notifyListWait(0xc426e02390, 0xc400000187) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc426e02380) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204ca0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351011 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc4386ce3c0, 0x18a0520, 0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/metric.(*Histogram).RecordValue(0xc426689080, 0x2968) /go/src/github.com/cockroachdb/cockroach/pkg/util/metric/metric.go:217 +0x3a github.com/cockroachdb/cockroach/pkg/storage.newReplica.func4(0x2968) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:603 +0x46 github.com/cockroachdb/cockroach/pkg/util/syncutil.ThresholdLogger.func1(0x2968) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:53 +0x6c github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Unlock(0xc42c07e0f0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:97 +0xb6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tickRaftMuLocked(0xc42c07e000, 0x1c35d00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2720 +0x1d9 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tick(0xc42c07e000, 0xc429202c00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2673 +0x85 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0x16, 0xc4391adef0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204e60) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2392633 [semacquire]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc4386ce280, 0xc430e73860) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2351008 [running]: goroutine running on other thread; stack unavailable created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2350740 [chan receive]: github.com/cockroachdb/cockroach/pkg/storage/engine.(*RocksDB).open.func1(0xc432777e00) /go/src/github.com/cockroachdb/cockroach/pkg/storage/engine/rocksdb.go:373 +0x44 created by github.com/cockroachdb/cockroach/pkg/storage/engine.(*RocksDB).open /go/src/github.com/cockroachdb/cockroach/pkg/storage/engine/rocksdb.go:374 +0x702 goroutine 2351040 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).maybeAddToPurgatory.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:637 +0xaef github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42fde5800) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2392603 [semacquire]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc4386ce1c0, 0xc430e72d20) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2392645 [semacquire]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc4386ce320, 0xc4311702a0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2392658 [semacquire]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc4386ce3d0, 0xc431170e40) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2392577 [runnable]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc4386ce130, 0xc430e72540) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2350781 [select]: github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).printStatsLoop(0xc42c62e820, 0x2b67bf4ed6f8, 0xc4200943f8) /go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:209 +0xbd3 github.com/cockroachdb/cockroach/pkg/kv.NewTxnCoordSender.func1() /go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:193 +0x98 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc426a54d70) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351005 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc4386ce130, 0x18a0520, 0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/metric.(*Histogram).RecordValue(0xc426689080, 0x14820) /go/src/github.com/cockroachdb/cockroach/pkg/util/metric/metric.go:217 +0x3a github.com/cockroachdb/cockroach/pkg/storage.newReplica.func4(0x14820) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:603 +0x46 github.com/cockroachdb/cockroach/pkg/util/syncutil.ThresholdLogger.func1(0x14820) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:53 +0x6c github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Unlock(0xc422ae0ff0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:97 +0xb6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tickRaftMuLocked(0xc422ae0f00, 0x1c35d00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2720 +0x1d9 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tick(0xc422ae0f00, 0xc429202c00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2673 +0x85 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0xd, 0xc43972bf00) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204da0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2392627 [semacquire]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc4386ce220, 0xc430e73380) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2350995 [chan receive]: github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:171 +0x55 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204c40) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2350742 [chan receive]: github.com/cockroachdb/cockroach/pkg/rpc.NewContext.func1() /go/src/github.com/cockroachdb/cockroach/pkg/rpc/context.go:137 +0x76 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42328eb80) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2392644 [semacquire]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc4386ce310, 0xc4311701e0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2351098 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tryAddWriteCmd(0xc431ee0a00, 0x2b67bf5f1878, 0xc43073b530, 0x7b, 0x749, 0x100000001, 0x1, 0x1a, 0x0, 0xc42c9e2c00, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2021 +0xb17 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).addWriteCmd(0xc431ee0a00, 0x2b67bf5f1878, 0xc43073b530, 0x7b, 0x749, 0x100000001, 0x1, 0x1a, 0x0, 0xc42c9e2c00, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:1847 +0x90 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).Send(0xc431ee0a00, 0x2b67bf5f1878, 0xc43073b530, 0x7b, 0x749, 0x100000001, 0x1, 0x1a, 0x0, 0xc42c9e2c00, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:1299 +0x222 github.com/cockroachdb/cockroach/pkg/storage.(*Store).Send(0xc431d86700, 0x2b67bf5f1878, 0xc43073b440, 0x7b, 0x749, 0x100000001, 0x1, 0x1a, 0x0, 0xc42c9e2c00, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:2471 +0x6c5 github.com/cockroachdb/cockroach/pkg/storage.(*Stores).Send(0xc4266b2a20, 0x2b67bf5f1878, 0xc43073b2c0, 0x0, 0x0, 0x100000001, 0x1, 0x1a, 0x0, 0xc42c9e2c00, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/stores.go:187 +0x202 github.com/cockroachdb/cockroach/pkg/kv.(*senderTransport).SendNext(0xc433296aa0, 0xc431af3740) /go/src/github.com/cockroachdb/cockroach/pkg/kv/transport.go:309 +0x1f8 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendToReplicas(0xc4319bdc20, 0x2b67bf5f1800, 0xc43869aa80, 0x1dcd6500, 0xc42caebf80, 0xc4319bdc58, 0x4, 0xc431583840, 0x1, 0x1, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1142 +0x35d github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendRPC(0xc4319bdc20, 0x2b67bf5f1800, 0xc43869aa80, 0x4, 0xc431583840, 0x1, 0x1, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:410 +0x372 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendSingleRange(0xc4319bdc20, 0x2b67bf515760, 0xc42c7ccb40, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc42c9e2b40, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:479 +0x16e github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendPartialBatch(0xc4319bdc20, 0x2b67bf515760, 0xc42c7ccb40, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc42c9e2b40, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:931 +0x2c7 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges(0xc4319bdc20, 0x2b67bf515760, 0xc42c7ccb40, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc42c9e2b40, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:810 +0x41f github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send(0xc4319bdc20, 0x2b67bf515760, 0xc42c7ccb40, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc42c9e2a80, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:625 +0x2d8 github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send(0xc42c62e820, 0x2b67bf515760, 0xc42c7ccb40, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc42c9e29c0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:418 +0x7e9 github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).sendInternal(0xc42a7bde40, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc42a7bde88, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:587 +0xfd github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).send(0xc42a7bde40, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:711 +0x498 github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).(github.com/cockroachdb/cockroach/pkg/internal/client.send)-fm(0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:313 +0x58 github.com/cockroachdb/cockroach/pkg/internal/client.sendAndFill(0xc4326591b8, 0xc421d60600, 0x0, 0xc43883df50) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:418 +0x168 github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Run(0xc42a7bde40, 0xc421d60600, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:313 +0xc6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).adminSplitWithDescriptor.func1(0xc42a7bde40, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:2407 +0x27b github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn.func1(0xc42a7bde40, 0xc43883ded0, 0x4000000000000000, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:468 +0x2f github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Exec(0xc42a7bde40, 0xc438830101, 0x0, 0xc43883dec0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:520 +0x215 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn(0xc432dc26e0, 0x2b67bf515760, 0xc42c7ccb40, 0xc438670640, 0xc4326598b0, 0x2) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:469 +0x1fa github.com/cockroachdb/cockroach/pkg/storage.(*Replica).adminSplitWithDescriptor(0xc431ee0a00, 0x2b67bf515760, 0xc42c7ccb40, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc438766928, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:2449 +0xab4 github.com/cockroachdb/cockroach/pkg/storage.(*splitQueue).process(0xc4269e87a0, 0x2b67bf515760, 0xc42c7ccb40, 0xc4224373e0, 0xc431ee0a00, 0xc430826000, 0x1d, 0x20, 0x0, 0x2) /go/src/github.com/cockroachdb/cockroach/pkg/storage/split_queue.go:111 +0x290 github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processReplica(0xc4319bde00, 0x2b67bf5f1878, 0xc422437320, 0xc431ee0a00, 0xc42a27ff80, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:579 +0x3e6 github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1.2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:499 +0xf2 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunTask(0xc42dc082d0, 0xc432659e40, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:224 +0x100 github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:506 +0x34d github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42936dae0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2350732 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Replica).beginCmds(0xc42e294f00, 0x2b67bf5f1878, 0xc422437680, 0xc4323487e0, 0x7b, 0x74d, 0xc4325b2370) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:1513 +0x10da github.com/cockroachdb/cockroach/pkg/storage.(*Replica).addReadOnlyCmd(0xc42e294f00, 0x2b67bf5f1878, 0xc422437680, 0x7b, 0x6e8, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:1770 +0x6ee github.com/cockroachdb/cockroach/pkg/storage.(*Replica).Send(0xc42e294f00, 0x2b67bf5f1878, 0xc422437680, 0x7b, 0x6e8, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:1302 +0x48f github.com/cockroachdb/cockroach/pkg/storage.(*Store).Send(0xc431d86700, 0x2b67bf5f1878, 0xc4324c2840, 0x7b, 0x6e8, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:2471 +0x6c5 github.com/cockroachdb/cockroach/pkg/storage.(*Stores).Send(0xc4266b2a20, 0x2b67bf5f1878, 0xc4324c27b0, 0x0, 0x0, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/stores.go:187 +0x202 github.com/cockroachdb/cockroach/pkg/kv.(*senderTransport).SendNext(0xc432c435e0, 0xc4331d7560) /go/src/github.com/cockroachdb/cockroach/pkg/kv/transport.go:309 +0x1f8 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendToReplicas(0xc4319bdc20, 0x2b67bf5f1800, 0xc438a9ce40, 0x1dcd6500, 0xc42caebf80, 0xc4319bdc58, 0x1, 0xc42d7228e0, 0x1, 0x1, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1142 +0x35d github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendRPC(0xc4319bdc20, 0x2b67bf5f1800, 0xc438a9ce40, 0x1, 0xc42d7228e0, 0x1, 0x1, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:410 +0x372 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendSingleRange(0xc4319bdc20, 0x2b67bf5f1878, 0xc4324c23f0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:479 +0x16e github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendPartialBatch(0xc4319bdc20, 0x2b67bf5f1878, 0xc4324c23f0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:931 +0x2c7 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges(0xc4319bdc20, 0x2b67bf5f1878, 0xc4324c23f0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:810 +0x41f github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send(0xc4319bdc20, 0x2b67bf5f1878, 0xc4324c23f0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:625 +0x2d8 github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send(0xc42c62e820, 0x2b67bf5f1878, 0xc4324c23f0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:418 +0x7e9 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).send(0xc432dc26e0, 0x2b67bf4ed6f8, 0xc420094410, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:513 +0xbf github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Run.func1(0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:447 +0x6e github.com/cockroachdb/cockroach/pkg/internal/client.sendAndFill(0xc42d722780, 0xc432f7ac00, 0x0, 0xc42d722760) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:418 +0x168 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Run(0xc432dc26e0, 0x2b67bf4ed6f8, 0xc420094410, 0xc432f7ac00, 0xc42d722760, 0xc42d92ae80) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:449 +0xac github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).scan(0xc432dc26e0, 0x2b67bf4ed6f8, 0xc420094410, 0x195bee0, 0xc42d722740, 0x195bee0, 0xc42d722760, 0x0, 0x2d722700, 0xc4345e1968, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:307 +0x105 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Scan(0xc432dc26e0, 0x2b67bf4ed6f8, 0xc420094410, 0x195bee0, 0xc42d722740, 0x195bee0, 0xc42d722760, 0x0, 0x1c2eadd, 0xc41f3a8225, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:318 +0x9a github.com/cockroachdb/cockroach/pkg/storage_test.TestStoreRangeSystemSplits.func2.1(0xecfdc7d23, 0xc43ac4d307) /go/src/github.com/cockroachdb/cockroach/pkg/storage/client_split_test.go:867 +0x15e github.com/cockroachdb/cockroach/pkg/util.RetryForDuration(0xa7a358200, 0xc4345e1c80, 0xc4345e1bb8, 0x270) /go/src/github.com/cockroachdb/cockroach/pkg/util/retry.go:31 +0xcd github.com/cockroachdb/cockroach/pkg/testutils.SucceedsSoonDepth(0x1, 0x26294a0, 0xc4311069c0, 0xc4345e1c80) /go/src/github.com/cockroachdb/cockroach/pkg/testutils/soon.go:41 +0x41 github.com/cockroachdb/cockroach/pkg/storage_test.TestStoreRangeSystemSplits.func2(0x16) /go/src/github.com/cockroachdb/cockroach/pkg/storage/client_split_test.go:879 +0x67d github.com/cockroachdb/cockroach/pkg/storage_test.TestStoreRangeSystemSplits(0xc4311069c0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/client_split_test.go:897 +0x448 testing.tRunner(0xc4311069c0, 0x1c35998) /usr/local/go/src/testing/testing.go:610 +0x81 created by testing.(*T).Run /usr/local/go/src/testing/testing.go:646 +0x2ec goroutine 2392148 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tryAddWriteCmd(0xc42e294f00, 0x2b67bf5f1878, 0xc43073ad50, 0x7b, 0x745, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2021 +0xb17 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).addWriteCmd(0xc42e294f00, 0x2b67bf5f1878, 0xc43073ad50, 0x7b, 0x745, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:1847 +0x90 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).Send(0xc42e294f00, 0x2b67bf5f1878, 0xc43073ad50, 0x7b, 0x745, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:1299 +0x222 github.com/cockroachdb/cockroach/pkg/storage.(*Store).Send(0xc431d86700, 0x2b67bf5f1878, 0xc43073acc0, 0x7b, 0x745, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:2471 +0x6c5 github.com/cockroachdb/cockroach/pkg/storage.(*Stores).Send(0xc4266b2a20, 0x2b67bf5f1878, 0xc43073ac00, 0x0, 0x0, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/stores.go:187 +0x202 github.com/cockroachdb/cockroach/pkg/kv.(*senderTransport).SendNext(0xc4332966e0, 0xc43074e780) /go/src/github.com/cockroachdb/cockroach/pkg/kv/transport.go:309 +0x1f8 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendToReplicas(0xc4319bdc20, 0x2b67bf5f1800, 0xc43889b780, 0x1dcd6500, 0xc42caebf80, 0xc4319bdc58, 0x1, 0xc43053f3e0, 0x1, 0x1, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1142 +0x35d github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendRPC(0xc4319bdc20, 0x2b67bf5f1800, 0xc43889b780, 0x1, 0xc43053f3e0, 0x1, 0x1, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:410 +0x372 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendSingleRange(0xc4319bdc20, 0x2b67bf5f1878, 0xc4392719e0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:479 +0x16e github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendPartialBatch(0xc4319bdc20, 0x2b67bf5f1878, 0xc4392719e0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:931 +0x2c7 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges(0xc4319bdc20, 0x2b67bf5f1878, 0xc4392719e0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:810 +0x41f github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send(0xc4319bdc20, 0x2b67bf5f1878, 0xc4392719e0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:625 +0x2d8 github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send(0xc42c62e820, 0x2b67bf5f1878, 0xc4392719e0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:418 +0x7e9 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).send(0xc432dc26e0, 0x2b67bf5f1878, 0xc4392719e0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:513 +0xbf github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Run.func1(0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:447 +0x6e github.com/cockroachdb/cockroach/pkg/internal/client.sendAndFill(0xc43053f2a0, 0xc431fd2600, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:418 +0x168 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Run(0xc432dc26e0, 0x2b67bf5f1878, 0xc4392719e0, 0xc431fd2600, 0x2625a01, 0xc439271a70) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:449 +0xac github.com/cockroachdb/cockroach/pkg/storage.(*intentResolver).resolveIntents.func1(0x19516a0, 0xc439271a70) /go/src/github.com/cockroachdb/cockroach/pkg/storage/intent_resolver.go:450 +0x4c github.com/cockroachdb/cockroach/pkg/storage.(*intentResolver).resolveIntents(0xc423226c40, 0x2b67bf5f1878, 0xc4392719e0, 0xc42a2dc7e0, 0x2, 0x2, 0x18a0001, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/intent_resolver.go:461 +0x5de github.com/cockroachdb/cockroach/pkg/storage.(*intentResolver).processIntents(0xc423226c40, 0x2b67bf4ed6f8, 0xc420094410, 0xc42c07ea00, 0x2625560, 0xc427c9d0e0, 0xc42a2dc7e0, 0x2, 0x2, 0x7b, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/intent_resolver.go:345 +0x449 github.com/cockroachdb/cockroach/pkg/storage.(*intentResolver).processIntentsAsync.func1(0x2b67bf4ed6f8, 0xc420094410) /go/src/github.com/cockroachdb/cockroach/pkg/storage/intent_resolver.go:277 +0x7f github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunLimitedAsyncTask.func1(0xc42dc082d0, 0x230c20d, 0x1a, 0x116, 0xc425098c00, 0x0, 0x0, 0xc43073a8a0, 0x2b67bf4ed6f8, 0xc420094410) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:326 +0x10c created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunLimitedAsyncTask /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:327 +0x3f0 goroutine 2351051 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Store).startGossip.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:1287 +0x39b github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42ac14690) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2392610 [semacquire]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc438767660, 0xc422481c80) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2350996 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc4386ce3d0, 0x18a0520, 0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/metric.(*Histogram).RecordValue(0xc426689080, 0x5999a) /go/src/github.com/cockroachdb/cockroach/pkg/util/metric/metric.go:217 +0x3a github.com/cockroachdb/cockroach/pkg/storage.newReplica.func4(0x5999a) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:603 +0x46 github.com/cockroachdb/cockroach/pkg/util/syncutil.ThresholdLogger.func1(0x5999a) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:53 +0x6c github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Unlock(0xc4232919f0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:97 +0xb6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tickRaftMuLocked(0xc423291900, 0x1c35d00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2720 +0x1d9 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tick(0xc423291900, 0xc429202c00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2673 +0x85 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0x12, 0xc42f981ef0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204c60) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351049 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Store).Start.func3() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:1201 +0x165 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42936d8a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2350997 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc4386ce280, 0x18a0520, 0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/metric.(*Histogram).RecordValue(0xc426689080, 0x2bc0) /go/src/github.com/cockroachdb/cockroach/pkg/util/metric/metric.go:217 +0x3a github.com/cockroachdb/cockroach/pkg/storage.newReplica.func4(0x2bc0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:603 +0x46 github.com/cockroachdb/cockroach/pkg/util/syncutil.ThresholdLogger.func1(0x2bc0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:53 +0x6c github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Unlock(0xc42fe30ff0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:97 +0xb6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tickRaftMuLocked(0xc42fe30f00, 0x1c35d00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2720 +0x1d9 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tick(0xc42fe30f00, 0xc429202c00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2673 +0x85 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0xf, 0xc43951bef0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204c80) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351012 [semacquire]: sync.runtime_Semacquire(0xc421bd0db4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*WaitGroup).Wait(0xc421bd0da8) /usr/local/go/src/sync/waitgroup.go:131 +0x97 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Wait(0xc421bd0d10) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:187 +0x34 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processRaft.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3357 +0x3a github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc4269a40d0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351007 [runnable]: sync.runtime_Semacquire(0xc4201522a4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4201522a0) /usr/local/go/src/sync/mutex.go:85 +0xd0 github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*lockOrder).PreLock(0xc4201522a0, 0x4, 0x18a0520, 0xc4269e8640) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:243 +0xc4 github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.PreLock(0x4, 0x18a0520, 0xc4269e8640) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:129 +0x4b github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc4386ce0e0, 0x18a0520, 0xc4269e8640) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:145 +0x93 github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4269e8640) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Lock(0xc431d869d8) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:86 +0x34 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0x9, 0xc431bdff00) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3332 +0x76 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204de0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351000 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc4386ce320, 0x18a0520, 0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/metric.(*Histogram).RecordValue(0xc426689080, 0x4b191) /go/src/github.com/cockroachdb/cockroach/pkg/util/metric/metric.go:217 +0x3a github.com/cockroachdb/cockroach/pkg/storage.newReplica.func4(0x4b191) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:603 +0x46 github.com/cockroachdb/cockroach/pkg/util/syncutil.ThresholdLogger.func1(0x4b191) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:53 +0x6c github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Unlock(0xc4321714f0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:97 +0xb6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tickRaftMuLocked(0xc432171400, 0x1c35d00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2720 +0x1d9 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tick(0xc432171400, 0xc429202c00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2673 +0x85 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0x2, 0xc439745f00) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204ce0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351050 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Store).startGossip.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:1287 +0x39b github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42ac14660) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351004 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc438766db0, 0x18a0520, 0xc4202c8560) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4202c8560) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/storage.TrackRaftProtos.func1(0x2619c00, 0xc434ba2280) /go/src/github.com/cockroachdb/cockroach/pkg/storage/track_raft_protos.go:52 +0xdb github.com/cockroachdb/cockroach/pkg/roachpb.(*Value).SetProto(0xc4392133f8, 0x2619c00, 0xc434ba2280, 0x10, 0x20) /go/src/github.com/cockroachdb/cockroach/pkg/roachpb/data.go:369 +0x118 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).append(0xc431ee0a00, 0x2b67bf5f1878, 0xc422437830, 0x2626560, 0xc43021a478, 0xb, 0x2a, 0xc434ba2280, 0x1, 0x1, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_raftstorage.go:504 +0x144 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).handleRaftReadyRaftMuLocked(0xc431ee0a00, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2561 +0x1028 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).handleRaftReady(0xc431ee0a00, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2453 +0xc5 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processReady(0xc431d86700, 0x1a) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3296 +0x137 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:230 +0x28c github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204d80) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351097 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:475 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42936dac0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351100 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:475 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42936db20) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2392651 [semacquire]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc4386ce370, 0xc4311707e0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2351006 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc4386ce2e0, 0x18a0520, 0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/metric.(*Histogram).RecordValue(0xc426689080, 0x4fb0) /go/src/github.com/cockroachdb/cockroach/pkg/util/metric/metric.go:217 +0x3a github.com/cockroachdb/cockroach/pkg/storage.newReplica.func4(0x4fb0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:603 +0x46 github.com/cockroachdb/cockroach/pkg/util/syncutil.ThresholdLogger.func1(0x4fb0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:53 +0x6c github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Unlock(0xc42c6680f0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:97 +0xb6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tickRaftMuLocked(0xc42c668000, 0x1c35d00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2723 +0x221 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tick(0xc42c668000, 0xc429202c00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2673 +0x85 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0x18, 0xc42edcbef0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204dc0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351099 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:475 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42936db00) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351424 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*idAllocator).start.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/id_alloc.go:133 +0x7c8 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc4268d2fc0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351101 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:475 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42936db40) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351001 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc4386ce310, 0x18a0520, 0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/metric.(*Histogram).RecordValue(0xc426689080, 0x2774) /go/src/github.com/cockroachdb/cockroach/pkg/util/metric/metric.go:217 +0x3a github.com/cockroachdb/cockroach/pkg/storage.newReplica.func4(0x2774) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:603 +0x46 github.com/cockroachdb/cockroach/pkg/util/syncutil.ThresholdLogger.func1(0x2774) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:53 +0x6c github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Unlock(0xc42fe30af0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:97 +0xb6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tickRaftMuLocked(0xc42fe30a00, 0x1c35d00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2720 +0x1d9 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tick(0xc42fe30a00, 0xc429202c00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2673 +0x85 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0xe, 0xc4397c9f00) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204d20) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2392441 [semacquire]: sync.runtime_Semacquire(0xc4202c8564) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4202c8560) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc438766db0, 0xc430616e40) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2392657 [semacquire]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc4386ce3c0, 0xc431170d80) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2351102 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:475 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc42936db60) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351010 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc438767660, 0x18a0520, 0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/metric.(*Histogram).RecordValue(0xc426689080, 0x2bc0) /go/src/github.com/cockroachdb/cockroach/pkg/util/metric/metric.go:217 +0x3a github.com/cockroachdb/cockroach/pkg/storage.newReplica.func4(0x2bc0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:603 +0x46 github.com/cockroachdb/cockroach/pkg/util/syncutil.ThresholdLogger.func1(0x2bc0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:53 +0x6c github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Unlock(0xc42ffbaaf0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:97 +0xb6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tickRaftMuLocked(0xc42ffbaa00, 0x1c35d00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2720 +0x1d9 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tick(0xc42ffbaa00, 0xc429202c00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2673 +0x85 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0x5, 0xc430ffff00) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204e40) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2392640 [semacquire]: sync.runtime_Semacquire(0xc4266890c4) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*Mutex).Lock(0xc4266890c0) /usr/local/go/src/sync/mutex.go:85 +0xd0 sync.(*Mutex).Lock-fm() /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x2a github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock.func1(0xc4386ce2e0, 0xc430e73e00) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:151 +0x27 created by github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:153 +0xf3 goroutine 2351013 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Store).raftTickLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3383 +0x334 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc4269a4140) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351014 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Store).startCoalescedHeartbeatsLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3414 +0x167 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc4269a4150) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351003 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc4386ce220, 0x18a0520, 0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/metric.(*Histogram).RecordValue(0xc426689080, 0x283c) /go/src/github.com/cockroachdb/cockroach/pkg/util/metric/metric.go:217 +0x3a github.com/cockroachdb/cockroach/pkg/storage.newReplica.func4(0x283c) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:603 +0x46 github.com/cockroachdb/cockroach/pkg/util/syncutil.ThresholdLogger.func1(0x283c) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:53 +0x6c github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Unlock(0xc427e76af0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:97 +0xb6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tickRaftMuLocked(0xc427e76a00, 0x1c35d00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2720 +0x1d9 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tick(0xc427e76a00, 0xc429202c00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2673 +0x85 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0xa, 0xc432f9ff00) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204d60) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351009 [semacquire]: sync.runtime_notifyListWait(0xc426e02390, 0xc400000186) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc426e02380) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204e20) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 2351002 [select]: github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.lock(0xc4386ce370, 0x18a0520, 0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:157 +0xc5d github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock.(*Mutex).Lock(0xc4266890c0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/sasha-s/go-deadlock/deadlock.go:55 +0x78 github.com/cockroachdb/cockroach/pkg/util/metric.(*Histogram).RecordValue(0xc426689080, 0x28a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/metric/metric.go:217 +0x3a github.com/cockroachdb/cockroach/pkg/storage.newReplica.func4(0x28a0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:603 +0x46 github.com/cockroachdb/cockroach/pkg/util/syncutil.ThresholdLogger.func1(0x28a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:53 +0x6c github.com/cockroachdb/cockroach/pkg/util/syncutil.(*TimedMutex).Unlock(0xc436e3aaf0) /go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/timedmutex.go:97 +0xb6 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tickRaftMuLocked(0xc436e3aa00, 0x1c35d00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2720 +0x1d9 github.com/cockroachdb/cockroach/pkg/storage.(*Replica).tick(0xc436e3aa00, 0xc429202c00, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/replica.go:2673 +0x85 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processTick(0xc431d86700, 0x11, 0xc4322a5f00) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3339 +0xfb github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc421bd0d10, 0xc42dc082d0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:225 +0x2b6 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc42dc082d0, 0xc423204d40) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 ERROR: exit status 2 make: *** [stress] Error 1 1 runs completed, 1 failures, over 1m1s Makefile:138: recipe for target 'stress' failed ```
test
github com cockroachdb cockroach pkg storage teststorerangesystemsplits failed under stress sha parameters cockroach proposer evaluated kv false tags deadlock goflags stress build found a failed test github com cockroachdb cockroach pkg storage store processtick go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock mutex lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach pkg util metric histogram recordvalue go src github com cockroachdb cockroach pkg util metric metric go github com cockroachdb cockroach pkg storage newreplica go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg util syncutil thresholdlogger go src github com cockroachdb cockroach pkg util syncutil timedmutex go github com cockroachdb cockroach pkg util syncutil timedmutex unlock go src github com cockroachdb cockroach pkg util syncutil timedmutex go github com cockroachdb cockroach pkg storage replica tickraftmulocked go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage replica tick go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage store processtick go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime semacquire usr local go src runtime sema go sync mutex lock usr local go src sync mutex go sync mutex lock fm go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go created by github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go goroutine goroutine running on other thread stack unavailable created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage engine rocksdb open go src github com cockroachdb cockroach pkg storage engine rocksdb go created by github com cockroachdb cockroach pkg storage engine rocksdb open go src github com cockroachdb cockroach pkg storage engine rocksdb go goroutine github com cockroachdb cockroach pkg storage basequeue maybeaddtopurgatory go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime semacquire usr local go src runtime sema go sync mutex lock usr local go src sync mutex go sync mutex lock fm go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go created by github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go goroutine sync runtime semacquire usr local go src runtime sema go sync mutex lock usr local go src sync mutex go sync mutex lock fm go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go created by github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go goroutine sync runtime semacquire usr local go src runtime sema go sync mutex lock usr local go src sync mutex go sync mutex lock fm go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go created by github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go goroutine sync runtime semacquire usr local go src runtime sema go sync mutex lock usr local go src sync mutex go sync mutex lock fm go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go created by github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go goroutine github com cockroachdb cockroach pkg kv txncoordsender printstatsloop go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg kv newtxncoordsender go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock mutex lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach pkg util metric histogram recordvalue go src github com cockroachdb cockroach pkg util metric metric go github com cockroachdb cockroach pkg storage newreplica go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg util syncutil thresholdlogger go src github com cockroachdb cockroach pkg util syncutil timedmutex go github com cockroachdb cockroach pkg util syncutil timedmutex unlock go src github com cockroachdb cockroach pkg util syncutil timedmutex go github com cockroachdb cockroach pkg storage replica tickraftmulocked go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage replica tick go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage store processtick go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime semacquire usr local go src runtime sema go sync mutex lock usr local go src sync mutex go sync mutex lock fm go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go created by github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go goroutine github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg rpc newcontext go src github com cockroachdb cockroach pkg rpc context go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime semacquire usr local go src runtime sema go sync mutex lock usr local go src sync mutex go sync mutex lock fm go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go created by github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go goroutine github com cockroachdb cockroach pkg storage replica tryaddwritecmd go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage replica addwritecmd go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage replica send go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage store send go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg storage stores send go src github com cockroachdb cockroach pkg storage stores go github com cockroachdb cockroach pkg kv sendertransport sendnext go src github com cockroachdb cockroach pkg kv transport go github com cockroachdb cockroach pkg kv distsender sendtoreplicas go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendrpc go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendsinglerange go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendpartialbatch go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender divideandsendbatchtoranges go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender send go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv txncoordsender send go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg internal client txn sendinternal go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client txn send go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client txn github com cockroachdb cockroach pkg internal client send fm go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client sendandfill go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client txn run go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg storage replica adminsplitwithdescriptor go src github com cockroachdb cockroach pkg storage replica command go github com cockroachdb cockroach pkg internal client db txn go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client txn exec go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client db txn go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg storage replica adminsplitwithdescriptor go src github com cockroachdb cockroach pkg storage replica command go github com cockroachdb cockroach pkg storage splitqueue process go src github com cockroachdb cockroach pkg storage split queue go github com cockroachdb cockroach pkg storage basequeue processreplica go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg storage basequeue processloop go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runtask go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg storage basequeue processloop go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage replica begincmds go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage replica addreadonlycmd go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage replica send go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage store send go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg storage stores send go src github com cockroachdb cockroach pkg storage stores go github com cockroachdb cockroach pkg kv sendertransport sendnext go src github com cockroachdb cockroach pkg kv transport go github com cockroachdb cockroach pkg kv distsender sendtoreplicas go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendrpc go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendsinglerange go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendpartialbatch go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender divideandsendbatchtoranges go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender send go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv txncoordsender send go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg internal client db send go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client db run go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client sendandfill go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client db run go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client db scan go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client db scan go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg storage test teststorerangesystemsplits go src github com cockroachdb cockroach pkg storage client split test go github com cockroachdb cockroach pkg util retryforduration go src github com cockroachdb cockroach pkg util retry go github com cockroachdb cockroach pkg testutils succeedssoondepth go src github com cockroachdb cockroach pkg testutils soon go github com cockroachdb cockroach pkg storage test teststorerangesystemsplits go src github com cockroachdb cockroach pkg storage client split test go github com cockroachdb cockroach pkg storage test teststorerangesystemsplits go src github com cockroachdb cockroach pkg storage client split test go testing trunner usr local go src testing testing go created by testing t run usr local go src testing testing go goroutine github com cockroachdb cockroach pkg storage replica tryaddwritecmd go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage replica addwritecmd go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage replica send go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage store send go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg storage stores send go src github com cockroachdb cockroach pkg storage stores go github com cockroachdb cockroach pkg kv sendertransport sendnext go src github com cockroachdb cockroach pkg kv transport go github com cockroachdb cockroach pkg kv distsender sendtoreplicas go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendrpc go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendsinglerange go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendpartialbatch go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender divideandsendbatchtoranges go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender send go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv txncoordsender send go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg internal client db send go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client db run go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client sendandfill go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client db run go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg storage intentresolver resolveintents go src github com cockroachdb cockroach pkg storage intent resolver go github com cockroachdb cockroach pkg storage intentresolver resolveintents go src github com cockroachdb cockroach pkg storage intent resolver go github com cockroachdb cockroach pkg storage intentresolver processintents go src github com cockroachdb cockroach pkg storage intent resolver go github com cockroachdb cockroach pkg storage intentresolver processintentsasync go src github com cockroachdb cockroach pkg storage intent resolver go github com cockroachdb cockroach pkg util stop stopper runlimitedasynctask go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runlimitedasynctask go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage store startgossip go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime semacquire usr local go src runtime sema go sync mutex lock usr local go src sync mutex go sync mutex lock fm go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go created by github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go goroutine github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock mutex lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach pkg util metric histogram recordvalue go src github com cockroachdb cockroach pkg util metric metric go github com cockroachdb cockroach pkg storage newreplica go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg util syncutil thresholdlogger go src github com cockroachdb cockroach pkg util syncutil timedmutex go github com cockroachdb cockroach pkg util syncutil timedmutex unlock go src github com cockroachdb cockroach pkg util syncutil timedmutex go github com cockroachdb cockroach pkg storage replica tickraftmulocked go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage replica tick go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage store processtick go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage store start go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock mutex lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach pkg util metric histogram recordvalue go src github com cockroachdb cockroach pkg util metric metric go github com cockroachdb cockroach pkg storage newreplica go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg util syncutil thresholdlogger go src github com cockroachdb cockroach pkg util syncutil timedmutex go github com cockroachdb cockroach pkg util syncutil timedmutex unlock go src github com cockroachdb cockroach pkg util syncutil timedmutex go github com cockroachdb cockroach pkg storage replica tickraftmulocked go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage replica tick go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage store processtick go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime semacquire usr local go src runtime sema go sync waitgroup wait usr local go src sync waitgroup go github com cockroachdb cockroach pkg storage raftscheduler wait go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage store processraft go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime semacquire usr local go src runtime sema go sync mutex lock usr local go src sync mutex go github com cockroachdb cockroach vendor github com sasha s go deadlock lockorder prelock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock prelock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock mutex lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach pkg util syncutil timedmutex lock go src github com cockroachdb cockroach pkg util syncutil timedmutex go github com cockroachdb cockroach pkg storage store processtick go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock mutex lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach pkg util metric histogram recordvalue go src github com cockroachdb cockroach pkg util metric metric go github com cockroachdb cockroach pkg storage newreplica go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg util syncutil thresholdlogger go src github com cockroachdb cockroach pkg util syncutil timedmutex go github com cockroachdb cockroach pkg util syncutil timedmutex unlock go src github com cockroachdb cockroach pkg util syncutil timedmutex go github com cockroachdb cockroach pkg storage replica tickraftmulocked go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage replica tick go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage store processtick go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage store startgossip go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock mutex lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach pkg storage trackraftprotos go src github com cockroachdb cockroach pkg storage track raft protos go github com cockroachdb cockroach pkg roachpb value setproto go src github com cockroachdb cockroach pkg roachpb data go github com cockroachdb cockroach pkg storage replica append go src github com cockroachdb cockroach pkg storage replica raftstorage go github com cockroachdb cockroach pkg storage replica handleraftreadyraftmulocked go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage replica handleraftready go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage store processready go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage basequeue processloop go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage basequeue processloop go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime semacquire usr local go src runtime sema go sync mutex lock usr local go src sync mutex go sync mutex lock fm go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go created by github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go goroutine github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock mutex lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach pkg util metric histogram recordvalue go src github com cockroachdb cockroach pkg util metric metric go github com cockroachdb cockroach pkg storage newreplica go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg util syncutil thresholdlogger go src github com cockroachdb cockroach pkg util syncutil timedmutex go github com cockroachdb cockroach pkg util syncutil timedmutex unlock go src github com cockroachdb cockroach pkg util syncutil timedmutex go github com cockroachdb cockroach pkg storage replica tickraftmulocked go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage replica tick go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage store processtick go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage basequeue processloop go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage idallocator start go src github com cockroachdb cockroach pkg storage id alloc go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage basequeue processloop go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock mutex lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach pkg util metric histogram recordvalue go src github com cockroachdb cockroach pkg util metric metric go github com cockroachdb cockroach pkg storage newreplica go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg util syncutil thresholdlogger go src github com cockroachdb cockroach pkg util syncutil timedmutex go github com cockroachdb cockroach pkg util syncutil timedmutex unlock go src github com cockroachdb cockroach pkg util syncutil timedmutex go github com cockroachdb cockroach pkg storage replica tickraftmulocked go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage replica tick go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage store processtick go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime semacquire usr local go src runtime sema go sync mutex lock usr local go src sync mutex go sync mutex lock fm go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go created by github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go goroutine sync runtime semacquire usr local go src runtime sema go sync mutex lock usr local go src sync mutex go sync mutex lock fm go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go created by github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go goroutine github com cockroachdb cockroach pkg storage basequeue processloop go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock mutex lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach pkg util metric histogram recordvalue go src github com cockroachdb cockroach pkg util metric metric go github com cockroachdb cockroach pkg storage newreplica go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg util syncutil thresholdlogger go src github com cockroachdb cockroach pkg util syncutil timedmutex go github com cockroachdb cockroach pkg util syncutil timedmutex unlock go src github com cockroachdb cockroach pkg util syncutil timedmutex go github com cockroachdb cockroach pkg storage replica tickraftmulocked go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage replica tick go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage store processtick go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime semacquire usr local go src runtime sema go sync mutex lock usr local go src sync mutex go sync mutex lock fm go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go created by github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go goroutine github com cockroachdb cockroach pkg storage store rafttickloop go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage store startcoalescedheartbeatsloop go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock mutex lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach pkg util metric histogram recordvalue go src github com cockroachdb cockroach pkg util metric metric go github com cockroachdb cockroach pkg storage newreplica go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg util syncutil thresholdlogger go src github com cockroachdb cockroach pkg util syncutil timedmutex go github com cockroachdb cockroach pkg util syncutil timedmutex unlock go src github com cockroachdb cockroach pkg util syncutil timedmutex go github com cockroachdb cockroach pkg storage replica tickraftmulocked go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage replica tick go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage store processtick go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach vendor github com sasha s go deadlock lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach vendor github com sasha s go deadlock mutex lock go src github com cockroachdb cockroach vendor github com sasha s go deadlock deadlock go github com cockroachdb cockroach pkg util metric histogram recordvalue go src github com cockroachdb cockroach pkg util metric metric go github com cockroachdb cockroach pkg storage newreplica go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg util syncutil thresholdlogger go src github com cockroachdb cockroach pkg util syncutil timedmutex go github com cockroachdb cockroach pkg util syncutil timedmutex unlock go src github com cockroachdb cockroach pkg util syncutil timedmutex go github com cockroachdb cockroach pkg storage replica tickraftmulocked go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage replica tick go src github com cockroachdb cockroach pkg storage replica go github com cockroachdb cockroach pkg storage store processtick go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go error exit status make error runs completed failures over makefile recipe for target stress failed
1
148,930
11,872,194,186
IssuesEvent
2020-03-26 15:27:56
NetsBlox/NetsBlox
https://api.github.com/repos/NetsBlox/NetsBlox
closed
XML loading tests need to be updated
bug testing
They currently rely on a function that has been changed on the client during a recent refactor.
1.0
XML loading tests need to be updated - They currently rely on a function that has been changed on the client during a recent refactor.
test
xml loading tests need to be updated they currently rely on a function that has been changed on the client during a recent refactor
1
180,282
13,928,864,768
IssuesEvent
2020-10-21 22:18:24
DuongTyler/haachama-coco
https://api.github.com/repos/DuongTyler/haachama-coco
opened
Database Backend Integration Tests
Tests backend good first issue
Integration tests from the API to the database to make sure the backend system works as we expect it to.
1.0
Database Backend Integration Tests - Integration tests from the API to the database to make sure the backend system works as we expect it to.
test
database backend integration tests integration tests from the api to the database to make sure the backend system works as we expect it to
1
217,480
16,855,783,539
IssuesEvent
2021-06-21 06:25:44
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
raftstore::test_conf_change::test_server_transfer_leader_safe failed
component/test-bench
raftstore::test_conf_change::test_server_transfer_leader_safe Latest failed builds: https://internal.pingcap.net/idc-jenkins/job/tikv_ghpr_test/20334/consoleFull
1.0
raftstore::test_conf_change::test_server_transfer_leader_safe failed - raftstore::test_conf_change::test_server_transfer_leader_safe Latest failed builds: https://internal.pingcap.net/idc-jenkins/job/tikv_ghpr_test/20334/consoleFull
test
raftstore test conf change test server transfer leader safe failed raftstore test conf change test server transfer leader safe latest failed builds
1
70,545
7,190,651,147
IssuesEvent
2018-02-02 18:01:27
openshift/origin
https://api.github.com/repos/openshift/origin
closed
TASK [Check latest available OpenShift RPM version] Could not parse metalink
area/tests kind/test-flake priority/P3
``` TASK [Check latest available OpenShift RPM version] **************************** task path: /usr/share/ansible/openshift-ansible/playbooks/common/openshift-cluster/upgrades/pre/verify_upgrade_targets.yml:16 fatal: [localhost]: FAILED! => { "changed": false, "failed": true, "generated_timestamp": "2017-07-04 16:15:55.100249", "msg": { "cmd": "/usr/bin/repoquery --plugins --quiet --pkgnarrow=repos --queryformat=%{version}|%{release}|%{arch}|%{repo}|%{version}-%{release} --config=/tmp/tmp4WjReN origin", "package_found": false, "results": {}, "returncode": 1, "stderr": "Repository centos-openshift-origin is listed more than once in the configuration\nRepository centos-openshift-origin-testing is listed more than once in the configuration\nRepository centos-openshift-origin-debuginfo is listed more than once in the configuration\nRepository centos-openshift-origin-source is listed more than once in the configuration\nFile /var/cache/yum/x86_64/7Server/epel/metalink.xml does not exist\n", "stdout": "Could not parse metalink https://mirrors.fedoraproject.org/metalink?repo=epel-7&arch=x86_64 error was \nNo timestamp for file\n" } } [WARNING]: Could not create retry file '/usr/share/ansible/openshift- ansible/playbooks/byo/openshift-cluster/upgrades/v3_6/upgrade.retry'. [Errno 13] Permission denied: u'/usr/share/ansible/openshift- ansible/playbooks/byo/openshift-cluster/upgrades/v3_6/upgrade.retry' PLAY RECAP ********************************************************************* localhost : ok=127 changed=5 unreachable=0 failed=1 ++ export status=FAILURE ++ status=FAILURE + set +o xtrace ########## FINISHED STAGE: FAILURE: UPDATE ORIGIN TO RELEASE [00h 01m 15s] ########## ``` https://ci.openshift.redhat.com/jenkins/job/merge_pull_request_origin/1227/consoleFull#32866359658b6e51eb7608a5981914356 @stevekuznetsov @sdodson
2.0
TASK [Check latest available OpenShift RPM version] Could not parse metalink - ``` TASK [Check latest available OpenShift RPM version] **************************** task path: /usr/share/ansible/openshift-ansible/playbooks/common/openshift-cluster/upgrades/pre/verify_upgrade_targets.yml:16 fatal: [localhost]: FAILED! => { "changed": false, "failed": true, "generated_timestamp": "2017-07-04 16:15:55.100249", "msg": { "cmd": "/usr/bin/repoquery --plugins --quiet --pkgnarrow=repos --queryformat=%{version}|%{release}|%{arch}|%{repo}|%{version}-%{release} --config=/tmp/tmp4WjReN origin", "package_found": false, "results": {}, "returncode": 1, "stderr": "Repository centos-openshift-origin is listed more than once in the configuration\nRepository centos-openshift-origin-testing is listed more than once in the configuration\nRepository centos-openshift-origin-debuginfo is listed more than once in the configuration\nRepository centos-openshift-origin-source is listed more than once in the configuration\nFile /var/cache/yum/x86_64/7Server/epel/metalink.xml does not exist\n", "stdout": "Could not parse metalink https://mirrors.fedoraproject.org/metalink?repo=epel-7&arch=x86_64 error was \nNo timestamp for file\n" } } [WARNING]: Could not create retry file '/usr/share/ansible/openshift- ansible/playbooks/byo/openshift-cluster/upgrades/v3_6/upgrade.retry'. [Errno 13] Permission denied: u'/usr/share/ansible/openshift- ansible/playbooks/byo/openshift-cluster/upgrades/v3_6/upgrade.retry' PLAY RECAP ********************************************************************* localhost : ok=127 changed=5 unreachable=0 failed=1 ++ export status=FAILURE ++ status=FAILURE + set +o xtrace ########## FINISHED STAGE: FAILURE: UPDATE ORIGIN TO RELEASE [00h 01m 15s] ########## ``` https://ci.openshift.redhat.com/jenkins/job/merge_pull_request_origin/1227/consoleFull#32866359658b6e51eb7608a5981914356 @stevekuznetsov @sdodson
test
task could not parse metalink task task path usr share ansible openshift ansible playbooks common openshift cluster upgrades pre verify upgrade targets yml fatal failed changed false failed true generated timestamp msg cmd usr bin repoquery plugins quiet pkgnarrow repos queryformat version release arch repo version release config tmp origin package found false results returncode stderr repository centos openshift origin is listed more than once in the configuration nrepository centos openshift origin testing is listed more than once in the configuration nrepository centos openshift origin debuginfo is listed more than once in the configuration nrepository centos openshift origin source is listed more than once in the configuration nfile var cache yum epel metalink xml does not exist n stdout could not parse metalink error was nno timestamp for file n could not create retry file usr share ansible openshift ansible playbooks byo openshift cluster upgrades upgrade retry permission denied u usr share ansible openshift ansible playbooks byo openshift cluster upgrades upgrade retry play recap localhost ok changed unreachable failed export status failure status failure set o xtrace finished stage failure update origin to release stevekuznetsov sdodson
1
85,527
7,975,679,507
IssuesEvent
2018-07-17 10:05:19
frappe/erpnext
https://api.github.com/repos/frappe/erpnext
closed
[v11][Testing] Reports
testing
- [ ] Customisable Cash Flow Reports - [ ] Employee Advance Summary - [ ] Presentation currency in General Ledger - [ ] Exploded view for BOM Stock Report - [ ] General Ledger - Group by party and Filter based on multiple party - [ ] Consolidated Financial Statements
1.0
[v11][Testing] Reports - - [ ] Customisable Cash Flow Reports - [ ] Employee Advance Summary - [ ] Presentation currency in General Ledger - [ ] Exploded view for BOM Stock Report - [ ] General Ledger - Group by party and Filter based on multiple party - [ ] Consolidated Financial Statements
test
reports customisable cash flow reports employee advance summary presentation currency in general ledger exploded view for bom stock report general ledger group by party and filter based on multiple party consolidated financial statements
1
803,646
29,186,187,440
IssuesEvent
2023-05-19 15:37:19
harvard-lil/perma
https://api.github.com/repos/harvard-lil/perma
closed
Use webpack to generate maintenance.js for Cloudflare service worker?
enhancement devops priority-tbd
See https://blog.cloudflare.com/introducing-cloudflare-workers/ This should include base64-encoded fonts. Sample output is attached at [maintenance.js.txt](https://github.com/harvard-lil/perma/files/2068604/maintenance.js.txt).
1.0
Use webpack to generate maintenance.js for Cloudflare service worker? - See https://blog.cloudflare.com/introducing-cloudflare-workers/ This should include base64-encoded fonts. Sample output is attached at [maintenance.js.txt](https://github.com/harvard-lil/perma/files/2068604/maintenance.js.txt).
non_test
use webpack to generate maintenance js for cloudflare service worker see this should include encoded fonts sample output is attached at
0
260,856
8,216,025,913
IssuesEvent
2018-09-05 07:59:07
geosolutions-it/smb-portal
https://api.github.com/repos/geosolutions-it/smb-portal
closed
Tracks segments geometries as GeoJSON
Priority: High pending review review
For the segments elements inside the track response, geometry (and only geometry) shoould be returned as GeoJSON instead of WKT. It will simplify the work on the app side...
1.0
Tracks segments geometries as GeoJSON - For the segments elements inside the track response, geometry (and only geometry) shoould be returned as GeoJSON instead of WKT. It will simplify the work on the app side...
non_test
tracks segments geometries as geojson for the segments elements inside the track response geometry and only geometry shoould be returned as geojson instead of wkt it will simplify the work on the app side
0
46,683
5,826,757,070
IssuesEvent
2017-05-08 06:33:38
myzenon/J3B
https://api.github.com/repos/myzenon/J3B
closed
Update PlantTest Class
Unit Test
Update PlantTest Class follows the Plant Class Implementation Task: GH-72
1.0
Update PlantTest Class - Update PlantTest Class follows the Plant Class Implementation Task: GH-72
test
update planttest class update planttest class follows the plant class implementation task gh
1
769,371
27,003,421,173
IssuesEvent
2023-02-10 09:44:53
feast-dev/feast
https://api.github.com/repos/feast-dev/feast
opened
Postgresql support for the Registry
kind/bug priority/p2
## Expected Behavior Postgresql as a storage for the Registry ## Current Behavior feast-feature-server crash with the error ``` Exception in thread "main" com.google.inject.ProvisionException: Unable to provision, see the following errors: 1) [Guice/ErrorInCustomProvider]: RuntimeException: Registry storage postgresql is unsupported at RegistryConfigModule.registryFile(RegistryConfigModule.java:52) at RegistryConfigModule.registryRepository(RegistryConfigModule.java:72) \_ for 1st parameter at RegistryConfigModule.registryRepository(RegistryConfigModule.java:72) at ServingServiceV2Module.registryBasedServingServiceV2(ServingServiceV2Module.java:40) \_ for 2nd parameter at ServingServiceV2Module.registryBasedServingServiceV2(ServingServiceV2Module.java:40) at OnlineServingGrpcServiceV2.<init>(OnlineServingGrpcServiceV2.java:33) \_ for 1st parameter at ServerModule.provideGrpcServer(ServerModule.java:43) \_ for 2nd parameter at ServerModule.provideGrpcServer(ServerModule.java:43) while locating Server Learn more: https://github.com/google/guice/wiki/ERROR_IN_CUSTOM_PROVIDER ``` ## Steps to reproduce Try to deploy Feast via Helm from the feast-charts/feast with the next values ``` global: registry: registry_type: sql path: "postgresql://postgres:password@feast-aurora.cluster-**.rds.amazonaws.com/postgres" project: feast_demo feature-server: application-override.yaml: enabled: true feast: active_store: online stores: - name: online type: REDIS config: host: feast-elasticache.***.cache.amazonaws.com port: 6379 - name: offline type: postgres config: host: feast-aurora.cluster-**.rds.amazonaws.com port: 5432 database: postgres db_schema: public user: postgresql password: password entityKeySerializationVersion: 2 ``` ### Specifications - Version: [Helm Chart v0.29](https://github.com/feast-dev/feast/tree/master/infra/charts/feast) - Platform: AWS - Subsystem: EKS ## Possible Solution As I understand, according to the source code of the registry.py, postgres as a storage for the Registry can't work. https://github.com/feast-dev/feast/blob/2f7c4ede8f9e66703714261f1152f78526d4bf43/sdk/python/feast/infra/registry/registry.py#L57-L70
1.0
Postgresql support for the Registry - ## Expected Behavior Postgresql as a storage for the Registry ## Current Behavior feast-feature-server crash with the error ``` Exception in thread "main" com.google.inject.ProvisionException: Unable to provision, see the following errors: 1) [Guice/ErrorInCustomProvider]: RuntimeException: Registry storage postgresql is unsupported at RegistryConfigModule.registryFile(RegistryConfigModule.java:52) at RegistryConfigModule.registryRepository(RegistryConfigModule.java:72) \_ for 1st parameter at RegistryConfigModule.registryRepository(RegistryConfigModule.java:72) at ServingServiceV2Module.registryBasedServingServiceV2(ServingServiceV2Module.java:40) \_ for 2nd parameter at ServingServiceV2Module.registryBasedServingServiceV2(ServingServiceV2Module.java:40) at OnlineServingGrpcServiceV2.<init>(OnlineServingGrpcServiceV2.java:33) \_ for 1st parameter at ServerModule.provideGrpcServer(ServerModule.java:43) \_ for 2nd parameter at ServerModule.provideGrpcServer(ServerModule.java:43) while locating Server Learn more: https://github.com/google/guice/wiki/ERROR_IN_CUSTOM_PROVIDER ``` ## Steps to reproduce Try to deploy Feast via Helm from the feast-charts/feast with the next values ``` global: registry: registry_type: sql path: "postgresql://postgres:password@feast-aurora.cluster-**.rds.amazonaws.com/postgres" project: feast_demo feature-server: application-override.yaml: enabled: true feast: active_store: online stores: - name: online type: REDIS config: host: feast-elasticache.***.cache.amazonaws.com port: 6379 - name: offline type: postgres config: host: feast-aurora.cluster-**.rds.amazonaws.com port: 5432 database: postgres db_schema: public user: postgresql password: password entityKeySerializationVersion: 2 ``` ### Specifications - Version: [Helm Chart v0.29](https://github.com/feast-dev/feast/tree/master/infra/charts/feast) - Platform: AWS - Subsystem: EKS ## Possible Solution As I understand, according to the source code of the registry.py, postgres as a storage for the Registry can't work. https://github.com/feast-dev/feast/blob/2f7c4ede8f9e66703714261f1152f78526d4bf43/sdk/python/feast/infra/registry/registry.py#L57-L70
non_test
postgresql support for the registry expected behavior postgresql as a storage for the registry current behavior feast feature server crash with the error exception in thread main com google inject provisionexception unable to provision see the following errors runtimeexception registry storage postgresql is unsupported at registryconfigmodule registryfile registryconfigmodule java at registryconfigmodule registryrepository registryconfigmodule java for parameter at registryconfigmodule registryrepository registryconfigmodule java at java for parameter at java at java for parameter at servermodule providegrpcserver servermodule java for parameter at servermodule providegrpcserver servermodule java while locating server learn more steps to reproduce try to deploy feast via helm from the feast charts feast with the next values global registry registry type sql path postgresql postgres password feast aurora cluster rds amazonaws com postgres project feast demo feature server application override yaml enabled true feast active store online stores name online type redis config host feast elasticache cache amazonaws com port name offline type postgres config host feast aurora cluster rds amazonaws com port database postgres db schema public user postgresql password password entitykeyserializationversion specifications version platform aws subsystem eks possible solution as i understand according to the source code of the registry py postgres as a storage for the registry can t work
0