Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 4 112 | repo_url stringlengths 33 141 | action stringclasses 3 values | title stringlengths 1 1.02k | labels stringlengths 4 1.54k | body stringlengths 1 262k | index stringclasses 17 values | text_combine stringlengths 95 262k | label stringclasses 2 values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
77,222 | 9,984,904,159 | IssuesEvent | 2019-07-10 15:24:29 | publiclab/leaflet-environmental-layers | https://api.github.com/repos/publiclab/leaflet-environmental-layers | closed | Add instructions regarding testing to readme.md | documentation | ## Description
Recently Jasmine was set up in LEL and we are writing tests for the different layers.
We need to document this process atleast briefly for now in
Readme.md and contributing.md files. | 1.0 | Add instructions regarding testing to readme.md - ## Description
Recently Jasmine was set up in LEL and we are writing tests for the different layers.
We need to document this process atleast briefly for now in
Readme.md and contributing.md files. | non_test | add instructions regarding testing to readme md description recently jasmine was set up in lel and we are writing tests for the different layers we need to document this process atleast briefly for now in readme md and contributing md files | 0 |
291,800 | 25,176,204,053 | IssuesEvent | 2022-11-11 09:28:51 | cosmos/interchain-security | https://api.github.com/repos/cosmos/interchain-security | closed | Add MaxValidators constraint to testnet plan | testing | I think we should make sure that the testnet plan(s) include scenarios where the system param `params.MaxValidators` is _less_ than the actual number of validators in the system. | 1.0 | Add MaxValidators constraint to testnet plan - I think we should make sure that the testnet plan(s) include scenarios where the system param `params.MaxValidators` is _less_ than the actual number of validators in the system. | test | add maxvalidators constraint to testnet plan i think we should make sure that the testnet plan s include scenarios where the system param params maxvalidators is less than the actual number of validators in the system | 1 |
148,520 | 11,854,887,236 | IssuesEvent | 2020-03-25 02:21:25 | microsoft/azure-tools-for-java | https://api.github.com/repos/microsoft/azure-tools-for-java | closed | [IntelliJ][ESPHIB] Gen2 root path could not be edit after saving with invalid value | HDInsight IntelliJ Internal Test fixed | Build: azure-toolkit-for-intellij-2019.2.4.develop.11606681.02-29-2020
Repro Steps:
1. Link one esphib cluster
2. Create one HDI config, select the linked esphib cluster
3. In upload field, input one invalid root path, save config
4. Re-open the config, try to edit the root path
Result:
Could not edit the root path

| 1.0 | [IntelliJ][ESPHIB] Gen2 root path could not be edit after saving with invalid value - Build: azure-toolkit-for-intellij-2019.2.4.develop.11606681.02-29-2020
Repro Steps:
1. Link one esphib cluster
2. Create one HDI config, select the linked esphib cluster
3. In upload field, input one invalid root path, save config
4. Re-open the config, try to edit the root path
Result:
Could not edit the root path

| test | root path could not be edit after saving with invalid value build azure toolkit for intellij develop repro steps link one esphib cluster create one hdi config select the linked esphib cluster in upload field input one invalid root path save config re open the config try to edit the root path result could not edit the root path | 1 |
134,191 | 19,097,108,754 | IssuesEvent | 2021-11-29 17:50:57 | plotn/coolreader | https://api.github.com/repos/plotn/coolreader | closed | действия по короткому и длинному тапам идут одним размером и начертанием шрифта, что сильно ухудшает восприятие информации. Возможно ли основное действие сделать жирным, а по длинному тапу - курсивом | enhancement design | Это в кнопках панели | 1.0 | действия по короткому и длинному тапам идут одним размером и начертанием шрифта, что сильно ухудшает восприятие информации. Возможно ли основное действие сделать жирным, а по длинному тапу - курсивом - Это в кнопках панели | non_test | действия по короткому и длинному тапам идут одним размером и начертанием шрифта что сильно ухудшает восприятие информации возможно ли основное действие сделать жирным а по длинному тапу курсивом это в кнопках панели | 0 |
32,749 | 13,924,882,103 | IssuesEvent | 2020-10-21 16:06:17 | cityofaustin/atd-data-tech | https://api.github.com/repos/cityofaustin/atd-data-tech | closed | Small tweaks to Scheduler interface | Need: 3-Could Have Product: Vision Zero in Action Service: Apps Workgroup: AMD Workgroup: VZ | Could-haves for the future:

- [ ] Change the date range calendar heading to use the full month names and add a comma after the date range. E.g. "July 6 — 12, 2020" instead of "Jul 6 — 12 2020"
- [ ] Hide hours where there are never VZA shifts: 12-6am and 10-12pm (I think? Or does Lewis occasionally schedule shifts during that time? | 1.0 | Small tweaks to Scheduler interface - Could-haves for the future:

- [ ] Change the date range calendar heading to use the full month names and add a comma after the date range. E.g. "July 6 — 12, 2020" instead of "Jul 6 — 12 2020"
- [ ] Hide hours where there are never VZA shifts: 12-6am and 10-12pm (I think? Or does Lewis occasionally schedule shifts during that time? | non_test | small tweaks to scheduler interface could haves for the future change the date range calendar heading to use the full month names and add a comma after the date range e g july — instead of jul — hide hours where there are never vza shifts and i think or does lewis occasionally schedule shifts during that time | 0 |
167,328 | 13,020,697,638 | IssuesEvent | 2020-07-27 04:00:27 | goodboy/tractor | https://api.github.com/repos/goodboy/tractor | closed | Settle the flaky tests.. | bug testing | On `mp` we'll get [hangs on the streaming tests](https://travis-ci.org/goodboy/tractor/jobs/648528123?utm_medium=notification&utm_source=github_status) and on `trip` we sometimes get [strange bails](https://travis-ci.org/goodboy/tractor/jobs/648528126?utm_medium=notification&utm_source=github_status) on the `test_nested_multierrors[trio_run_in_process]` test. | 1.0 | Settle the flaky tests.. - On `mp` we'll get [hangs on the streaming tests](https://travis-ci.org/goodboy/tractor/jobs/648528123?utm_medium=notification&utm_source=github_status) and on `trip` we sometimes get [strange bails](https://travis-ci.org/goodboy/tractor/jobs/648528126?utm_medium=notification&utm_source=github_status) on the `test_nested_multierrors[trio_run_in_process]` test. | test | settle the flaky tests on mp we ll get and on trip we sometimes get on the test nested multierrors test | 1 |
267,652 | 23,312,726,997 | IssuesEvent | 2022-08-08 09:42:39 | openSUSE/open-build-service | https://api.github.com/repos/openSUSE/open-build-service | closed | `uninitialized constant Suse::ValidationError` errors in PackageService::FileVerifier specs | Frontend Test Suite / CI :syringe: Flickering :cyclone: | RSpec tests for PackageService::FileVerifier fail from time to time, with `uninitialized constant Suse::ValidationError.
Expected Result
============
The spec tests don't fail.
How to Reproduce
===============
1. Run `bin/rspec --seed 18030 spec/services/package_service/file_verifier_spec.rb` in a development environment.
1. See error:
```
1) PackageService::FileVerifier.call invalid service file
Failure/Error: it { expect { subject }.to raise_error(Suse::ValidationError) }
NameError:
uninitialized constant Suse::ValidationError
Did you mean? Suse::ValidationError
ValidationHelper
# ./spec/services/package_service/file_verifier_spec.rb:50:in `block (4 levels) in <top (required)>'
2) PackageService::FileVerifier.call invalid constraints
Failure/Error: it { expect { subject }.to raise_error(Suse::ValidationError) }
NameError:
uninitialized constant Suse::ValidationError
Did you mean? Suse::ValidationError
ValidationHelper
# ./spec/services/package_service/file_verifier_spec.rb:41:in `block (4 levels) in <top (required)>'
``` | 1.0 | `uninitialized constant Suse::ValidationError` errors in PackageService::FileVerifier specs - RSpec tests for PackageService::FileVerifier fail from time to time, with `uninitialized constant Suse::ValidationError.
Expected Result
============
The spec tests don't fail.
How to Reproduce
===============
1. Run `bin/rspec --seed 18030 spec/services/package_service/file_verifier_spec.rb` in a development environment.
1. See error:
```
1) PackageService::FileVerifier.call invalid service file
Failure/Error: it { expect { subject }.to raise_error(Suse::ValidationError) }
NameError:
uninitialized constant Suse::ValidationError
Did you mean? Suse::ValidationError
ValidationHelper
# ./spec/services/package_service/file_verifier_spec.rb:50:in `block (4 levels) in <top (required)>'
2) PackageService::FileVerifier.call invalid constraints
Failure/Error: it { expect { subject }.to raise_error(Suse::ValidationError) }
NameError:
uninitialized constant Suse::ValidationError
Did you mean? Suse::ValidationError
ValidationHelper
# ./spec/services/package_service/file_verifier_spec.rb:41:in `block (4 levels) in <top (required)>'
``` | test | uninitialized constant suse validationerror errors in packageservice fileverifier specs rspec tests for packageservice fileverifier fail from time to time with uninitialized constant suse validationerror expected result the spec tests don t fail how to reproduce run bin rspec seed spec services package service file verifier spec rb in a development environment see error packageservice fileverifier call invalid service file failure error it expect subject to raise error suse validationerror nameerror uninitialized constant suse validationerror did you mean suse validationerror validationhelper spec services package service file verifier spec rb in block levels in packageservice fileverifier call invalid constraints failure error it expect subject to raise error suse validationerror nameerror uninitialized constant suse validationerror did you mean suse validationerror validationhelper spec services package service file verifier spec rb in block levels in | 1 |
326,073 | 27,975,357,197 | IssuesEvent | 2023-03-25 14:13:41 | dudykr/stc | https://api.github.com/repos/dudykr/stc | opened | Fix unit test: `tests/pass-only/conformance/controlwFlow/controlFlowForOfStatement/.1.ts` | tsc-unit-test |
---
Related test: https://github.com/dudykr/stc/blob/main/crates/stc_ts_file_analyzer/tests/pass-only/conformance/controlwFlow/controlFlowForOfStatement/.1.ts
---
This issue is created by sync script.
| 1.0 | Fix unit test: `tests/pass-only/conformance/controlwFlow/controlFlowForOfStatement/.1.ts` -
---
Related test: https://github.com/dudykr/stc/blob/main/crates/stc_ts_file_analyzer/tests/pass-only/conformance/controlwFlow/controlFlowForOfStatement/.1.ts
---
This issue is created by sync script.
| test | fix unit test tests pass only conformance controlwflow controlflowforofstatement ts related test this issue is created by sync script | 1 |
248,332 | 21,011,954,413 | IssuesEvent | 2022-03-30 07:33:44 | WordPress/gutenberg | https://api.github.com/repos/WordPress/gutenberg | closed | @wordpress/e2e-test-utils activateTheme() creates issues due to not awaiting (or returning) when calling `switchUserToTest()` | [Type] Bug [Status] In Progress [Package] E2E Test Utils | I have an e2e test that switches the active theme using `activateTheme()`. When the theme is *already set* to the requested theme (for example if I request to switch to 2022 and the active theme is already 2022) the following condition is met in the function:
https://github.com/WordPress/gutenberg/blob/f594f6cff6db36e27df4ac44e78c448bb8a173d4/packages/e2e-test-utils/src/activate-theme.js#L20-L23
I believe that L21 should be `await switchUserToTest()`. Or L21-22 should be `return switchUserToTest()` so that when calling the method an await can be added as needed.
The error I'm encountering as a result is intermittent errors being reported from apiFetch, "You are probably offline". After quite a long while trying to trace back the source of this issue I've settled on rerunning the offending test with `jest.retryTimes(2)` however when doing this I encounter the following error:
```
/home/runner/work/lifterlms/lifterlms/node_modules/puppeteer-core/lib/cjs/puppeteer/common/Connection.js:230
error: new Errors_js_1.ProtocolError(),
^
ProtocolError: Protocol error (Network.getCookies): Target closed.
at /home/runner/work/lifterlms/lifterlms/node_modules/puppeteer-core/src/common/Connection.ts:304:16
at new Promise (<anonymous>)
at CDPSession.send (/home/runner/work/lifterlms/lifterlms/node_modules/puppeteer-core/src/common/Connection.ts:300:12)
at Page.cookies (/home/runner/work/lifterlms/lifterlms/node_modules/puppeteer-core/src/common/Page.ts:1245:26)
at getCurrentUser (/home/runner/work/lifterlms/lifterlms/node_modules/@wordpress/e2e-test-utils/build/@wordpress/e2e-test-utils/src/get-current-user.js:7:29)
at switchUserToTest (/home/runner/work/lifterlms/lifterlms/node_modules/@wordpress/e2e-test-utils/build/@wordpress/e2e-test-utils/src/switch-user-to-test.js:13:15)
at activateTheme (/home/runner/work/lifterlms/lifterlms/node_modules/@wordpress/e2e-test-utils/build/@wordpress/e2e-test-utils/src/activate-theme.js:[21](https://github.com/gocodebox/lifterlms/runs/5743094473?check_suite_focus=true#step:4:21):3)
at runMicrotasks (<anonymous>)
at processTicksAndRejections (node:internal/process/task_queues:96:5)
at Object.<anonymous> (/home/runner/work/lifterlms/lifterlms/tests/e2e/tests/activate/bootstrap.test.js:17:5) {
originalMessage: ''
}
```
I believe that this is ultimately resulting from the same (somewhat misleading) fetch_error ("You are probably offline."). Everything I can find points to missing `await` in some place so that the browser is closed with an open connection which would make sense if the async `switchUserToTest()` isn't completed yet.
I'm struggling to provide a reduced test case to display the error as I can't recreate this locally, it only happens when running my tests in a CI environment.
You can see my latest failing tests at https://github.com/gocodebox/lifterlms/actions/runs/2060616521
You'll notice that some tests fail and some succeed. | 1.0 | @wordpress/e2e-test-utils activateTheme() creates issues due to not awaiting (or returning) when calling `switchUserToTest()` - I have an e2e test that switches the active theme using `activateTheme()`. When the theme is *already set* to the requested theme (for example if I request to switch to 2022 and the active theme is already 2022) the following condition is met in the function:
https://github.com/WordPress/gutenberg/blob/f594f6cff6db36e27df4ac44e78c448bb8a173d4/packages/e2e-test-utils/src/activate-theme.js#L20-L23
I believe that L21 should be `await switchUserToTest()`. Or L21-22 should be `return switchUserToTest()` so that when calling the method an await can be added as needed.
The error I'm encountering as a result is intermittent errors being reported from apiFetch, "You are probably offline". After quite a long while trying to trace back the source of this issue I've settled on rerunning the offending test with `jest.retryTimes(2)` however when doing this I encounter the following error:
```
/home/runner/work/lifterlms/lifterlms/node_modules/puppeteer-core/lib/cjs/puppeteer/common/Connection.js:230
error: new Errors_js_1.ProtocolError(),
^
ProtocolError: Protocol error (Network.getCookies): Target closed.
at /home/runner/work/lifterlms/lifterlms/node_modules/puppeteer-core/src/common/Connection.ts:304:16
at new Promise (<anonymous>)
at CDPSession.send (/home/runner/work/lifterlms/lifterlms/node_modules/puppeteer-core/src/common/Connection.ts:300:12)
at Page.cookies (/home/runner/work/lifterlms/lifterlms/node_modules/puppeteer-core/src/common/Page.ts:1245:26)
at getCurrentUser (/home/runner/work/lifterlms/lifterlms/node_modules/@wordpress/e2e-test-utils/build/@wordpress/e2e-test-utils/src/get-current-user.js:7:29)
at switchUserToTest (/home/runner/work/lifterlms/lifterlms/node_modules/@wordpress/e2e-test-utils/build/@wordpress/e2e-test-utils/src/switch-user-to-test.js:13:15)
at activateTheme (/home/runner/work/lifterlms/lifterlms/node_modules/@wordpress/e2e-test-utils/build/@wordpress/e2e-test-utils/src/activate-theme.js:[21](https://github.com/gocodebox/lifterlms/runs/5743094473?check_suite_focus=true#step:4:21):3)
at runMicrotasks (<anonymous>)
at processTicksAndRejections (node:internal/process/task_queues:96:5)
at Object.<anonymous> (/home/runner/work/lifterlms/lifterlms/tests/e2e/tests/activate/bootstrap.test.js:17:5) {
originalMessage: ''
}
```
I believe that this is ultimately resulting from the same (somewhat misleading) fetch_error ("You are probably offline."). Everything I can find points to missing `await` in some place so that the browser is closed with an open connection which would make sense if the async `switchUserToTest()` isn't completed yet.
I'm struggling to provide a reduced test case to display the error as I can't recreate this locally, it only happens when running my tests in a CI environment.
You can see my latest failing tests at https://github.com/gocodebox/lifterlms/actions/runs/2060616521
You'll notice that some tests fail and some succeed. | test | wordpress test utils activatetheme creates issues due to not awaiting or returning when calling switchusertotest i have an test that switches the active theme using activatetheme when the theme is already set to the requested theme for example if i request to switch to and the active theme is already the following condition is met in the function i believe that should be await switchusertotest or should be return switchusertotest so that when calling the method an await can be added as needed the error i m encountering as a result is intermittent errors being reported from apifetch you are probably offline after quite a long while trying to trace back the source of this issue i ve settled on rerunning the offending test with jest retrytimes however when doing this i encounter the following error home runner work lifterlms lifterlms node modules puppeteer core lib cjs puppeteer common connection js error new errors js protocolerror protocolerror protocol error network getcookies target closed at home runner work lifterlms lifterlms node modules puppeteer core src common connection ts at new promise at cdpsession send home runner work lifterlms lifterlms node modules puppeteer core src common connection ts at page cookies home runner work lifterlms lifterlms node modules puppeteer core src common page ts at getcurrentuser home runner work lifterlms lifterlms node modules wordpress test utils build wordpress test utils src get current user js at switchusertotest home runner work lifterlms lifterlms node modules wordpress test utils build wordpress test utils src switch user to test js at activatetheme home runner work lifterlms lifterlms node modules wordpress test utils build wordpress test utils src activate theme js at runmicrotasks at processticksandrejections node internal process task queues at object home runner work lifterlms lifterlms tests tests activate bootstrap test js originalmessage i believe that this is ultimately resulting from the same somewhat misleading fetch error you are probably offline everything i can find points to missing await in some place so that the browser is closed with an open connection which would make sense if the async switchusertotest isn t completed yet i m struggling to provide a reduced test case to display the error as i can t recreate this locally it only happens when running my tests in a ci environment you can see my latest failing tests at you ll notice that some tests fail and some succeed | 1 |
228,963 | 18,273,304,368 | IssuesEvent | 2021-10-04 15:52:08 | froescayo/prog-acad-web | https://api.github.com/repos/froescayo/prog-acad-web | closed | Atualizar diagrama do Back-End | Gestão Documentação Testes Back-End | Construir o diagrama de arquitetura do sistema.
Segue o link da arquitetura atual: https://whimsical.com/arquitetura-da-aplicacao-72ThrruL54ybvGVZX3mMfc | 1.0 | Atualizar diagrama do Back-End - Construir o diagrama de arquitetura do sistema.
Segue o link da arquitetura atual: https://whimsical.com/arquitetura-da-aplicacao-72ThrruL54ybvGVZX3mMfc | test | atualizar diagrama do back end construir o diagrama de arquitetura do sistema segue o link da arquitetura atual | 1 |
14,018 | 5,536,231,273 | IssuesEvent | 2017-03-21 19:07:09 | mobdata/replication | https://api.github.com/repos/mobdata/replication | closed | (Parent #232) Initial nodejs setup | component : build ready Subtask | Prereq: #233
Once the "Hello, World" app is working there are a couple more things needed before we make an rpm from it.
What to do:
* Must support SSL out of the box
* Document what's needed to get SSL working | 1.0 | (Parent #232) Initial nodejs setup - Prereq: #233
Once the "Hello, World" app is working there are a couple more things needed before we make an rpm from it.
What to do:
* Must support SSL out of the box
* Document what's needed to get SSL working | non_test | parent initial nodejs setup prereq once the hello world app is working there are a couple more things needed before we make an rpm from it what to do must support ssl out of the box document what s needed to get ssl working | 0 |
273,283 | 23,743,626,371 | IssuesEvent | 2022-08-31 14:19:37 | elastic/elasticsearch | https://api.github.com/repos/elastic/elasticsearch | closed | [CI] XPackRestIT test {p0=transform/transforms_start_stop/Test stop all} failing | >test-failure :ml/Transform Team:ML | **Build scan:**
https://gradle-enterprise.elastic.co/s/pky3bwfrwrpjs/tests/:x-pack:plugin:yamlRestTest/org.elasticsearch.xpack.test.rest.XPackRestIT/test%20%7Bp0=transform%2Ftransforms_start_stop%2FTest%20stop%20all%7D
**Reproduction line:**
`./gradlew ':x-pack:plugin:yamlRestTest' --tests "org.elasticsearch.xpack.test.rest.XPackRestIT.test {p0=transform/transforms_start_stop/Test stop all}" -Dtests.seed=1F8B9F800B9DEB00 -Dtests.locale=he-IL -Dtests.timezone=US/Michigan -Druntime.java=17`
**Applicable branches:**
main
**Reproduces locally?:**
No
**Failure history:**
https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.test.rest.XPackRestIT&tests.test=test%20%7Bp0%3Dtransform/transforms_start_stop/Test%20stop%20all%7D
**Failure excerpt:**
```
java.lang.AssertionError: Failure at [transform/transforms_start_stop:373]: expected [2xx] status code but api [transform.stop_transform] returned [409 Conflict] [{"error":{"root_cause":[{"type":"status_exception","reason":"Failed to update transform task [airline-transform-start-stop-continuous] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [airline-transform-start-stop-continuous] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:323)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:299)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1380)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1465)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1445)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onResponse(TransportTasksAction.java:108)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onResponse(TransportTasksAction.java:104)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$2(TransportStopTransformAction.java:241)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:808)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"}],"type":"status_exception","reason":"Failed to update transform task [airline-transform-start-stop-continuous] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [airline-transform-start-stop-continuous] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:323)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:299)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1380)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1465)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1445)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onResponse(TransportTasksAction.java:108)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onResponse(TransportTasksAction.java:104)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$2(TransportStopTransformAction.java:241)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:808)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"},"status":409}]
at __randomizedtesting.SeedInfo.seed([1F8B9F800B9DEB00:97DFA05AA56186F8]:0)
at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.executeSection(ESClientYamlSuiteTestCase.java:510)
at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.test(ESClientYamlSuiteTestCase.java:479)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:568)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850)
at java.lang.Thread.run(Thread.java:833)
Caused by: java.lang.AssertionError: expected [2xx] status code but api [transform.stop_transform] returned [409 Conflict] [{"error":{"root_cause":[{"type":"status_exception","reason":"Failed to update transform task [airline-transform-start-stop-continuous] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [airline-transform-start-stop-continuous] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:323)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:299)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1380)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1465)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1445)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onResponse(TransportTasksAction.java:108)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onResponse(TransportTasksAction.java:104)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$2(TransportStopTransformAction.java:241)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:808)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"}],"type":"status_exception","reason":"Failed to update transform task [airline-transform-start-stop-continuous] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [airline-transform-start-stop-continuous] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:323)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:299)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1380)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1465)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1445)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onResponse(TransportTasksAction.java:108)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onResponse(TransportTasksAction.java:104)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$2(TransportStopTransformAction.java:241)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:808)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"},"status":409}]
at org.junit.Assert.fail(Assert.java:88)
at org.elasticsearch.test.rest.yaml.section.DoSection.execute(DoSection.java:373)
at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.executeSection(ESClientYamlSuiteTestCase.java:499)
at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.test(ESClientYamlSuiteTestCase.java:479)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:568)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850)
at java.lang.Thread.run(Thread.java:833)
``` | 1.0 | [CI] XPackRestIT test {p0=transform/transforms_start_stop/Test stop all} failing - **Build scan:**
https://gradle-enterprise.elastic.co/s/pky3bwfrwrpjs/tests/:x-pack:plugin:yamlRestTest/org.elasticsearch.xpack.test.rest.XPackRestIT/test%20%7Bp0=transform%2Ftransforms_start_stop%2FTest%20stop%20all%7D
**Reproduction line:**
`./gradlew ':x-pack:plugin:yamlRestTest' --tests "org.elasticsearch.xpack.test.rest.XPackRestIT.test {p0=transform/transforms_start_stop/Test stop all}" -Dtests.seed=1F8B9F800B9DEB00 -Dtests.locale=he-IL -Dtests.timezone=US/Michigan -Druntime.java=17`
**Applicable branches:**
main
**Reproduces locally?:**
No
**Failure history:**
https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.test.rest.XPackRestIT&tests.test=test%20%7Bp0%3Dtransform/transforms_start_stop/Test%20stop%20all%7D
**Failure excerpt:**
```
java.lang.AssertionError: Failure at [transform/transforms_start_stop:373]: expected [2xx] status code but api [transform.stop_transform] returned [409 Conflict] [{"error":{"root_cause":[{"type":"status_exception","reason":"Failed to update transform task [airline-transform-start-stop-continuous] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [airline-transform-start-stop-continuous] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:323)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:299)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1380)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1465)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1445)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onResponse(TransportTasksAction.java:108)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onResponse(TransportTasksAction.java:104)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$2(TransportStopTransformAction.java:241)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:808)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"}],"type":"status_exception","reason":"Failed to update transform task [airline-transform-start-stop-continuous] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [airline-transform-start-stop-continuous] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:323)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:299)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1380)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1465)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1445)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onResponse(TransportTasksAction.java:108)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onResponse(TransportTasksAction.java:104)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$2(TransportStopTransformAction.java:241)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:808)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"},"status":409}]
at __randomizedtesting.SeedInfo.seed([1F8B9F800B9DEB00:97DFA05AA56186F8]:0)
at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.executeSection(ESClientYamlSuiteTestCase.java:510)
at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.test(ESClientYamlSuiteTestCase.java:479)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:568)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850)
at java.lang.Thread.run(Thread.java:833)
Caused by: java.lang.AssertionError: expected [2xx] status code but api [transform.stop_transform] returned [409 Conflict] [{"error":{"root_cause":[{"type":"status_exception","reason":"Failed to update transform task [airline-transform-start-stop-continuous] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [airline-transform-start-stop-continuous] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:323)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:299)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1380)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1465)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1445)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onResponse(TransportTasksAction.java:108)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onResponse(TransportTasksAction.java:104)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$2(TransportStopTransformAction.java:241)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:808)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"}],"type":"status_exception","reason":"Failed to update transform task [airline-transform-start-stop-continuous] state value should_stop_at_checkpoint from [false] to [false]","stack_trace":"org.elasticsearch.ElasticsearchStatusException: Failed to update transform task [airline-transform-start-stop-continuous] state value should_stop_at_checkpoint from [false] to [false]\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.buildException(TransportStopTransformAction.java:323)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$waitForStopListener$9(TransportStopTransformAction.java:299)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.finishHim(TransportTasksAction.java:323)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction.onOperation(TransportTasksAction.java:297)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:277)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$AsyncAction$1.handleResponse(TransportTasksAction.java:269)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1380)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.processResponse(TransportService.java:1465)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TransportService$DirectResponseChannel.sendResponse(TransportService.java:1445)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.transport.TaskTransportChannel.sendResponse(TaskTransportChannel.java:41)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.respondIfFinished(TransportTasksAction.java:134)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onResponse(TransportTasksAction.java:108)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.support.tasks.TransportTasksAction$1.onResponse(TransportTasksAction.java:104)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$2(TransportStopTransformAction.java:241)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.action.ActionListener$2.onResponse(ActionListener.java:162)\n\tat org.elasticsearch.xpack.transform.transforms.TransformTask.setShouldStopAtCheckpoint(TransformTask.java:319)\n\tat org.elasticsearch.xpack.transform.action.TransportStopTransformAction.lambda$taskOperation$4(TransportStopTransformAction.java:238)\n\tat org.elasticsearch.server@8.5.0-SNAPSHOT/org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:808)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)\n\tat java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)\n\tat java.base/java.lang.Thread.run(Thread.java:833)\n"},"status":409}]
at org.junit.Assert.fail(Assert.java:88)
at org.elasticsearch.test.rest.yaml.section.DoSection.execute(DoSection.java:373)
at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.executeSection(ESClientYamlSuiteTestCase.java:499)
at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.test(ESClientYamlSuiteTestCase.java:479)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:568)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850)
at java.lang.Thread.run(Thread.java:833)
``` | test | xpackrestit test transform transforms start stop test stop all failing build scan reproduction line gradlew x pack plugin yamlresttest tests org elasticsearch xpack test rest xpackrestit test transform transforms start stop test stop all dtests seed dtests locale he il dtests timezone us michigan druntime java applicable branches main reproduces locally no failure history failure excerpt java lang assertionerror failure at expected status code but api returned state value should stop at checkpoint from to stack trace org elasticsearch elasticsearchstatusexception failed to update transform task state value should stop at checkpoint from to n tat org elasticsearch xpack transform action transportstoptransformaction buildexception transportstoptransformaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda waitforstoplistener transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction finishhim transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction onoperation transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch transport transportservice contextrestoreresponsehandler handleresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel processresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel sendresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport tasktransportchannel sendresponse tasktransportchannel java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction respondiffinished transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction onresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction onresponse transporttasksaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch xpack transform transforms transformtask setshouldstopatcheckpoint transformtask java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch common util concurrent threadcontext contextpreservingrunnable run threadcontext java n tat java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java n tat java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java n tat java base java lang thread run thread java n type status exception reason failed to update transform task state value should stop at checkpoint from to stack trace org elasticsearch elasticsearchstatusexception failed to update transform task state value should stop at checkpoint from to n tat org elasticsearch xpack transform action transportstoptransformaction buildexception transportstoptransformaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda waitforstoplistener transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction finishhim transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction onoperation transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch transport transportservice contextrestoreresponsehandler handleresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel processresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel sendresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport tasktransportchannel sendresponse tasktransportchannel java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction respondiffinished transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction onresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction onresponse transporttasksaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch xpack transform transforms transformtask setshouldstopatcheckpoint transformtask java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch common util concurrent threadcontext contextpreservingrunnable run threadcontext java n tat java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java n tat java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java n tat java base java lang thread run thread java n status at randomizedtesting seedinfo seed at org elasticsearch test rest yaml esclientyamlsuitetestcase executesection esclientyamlsuitetestcase java at org elasticsearch test rest yaml esclientyamlsuitetestcase test esclientyamlsuitetestcase java at jdk internal reflect nativemethodaccessorimpl nativemethodaccessorimpl java at jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulesetupteardownchained evaluate testrulesetupteardownchained java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulethreadandtestname evaluate testrulethreadandtestname java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol forktimeoutingtask threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol evaluate threadleakcontrol java at com carrotsearch randomizedtesting randomizedrunner runsingletest randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulestoreclassname evaluate testrulestoreclassname java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testruleassertionsrequired evaluate testruleassertionsrequired java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testruleignoretestsuites evaluate testruleignoretestsuites java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol lambda forktimeoutingtask threadleakcontrol java at java lang thread run thread java caused by java lang assertionerror expected status code but api returned state value should stop at checkpoint from to stack trace org elasticsearch elasticsearchstatusexception failed to update transform task state value should stop at checkpoint from to n tat org elasticsearch xpack transform action transportstoptransformaction buildexception transportstoptransformaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda waitforstoplistener transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction finishhim transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction onoperation transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch transport transportservice contextrestoreresponsehandler handleresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel processresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel sendresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport tasktransportchannel sendresponse tasktransportchannel java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction respondiffinished transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction onresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction onresponse transporttasksaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch xpack transform transforms transformtask setshouldstopatcheckpoint transformtask java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch common util concurrent threadcontext contextpreservingrunnable run threadcontext java n tat java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java n tat java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java n tat java base java lang thread run thread java n type status exception reason failed to update transform task state value should stop at checkpoint from to stack trace org elasticsearch elasticsearchstatusexception failed to update transform task state value should stop at checkpoint from to n tat org elasticsearch xpack transform action transportstoptransformaction buildexception transportstoptransformaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda waitforstoplistener transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction finishhim transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction onoperation transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction asyncaction handleresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch transport transportservice contextrestoreresponsehandler handleresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel processresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport transportservice directresponsechannel sendresponse transportservice java n tat org elasticsearch server snapshot org elasticsearch transport tasktransportchannel sendresponse tasktransportchannel java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction respondiffinished transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction onresponse transporttasksaction java n tat org elasticsearch server snapshot org elasticsearch action support tasks transporttasksaction onresponse transporttasksaction java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch action actionlistener onresponse actionlistener java n tat org elasticsearch xpack transform transforms transformtask setshouldstopatcheckpoint transformtask java n tat org elasticsearch xpack transform action transportstoptransformaction lambda taskoperation transportstoptransformaction java n tat org elasticsearch server snapshot org elasticsearch common util concurrent threadcontext contextpreservingrunnable run threadcontext java n tat java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java n tat java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java n tat java base java lang thread run thread java n status at org junit assert fail assert java at org elasticsearch test rest yaml section dosection execute dosection java at org elasticsearch test rest yaml esclientyamlsuitetestcase executesection esclientyamlsuitetestcase java at org elasticsearch test rest yaml esclientyamlsuitetestcase test esclientyamlsuitetestcase java at jdk internal reflect nativemethodaccessorimpl nativemethodaccessorimpl java at jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulesetupteardownchained evaluate testrulesetupteardownchained java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulethreadandtestname evaluate testrulethreadandtestname java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol forktimeoutingtask threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol evaluate threadleakcontrol java at com carrotsearch randomizedtesting randomizedrunner runsingletest randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulestoreclassname evaluate testrulestoreclassname java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testruleassertionsrequired evaluate testruleassertionsrequired java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testruleignoretestsuites evaluate testruleignoretestsuites java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol lambda forktimeoutingtask threadleakcontrol java at java lang thread run thread java | 1 |
27,683 | 4,325,755,522 | IssuesEvent | 2016-07-26 01:37:47 | nwjs/nw.js | https://api.github.com/repos/nwjs/nw.js | closed | NWJS crashes if <iframe> has JavaScript error | test-todo | Application folder has 3 files:
- `package.json` has `{ "name": "name", "main": "main.html"}`
- `main.html` has `<iframe src="frame.html" nwdisable></iframe>`
- `frame.html` has `<script>error</script>`
Application starts and instantly crashes on Windows 10.
`iframe`'s `src` can point to some site that has errors in console, result is the same. `nwdisable` is not necessary in this case, for example: `<iframe src="https://www.microsoft.com/en-us/windows/windows-10-upgrade"></iframe>`
NW.js is v0.16.0 | 1.0 | NWJS crashes if <iframe> has JavaScript error - Application folder has 3 files:
- `package.json` has `{ "name": "name", "main": "main.html"}`
- `main.html` has `<iframe src="frame.html" nwdisable></iframe>`
- `frame.html` has `<script>error</script>`
Application starts and instantly crashes on Windows 10.
`iframe`'s `src` can point to some site that has errors in console, result is the same. `nwdisable` is not necessary in this case, for example: `<iframe src="https://www.microsoft.com/en-us/windows/windows-10-upgrade"></iframe>`
NW.js is v0.16.0 | test | nwjs crashes if has javascript error application folder has files package json has name name main main html main html has frame html has error application starts and instantly crashes on windows iframe s src can point to some site that has errors in console result is the same nwdisable is not necessary in this case for example iframe src nw js is | 1 |
165,764 | 12,879,871,615 | IssuesEvent | 2020-07-12 01:27:14 | osquery/osquery | https://api.github.com/repos/osquery/osquery | closed | Create tests for the table `authorized_keys` | FreeBSD Linux good-first-issue macOS test | ## Feature request
### What new feature do you want?
Create tests for the table `authorized_keys`
- Create header file for the table implementation, if one is not exists.
- In test, query the table and check if retrieved columns (name and types) match the columns from table spec.
- If there is any guarantee to number of rows (e.g. only 1 record in every query result, more than 3 records or something else) check it.
- Test the implementation details of the table, if it possible.
Table spec: `specs/posix/authorized_keys.table`
Source files:
- `osquery/tables/system/posix/authorized_keys.cpp`
Table generating function: `getAuthorizedKeys()`
### How is this new feature useful?
esting improves the quality of the code, helps to find bugs. But the biggest value is it allows us to modify code and make sure everything works correctly.
### How can this be implemented?
If you need an example how to do it use `osquery/tables/system/posix/tests/shell_history_tests.cpp` or any existing test for the table implementation.
Labels: `good-first-issue` `build/test` `linux` `freebsd` `macos`
| 1.0 | Create tests for the table `authorized_keys` - ## Feature request
### What new feature do you want?
Create tests for the table `authorized_keys`
- Create header file for the table implementation, if one is not exists.
- In test, query the table and check if retrieved columns (name and types) match the columns from table spec.
- If there is any guarantee to number of rows (e.g. only 1 record in every query result, more than 3 records or something else) check it.
- Test the implementation details of the table, if it possible.
Table spec: `specs/posix/authorized_keys.table`
Source files:
- `osquery/tables/system/posix/authorized_keys.cpp`
Table generating function: `getAuthorizedKeys()`
### How is this new feature useful?
esting improves the quality of the code, helps to find bugs. But the biggest value is it allows us to modify code and make sure everything works correctly.
### How can this be implemented?
If you need an example how to do it use `osquery/tables/system/posix/tests/shell_history_tests.cpp` or any existing test for the table implementation.
Labels: `good-first-issue` `build/test` `linux` `freebsd` `macos`
| test | create tests for the table authorized keys feature request what new feature do you want create tests for the table authorized keys create header file for the table implementation if one is not exists in test query the table and check if retrieved columns name and types match the columns from table spec if there is any guarantee to number of rows e g only record in every query result more than records or something else check it test the implementation details of the table if it possible table spec specs posix authorized keys table source files osquery tables system posix authorized keys cpp table generating function getauthorizedkeys how is this new feature useful esting improves the quality of the code helps to find bugs but the biggest value is it allows us to modify code and make sure everything works correctly how can this be implemented if you need an example how to do it use osquery tables system posix tests shell history tests cpp or any existing test for the table implementation labels good first issue build test linux freebsd macos | 1 |
36,356 | 17,633,545,081 | IssuesEvent | 2021-08-19 11:01:35 | dotnet/project-system | https://api.github.com/repos/dotnet/project-system | closed | Introduce ConfiguredProjectDataSourceJoinBlock | Feature-NuGet Tenet-Performance Feature-Up-to-date Performance-Scenario-General | Currently we use `UnwrapCollectionChainedProjectValueDataSource` to aggregate data across project configurations. Its API is a bit hard to understand.
We should use the newer `ConfiguredProjectDataSourceJoinBlock` instead.
Added in CPS PR: https://devdiv.visualstudio.com/DevDiv/_git/CPS/pullrequest/329048 | True | Introduce ConfiguredProjectDataSourceJoinBlock - Currently we use `UnwrapCollectionChainedProjectValueDataSource` to aggregate data across project configurations. Its API is a bit hard to understand.
We should use the newer `ConfiguredProjectDataSourceJoinBlock` instead.
Added in CPS PR: https://devdiv.visualstudio.com/DevDiv/_git/CPS/pullrequest/329048 | non_test | introduce configuredprojectdatasourcejoinblock currently we use unwrapcollectionchainedprojectvaluedatasource to aggregate data across project configurations its api is a bit hard to understand we should use the newer configuredprojectdatasourcejoinblock instead added in cps pr | 0 |
613,673 | 19,096,304,261 | IssuesEvent | 2021-11-29 17:01:45 | chaotic-aur/packages | https://api.github.com/repos/chaotic-aur/packages | closed | [Request] gtksourceview-git | request:new-pkg priority:low | ## 👶 requesting new package: `gtksourceview-git`
- Link to the package(s) in AUR (**_git_**): [gtksourceview-git](https://aur.archlinux.org/gtksourceview-git.git)
- Link to the package(s) in AUR (**_webpage_**): [gtksourceview-git](https://aur.archlinux.org/packages/gtksourceview-git)
- Utility this package has for you:
Provides the latest development version of gtksourceview and in result, makes other bleeding edge versions of GNOME apps buildable/usable (most notably gnome-text-editor-git)
- Do you consider this package(s) to be useful for **every** chaotic user?:
- [ ] YES
- [ ] No, but yes for a great amount.
- [x] No, but yes for a few.
- [ ] No, it's useful only for me.
- Do you consider this package(s) to be useful for feature testing/preview (e.g: mesa-aco, wine-wayland)?:
- [ ] YES
- [ ] NO
- [x] IDK. Kinda
- Are you sure we don't have this package already (test with `pacman -Ss <pkgname>`)?:
- [x] YES
- Have you tested if this package builds in a clean chroot?:
- [ ] YES
- [x] NO
- Does the package's license allows us to redistribute it?:
- [x] YES
- [ ] No clue.
- [ ] No, but the author doesn't really care, it's just for bureaucracy.
- Have you searched the [issues](https://github.com/chaotic-aur/packages/issues) to ensure this request is new (not duplicated)?:
- [x] YES
- Have you read the [README](https://github.com/chaotic-aur/packages#banished-and-rejected-packages) to ensure this package is not banned?:
- [x] YES
| 1.0 | [Request] gtksourceview-git - ## 👶 requesting new package: `gtksourceview-git`
- Link to the package(s) in AUR (**_git_**): [gtksourceview-git](https://aur.archlinux.org/gtksourceview-git.git)
- Link to the package(s) in AUR (**_webpage_**): [gtksourceview-git](https://aur.archlinux.org/packages/gtksourceview-git)
- Utility this package has for you:
Provides the latest development version of gtksourceview and in result, makes other bleeding edge versions of GNOME apps buildable/usable (most notably gnome-text-editor-git)
- Do you consider this package(s) to be useful for **every** chaotic user?:
- [ ] YES
- [ ] No, but yes for a great amount.
- [x] No, but yes for a few.
- [ ] No, it's useful only for me.
- Do you consider this package(s) to be useful for feature testing/preview (e.g: mesa-aco, wine-wayland)?:
- [ ] YES
- [ ] NO
- [x] IDK. Kinda
- Are you sure we don't have this package already (test with `pacman -Ss <pkgname>`)?:
- [x] YES
- Have you tested if this package builds in a clean chroot?:
- [ ] YES
- [x] NO
- Does the package's license allows us to redistribute it?:
- [x] YES
- [ ] No clue.
- [ ] No, but the author doesn't really care, it's just for bureaucracy.
- Have you searched the [issues](https://github.com/chaotic-aur/packages/issues) to ensure this request is new (not duplicated)?:
- [x] YES
- Have you read the [README](https://github.com/chaotic-aur/packages#banished-and-rejected-packages) to ensure this package is not banned?:
- [x] YES
| non_test | gtksourceview git 👶 requesting new package gtksourceview git link to the package s in aur git link to the package s in aur webpage utility this package has for you provides the latest development version of gtksourceview and in result makes other bleeding edge versions of gnome apps buildable usable most notably gnome text editor git do you consider this package s to be useful for every chaotic user yes no but yes for a great amount no but yes for a few no it s useful only for me do you consider this package s to be useful for feature testing preview e g mesa aco wine wayland yes no idk kinda are you sure we don t have this package already test with pacman ss yes have you tested if this package builds in a clean chroot yes no does the package s license allows us to redistribute it yes no clue no but the author doesn t really care it s just for bureaucracy have you searched the to ensure this request is new not duplicated yes have you read the to ensure this package is not banned yes | 0 |
136,647 | 30,568,060,829 | IssuesEvent | 2023-07-20 19:28:06 | h4sh5/pypi-auto-scanner | https://api.github.com/repos/h4sh5/pypi-auto-scanner | opened | ddtrace 1.17.0rc1 has 2 GuardDog issues | guarddog code-execution exec-base64 | https://pypi.org/project/ddtrace
https://inspector.pypi.io/project/ddtrace
```{
"dependency": "ddtrace",
"version": "1.17.0rc1",
"result": {
"issues": 2,
"errors": {},
"results": {
"exec-base64": [
{
"location": "ddtrace-1.17.0rc1/setup.py:347",
"code": " subprocess.run(cmake_cmd_with_args, cwd=tmp_iast_path, check=True)",
"message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n"
}
],
"code-execution": [
{
"location": "ddtrace-1.17.0rc1/setup.py:350",
"code": " subprocess.run(build_command, cwd=tmp_iast_path, check=True)",
"message": "This package is executing OS commands in the setup.py file"
}
]
},
"path": "/tmp/tmplw156alw/ddtrace"
}
}``` | 1.0 | ddtrace 1.17.0rc1 has 2 GuardDog issues - https://pypi.org/project/ddtrace
https://inspector.pypi.io/project/ddtrace
```{
"dependency": "ddtrace",
"version": "1.17.0rc1",
"result": {
"issues": 2,
"errors": {},
"results": {
"exec-base64": [
{
"location": "ddtrace-1.17.0rc1/setup.py:347",
"code": " subprocess.run(cmake_cmd_with_args, cwd=tmp_iast_path, check=True)",
"message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n"
}
],
"code-execution": [
{
"location": "ddtrace-1.17.0rc1/setup.py:350",
"code": " subprocess.run(build_command, cwd=tmp_iast_path, check=True)",
"message": "This package is executing OS commands in the setup.py file"
}
]
},
"path": "/tmp/tmplw156alw/ddtrace"
}
}``` | non_test | ddtrace has guarddog issues dependency ddtrace version result issues errors results exec location ddtrace setup py code subprocess run cmake cmd with args cwd tmp iast path check true message this package contains a call to the eval function with a encoded string as argument nthis is a common method used to hide a malicious payload in a module as static analysis will not decode the nstring n code execution location ddtrace setup py code subprocess run build command cwd tmp iast path check true message this package is executing os commands in the setup py file path tmp ddtrace | 0 |
249,389 | 7,961,301,116 | IssuesEvent | 2018-07-13 10:15:27 | Stivius/XiboLinuxStack | https://api.github.com/repos/Stivius/XiboLinuxStack | closed | Investigate XMDS | medium priority task | - [x] Find out about XMDS in Xibo Docs
- [x] Check libs for implementing SOAP client | 1.0 | Investigate XMDS - - [x] Find out about XMDS in Xibo Docs
- [x] Check libs for implementing SOAP client | non_test | investigate xmds find out about xmds in xibo docs check libs for implementing soap client | 0 |
269,457 | 23,444,006,821 | IssuesEvent | 2022-08-15 17:40:43 | pvlib/pvlib-python | https://api.github.com/repos/pvlib/pvlib-python | closed | Changes to PVGIS in version 5.2 | testing io | **Describe the bug**
PVGIS is releasing version 5.2 on March 1st, 2022. In the new version, the irradiance databases CMSAF and COSMO have been removed. Additionally, the ERA5 coverage has been extended to cover almost the entire globe and a new database, SARAH-2, has been added.
The full description can be found [here](https://joint-research-centre.ec.europa.eu/pvgis-photovoltaic-geographical-information-system/pvgis-releases/pvgis-52_en).
The new default databases can be seen here:

**Additional context**
There is nothing that indicates any changes to the API, so the pvlib.iootools functions ``get_pvgis_tmy`` and ``get_pvgis_hourly`` should continue to work.
However, some of the test files used in `test_pvgis.py` are from the CMSAF database. Since those tests seem to be either based on local files or mock tests, I do not expect test failures. However, it would probably make sense to update the tests regardless.
| 1.0 | Changes to PVGIS in version 5.2 - **Describe the bug**
PVGIS is releasing version 5.2 on March 1st, 2022. In the new version, the irradiance databases CMSAF and COSMO have been removed. Additionally, the ERA5 coverage has been extended to cover almost the entire globe and a new database, SARAH-2, has been added.
The full description can be found [here](https://joint-research-centre.ec.europa.eu/pvgis-photovoltaic-geographical-information-system/pvgis-releases/pvgis-52_en).
The new default databases can be seen here:

**Additional context**
There is nothing that indicates any changes to the API, so the pvlib.iootools functions ``get_pvgis_tmy`` and ``get_pvgis_hourly`` should continue to work.
However, some of the test files used in `test_pvgis.py` are from the CMSAF database. Since those tests seem to be either based on local files or mock tests, I do not expect test failures. However, it would probably make sense to update the tests regardless.
| test | changes to pvgis in version describe the bug pvgis is releasing version on march in the new version the irradiance databases cmsaf and cosmo have been removed additionally the coverage has been extended to cover almost the entire globe and a new database sarah has been added the full description can be found the new default databases can be seen here additional context there is nothing that indicates any changes to the api so the pvlib iootools functions get pvgis tmy and get pvgis hourly should continue to work however some of the test files used in test pvgis py are from the cmsaf database since those tests seem to be either based on local files or mock tests i do not expect test failures however it would probably make sense to update the tests regardless | 1 |
773,496 | 27,159,539,180 | IssuesEvent | 2023-02-17 10:40:59 | ethereum-push-notification-service/push-dapp | https://api.github.com/repos/ethereum-push-notification-service/push-dapp | closed | Changes required in channel creation process for Time-Bound channels | Priority 1 | - A new type of Channel that is **TIME BOUND CHANNELS** are to be created. Check Details here.
- For such channels, users will have the option to toggle and select if their channels are timebound.
- If they are, then the DAPP must provide a way for them to pass the Deadline for the Channel. | 1.0 | Changes required in channel creation process for Time-Bound channels - - A new type of Channel that is **TIME BOUND CHANNELS** are to be created. Check Details here.
- For such channels, users will have the option to toggle and select if their channels are timebound.
- If they are, then the DAPP must provide a way for them to pass the Deadline for the Channel. | non_test | changes required in channel creation process for time bound channels a new type of channel that is time bound channels are to be created check details here for such channels users will have the option to toggle and select if their channels are timebound if they are then the dapp must provide a way for them to pass the deadline for the channel | 0 |
185,210 | 15,014,512,101 | IssuesEvent | 2021-02-01 06:45:26 | auxmoney/OpentracingBundle-core | https://api.github.com/repos/auxmoney/OpentracingBundle-core | opened | refurbish README | documentation | update README reflecting all changes (correct CI badges, remove some quality badges, link to tutorial, ...) | 1.0 | refurbish README - update README reflecting all changes (correct CI badges, remove some quality badges, link to tutorial, ...) | non_test | refurbish readme update readme reflecting all changes correct ci badges remove some quality badges link to tutorial | 0 |
367,814 | 25,763,524,961 | IssuesEvent | 2022-12-08 22:55:30 | bounswe/bounswe2022group2 | https://api.github.com/repos/bounswe/bounswe2022group2 | opened | Milestone 2: Individual Report | priority-high type-documentation status-inprogress milestone | ### Issue Description
As mentioned in #668, the Individual contribution report of @bahricanyesil needs to be documented. The report will include my review of our second demo, unit tests I have implemented so far, and issues & pull requests I have created for Milestone-2.
### Step Details
Steps that will be performed:
- [ ] Document my review
- [ ] Link issues and pull requests
- [ ] Include unit tests
### Final Actions
The related PR review and merge will close this issue.
### Deadline of the Issue
09.12.2022 - 14.00
### Reviewer
Onur Kömürcü
### Deadline for the Review
09.12.2022 - 16.00 | 1.0 | Milestone 2: Individual Report - ### Issue Description
As mentioned in #668, the Individual contribution report of @bahricanyesil needs to be documented. The report will include my review of our second demo, unit tests I have implemented so far, and issues & pull requests I have created for Milestone-2.
### Step Details
Steps that will be performed:
- [ ] Document my review
- [ ] Link issues and pull requests
- [ ] Include unit tests
### Final Actions
The related PR review and merge will close this issue.
### Deadline of the Issue
09.12.2022 - 14.00
### Reviewer
Onur Kömürcü
### Deadline for the Review
09.12.2022 - 16.00 | non_test | milestone individual report issue description as mentioned in the individual contribution report of bahricanyesil needs to be documented the report will include my review of our second demo unit tests i have implemented so far and issues pull requests i have created for milestone step details steps that will be performed document my review link issues and pull requests include unit tests final actions the related pr review and merge will close this issue deadline of the issue reviewer onur kömürcü deadline for the review | 0 |
148,688 | 11,861,820,281 | IssuesEvent | 2020-03-25 16:55:53 | openshift/odo | https://api.github.com/repos/openshift/odo | closed | Template scripts for unit tests | area/testing kind/feature priority/Low | [kind/Enhancement]
<!--
Welcome! - We kindly ask you to:
1. Fill out the issue template below
2. Use the chat and talk to us if you have a question rather than a bug or feature request.
The chat room is at: https://chat.openshift.io/developers/channels/odo
Thanks for understanding, and for contributing to the project!
-->
## Which functionality do you think we should update/improve?
We should add scripts to generate unit tests from a template (preferably bash file or linked into Makefile)
## Why is this needed?
Write tests faster ;)
| 1.0 | Template scripts for unit tests - [kind/Enhancement]
<!--
Welcome! - We kindly ask you to:
1. Fill out the issue template below
2. Use the chat and talk to us if you have a question rather than a bug or feature request.
The chat room is at: https://chat.openshift.io/developers/channels/odo
Thanks for understanding, and for contributing to the project!
-->
## Which functionality do you think we should update/improve?
We should add scripts to generate unit tests from a template (preferably bash file or linked into Makefile)
## Why is this needed?
Write tests faster ;)
| test | template scripts for unit tests welcome we kindly ask you to fill out the issue template below use the chat and talk to us if you have a question rather than a bug or feature request the chat room is at thanks for understanding and for contributing to the project which functionality do you think we should update improve we should add scripts to generate unit tests from a template preferably bash file or linked into makefile why is this needed write tests faster | 1 |
211,339 | 16,202,815,411 | IssuesEvent | 2021-05-05 00:31:26 | abaoa/abaoa-blog-comments | https://api.github.com/repos/abaoa/abaoa-blog-comments | opened | 对象内存结构及虚函数表分析 | /abaoa/2021/05/02/test.html gitment | https://abaoa.ltd/abaoa/2021/05/02/test.html
最近抽空浏览了一遍《COM 原理与应用》,一本老书了,COM 技术在我工作中运用得不多,但是接口设计规范和标准这一套东西还是能带给我一些有用的实践经验和启发的。在读到第二章《COM 对象和接口》的时候,看到虚函数表的一些相关知识,这些之前倒是也都知道,但是从来没有试着自己描述过,所以老觉得理解得不够彻底,那么……就试着结合一些小的代码段描述一下看,权当笔记加深记忆。 | 1.0 | 对象内存结构及虚函数表分析 - https://abaoa.ltd/abaoa/2021/05/02/test.html
最近抽空浏览了一遍《COM 原理与应用》,一本老书了,COM 技术在我工作中运用得不多,但是接口设计规范和标准这一套东西还是能带给我一些有用的实践经验和启发的。在读到第二章《COM 对象和接口》的时候,看到虚函数表的一些相关知识,这些之前倒是也都知道,但是从来没有试着自己描述过,所以老觉得理解得不够彻底,那么……就试着结合一些小的代码段描述一下看,权当笔记加深记忆。 | test | 对象内存结构及虚函数表分析 最近抽空浏览了一遍《com 原理与应用》,一本老书了,com 技术在我工作中运用得不多,但是接口设计规范和标准这一套东西还是能带给我一些有用的实践经验和启发的。在读到第二章《com 对象和接口》的时候,看到虚函数表的一些相关知识,这些之前倒是也都知道,但是从来没有试着自己描述过,所以老觉得理解得不够彻底,那么……就试着结合一些小的代码段描述一下看,权当笔记加深记忆。 | 1 |
95,748 | 10,887,269,832 | IssuesEvent | 2019-11-18 14:15:05 | keepassxreboot/keepassxc | https://api.github.com/repos/keepassxreboot/keepassxc | reopened | Update INSTALL.md to include xvfb testing instructions | documentation :bookmark_tabs: not a bug | On a Debian oldstable based distribution x64, I tried to build the last version (2.5.1).
The options for cmake were to keep from plugins, only the update check:
`cmake -DCMAKE_VERBOSE_MAKEFILE=OFF -DCMAKE_BUILD_TYPE=Release -DWITH_APP_BUNDLE=OFF -DWITH_XC_AUTOTYPE=OFF -DWITH_XC_NETWORKING=ON -DWITH_XC_UPDATECHECK=ON -DWITH_GUI_TESTS=ON ..`
Cmake and make completed successfully.
However, tests failed:
```
94% tests passed, 2 tests failed out of 34
Total Test time (real) = 63.37 sec
The following tests FAILED:
32 - testcli (Failed)
33 - testgui (Failed)
Errors while running CTest
Makefile:116: recipe for target 'test' failed
make: *** [test] Error 8
```
| 1.0 | Update INSTALL.md to include xvfb testing instructions - On a Debian oldstable based distribution x64, I tried to build the last version (2.5.1).
The options for cmake were to keep from plugins, only the update check:
`cmake -DCMAKE_VERBOSE_MAKEFILE=OFF -DCMAKE_BUILD_TYPE=Release -DWITH_APP_BUNDLE=OFF -DWITH_XC_AUTOTYPE=OFF -DWITH_XC_NETWORKING=ON -DWITH_XC_UPDATECHECK=ON -DWITH_GUI_TESTS=ON ..`
Cmake and make completed successfully.
However, tests failed:
```
94% tests passed, 2 tests failed out of 34
Total Test time (real) = 63.37 sec
The following tests FAILED:
32 - testcli (Failed)
33 - testgui (Failed)
Errors while running CTest
Makefile:116: recipe for target 'test' failed
make: *** [test] Error 8
```
| non_test | update install md to include xvfb testing instructions on a debian oldstable based distribution i tried to build the last version the options for cmake were to keep from plugins only the update check cmake dcmake verbose makefile off dcmake build type release dwith app bundle off dwith xc autotype off dwith xc networking on dwith xc updatecheck on dwith gui tests on cmake and make completed successfully however tests failed tests passed tests failed out of total test time real sec the following tests failed testcli failed testgui failed errors while running ctest makefile recipe for target test failed make error | 0 |
56,246 | 6,510,562,918 | IssuesEvent | 2017-08-25 04:26:50 | sagemathinc/cocalc | https://api.github.com/repos/sagemathinc/cocalc | closed | kucalc sagews pytest failure | A-backend A-kucalc A-testing I-bug | sagews pytest file fails in kucalc (passes in classic with `53 passed, 5 skipped in 40.65 seconds`)
in kucalc:
```
$ grep "^__version__" /usr/local/lib/python2.7/dist-packages/OpenSSL/version.py
__version__ = '0.15.1'
```
in classic:
```
$ grep "^__version__" /usr/local/lib/python2.7/dist-packages/OpenSSL/version.py
__version__ = "16.2.0"
```
```
~/cocalc/src/smc_sagews/smc_sagews/tests$ python -m pytest test_sagews.py
Traceback (most recent call last):
File "/usr/lib/python2.7/runpy.py", line 174, in _run_module_as_main
"__main__", fname, loader, pkg_name)
File "/usr/lib/python2.7/runpy.py", line 72, in _run_code
exec code in run_globals
File "/usr/local/lib/python2.7/dist-packages/pytest.py", line 73, in <module>
raise SystemExit(pytest.main())
File "/usr/local/lib/python2.7/dist-packages/_pytest/config.py", line 49, in main
config = _prepareconfig(args, plugins)
File "/usr/local/lib/python2.7/dist-packages/_pytest/config.py", line 168, in _prepareconfig
pluginmanager=pluginmanager, args=args)
File "/usr/local/lib/python2.7/dist-packages/_pytest/vendored_packages/pluggy.py", line 745, in __call__
return self._hookexec(self, self._nonwrappers + self._wrappers, kwargs)
File "/usr/local/lib/python2.7/dist-packages/_pytest/vendored_packages/pluggy.py", line 339, in _hookexec
return self._inner_hookexec(hook, methods, kwargs)
File "/usr/local/lib/python2.7/dist-packages/_pytest/vendored_packages/pluggy.py", line 334, in <lambda>
_MultiCall(methods, kwargs, hook.spec_opts).execute()
File "/usr/local/lib/python2.7/dist-packages/_pytest/vendored_packages/pluggy.py", line 613, in execute
return _wrapped_call(hook_impl.function(*args), self.execute)
File "/usr/local/lib/python2.7/dist-packages/_pytest/vendored_packages/pluggy.py", line 250, in _wrapped_call
wrap_controller.send(call_outcome)
File "/usr/local/lib/python2.7/dist-packages/_pytest/helpconfig.py", line 68, in pytest_cmdline_parse
config = outcome.get_result()
File "/usr/local/lib/python2.7/dist-packages/_pytest/vendored_packages/pluggy.py", line 280, in get_result
_reraise(*ex) # noqa
File "/usr/local/lib/python2.7/dist-packages/_pytest/vendored_packages/pluggy.py", line 265, in __init__
self.result = func()
File "/usr/local/lib/python2.7/dist-packages/_pytest/vendored_packages/pluggy.py", line 614, in execute
res = hook_impl.function(*args)
File "/usr/local/lib/python2.7/dist-packages/_pytest/config.py", line 957, in pytest_cmdline_parse
self.parse(args)
File "/usr/local/lib/python2.7/dist-packages/_pytest/vendored_packages/pluggy.py", line 510, in load_setuptools_entrypoints
plugin = ep.load()
File "/usr/local/lib/python2.7/dist-packages/pkg_resources/__init__.py", line 2316, in load
return self.resolve()
File "/usr/local/lib/python2.7/dist-packages/pkg_resources/__init__.py", line 2322, in resolve
module = __import__(self.module_name, fromlist=['__name__'], level=0)
File "/usr/local/lib/python2.7/dist-packages/_pytest/assertion/rewrite.py", line 212, in load_module
py.builtin.exec_(co, mod.__dict__)
File "/usr/local/lib/python2.7/dist-packages/py/_builtin.py", line 221, in exec_
exec2(obj, globals, locals)
File "<string>", line 7, in exec2
File "/usr/local/lib/python2.7/dist-packages/pytest_selenium/drivers/crossbrowsertesting.py", line 7, in <module>
import requests
File "/usr/local/lib/python2.7/dist-packages/requests/__init__.py", line 84, in <module>
from urllib3.contrib import pyopenssl
File "/usr/local/lib/python2.7/dist-packages/urllib3/contrib/pyopenssl.py", line 46, in <module>
import OpenSSL.SSL
File "/usr/local/lib/python2.7/dist-packages/OpenSSL/__init__.py", line 8, in <module>
from OpenSSL import rand, crypto, SSL
File "/usr/local/lib/python2.7/dist-packages/OpenSSL/SSL.py", line 118, in <module>
SSL_ST_INIT = _lib.SSL_ST_INIT
AttributeError: 'module' object has no attribute 'SSL_ST_INIT'
``` | 1.0 | kucalc sagews pytest failure - sagews pytest file fails in kucalc (passes in classic with `53 passed, 5 skipped in 40.65 seconds`)
in kucalc:
```
$ grep "^__version__" /usr/local/lib/python2.7/dist-packages/OpenSSL/version.py
__version__ = '0.15.1'
```
in classic:
```
$ grep "^__version__" /usr/local/lib/python2.7/dist-packages/OpenSSL/version.py
__version__ = "16.2.0"
```
```
~/cocalc/src/smc_sagews/smc_sagews/tests$ python -m pytest test_sagews.py
Traceback (most recent call last):
File "/usr/lib/python2.7/runpy.py", line 174, in _run_module_as_main
"__main__", fname, loader, pkg_name)
File "/usr/lib/python2.7/runpy.py", line 72, in _run_code
exec code in run_globals
File "/usr/local/lib/python2.7/dist-packages/pytest.py", line 73, in <module>
raise SystemExit(pytest.main())
File "/usr/local/lib/python2.7/dist-packages/_pytest/config.py", line 49, in main
config = _prepareconfig(args, plugins)
File "/usr/local/lib/python2.7/dist-packages/_pytest/config.py", line 168, in _prepareconfig
pluginmanager=pluginmanager, args=args)
File "/usr/local/lib/python2.7/dist-packages/_pytest/vendored_packages/pluggy.py", line 745, in __call__
return self._hookexec(self, self._nonwrappers + self._wrappers, kwargs)
File "/usr/local/lib/python2.7/dist-packages/_pytest/vendored_packages/pluggy.py", line 339, in _hookexec
return self._inner_hookexec(hook, methods, kwargs)
File "/usr/local/lib/python2.7/dist-packages/_pytest/vendored_packages/pluggy.py", line 334, in <lambda>
_MultiCall(methods, kwargs, hook.spec_opts).execute()
File "/usr/local/lib/python2.7/dist-packages/_pytest/vendored_packages/pluggy.py", line 613, in execute
return _wrapped_call(hook_impl.function(*args), self.execute)
File "/usr/local/lib/python2.7/dist-packages/_pytest/vendored_packages/pluggy.py", line 250, in _wrapped_call
wrap_controller.send(call_outcome)
File "/usr/local/lib/python2.7/dist-packages/_pytest/helpconfig.py", line 68, in pytest_cmdline_parse
config = outcome.get_result()
File "/usr/local/lib/python2.7/dist-packages/_pytest/vendored_packages/pluggy.py", line 280, in get_result
_reraise(*ex) # noqa
File "/usr/local/lib/python2.7/dist-packages/_pytest/vendored_packages/pluggy.py", line 265, in __init__
self.result = func()
File "/usr/local/lib/python2.7/dist-packages/_pytest/vendored_packages/pluggy.py", line 614, in execute
res = hook_impl.function(*args)
File "/usr/local/lib/python2.7/dist-packages/_pytest/config.py", line 957, in pytest_cmdline_parse
self.parse(args)
File "/usr/local/lib/python2.7/dist-packages/_pytest/vendored_packages/pluggy.py", line 510, in load_setuptools_entrypoints
plugin = ep.load()
File "/usr/local/lib/python2.7/dist-packages/pkg_resources/__init__.py", line 2316, in load
return self.resolve()
File "/usr/local/lib/python2.7/dist-packages/pkg_resources/__init__.py", line 2322, in resolve
module = __import__(self.module_name, fromlist=['__name__'], level=0)
File "/usr/local/lib/python2.7/dist-packages/_pytest/assertion/rewrite.py", line 212, in load_module
py.builtin.exec_(co, mod.__dict__)
File "/usr/local/lib/python2.7/dist-packages/py/_builtin.py", line 221, in exec_
exec2(obj, globals, locals)
File "<string>", line 7, in exec2
File "/usr/local/lib/python2.7/dist-packages/pytest_selenium/drivers/crossbrowsertesting.py", line 7, in <module>
import requests
File "/usr/local/lib/python2.7/dist-packages/requests/__init__.py", line 84, in <module>
from urllib3.contrib import pyopenssl
File "/usr/local/lib/python2.7/dist-packages/urllib3/contrib/pyopenssl.py", line 46, in <module>
import OpenSSL.SSL
File "/usr/local/lib/python2.7/dist-packages/OpenSSL/__init__.py", line 8, in <module>
from OpenSSL import rand, crypto, SSL
File "/usr/local/lib/python2.7/dist-packages/OpenSSL/SSL.py", line 118, in <module>
SSL_ST_INIT = _lib.SSL_ST_INIT
AttributeError: 'module' object has no attribute 'SSL_ST_INIT'
``` | test | kucalc sagews pytest failure sagews pytest file fails in kucalc passes in classic with passed skipped in seconds in kucalc grep version usr local lib dist packages openssl version py version in classic grep version usr local lib dist packages openssl version py version cocalc src smc sagews smc sagews tests python m pytest test sagews py traceback most recent call last file usr lib runpy py line in run module as main main fname loader pkg name file usr lib runpy py line in run code exec code in run globals file usr local lib dist packages pytest py line in raise systemexit pytest main file usr local lib dist packages pytest config py line in main config prepareconfig args plugins file usr local lib dist packages pytest config py line in prepareconfig pluginmanager pluginmanager args args file usr local lib dist packages pytest vendored packages pluggy py line in call return self hookexec self self nonwrappers self wrappers kwargs file usr local lib dist packages pytest vendored packages pluggy py line in hookexec return self inner hookexec hook methods kwargs file usr local lib dist packages pytest vendored packages pluggy py line in multicall methods kwargs hook spec opts execute file usr local lib dist packages pytest vendored packages pluggy py line in execute return wrapped call hook impl function args self execute file usr local lib dist packages pytest vendored packages pluggy py line in wrapped call wrap controller send call outcome file usr local lib dist packages pytest helpconfig py line in pytest cmdline parse config outcome get result file usr local lib dist packages pytest vendored packages pluggy py line in get result reraise ex noqa file usr local lib dist packages pytest vendored packages pluggy py line in init self result func file usr local lib dist packages pytest vendored packages pluggy py line in execute res hook impl function args file usr local lib dist packages pytest config py line in pytest cmdline parse self parse args file usr local lib dist packages pytest vendored packages pluggy py line in load setuptools entrypoints plugin ep load file usr local lib dist packages pkg resources init py line in load return self resolve file usr local lib dist packages pkg resources init py line in resolve module import self module name fromlist level file usr local lib dist packages pytest assertion rewrite py line in load module py builtin exec co mod dict file usr local lib dist packages py builtin py line in exec obj globals locals file line in file usr local lib dist packages pytest selenium drivers crossbrowsertesting py line in import requests file usr local lib dist packages requests init py line in from contrib import pyopenssl file usr local lib dist packages contrib pyopenssl py line in import openssl ssl file usr local lib dist packages openssl init py line in from openssl import rand crypto ssl file usr local lib dist packages openssl ssl py line in ssl st init lib ssl st init attributeerror module object has no attribute ssl st init | 1 |
67,763 | 3,281,887,251 | IssuesEvent | 2015-10-28 01:09:30 | sceene/sprint-planning | https://api.github.com/repos/sceene/sprint-planning | closed | 3. HOME FEED | MEDIUM PRIORITY | ## Home Feed
- [x] 1. Update Home Feed (Check with Steve this is latest, it should be)
http://api.sceene.com:5000/v1/home/feed?key=oPKlL9P3XOdBWr2isnboRav27KKpMVrXNKtCARGf&UUID=12010881061059250338
- [ ] 2. Make sure the avatars are being used and resized correctly via imgix
- [ ] 3. When posting a comment the user avatar isn't working
- [ ] Is this being faked in the app, or not done through the API? If so, remove that improper comment posting process.
- [ ] 4. When liking the card it's blowing the header and comments section up.
- [ ] 5. We're going to go over the new home card layout separately. This will be it's own issue.
## 3. Comment Upload

## 5. Liking Issue
 | 1.0 | 3. HOME FEED - ## Home Feed
- [x] 1. Update Home Feed (Check with Steve this is latest, it should be)
http://api.sceene.com:5000/v1/home/feed?key=oPKlL9P3XOdBWr2isnboRav27KKpMVrXNKtCARGf&UUID=12010881061059250338
- [ ] 2. Make sure the avatars are being used and resized correctly via imgix
- [ ] 3. When posting a comment the user avatar isn't working
- [ ] Is this being faked in the app, or not done through the API? If so, remove that improper comment posting process.
- [ ] 4. When liking the card it's blowing the header and comments section up.
- [ ] 5. We're going to go over the new home card layout separately. This will be it's own issue.
## 3. Comment Upload

## 5. Liking Issue
 | non_test | home feed home feed update home feed check with steve this is latest it should be make sure the avatars are being used and resized correctly via imgix when posting a comment the user avatar isn t working is this being faked in the app or not done through the api if so remove that improper comment posting process when liking the card it s blowing the header and comments section up we re going to go over the new home card layout separately this will be it s own issue comment upload liking issue | 0 |
195,183 | 14,706,450,696 | IssuesEvent | 2021-01-04 19:52:27 | Azure/azure-sdk-for-js | https://api.github.com/repos/Azure/azure-sdk-for-js | closed | Azure Tables Samples Issues | Client Docs Tables bug test-manual-pass | 1.
Section [Link1](https://github.com/Azure/azure-sdk-for-js/blob/master/sdk/tables/data-tables/samples/javascript/src/authenticationMethods.js), [Link2](https://github.com/Azure/azure-sdk-for-js/blob/master/sdk/tables/data-tables/samples/javascript/sample.env)


Reason:
The variable TABLES_URL and SAS_TOKEN not exist in the file "sample.env".
Suggestion:
Add the variable TABLES_URL and SAS_TOKEN:
```
TABLES_URL=
SAS_TOKEN=
```
@jongio for notification. | 1.0 | Azure Tables Samples Issues - 1.
Section [Link1](https://github.com/Azure/azure-sdk-for-js/blob/master/sdk/tables/data-tables/samples/javascript/src/authenticationMethods.js), [Link2](https://github.com/Azure/azure-sdk-for-js/blob/master/sdk/tables/data-tables/samples/javascript/sample.env)


Reason:
The variable TABLES_URL and SAS_TOKEN not exist in the file "sample.env".
Suggestion:
Add the variable TABLES_URL and SAS_TOKEN:
```
TABLES_URL=
SAS_TOKEN=
```
@jongio for notification. | test | azure tables samples issues section reason the variable tables url and sas token not exist in the file sample env suggestion add the variable tables url and sas token tables url sas token jongio for notification | 1 |
217,272 | 16,848,847,985 | IssuesEvent | 2021-06-20 04:17:07 | hakehuang/infoflow | https://api.github.com/repos/hakehuang/infoflow | opened |
tests-ci :kernel.memory_protection.new_user_thread_with_in_use_stack_obj : zephyr-v2.6.0-286-g46029914a7ac: lpcxpresso55s28: test Timeout
| area: Tests |
**Describe the bug**
kernel.memory_protection.new_user_thread_with_in_use_stack_obj test is Timeout on zephyr-v2.6.0-286-g46029914a7ac on lpcxpresso55s28
see logs for details
**To Reproduce**
1.
```
scripts/twister --device-testing --device-serial /dev/ttyACM0 -p lpcxpresso55s28 --testcase-root tests --sub-test kernel.memory_protection
```
2. See error
**Expected behavior**
test pass
**Impact**
**Logs and console output**
```
-
*** Booting Zephyr OS build zephyr-v2.6.0-286-g46029914a7ac ***
Running test suite memory_protection_test_suite
===================================================================
START - test_permission_inheritance
ASSERTION FAIL [esf != ((void *)0)] @ WEST_TOPDIR/zephyr/arch/arm/core/aarch32/cortex_m/fault.c:993
ESF could not be retrieved successfully. Shall never occur.
ASSERTION FAIL [esf != ((void *)0)] @ WEST_TOPDIR/zephyr/arch/arm/core/aarch32/cortex_m/fault.c:993
ESF could not be retrieved successfully. Shall never occur.
```
**Environment (please complete the following information):**
- OS: (e.g. Linux )
- Toolchain (e.g Zephyr SDK)
- Commit SHA or Version used: zephyr-v2.6.0-286-g46029914a7ac
| 1.0 |
tests-ci :kernel.memory_protection.new_user_thread_with_in_use_stack_obj : zephyr-v2.6.0-286-g46029914a7ac: lpcxpresso55s28: test Timeout
-
**Describe the bug**
kernel.memory_protection.new_user_thread_with_in_use_stack_obj test is Timeout on zephyr-v2.6.0-286-g46029914a7ac on lpcxpresso55s28
see logs for details
**To Reproduce**
1.
```
scripts/twister --device-testing --device-serial /dev/ttyACM0 -p lpcxpresso55s28 --testcase-root tests --sub-test kernel.memory_protection
```
2. See error
**Expected behavior**
test pass
**Impact**
**Logs and console output**
```
-
*** Booting Zephyr OS build zephyr-v2.6.0-286-g46029914a7ac ***
Running test suite memory_protection_test_suite
===================================================================
START - test_permission_inheritance
ASSERTION FAIL [esf != ((void *)0)] @ WEST_TOPDIR/zephyr/arch/arm/core/aarch32/cortex_m/fault.c:993
ESF could not be retrieved successfully. Shall never occur.
ASSERTION FAIL [esf != ((void *)0)] @ WEST_TOPDIR/zephyr/arch/arm/core/aarch32/cortex_m/fault.c:993
ESF could not be retrieved successfully. Shall never occur.
```
**Environment (please complete the following information):**
- OS: (e.g. Linux )
- Toolchain (e.g Zephyr SDK)
- Commit SHA or Version used: zephyr-v2.6.0-286-g46029914a7ac
| test | tests ci kernel memory protection new user thread with in use stack obj zephyr test timeout describe the bug kernel memory protection new user thread with in use stack obj test is timeout on zephyr on see logs for details to reproduce scripts twister device testing device serial dev p testcase root tests sub test kernel memory protection see error expected behavior test pass impact logs and console output booting zephyr os build zephyr running test suite memory protection test suite start test permission inheritance assertion fail west topdir zephyr arch arm core cortex m fault c esf could not be retrieved successfully shall never occur assertion fail west topdir zephyr arch arm core cortex m fault c esf could not be retrieved successfully shall never occur environment please complete the following information os e g linux toolchain e g zephyr sdk commit sha or version used zephyr | 1 |
302,585 | 26,152,057,578 | IssuesEvent | 2022-12-30 15:12:26 | department-of-veterans-affairs/va.gov-cms | https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms | closed | Testing: Add test for Lovell listing pages to contain both federal events and events specific to either VA or Tricare. | VA.gov frontend ⭐️ Facilities testing VA Lovell | Since we can not guarantee that naturally occurring data will always have Lovell Federal, TRICARE and VA, we have to use mock data for this.
Seems like a u nit test, not end to end test
## ACs
- [ ] Test ensures Lovell VA Events page has content from Lovell Federal and Lovell VA
- [ ] Test ensures Lovell TRICARE Events page has content from Lovell Federal and Lovell TRICARE
- [ ] Test ensures Lovell VA Stories page has content from Lovell Federal and Lovell VA
- [ ] Test ensures Lovell TRICARE Stories page has content from Lovell Federal and Lovell TRICARE
- [ ] Test ensures Lovell VA News page has content from Lovell Federal and Lovell VA
- [ ] Test ensures Lovell TRICARE News page has content from Lovell Federal and Lovell TRICARE | 1.0 | Testing: Add test for Lovell listing pages to contain both federal events and events specific to either VA or Tricare. - Since we can not guarantee that naturally occurring data will always have Lovell Federal, TRICARE and VA, we have to use mock data for this.
Seems like a u nit test, not end to end test
## ACs
- [ ] Test ensures Lovell VA Events page has content from Lovell Federal and Lovell VA
- [ ] Test ensures Lovell TRICARE Events page has content from Lovell Federal and Lovell TRICARE
- [ ] Test ensures Lovell VA Stories page has content from Lovell Federal and Lovell VA
- [ ] Test ensures Lovell TRICARE Stories page has content from Lovell Federal and Lovell TRICARE
- [ ] Test ensures Lovell VA News page has content from Lovell Federal and Lovell VA
- [ ] Test ensures Lovell TRICARE News page has content from Lovell Federal and Lovell TRICARE | test | testing add test for lovell listing pages to contain both federal events and events specific to either va or tricare since we can not guarantee that naturally occurring data will always have lovell federal tricare and va we have to use mock data for this seems like a u nit test not end to end test acs test ensures lovell va events page has content from lovell federal and lovell va test ensures lovell tricare events page has content from lovell federal and lovell tricare test ensures lovell va stories page has content from lovell federal and lovell va test ensures lovell tricare stories page has content from lovell federal and lovell tricare test ensures lovell va news page has content from lovell federal and lovell va test ensures lovell tricare news page has content from lovell federal and lovell tricare | 1 |
70,407 | 8,531,919,305 | IssuesEvent | 2018-11-04 16:51:12 | the-tale/the-tale | https://api.github.com/repos/the-tale/the-tale | closed | При заточке артефакта, его максимальная целостность не может упасть ниже 49 | comp_game_logic cont_game_designe decision required est_simple good first issue type_bug | Возможно, следует разрешить снижать её до минимума (до 1). | 1.0 | При заточке артефакта, его максимальная целостность не может упасть ниже 49 - Возможно, следует разрешить снижать её до минимума (до 1). | non_test | при заточке артефакта его максимальная целостность не может упасть ниже возможно следует разрешить снижать её до минимума до | 0 |
725,591 | 24,967,463,085 | IssuesEvent | 2022-11-01 20:44:11 | Greenstand/treetracker-admin-api | https://api.github.com/repos/Greenstand/treetracker-admin-api | closed | Species: Only query capture count if user has permission to manage species | good first issue priority tool: Species size: small type: enhancement | We currently augment every species list query with the capture counts for each species:
https://github.com/Greenstand/treetracker-admin-api/blob/a400e8f2fa894de5d1bc11dec545e2e37037d578/src/controllers/species.controller.ts#L68
This is much more efficient than having the client query for each species individually, but adds several seconds onto the initial query.
Currently, the Species tool is only available to users with the `super_permission` policy in their roles (although soon that will include those with `manage_species`), which is the only place the counts are displayed in the client.
Therefore, there's no need to include these counts when the user doesn't have one of those policies.
That would significantly reduce the loading time for Organization users in particular. | 1.0 | Species: Only query capture count if user has permission to manage species - We currently augment every species list query with the capture counts for each species:
https://github.com/Greenstand/treetracker-admin-api/blob/a400e8f2fa894de5d1bc11dec545e2e37037d578/src/controllers/species.controller.ts#L68
This is much more efficient than having the client query for each species individually, but adds several seconds onto the initial query.
Currently, the Species tool is only available to users with the `super_permission` policy in their roles (although soon that will include those with `manage_species`), which is the only place the counts are displayed in the client.
Therefore, there's no need to include these counts when the user doesn't have one of those policies.
That would significantly reduce the loading time for Organization users in particular. | non_test | species only query capture count if user has permission to manage species we currently augment every species list query with the capture counts for each species this is much more efficient than having the client query for each species individually but adds several seconds onto the initial query currently the species tool is only available to users with the super permission policy in their roles although soon that will include those with manage species which is the only place the counts are displayed in the client therefore there s no need to include these counts when the user doesn t have one of those policies that would significantly reduce the loading time for organization users in particular | 0 |
426,182 | 29,511,664,570 | IssuesEvent | 2023-06-04 01:45:33 | karmada-io/karmada | https://api.github.com/repos/karmada-io/karmada | closed | kubectl-karmada init failing with error | kind/documentation kind/cleanup | <!-- Please use this template while reporting a bug and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks!
-->
**What happened**:
`kubectl-karmada` when invoking `init` is failing with error **error: failed to create directory: /etc/karmada, error: mkdir /etc/karmada: permission denied**
**What you expected to happen**:
`init` command should successfully install Karmada control plane
**How to reproduce it (as minimally and precisely as possible)**:
1. Install single node k3s cluster
2. Install `kubectl-karmada`
3. Run `kubectl karmada init`
**Anything else we need to know?**:
**Environment**:
- kubectl-karmada or karmadactl version (the result of `kubectl-karmada version` or `karmadactl version`):
```sh
kubectl karmada version: version.Info{GitVersion:"v1.6.0", GitCommit:"6eb79b38949e480cf7a2e12cfa56fef47bda81ea", GitTreeState:"clean", BuildDate:"2023-05-31T09:55:31Z", GoVersion:"go1.20.4", Compiler:"gc", Platform:"linux/amd64"}
```
| 1.0 | kubectl-karmada init failing with error - <!-- Please use this template while reporting a bug and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks!
-->
**What happened**:
`kubectl-karmada` when invoking `init` is failing with error **error: failed to create directory: /etc/karmada, error: mkdir /etc/karmada: permission denied**
**What you expected to happen**:
`init` command should successfully install Karmada control plane
**How to reproduce it (as minimally and precisely as possible)**:
1. Install single node k3s cluster
2. Install `kubectl-karmada`
3. Run `kubectl karmada init`
**Anything else we need to know?**:
**Environment**:
- kubectl-karmada or karmadactl version (the result of `kubectl-karmada version` or `karmadactl version`):
```sh
kubectl karmada version: version.Info{GitVersion:"v1.6.0", GitCommit:"6eb79b38949e480cf7a2e12cfa56fef47bda81ea", GitTreeState:"clean", BuildDate:"2023-05-31T09:55:31Z", GoVersion:"go1.20.4", Compiler:"gc", Platform:"linux/amd64"}
```
| non_test | kubectl karmada init failing with error please use this template while reporting a bug and provide as much info as possible not doing so may result in your bug not being addressed in a timely manner thanks what happened kubectl karmada when invoking init is failing with error error failed to create directory etc karmada error mkdir etc karmada permission denied what you expected to happen init command should successfully install karmada control plane how to reproduce it as minimally and precisely as possible install single node cluster install kubectl karmada run kubectl karmada init anything else we need to know environment kubectl karmada or karmadactl version the result of kubectl karmada version or karmadactl version sh kubectl karmada version version info gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux | 0 |
246,204 | 20,828,804,723 | IssuesEvent | 2022-03-19 04:21:36 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | System.Threading.ThreadPools.Tests.ThreadPoolTests.CooperativeBlockingCanCreateThreadsFaster has failed | area-System.Threading in-pr test-failure | Failing in rolling build: Libraries Test Run release coreclr OSX x64 Release, but not every time. In [20220315.1](https://dev.azure.com/dnceng/public/_build/results?buildId=1664004&view=logs&j=05e92ac1-194e-59cf-664a-fa72d1cdd19b&t=caea1d4b-c90c-5be1-e57d-c4635079c333) and [20220315.4](https://dev.azure.com/dnceng/public/_build/results?buildId=1665195&view=logs&j=05e92ac1-194e-59cf-664a-fa72d1cdd19b&t=caea1d4b-c90c-5be1-e57d-c4635079c333) it failed, then in [20220316.1](https://dev.azure.com/dnceng/public/_build/results?buildId=1666244&view=logs&j=05e92ac1-194e-59cf-664a-fa72d1cdd19b&t=caea1d4b-c90c-5be1-e57d-c4635079c333) it passed and in [20220316.4](https://dev.azure.com/dnceng/public/_build/results?buildId=1667233&view=logs&j=05e92ac1-194e-59cf-664a-fa72d1cdd19b&t=caea1d4b-c90c-5be1-e57d-c4635079c333) it failed again.
The error:
```
Unhandled exception. Xunit.Sdk.TrueException: Assert.True() Failure
Expected: True
Actual: False
at System.Threading.ThreadPools.Tests.ThreadPoolTests.<>c.<CooperativeBlockingCanCreateThreadsFaster>b__38_0() in /_/src/libraries/System.Threading.ThreadPool/tests/ThreadPoolTests.cs:line 953
--- End of stack trace from previous location ---
at Microsoft.DotNet.RemoteExecutor.Program.Main(String[] args) in /_/src/Microsoft.DotNet.RemoteExecutor/src/Program.cs:line 94
System.Threading.ThreadPools.Tests.ThreadPoolTests.CooperativeBlockingCanCreateThreadsFaster [FAIL]
Microsoft.DotNet.RemoteExecutor.RemoteExecutionException : Remote process failed with an unhandled exception.
Stack Trace:
Child exception:
Xunit.Sdk.TrueException: Assert.True() Failure
Expected: True
Actual: False
/_/src/libraries/System.Threading.ThreadPool/tests/ThreadPoolTests.cs(953,0): at System.Threading.ThreadPools.Tests.ThreadPoolTests.<>c.<CooperativeBlockingCanCreateThreadsFaster>b__38_0()
Child process:
System.Threading.ThreadPool.Tests, Version=7.0.0.0, Culture=neutral, PublicKeyToken=cc7b13ffcd2ddd51 System.Threading.ThreadPools.Tests.ThreadPoolTests+<>c Void <CooperativeBlockingCanCreateThreadsFaster>b__38_0()
```
| 1.0 | System.Threading.ThreadPools.Tests.ThreadPoolTests.CooperativeBlockingCanCreateThreadsFaster has failed - Failing in rolling build: Libraries Test Run release coreclr OSX x64 Release, but not every time. In [20220315.1](https://dev.azure.com/dnceng/public/_build/results?buildId=1664004&view=logs&j=05e92ac1-194e-59cf-664a-fa72d1cdd19b&t=caea1d4b-c90c-5be1-e57d-c4635079c333) and [20220315.4](https://dev.azure.com/dnceng/public/_build/results?buildId=1665195&view=logs&j=05e92ac1-194e-59cf-664a-fa72d1cdd19b&t=caea1d4b-c90c-5be1-e57d-c4635079c333) it failed, then in [20220316.1](https://dev.azure.com/dnceng/public/_build/results?buildId=1666244&view=logs&j=05e92ac1-194e-59cf-664a-fa72d1cdd19b&t=caea1d4b-c90c-5be1-e57d-c4635079c333) it passed and in [20220316.4](https://dev.azure.com/dnceng/public/_build/results?buildId=1667233&view=logs&j=05e92ac1-194e-59cf-664a-fa72d1cdd19b&t=caea1d4b-c90c-5be1-e57d-c4635079c333) it failed again.
The error:
```
Unhandled exception. Xunit.Sdk.TrueException: Assert.True() Failure
Expected: True
Actual: False
at System.Threading.ThreadPools.Tests.ThreadPoolTests.<>c.<CooperativeBlockingCanCreateThreadsFaster>b__38_0() in /_/src/libraries/System.Threading.ThreadPool/tests/ThreadPoolTests.cs:line 953
--- End of stack trace from previous location ---
at Microsoft.DotNet.RemoteExecutor.Program.Main(String[] args) in /_/src/Microsoft.DotNet.RemoteExecutor/src/Program.cs:line 94
System.Threading.ThreadPools.Tests.ThreadPoolTests.CooperativeBlockingCanCreateThreadsFaster [FAIL]
Microsoft.DotNet.RemoteExecutor.RemoteExecutionException : Remote process failed with an unhandled exception.
Stack Trace:
Child exception:
Xunit.Sdk.TrueException: Assert.True() Failure
Expected: True
Actual: False
/_/src/libraries/System.Threading.ThreadPool/tests/ThreadPoolTests.cs(953,0): at System.Threading.ThreadPools.Tests.ThreadPoolTests.<>c.<CooperativeBlockingCanCreateThreadsFaster>b__38_0()
Child process:
System.Threading.ThreadPool.Tests, Version=7.0.0.0, Culture=neutral, PublicKeyToken=cc7b13ffcd2ddd51 System.Threading.ThreadPools.Tests.ThreadPoolTests+<>c Void <CooperativeBlockingCanCreateThreadsFaster>b__38_0()
```
| test | system threading threadpools tests threadpooltests cooperativeblockingcancreatethreadsfaster has failed failing in rolling build libraries test run release coreclr osx release but not every time in and it failed then in it passed and in it failed again the error unhandled exception xunit sdk trueexception assert true failure expected true actual false at system threading threadpools tests threadpooltests c b in src libraries system threading threadpool tests threadpooltests cs line end of stack trace from previous location at microsoft dotnet remoteexecutor program main string args in src microsoft dotnet remoteexecutor src program cs line system threading threadpools tests threadpooltests cooperativeblockingcancreatethreadsfaster microsoft dotnet remoteexecutor remoteexecutionexception remote process failed with an unhandled exception stack trace child exception xunit sdk trueexception assert true failure expected true actual false src libraries system threading threadpool tests threadpooltests cs at system threading threadpools tests threadpooltests c b child process system threading threadpool tests version culture neutral publickeytoken system threading threadpools tests threadpooltests c void b | 1 |
220,522 | 17,201,072,621 | IssuesEvent | 2021-07-17 08:29:17 | rancher/dashboard | https://api.github.com/repos/rancher/dashboard | closed | Show wechat info if zh-hans locale | [zube]: To Test | - In ember, when locale is zh-hans, an additional 'forum' is shown for wechat. Clicking on link shows a modal containing a 2d barcode
- Need to replicate this functionality somehow
| 1.0 | Show wechat info if zh-hans locale - - In ember, when locale is zh-hans, an additional 'forum' is shown for wechat. Clicking on link shows a modal containing a 2d barcode
- Need to replicate this functionality somehow
| test | show wechat info if zh hans locale in ember when locale is zh hans an additional forum is shown for wechat clicking on link shows a modal containing a barcode need to replicate this functionality somehow | 1 |
31,663 | 4,713,754,858 | IssuesEvent | 2016-10-14 21:11:24 | rancher/rancher | https://api.github.com/repos/rancher/rancher | closed | Failed to deploy pods in kubernetes 1.4 | area/kubernetes kind/bug status/blocker status/resolved status/to-test | **Rancher Version:**
master image
**Docker Version:**
1.12.2
**OS and where are the hosts located? (cloud, bare metal, etc):**
Ubuntu 16.04
**Setup Details: (single node rancher vs. HA rancher, internal DB vs. external DB)**
Single node rancher
**Environment Type: (Cattle/Kubernetes/Swarm/Mesos)**
Kubernetes 1.4
**Steps to Reproduce:**
Install kubernetes 1.4 from infra catalog
try to deploy any pod
**Results:**
The pods are stuck in containercreating state and producing this error:
```
$ kubectl describe pods --namespace=testingress1
Name: k8testrc1-8fcza
Namespace: testingress1
Node: hussein-10acre-1/104.197.240.246
Start Time: Thu, 13 Oct 2016 19:22:37 +0200
Labels: k8s-app=k8test1-service
Status: Pending
IP:
Controllers: ReplicationController/k8testrc1
Containers:
testcontainer:
Container ID:
Image: sangeetha/testnewhostrouting
Image ID:
Port: 81/TCP
State: Waiting
Reason: ContainerCreating
Ready: False
Restart Count: 0
Volume Mounts:
/var/run/secrets/kubernetes.io/serviceaccount from default-token-1pc7p (ro)
Environment Variables: <none>
Conditions:
Type Status
Initialized True
Ready False
PodScheduled True
Volumes:
default-token-1pc7p:
Type: Secret (a volume populated by a Secret)
SecretName: default-token-1pc7p
QoS Class: BestEffort
Tolerations: <none>
Events:
FirstSeen LastSeen Count From SubobjectPath Type Reason Message
--------- -------- ----- ---- ------------- -------- ------ -------
56m 56m 1 {default-scheduler } Normal Scheduled Successfully assigned k8testrc1-8fcza to hussein-10acre-1
56m 4s 142 {kubelet hussein-10acre-1} Warning FailedSync Error syncing pod, skipping: failed to "SetupNetwork" for "k8testrc1-8fcza_testingress1" with SetupNetworkError: "Failed to setup network for pod \"k8testrc1-8fcza_testingress1(a2c3b262-9169-11e6-a468-02bcb92f6900)\" using network plugins \"cni\": failed to add IP addr to \"eth0\": file exists; Skipping pod"
Name: k8testrc1-wn0yy
Namespace: testingress1
Node: hussein-10acre-3/104.198.54.251
Start Time: Thu, 13 Oct 2016 19:22:37 +0200
Labels: k8s-app=k8test1-service
Status: Pending
IP:
Controllers: ReplicationController/k8testrc1
Containers:
testcontainer:
Container ID:
Image: sangeetha/testnewhostrouting
Image ID:
Port: 81/TCP
State: Waiting
Reason: ContainerCreating
Ready: False
Restart Count: 0
Volume Mounts:
/var/run/secrets/kubernetes.io/serviceaccount from default-token-1pc7p (ro)
Environment Variables: <none>
Conditions:
Type Status
Initialized True
Ready False
PodScheduled True
Volumes:
default-token-1pc7p:
Type: Secret (a volume populated by a Secret)
SecretName: default-token-1pc7p
QoS Class: BestEffort
Tolerations: <none>
Events:
FirstSeen LastSeen Count From SubobjectPath Type Reason Message
--------- -------- ----- ---- ------------- -------- ------ -------
56m 56m 1 {default-scheduler } Normal Scheduled Successfully assigned k8testrc1-wn0yy to hussein-10acre-3
56m 58s 146 {kubelet hussein-10acre-3} Warning FailedSync Error syncing pod, skipping: failed to "SetupNetwork" for "k8testrc1-wn0yy_testingress1" with SetupNetworkError: "Failed to setup network for pod \"k8testrc1-wn0yy_testingress1(a2c3e954-9169-11e6-a468-02bcb92f6900)\" using network plugins \"cni\": failed to add IP addr to \"eth0\": file exists; Skipping pod"
``` | 1.0 | Failed to deploy pods in kubernetes 1.4 - **Rancher Version:**
master image
**Docker Version:**
1.12.2
**OS and where are the hosts located? (cloud, bare metal, etc):**
Ubuntu 16.04
**Setup Details: (single node rancher vs. HA rancher, internal DB vs. external DB)**
Single node rancher
**Environment Type: (Cattle/Kubernetes/Swarm/Mesos)**
Kubernetes 1.4
**Steps to Reproduce:**
Install kubernetes 1.4 from infra catalog
try to deploy any pod
**Results:**
The pods are stuck in containercreating state and producing this error:
```
$ kubectl describe pods --namespace=testingress1
Name: k8testrc1-8fcza
Namespace: testingress1
Node: hussein-10acre-1/104.197.240.246
Start Time: Thu, 13 Oct 2016 19:22:37 +0200
Labels: k8s-app=k8test1-service
Status: Pending
IP:
Controllers: ReplicationController/k8testrc1
Containers:
testcontainer:
Container ID:
Image: sangeetha/testnewhostrouting
Image ID:
Port: 81/TCP
State: Waiting
Reason: ContainerCreating
Ready: False
Restart Count: 0
Volume Mounts:
/var/run/secrets/kubernetes.io/serviceaccount from default-token-1pc7p (ro)
Environment Variables: <none>
Conditions:
Type Status
Initialized True
Ready False
PodScheduled True
Volumes:
default-token-1pc7p:
Type: Secret (a volume populated by a Secret)
SecretName: default-token-1pc7p
QoS Class: BestEffort
Tolerations: <none>
Events:
FirstSeen LastSeen Count From SubobjectPath Type Reason Message
--------- -------- ----- ---- ------------- -------- ------ -------
56m 56m 1 {default-scheduler } Normal Scheduled Successfully assigned k8testrc1-8fcza to hussein-10acre-1
56m 4s 142 {kubelet hussein-10acre-1} Warning FailedSync Error syncing pod, skipping: failed to "SetupNetwork" for "k8testrc1-8fcza_testingress1" with SetupNetworkError: "Failed to setup network for pod \"k8testrc1-8fcza_testingress1(a2c3b262-9169-11e6-a468-02bcb92f6900)\" using network plugins \"cni\": failed to add IP addr to \"eth0\": file exists; Skipping pod"
Name: k8testrc1-wn0yy
Namespace: testingress1
Node: hussein-10acre-3/104.198.54.251
Start Time: Thu, 13 Oct 2016 19:22:37 +0200
Labels: k8s-app=k8test1-service
Status: Pending
IP:
Controllers: ReplicationController/k8testrc1
Containers:
testcontainer:
Container ID:
Image: sangeetha/testnewhostrouting
Image ID:
Port: 81/TCP
State: Waiting
Reason: ContainerCreating
Ready: False
Restart Count: 0
Volume Mounts:
/var/run/secrets/kubernetes.io/serviceaccount from default-token-1pc7p (ro)
Environment Variables: <none>
Conditions:
Type Status
Initialized True
Ready False
PodScheduled True
Volumes:
default-token-1pc7p:
Type: Secret (a volume populated by a Secret)
SecretName: default-token-1pc7p
QoS Class: BestEffort
Tolerations: <none>
Events:
FirstSeen LastSeen Count From SubobjectPath Type Reason Message
--------- -------- ----- ---- ------------- -------- ------ -------
56m 56m 1 {default-scheduler } Normal Scheduled Successfully assigned k8testrc1-wn0yy to hussein-10acre-3
56m 58s 146 {kubelet hussein-10acre-3} Warning FailedSync Error syncing pod, skipping: failed to "SetupNetwork" for "k8testrc1-wn0yy_testingress1" with SetupNetworkError: "Failed to setup network for pod \"k8testrc1-wn0yy_testingress1(a2c3e954-9169-11e6-a468-02bcb92f6900)\" using network plugins \"cni\": failed to add IP addr to \"eth0\": file exists; Skipping pod"
``` | test | failed to deploy pods in kubernetes rancher version master image docker version os and where are the hosts located cloud bare metal etc ubuntu setup details single node rancher vs ha rancher internal db vs external db single node rancher environment type cattle kubernetes swarm mesos kubernetes steps to reproduce install kubernetes from infra catalog try to deploy any pod results the pods are stuck in containercreating state and producing this error kubectl describe pods namespace name namespace node hussein start time thu oct labels app service status pending ip controllers replicationcontroller containers testcontainer container id image sangeetha testnewhostrouting image id port tcp state waiting reason containercreating ready false restart count volume mounts var run secrets kubernetes io serviceaccount from default token ro environment variables conditions type status initialized true ready false podscheduled true volumes default token type secret a volume populated by a secret secretname default token qos class besteffort tolerations events firstseen lastseen count from subobjectpath type reason message default scheduler normal scheduled successfully assigned to hussein kubelet hussein warning failedsync error syncing pod skipping failed to setupnetwork for with setupnetworkerror failed to setup network for pod using network plugins cni failed to add ip addr to file exists skipping pod name namespace node hussein start time thu oct labels app service status pending ip controllers replicationcontroller containers testcontainer container id image sangeetha testnewhostrouting image id port tcp state waiting reason containercreating ready false restart count volume mounts var run secrets kubernetes io serviceaccount from default token ro environment variables conditions type status initialized true ready false podscheduled true volumes default token type secret a volume populated by a secret secretname default token qos class besteffort tolerations events firstseen lastseen count from subobjectpath type reason message default scheduler normal scheduled successfully assigned to hussein kubelet hussein warning failedsync error syncing pod skipping failed to setupnetwork for with setupnetworkerror failed to setup network for pod using network plugins cni failed to add ip addr to file exists skipping pod | 1 |
29,270 | 8,315,454,950 | IssuesEvent | 2018-09-25 05:16:21 | GrangerHub/tremulous-game-logic | https://api.github.com/repos/GrangerHub/tremulous-game-logic | opened | Teammates don't take damage from exploding enemy buildings if a different teammate kills said building. | accepted bug buildable team | Teammates don't take damage from exploding enemy buildings if a different teammate kills said building. Likely solution is to have damage from exploding buildables applied to teammates of the buildable "attacker" treat the buildable as the attacker. | 1.0 | Teammates don't take damage from exploding enemy buildings if a different teammate kills said building. - Teammates don't take damage from exploding enemy buildings if a different teammate kills said building. Likely solution is to have damage from exploding buildables applied to teammates of the buildable "attacker" treat the buildable as the attacker. | non_test | teammates don t take damage from exploding enemy buildings if a different teammate kills said building teammates don t take damage from exploding enemy buildings if a different teammate kills said building likely solution is to have damage from exploding buildables applied to teammates of the buildable attacker treat the buildable as the attacker | 0 |
42,713 | 17,269,480,091 | IssuesEvent | 2021-07-22 17:45:01 | hashicorp/terraform-provider-aws | https://api.github.com/repos/hashicorp/terraform-provider-aws | closed | RDS PendingModifiedValues not removed when value modification is reverted | needs-triage service/rds stale | Repro:
1. Create an RDS database instance via the aws tf provider (for my example, postgres), on a version that can be upgraded from (e.g. 9.6.9). Set a maintenance window so that upgrades to the db engine version do not apply immediately. Apply.
2. Now, via terraform, upgrade the RDS db instance engine version (say to 9.6.11). Apply.
3. Observe via the aws cli that there is a PendingModifiedValues for the engine version (being set to 9.6.11 for example)
4. Attempt to undo the engine version change by reverting the terraform db instance enging version (say to 9.6.9). Apply.
5. Observe via the aws cli that the PendingModifiedValues for the engine version still is attempting to upgrade to 9.6.11
Expected:
When "downgrading" the engine version to 9.6.9 (step 4), the PendingModifiedValues upgrade to 9.6.11 is removed.
Impact:
This caused a production database of ours to be unexpectedly upgraded, since we saw the version in tf code at 9.6.9 and yet during the maintenance window the upgrade to 9.6.11 happened anyway.
Notes:
Confirmed with cloudtrail, terraform detects no update necessary in step 4 of the above repro and makes no calls to AWS. I expect the logic only checks the current version and doesn't look at pending changes.
Other details:
Terraform version: 0.11.11
AWS tf version: 1.60.0
| 1.0 | RDS PendingModifiedValues not removed when value modification is reverted - Repro:
1. Create an RDS database instance via the aws tf provider (for my example, postgres), on a version that can be upgraded from (e.g. 9.6.9). Set a maintenance window so that upgrades to the db engine version do not apply immediately. Apply.
2. Now, via terraform, upgrade the RDS db instance engine version (say to 9.6.11). Apply.
3. Observe via the aws cli that there is a PendingModifiedValues for the engine version (being set to 9.6.11 for example)
4. Attempt to undo the engine version change by reverting the terraform db instance enging version (say to 9.6.9). Apply.
5. Observe via the aws cli that the PendingModifiedValues for the engine version still is attempting to upgrade to 9.6.11
Expected:
When "downgrading" the engine version to 9.6.9 (step 4), the PendingModifiedValues upgrade to 9.6.11 is removed.
Impact:
This caused a production database of ours to be unexpectedly upgraded, since we saw the version in tf code at 9.6.9 and yet during the maintenance window the upgrade to 9.6.11 happened anyway.
Notes:
Confirmed with cloudtrail, terraform detects no update necessary in step 4 of the above repro and makes no calls to AWS. I expect the logic only checks the current version and doesn't look at pending changes.
Other details:
Terraform version: 0.11.11
AWS tf version: 1.60.0
| non_test | rds pendingmodifiedvalues not removed when value modification is reverted repro create an rds database instance via the aws tf provider for my example postgres on a version that can be upgraded from e g set a maintenance window so that upgrades to the db engine version do not apply immediately apply now via terraform upgrade the rds db instance engine version say to apply observe via the aws cli that there is a pendingmodifiedvalues for the engine version being set to for example attempt to undo the engine version change by reverting the terraform db instance enging version say to apply observe via the aws cli that the pendingmodifiedvalues for the engine version still is attempting to upgrade to expected when downgrading the engine version to step the pendingmodifiedvalues upgrade to is removed impact this caused a production database of ours to be unexpectedly upgraded since we saw the version in tf code at and yet during the maintenance window the upgrade to happened anyway notes confirmed with cloudtrail terraform detects no update necessary in step of the above repro and makes no calls to aws i expect the logic only checks the current version and doesn t look at pending changes other details terraform version aws tf version | 0 |
106,136 | 9,114,962,339 | IssuesEvent | 2019-02-22 02:35:38 | MicrosoftDocs/vsts-docs | https://api.github.com/repos/MicrosoftDocs/vsts-docs | reopened | Test Runner -- how to move this window or resize it | test | It seems to have locked itself to the leftmost part of my screen(s) and can't be moved. Is there a way to move this around my desktop? Please advise.
Thanks
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 985ecee3-e347-43bb-5ed2-08ce3aee7621
* Version Independent ID: fdf465fc-a9e1-0198-2e8c-e859c74252b3
* Content: [FAQs and problem solutions - Azure Test Plans](https://docs.microsoft.com/en-us/azure/devops/test/reference-qa?view=azure-devops)
* Content Source: [docs/test/reference-qa.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/test/reference-qa.md)
* Product: **devops**
* GitHub Login: @alexhomer1
* Microsoft Alias: **ahomer** | 1.0 | Test Runner -- how to move this window or resize it - It seems to have locked itself to the leftmost part of my screen(s) and can't be moved. Is there a way to move this around my desktop? Please advise.
Thanks
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 985ecee3-e347-43bb-5ed2-08ce3aee7621
* Version Independent ID: fdf465fc-a9e1-0198-2e8c-e859c74252b3
* Content: [FAQs and problem solutions - Azure Test Plans](https://docs.microsoft.com/en-us/azure/devops/test/reference-qa?view=azure-devops)
* Content Source: [docs/test/reference-qa.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/test/reference-qa.md)
* Product: **devops**
* GitHub Login: @alexhomer1
* Microsoft Alias: **ahomer** | test | test runner how to move this window or resize it it seems to have locked itself to the leftmost part of my screen s and can t be moved is there a way to move this around my desktop please advise thanks document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops github login microsoft alias ahomer | 1 |
152,126 | 12,091,733,186 | IssuesEvent | 2020-04-19 12:57:18 | yakrash/Billing | https://api.github.com/repos/yakrash/Billing | opened | Тестирование: Валидация транзакции | test | Необходимо добавить JUnit в проект.
И написать первый тест. В данном случае юнит тест, проверяющий валидацию.
Я б еще сделал validate => static | 1.0 | Тестирование: Валидация транзакции - Необходимо добавить JUnit в проект.
И написать первый тест. В данном случае юнит тест, проверяющий валидацию.
Я б еще сделал validate => static | test | тестирование валидация транзакции необходимо добавить junit в проект и написать первый тест в данном случае юнит тест проверяющий валидацию я б еще сделал validate static | 1 |
6,950 | 2,610,319,125 | IssuesEvent | 2015-02-26 19:43:00 | chrsmith/republic-at-war | https://api.github.com/repos/chrsmith/republic-at-war | closed | Text | auto-migrated Priority-Medium Type-Defect | ```
* Income increase upgrade for Space Station (Republic, Skirmish, Geonosis) missing description
* Prototype Designs 2 upgrade missing text (Space Skirmish, Republic)
* CIS Proton Bomb 2 upgrade text missing (CIS, Space, Korriban, Outer Rim Sieges)
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 6 May 2011 at 9:49 | 1.0 | Text - ```
* Income increase upgrade for Space Station (Republic, Skirmish, Geonosis) missing description
* Prototype Designs 2 upgrade missing text (Space Skirmish, Republic)
* CIS Proton Bomb 2 upgrade text missing (CIS, Space, Korriban, Outer Rim Sieges)
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 6 May 2011 at 9:49 | non_test | text income increase upgrade for space station republic skirmish geonosis missing description prototype designs upgrade missing text space skirmish republic cis proton bomb upgrade text missing cis space korriban outer rim sieges original issue reported on code google com by gmail com on may at | 0 |
120,408 | 10,115,043,463 | IssuesEvent | 2019-07-30 20:43:16 | NuGet/Home | https://api.github.com/repos/NuGet/Home | closed | Test: NuGet.PackageManagement.UI.Test.PackageItemLoaderTests.EmitsSearchTelemetryEvents is flaky | Area:Test Sprint 156 | Test `NuGet.PackageManagement.UI.Test.PackageItemLoaderTests.EmitsSearchTelemetryEvents` is flaky.
```
NuGet.PackageManagement.UI.Test.PackageItemLoaderTests.EmitsSearchTelemetryEvents [FAIL]
Assert.Equal() Failure
Expected: 4
Actual: 5
Stack Trace:
C:\git\NuGet.Client\test\NuGet.Clients.Tests\NuGet.PackageManagement.UI.Test\PackageItemLoaderTests.cs(107,0): at NuGet.PackageManagement.UI.Test.PackageItemLoaderTests.<EmitsSearchTelemetryEvents>d__2.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
``` | 1.0 | Test: NuGet.PackageManagement.UI.Test.PackageItemLoaderTests.EmitsSearchTelemetryEvents is flaky - Test `NuGet.PackageManagement.UI.Test.PackageItemLoaderTests.EmitsSearchTelemetryEvents` is flaky.
```
NuGet.PackageManagement.UI.Test.PackageItemLoaderTests.EmitsSearchTelemetryEvents [FAIL]
Assert.Equal() Failure
Expected: 4
Actual: 5
Stack Trace:
C:\git\NuGet.Client\test\NuGet.Clients.Tests\NuGet.PackageManagement.UI.Test\PackageItemLoaderTests.cs(107,0): at NuGet.PackageManagement.UI.Test.PackageItemLoaderTests.<EmitsSearchTelemetryEvents>d__2.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
``` | test | test nuget packagemanagement ui test packageitemloadertests emitssearchtelemetryevents is flaky test nuget packagemanagement ui test packageitemloadertests emitssearchtelemetryevents is flaky nuget packagemanagement ui test packageitemloadertests emitssearchtelemetryevents assert equal failure expected actual stack trace c git nuget client test nuget clients tests nuget packagemanagement ui test packageitemloadertests cs at nuget packagemanagement ui test packageitemloadertests d movenext end of stack trace from previous location where exception was thrown at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task end of stack trace from previous location where exception was thrown at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task end of stack trace from previous location where exception was thrown at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task | 1 |
158,188 | 12,405,531,010 | IssuesEvent | 2020-05-21 17:26:44 | Mercury-Leo/AR_Motion_matching | https://api.github.com/repos/Mercury-Leo/AR_Motion_matching | closed | Test: Scan screen with correct technique | App testing | Steps: Choose a technique and go to scan screen.
Result: Scan screen shall compare the users motion to the chosen technique | 1.0 | Test: Scan screen with correct technique - Steps: Choose a technique and go to scan screen.
Result: Scan screen shall compare the users motion to the chosen technique | test | test scan screen with correct technique steps choose a technique and go to scan screen result scan screen shall compare the users motion to the chosen technique | 1 |
815,792 | 30,571,621,178 | IssuesEvent | 2023-07-20 23:01:19 | googleapis/google-auth-library-nodejs | https://api.github.com/repos/googleapis/google-auth-library-nodejs | closed | AwsClient for environment variables retrieved tokens retrieveSubjectToken(): should reject when AWS region is not determined failed | type: bug priority: p1 flakybot: issue | Note: #1570 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: d8e5eb9994d2157a937543ed9f7d911b9e9db413
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/cff7842d-4e9c-4e55-b548-efbb59288c96), [Sponge](http://sponge2/cff7842d-4e9c-4e55-b548-efbb59288c96)
status: failed
<details><summary>Test output</summary><br><pre>Expected values to be strictly deep-equal:
+ actual - expected
+ Comparison {}
- Comparison {
- status: 500
- }
AssertionError [ERR_ASSERTION]: Expected values to be strictly deep-equal:
+ actual - expected
+ Comparison {}
- Comparison {
- status: 500
- }
at processTicksAndRejections (internal/process/task_queues.js:95:5)
at Context.<anonymous> (test/test.awsclient.ts:709:9)</pre></details> | 1.0 | AwsClient for environment variables retrieved tokens retrieveSubjectToken(): should reject when AWS region is not determined failed - Note: #1570 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: d8e5eb9994d2157a937543ed9f7d911b9e9db413
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/cff7842d-4e9c-4e55-b548-efbb59288c96), [Sponge](http://sponge2/cff7842d-4e9c-4e55-b548-efbb59288c96)
status: failed
<details><summary>Test output</summary><br><pre>Expected values to be strictly deep-equal:
+ actual - expected
+ Comparison {}
- Comparison {
- status: 500
- }
AssertionError [ERR_ASSERTION]: Expected values to be strictly deep-equal:
+ actual - expected
+ Comparison {}
- Comparison {
- status: 500
- }
at processTicksAndRejections (internal/process/task_queues.js:95:5)
at Context.<anonymous> (test/test.awsclient.ts:709:9)</pre></details> | non_test | awsclient for environment variables retrieved tokens retrievesubjecttoken should reject when aws region is not determined failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output expected values to be strictly deep equal actual expected comparison comparison status assertionerror expected values to be strictly deep equal actual expected comparison comparison status at processticksandrejections internal process task queues js at context test test awsclient ts | 0 |
29,129 | 13,050,473,373 | IssuesEvent | 2020-07-29 15:32:50 | cityofaustin/atd-data-tech | https://api.github.com/repos/cityofaustin/atd-data-tech | closed | Create AGOL Feature Layer | Product: TIA Module Product: Vision Zero Crash Data System Service: Geo Workgroup: TDSD Workgroup: VZ | To get a prototype of the VZ Intersection Polygons into AGOL, publish the polygon layer that Frank created. | 1.0 | Create AGOL Feature Layer - To get a prototype of the VZ Intersection Polygons into AGOL, publish the polygon layer that Frank created. | non_test | create agol feature layer to get a prototype of the vz intersection polygons into agol publish the polygon layer that frank created | 0 |
197,105 | 22,572,297,336 | IssuesEvent | 2022-06-28 02:12:55 | YJSoft/macos-plist | https://api.github.com/repos/YJSoft/macos-plist | opened | CVE-2021-42740 (High) detected in shell-quote-1.7.2.tgz | security vulnerability | ## CVE-2021-42740 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>shell-quote-1.7.2.tgz</b></p></summary>
<p>quote and parse shell commands</p>
<p>Library home page: <a href="https://registry.npmjs.org/shell-quote/-/shell-quote-1.7.2.tgz">https://registry.npmjs.org/shell-quote/-/shell-quote-1.7.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/shell-quote/package.json</p>
<p>
Dependency Hierarchy:
- browserify-16.5.2.tgz (Root Library)
- :x: **shell-quote-1.7.2.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The shell-quote package before 1.7.3 for Node.js allows command injection. An attacker can inject unescaped shell metacharacters through a regex designed to support Windows drive letters. If the output of this package is passed to a real shell as a quoted argument to a command with exec(), an attacker can inject arbitrary commands. This is because the Windows drive letter regex character class is {A-z] instead of the correct {A-Za-z]. Several shell metacharacters exist in the space between capital letter Z and lower case letter a, such as the backtick character.
<p>Publish Date: 2021-10-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-42740>CVE-2021-42740</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-42740">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-42740</a></p>
<p>Release Date: 2021-10-21</p>
<p>Fix Resolution: shell-quote - 1.7.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-42740 (High) detected in shell-quote-1.7.2.tgz - ## CVE-2021-42740 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>shell-quote-1.7.2.tgz</b></p></summary>
<p>quote and parse shell commands</p>
<p>Library home page: <a href="https://registry.npmjs.org/shell-quote/-/shell-quote-1.7.2.tgz">https://registry.npmjs.org/shell-quote/-/shell-quote-1.7.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/shell-quote/package.json</p>
<p>
Dependency Hierarchy:
- browserify-16.5.2.tgz (Root Library)
- :x: **shell-quote-1.7.2.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The shell-quote package before 1.7.3 for Node.js allows command injection. An attacker can inject unescaped shell metacharacters through a regex designed to support Windows drive letters. If the output of this package is passed to a real shell as a quoted argument to a command with exec(), an attacker can inject arbitrary commands. This is because the Windows drive letter regex character class is {A-z] instead of the correct {A-Za-z]. Several shell metacharacters exist in the space between capital letter Z and lower case letter a, such as the backtick character.
<p>Publish Date: 2021-10-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-42740>CVE-2021-42740</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-42740">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-42740</a></p>
<p>Release Date: 2021-10-21</p>
<p>Fix Resolution: shell-quote - 1.7.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve high detected in shell quote tgz cve high severity vulnerability vulnerable library shell quote tgz quote and parse shell commands library home page a href path to dependency file package json path to vulnerable library node modules shell quote package json dependency hierarchy browserify tgz root library x shell quote tgz vulnerable library found in base branch master vulnerability details the shell quote package before for node js allows command injection an attacker can inject unescaped shell metacharacters through a regex designed to support windows drive letters if the output of this package is passed to a real shell as a quoted argument to a command with exec an attacker can inject arbitrary commands this is because the windows drive letter regex character class is a z instead of the correct a za z several shell metacharacters exist in the space between capital letter z and lower case letter a such as the backtick character publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution shell quote step up your open source security game with mend | 0 |
276,637 | 20,993,384,598 | IssuesEvent | 2022-03-29 11:25:21 | Sitecore/developer-portal | https://api.github.com/repos/Sitecore/developer-portal | closed | Create DevOps guide for Managed Cloud Containers | documentation | Need an article added to the Getting Started section (`/learn/getting-started`) area that can pull together all the documentation and steps a developer may need to get started with container GitOps for Managed Cloud Containers.
We also don't have a Managed Cloud product area on the site yet, so that will need to be added as well so that this guide can be found. | 1.0 | Create DevOps guide for Managed Cloud Containers - Need an article added to the Getting Started section (`/learn/getting-started`) area that can pull together all the documentation and steps a developer may need to get started with container GitOps for Managed Cloud Containers.
We also don't have a Managed Cloud product area on the site yet, so that will need to be added as well so that this guide can be found. | non_test | create devops guide for managed cloud containers need an article added to the getting started section learn getting started area that can pull together all the documentation and steps a developer may need to get started with container gitops for managed cloud containers we also don t have a managed cloud product area on the site yet so that will need to be added as well so that this guide can be found | 0 |
174,454 | 14,483,172,461 | IssuesEvent | 2020-12-10 14:51:51 | nilearn/nilearn | https://api.github.com/repos/nilearn/nilearn | closed | Improve Documentation of masker family | Documentation | Perhaps the docstrings for the following classes could be improved:
```
- NiftiMasker
- NiftiLabelsMasker
- NiftiMapsMasker
- MultiNiftiMasker
```
More specifically, at the top of the respective class docstrings, one might add
- a more generous explanation of the purpose of the class
- its commonalities and differences comparing to the respective other masker classes
- provide one simple use case when this is typically useful
Apart from that, a new nilearn example might be useful that uses the various masker and, thus, gives an overview of this part of the API.
Edited by @AlexandreAbraham:
Here is the list of remarks made on _Nifti_Masker doc across issues:
- [ ] Specify which data is detrended (#418)
- [ ] Add intersect_masks to NiftiMasker's "related functions" (#436)
- [ ] Specify that only `mask_args` and `mask_strategy` are used for fitting. It's specified in the class docstring, but not on the fit function docstring (#436). Gael suggested to add it in the Notes section.
- [ ] Add in transform (#436):
- Parameters (detrend, standardize) are computed from the images passed to transform (i.e. not computed and stored from fit)
- Whether the inverse of these parameters is applied when inverse_transform is called (not sure if they are, for detrend, standardize, and confounds).
- [ ] NiftiMapsMasker and NiftiLabelsMasker don't have docstring for fit_transforms (#603).
| 1.0 | Improve Documentation of masker family - Perhaps the docstrings for the following classes could be improved:
```
- NiftiMasker
- NiftiLabelsMasker
- NiftiMapsMasker
- MultiNiftiMasker
```
More specifically, at the top of the respective class docstrings, one might add
- a more generous explanation of the purpose of the class
- its commonalities and differences comparing to the respective other masker classes
- provide one simple use case when this is typically useful
Apart from that, a new nilearn example might be useful that uses the various masker and, thus, gives an overview of this part of the API.
Edited by @AlexandreAbraham:
Here is the list of remarks made on _Nifti_Masker doc across issues:
- [ ] Specify which data is detrended (#418)
- [ ] Add intersect_masks to NiftiMasker's "related functions" (#436)
- [ ] Specify that only `mask_args` and `mask_strategy` are used for fitting. It's specified in the class docstring, but not on the fit function docstring (#436). Gael suggested to add it in the Notes section.
- [ ] Add in transform (#436):
- Parameters (detrend, standardize) are computed from the images passed to transform (i.e. not computed and stored from fit)
- Whether the inverse of these parameters is applied when inverse_transform is called (not sure if they are, for detrend, standardize, and confounds).
- [ ] NiftiMapsMasker and NiftiLabelsMasker don't have docstring for fit_transforms (#603).
| non_test | improve documentation of masker family perhaps the docstrings for the following classes could be improved niftimasker niftilabelsmasker niftimapsmasker multiniftimasker more specifically at the top of the respective class docstrings one might add a more generous explanation of the purpose of the class its commonalities and differences comparing to the respective other masker classes provide one simple use case when this is typically useful apart from that a new nilearn example might be useful that uses the various masker and thus gives an overview of this part of the api edited by alexandreabraham here is the list of remarks made on nifti masker doc across issues specify which data is detrended add intersect masks to niftimasker s related functions specify that only mask args and mask strategy are used for fitting it s specified in the class docstring but not on the fit function docstring gael suggested to add it in the notes section add in transform parameters detrend standardize are computed from the images passed to transform i e not computed and stored from fit whether the inverse of these parameters is applied when inverse transform is called not sure if they are for detrend standardize and confounds niftimapsmasker and niftilabelsmasker don t have docstring for fit transforms | 0 |
73,287 | 7,330,982,915 | IssuesEvent | 2018-03-05 11:52:19 | MachoThemes/modula-lite | https://api.github.com/repos/MachoThemes/modula-lite | closed | alt tags empty | enhancement need testing | Modula Gallery is not optimized for image referencing (SEO)
The "alt" tags are empty in the html code. | 1.0 | alt tags empty - Modula Gallery is not optimized for image referencing (SEO)
The "alt" tags are empty in the html code. | test | alt tags empty modula gallery is not optimized for image referencing seo the alt tags are empty in the html code | 1 |
10,266 | 13,113,079,995 | IssuesEvent | 2020-08-05 04:16:15 | googleapis/python-datastore | https://api.github.com/repos/googleapis/python-datastore | closed | 'TestClient.test_constructor_w_implicit_inputs' unit test failure | api: datastore priority: p0 testing type: process | From [this Kokoro build failure]():
```python
________________ TestClient.test_constructor_w_implicit_inputs _________________
self = <tests.unit.test_client.TestClient testMethod=test_constructor_w_implicit_inputs>
def test_constructor_w_implicit_inputs(self):
from google.cloud.datastore.client import _CLIENT_INFO
from google.cloud.datastore.client import _DATASTORE_BASE_URL
other = "other"
creds = _make_credentials()
klass = self._get_target_class()
patch1 = mock.patch(
"google.cloud.datastore.client._determine_default_project",
return_value=other,
)
patch2 = mock.patch("google.auth.default", return_value=(creds, None))
with patch1 as _determine_default_project:
with patch2 as default:
client = klass()
self.assertEqual(client.project, other)
self.assertIsNone(client.namespace)
self.assertIs(client._credentials, creds)
self.assertIs(client._client_info, _CLIENT_INFO)
self.assertIsNone(client._http_internal)
self.assertIsNone(client._client_options)
self.assertEqual(client.base_url, _DATASTORE_BASE_URL)
self.assertIsNone(client.current_batch)
self.assertIsNone(client.current_transaction)
> default.assert_called_once_with()
tests/unit/test_client.py:183:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
.nox/unit-2-7/lib/python2.7/site-packages/mock/mock.py:957: in assert_called_once_with
return self.assert_called_with(*args, **kwargs)
.nox/unit-2-7/lib/python2.7/site-packages/mock/mock.py:944: in assert_called_with
six.raise_from(AssertionError(_error_message(cause)), cause)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
value = AssertionError("expected call not found.\nExpected: default()\nActual: default(scopes=('https://www.googleapis.com/auth/datastore',))",)
from_value = None
def raise_from(value, from_value):
> raise value
E AssertionError: expected call not found.
E Expected: default()
E Actual: default(scopes=('https://www.googleapis.com/auth/datastore',))
``` | 1.0 | 'TestClient.test_constructor_w_implicit_inputs' unit test failure - From [this Kokoro build failure]():
```python
________________ TestClient.test_constructor_w_implicit_inputs _________________
self = <tests.unit.test_client.TestClient testMethod=test_constructor_w_implicit_inputs>
def test_constructor_w_implicit_inputs(self):
from google.cloud.datastore.client import _CLIENT_INFO
from google.cloud.datastore.client import _DATASTORE_BASE_URL
other = "other"
creds = _make_credentials()
klass = self._get_target_class()
patch1 = mock.patch(
"google.cloud.datastore.client._determine_default_project",
return_value=other,
)
patch2 = mock.patch("google.auth.default", return_value=(creds, None))
with patch1 as _determine_default_project:
with patch2 as default:
client = klass()
self.assertEqual(client.project, other)
self.assertIsNone(client.namespace)
self.assertIs(client._credentials, creds)
self.assertIs(client._client_info, _CLIENT_INFO)
self.assertIsNone(client._http_internal)
self.assertIsNone(client._client_options)
self.assertEqual(client.base_url, _DATASTORE_BASE_URL)
self.assertIsNone(client.current_batch)
self.assertIsNone(client.current_transaction)
> default.assert_called_once_with()
tests/unit/test_client.py:183:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
.nox/unit-2-7/lib/python2.7/site-packages/mock/mock.py:957: in assert_called_once_with
return self.assert_called_with(*args, **kwargs)
.nox/unit-2-7/lib/python2.7/site-packages/mock/mock.py:944: in assert_called_with
six.raise_from(AssertionError(_error_message(cause)), cause)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
value = AssertionError("expected call not found.\nExpected: default()\nActual: default(scopes=('https://www.googleapis.com/auth/datastore',))",)
from_value = None
def raise_from(value, from_value):
> raise value
E AssertionError: expected call not found.
E Expected: default()
E Actual: default(scopes=('https://www.googleapis.com/auth/datastore',))
``` | non_test | testclient test constructor w implicit inputs unit test failure from python testclient test constructor w implicit inputs self def test constructor w implicit inputs self from google cloud datastore client import client info from google cloud datastore client import datastore base url other other creds make credentials klass self get target class mock patch google cloud datastore client determine default project return value other mock patch google auth default return value creds none with as determine default project with as default client klass self assertequal client project other self assertisnone client namespace self assertis client credentials creds self assertis client client info client info self assertisnone client http internal self assertisnone client client options self assertequal client base url datastore base url self assertisnone client current batch self assertisnone client current transaction default assert called once with tests unit test client py nox unit lib site packages mock mock py in assert called once with return self assert called with args kwargs nox unit lib site packages mock mock py in assert called with six raise from assertionerror error message cause cause value assertionerror expected call not found nexpected default nactual default scopes from value none def raise from value from value raise value e assertionerror expected call not found e expected default e actual default scopes | 0 |
51,383 | 6,157,199,488 | IssuesEvent | 2017-06-28 18:25:15 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | teamcity: failed tests on master: testrace/TestStoreRangeCorruptionChangeReplicas, testrace/TestSystemZoneConfigs | Robot test-failure | The following tests appear to have failed:
[#282905](https://teamcity.cockroachdb.com/viewLog.html?buildId=282905):
```
--- FAIL: testrace/TestStoreRangeCorruptionChangeReplicas (0.900s)
client_raft_test.go:1210: [NotLeaseHolderError] r1: replica (n1,s1):1 not lease holder; current lease is repl=(n4,s4):4 start=0.000000123,427 exp=0.900000123,427 pro=0.000000123,428
------- Stdout: -------
I170628 18:18:42.039506 32093 gossip/gossip.go:297 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:54812" > attrs:<> locality:<>
W170628 18:18:42.051538 32093 gossip/gossip.go:1196 [n2] no incoming or outgoing connections
I170628 18:18:42.055676 32365 gossip/client.go:131 [n2] started gossip client to 127.0.0.1:54812
I170628 18:18:42.069941 32093 gossip/gossip.go:297 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:48975" > attrs:<> locality:<>
W170628 18:18:42.083864 32093 gossip/gossip.go:1196 [n3] no incoming or outgoing connections
I170628 18:18:42.093278 32428 gossip/client.go:131 [n3] started gossip client to 127.0.0.1:54812
I170628 18:18:42.103542 32093 storage/store.go:1265 [n3,s3]: failed initial metrics computation: [n3,s3]: system config not yet available
I170628 18:18:42.103846 32093 gossip/gossip.go:297 [n3] NodeDescriptor set to node_id:3 address:<network_field:"tcp" address_field:"127.0.0.1:34226" > attrs:<> locality:<>
W170628 18:18:42.127052 32093 gossip/gossip.go:1196 [n4] no incoming or outgoing connections
I170628 18:18:42.131932 32633 gossip/client.go:131 [n4] started gossip client to 127.0.0.1:54812
I170628 18:18:42.142942 32093 gossip/gossip.go:297 [n4] NodeDescriptor set to node_id:4 address:<network_field:"tcp" address_field:"127.0.0.1:40101" > attrs:<> locality:<>
W170628 18:18:42.182637 32093 gossip/gossip.go:1196 [n5] no incoming or outgoing connections
I170628 18:18:42.193550 32093 storage/store.go:1265 [n5,s5]: failed initial metrics computation: [n5,s5]: system config not yet available
I170628 18:18:42.193833 32093 gossip/gossip.go:297 [n5] NodeDescriptor set to node_id:5 address:<network_field:"tcp" address_field:"127.0.0.1:34547" > attrs:<> locality:<>
I170628 18:18:42.197140 32617 gossip/client.go:131 [n5] started gossip client to 127.0.0.1:54812
I170628 18:18:42.198251 32749 gossip/server.go:285 [n1] refusing gossip from node 5 (max 3 conns); forwarding to 3 ({tcp 127.0.0.1:34226})
I170628 18:18:42.206388 32617 gossip/client.go:136 [n5] closing client to node 1 (127.0.0.1:54812): received forward from node 1 to 3 (127.0.0.1:34226)
I170628 18:18:42.206818 32771 gossip/gossip.go:1210 [n5] node has connected to cluster via gossip
I170628 18:18:42.207910 32584 gossip/client.go:131 [n5] started gossip client to 127.0.0.1:34226
W170628 18:18:42.286554 32093 gossip/gossip.go:1196 [n6] no incoming or outgoing connections
I170628 18:18:42.288066 32190 gossip/client.go:131 [n6] started gossip client to 127.0.0.1:54812
I170628 18:18:42.289202 32696 gossip/server.go:285 [n1] refusing gossip from node 6 (max 3 conns); forwarding to 2 ({tcp 127.0.0.1:48975})
I170628 18:18:42.292831 32190 gossip/client.go:136 [n6] closing client to node 1 (127.0.0.1:54812): received forward from node 1 to 2 (127.0.0.1:48975)
I170628 18:18:42.294080 32916 gossip/gossip.go:1210 [n6] node has connected to cluster via gossip
I170628 18:18:42.295214 32855 gossip/client.go:131 [n6] started gossip client to 127.0.0.1:48975
I170628 18:18:42.313508 32093 gossip/gossip.go:297 [n6] NodeDescriptor set to node_id:6 address:<network_field:"tcp" address_field:"127.0.0.1:34349" > attrs:<> locality:<>
I170628 18:18:42.373654 32093 storage/client_test.go:410 gossip network initialized
I170628 18:18:42.374929 32093 storage/replica_raftstorage.go:442 [replicate,s1,r1/1:/M{in-ax}] generated preemptive snapshot ec6a1960 at index 18
I170628 18:18:42.389357 32093 storage/store.go:3416 [replicate,s1,r1/1:/M{in-ax}] streamed snapshot to (n6,s6):?: kv pairs: 36, log entries: 8, rate-limit: 8.0 MiB/sec, 14ms
I170628 18:18:42.390259 32938 storage/replica_raftstorage.go:639 [s6,r1/?:{-}] applying preemptive snapshot at index 18 (id=ec6a1960, encoded size=7114, 1 rocksdb batches, 8 log entries)
I170628 18:18:42.391805 32938 storage/replica_raftstorage.go:647 [s6,r1/?:/M{in-ax}] applied preemptive snapshot in 1ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:18:42.393956 32093 storage/replica_command.go:3615 [replicate,s1,r1/1:/M{in-ax}] change replicas (ADD_REPLICA (n6,s6):2): read existing descriptor r1:/M{in-ax} [(n1,s1):1, next=2]
I170628 18:18:42.398815 33092 storage/replica.go:2887 [replicate,s1,r1/1:/M{in-ax}] proposing ADD_REPLICA (n6,s6):2: [(n1,s1):1 (n6,s6):2]
I170628 18:18:42.402861 32093 storage/replica_raftstorage.go:442 [replicate,s1,r1/1:/M{in-ax}] generated preemptive snapshot 30e6642c at index 20
I170628 18:18:42.408045 32093 storage/store.go:3416 [replicate,s1,r1/1:/M{in-ax}] streamed snapshot to (n3,s3):?: kv pairs: 39, log entries: 10, rate-limit: 8.0 MiB/sec, 3ms
I170628 18:18:42.412534 33052 storage/replica_raftstorage.go:639 [s3,r1/?:{-}] applying preemptive snapshot at index 20 (id=30e6642c, encoded size=8680, 1 rocksdb batches, 10 log entries)
I170628 18:18:42.414655 33052 storage/replica_raftstorage.go:647 [s3,r1/?:/M{in-ax}] applied preemptive snapshot in 2ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:18:42.436549 32093 storage/replica_command.go:3615 [replicate,s1,r1/1:/M{in-ax}] change replicas (ADD_REPLICA (n3,s3):3): read existing descriptor r1:/M{in-ax} [(n1,s1):1, (n6,s6):2, next=3]
I170628 18:18:42.459365 33124 storage/raft_transport.go:456 raft transport stream to node 1 established
I170628 18:18:42.468188 33037 storage/replica.go:2887 [replicate,s1,r1/1:/M{in-ax}] proposing ADD_REPLICA (n3,s3):3: [(n1,s1):1 (n6,s6):2 (n3,s3):3]
E170628 18:18:42.491831 32925 storage/replica.go:4924 [s6,r1/2:/M{in-ax}] stalling replica due to: boom
I170628 18:18:42.498379 32093 storage/replica_command.go:3615 [replicate,s1,r1/1:/M{in-ax}] change replicas (REMOVE_REPLICA (n6,s6):2): read existing descriptor r1:/M{in-ax} [(n1,s1):1, (n6,s6):2, (n3,s3):3, next=4]
W170628 18:18:42.515704 33123 storage/store.go:3231 [s1] got error from r1, replica (n6,s6):2: replica corruption (processed=true): boom
W170628 18:18:42.519010 33123 storage/store.go:3231 [s1] got error from r1, replica (n6,s6):2: replica corruption (processed=true): boom
W170628 18:18:42.519165 33123 storage/store.go:3231 [s1] got error from r1, replica (n6,s6):2: replica corruption (processed=true): boom
I170628 18:18:42.523686 33073 storage/replica.go:2887 [replicate,s1,r1/1:/M{in-ax}] proposing REMOVE_REPLICA (n6,s6):2: [(n1,s1):1 (n3,s3):3]
W170628 18:18:42.530618 33123 storage/store.go:3231 [s1] got error from r1, replica (n6,s6):2: replica corruption (processed=true): boom
W170628 18:18:42.536108 33123 storage/store.go:3231 [s1] got error from r1, replica (n6,s6):2: replica corruption (processed=true): boom
I170628 18:18:42.540300 32093 storage/replica_raftstorage.go:442 [replicate,s1,r1/1:/M{in-ax}] generated preemptive snapshot a570ed02 at index 25
I170628 18:18:42.554709 32093 storage/store.go:3416 [replicate,s1,r1/1:/M{in-ax}] streamed snapshot to (n4,s4):?: kv pairs: 46, log entries: 15, rate-limit: 8.0 MiB/sec, 13ms
I170628 18:18:42.555791 33119 storage/replica_raftstorage.go:639 [s4,r1/?:{-}] applying preemptive snapshot at index 25 (id=a570ed02, encoded size=12373, 1 rocksdb batches, 15 log entries)
I170628 18:18:42.562230 33119 storage/replica_raftstorage.go:647 [s4,r1/?:/M{in-ax}] applied preemptive snapshot in 6ms [clear=0ms batch=0ms entries=1ms commit=4ms]
I170628 18:18:42.579036 32093 storage/replica_command.go:3615 [replicate,s1,r1/1:/M{in-ax}] change replicas (ADD_REPLICA (n4,s4):4): read existing descriptor r1:/M{in-ax} [(n1,s1):1, (n3,s3):3, next=4]
I170628 18:18:42.588160 33088 storage/replica.go:2887 [replicate,s1,r1/1:/M{in-ax}] proposing ADD_REPLICA (n4,s4):4: [(n1,s1):1 (n3,s3):3 (n4,s4):4]
E170628 18:18:42.600565 32518 storage/replica.go:4924 [s3,r1/3:/M{in-ax}] stalling replica due to: boom
I170628 18:18:42.615557 32093 storage/replica_command.go:3615 [replicate,s1,r1/1:/M{in-ax}] change replicas (REMOVE_REPLICA (n3,s3):3): read existing descriptor r1:/M{in-ax} [(n1,s1):1, (n3,s3):3, (n4,s4):4, next=5]
W170628 18:18:42.619227 33099 storage/store.go:3231 [s1] got error from r1, replica (n3,s3):3: replica corruption (processed=true): boom
I170628 18:18:42.625580 33235 storage/replica.go:2887 [replicate,s1,r1/1:/M{in-ax}] proposing REMOVE_REPLICA (n3,s3):3: [(n1,s1):1 (n4,s4):4]
W170628 18:18:42.628331 33099 storage/store.go:3231 [s1] got error from r1, replica (n3,s3):3: replica corruption (processed=true): boom
W170628 18:18:42.628508 33099 storage/store.go:3231 [s1] got error from r1, replica (n3,s3):3: replica corruption (processed=true): boom
W170628 18:18:42.631838 33099 storage/store.go:3231 [s1] got error from r1, replica (n3,s3):3: replica corruption (processed=true): boom
I170628 18:18:42.634528 32096 storage/replica_raftstorage.go:442 [replicate,s1,r1/1:/M{in-ax}] generated preemptive snapshot 3f71ce49 at index 30
I170628 18:18:42.639944 32096 storage/store.go:3416 [replicate,s1,r1/1:/M{in-ax}] streamed snapshot to (n2,s2):?: kv pairs: 53, log entries: 20, rate-limit: 8.0 MiB/sec, 3ms
I170628 18:18:42.643043 33224 storage/replica_raftstorage.go:639 [s2,r1/?:{-}] applying preemptive snapshot at index 30 (id=3f71ce49, encoded size=16066, 1 rocksdb batches, 20 log entries)
I170628 18:18:42.651484 33224 storage/replica_raftstorage.go:647 [s2,r1/?:/M{in-ax}] applied preemptive snapshot in 8ms [clear=0ms batch=0ms entries=7ms commit=0ms]
I170628 18:18:42.655067 32096 storage/replica_command.go:3615 [replicate,s1,r1/1:/M{in-ax}] change replicas (ADD_REPLICA (n2,s2):5): read existing descriptor r1:/M{in-ax} [(n1,s1):1, (n4,s4):4, next=5]
I170628 18:18:42.673879 33191 storage/replica.go:2887 [replicate,s1,r1/1:/M{in-ax}] proposing ADD_REPLICA (n2,s2):5: [(n1,s1):1 (n4,s4):4 (n2,s2):5]
I170628 18:18:42.687305 32093 storage/replica_raftstorage.go:442 [replicate,s1,r1/1:/M{in-ax}] generated preemptive snapshot f9c76c99 at index 32
I170628 18:18:42.691517 32093 storage/store.go:3416 [replicate,s1,r1/1:/M{in-ax}] streamed snapshot to (n5,s5):?: kv pairs: 56, log entries: 22, rate-limit: 2.0 MiB/sec, 4ms
I170628 18:18:42.692640 33165 storage/replica_raftstorage.go:639 [s5,r1/?:{-}] applying preemptive snapshot at index 32 (id=f9c76c99, encoded size=17696, 1 rocksdb batches, 22 log entries)
I170628 18:18:42.695242 33165 storage/replica_raftstorage.go:647 [s5,r1/?:/M{in-ax}] applied preemptive snapshot in 2ms [clear=0ms batch=0ms entries=2ms commit=0ms]
I170628 18:18:42.698120 32093 storage/replica_command.go:3615 [replicate,s1,r1/1:/M{in-ax}] change replicas (ADD_REPLICA (n5,s5):6): read existing descriptor r1:/M{in-ax} [(n1,s1):1, (n4,s4):4, (n2,s2):5, next=6]
I170628 18:18:42.709954 33259 storage/replica.go:2887 [replicate,s1,r1/1:/M{in-ax}] proposing ADD_REPLICA (n5,s5):6: [(n1,s1):1 (n4,s4):4 (n2,s2):5 (n5,s5):6]
E170628 18:18:42.770293 32785 storage/replica.go:4924 [s5,r1/?:/M{in-ax}] stalling replica due to: boom
E170628 18:18:42.771138 32785 storage/replica.go:4924 [s5,r1/?:/M{in-ax}] stalling replica due to: boom
E170628 18:18:42.772612 32785 storage/replica.go:4924 [s5,r1/6:/M{in-ax}] stalling replica due to: boom
W170628 18:18:42.792463 32342 storage/node_liveness.go:318 [hb] failed node liveness heartbeat: node unavailable; try another peer
W170628 18:18:42.795256 33139 storage/raft_transport.go:272 unable to accept Raft message from (n5,s5):6: no handler registered for (n1,s1):1
W170628 18:18:42.795945 33264 storage/store.go:3231 [s4] got error from r1, replica (n5,s5):6: replica corruption (processed=true): boom
W170628 18:18:42.796387 33126 storage/store.go:3227 [s5] raft error: node 1 claims to not contain store 1 for replica (n1,s1):1: store 1 was not found
W170628 18:18:42.796559 33139 storage/raft_transport.go:272 unable to accept Raft message from (n4,s4):4: no handler registered for (n1,s1):1
W170628 18:18:42.796716 33124 storage/raft_transport.go:462 raft transport stream to node 1 failed: store 1 was not found
W170628 18:18:42.800187 33264 storage/store.go:3231 [s4] got error from r1, replica (n5,s5):6: replica corruption (processed=true): boom
W170628 18:18:42.800756 33244 storage/raft_transport.go:272 unable to accept Raft message from (n4,s4):4: no handler registered for (n2,s2):5
W170628 18:18:42.801735 33268 storage/store.go:3227 [s4] raft error: node 2 claims to not contain store 2 for replica (n2,s2):5: store 2 was not found
W170628 18:18:42.801965 33266 storage/raft_transport.go:462 raft transport stream to node 2 failed: store 2 was not found
W170628 18:18:42.884557 32916 gossip/gossip.go:1196 [n6] no incoming or outgoing connections
W170628 18:18:42.885603 33057 storage/raft_transport.go:462 raft transport stream to node 6 failed: rpc error: code = Internal desc = transport is closing
I170628 18:18:42.886036 32614 vendor/google.golang.org/grpc/transport/http2_server.go:399 transport: http2Server.HandleStreams failed to read frame: read tcp 127.0.0.1:34547->127.0.0.1:60106: use of closed network connection
I170628 18:18:42.887097 32425 vendor/google.golang.org/grpc/transport/http2_server.go:399 transport: http2Server.HandleStreams failed to read frame: read tcp 127.0.0.1:34226->127.0.0.1:34086: use of closed network connection
I170628 18:18:42.887865 32370 vendor/google.golang.org/grpc/transport/http2_server.go:399 transport: http2Server.HandleStreams failed to read frame: read tcp 127.0.0.1:48975->127.0.0.1:47604: use of closed network connection
I170628 18:18:42.888284 32160 vendor/google.golang.org/grpc/transport/http2_server.go:399 transport: http2Server.HandleStreams failed to read frame: read tcp 127.0.0.1:54812->127.0.0.1:41161: use of closed network connection
I170628 18:18:42.888693 32377 vendor/google.golang.org/grpc/transport/http2_server.go:399 transport: http2Server.HandleStreams failed to read frame: read tcp 127.0.0.1:40101->127.0.0.1:57971: use of closed network connection
W170628 18:18:42.889788 33262 storage/raft_transport.go:462 raft transport stream to node 5 failed: rpc error: code = Internal desc = transport is closing
--- FAIL: testrace/TestSystemZoneConfigs (54.980s)
<autogenerated>:14: storage/client_replica_test.go:1575, condition failed to evaluate within 45s: got 41 replicas, want 39; details: map[5:3 12:3 1:3 8:3 13:3 3:3 4:4 6:3 7:4 9:3 10:3 11:3 2:3]
------- Stdout: -------
W170628 18:19:26.872997 59748 server/status/runtime.go:111 Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I170628 18:19:26.890981 59748 server/config.go:479 [n?] 1 storage engine initialized
I170628 18:19:26.891063 59748 server/config.go:481 [n?] RocksDB cache size: 512 MiB
I170628 18:19:26.891106 59748 server/config.go:481 [n?] store 0: in-memory, size 100 MiB
I170628 18:19:26.892471 59748 server/node.go:450 [n?] store [n0,s0] not bootstrapped
I170628 18:19:26.918588 59748 server/node.go:385 [n?] **** cluster 1af3ddf7-bc64-4eaa-96da-7ae509382d89 has been created
I170628 18:19:26.918683 59748 server/node.go:386 [n?] **** add additional nodes by specifying --join=127.0.0.1:60907
I170628 18:19:26.954622 59748 server/node.go:463 [n1] initialized store [n1,s1]: {Capacity:536870912 Available:536870912 RangeCount:1 LeaseCount:1 WritesPerSecond:58.1946861908287}
I170628 18:19:26.954878 59748 server/node.go:347 [n1] node ID 1 initialized
I170628 18:19:26.955096 59748 gossip/gossip.go:297 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:60907" > attrs:<> locality:<>
I170628 18:19:26.955642 59748 storage/stores.go:296 [n1] read 0 node addresses from persistent storage
I170628 18:19:26.955896 59748 server/node.go:604 [n1] connecting to gossip network to verify cluster ID...
I170628 18:19:26.955998 59748 server/node.go:629 [n1] node connected via gossip and verified as part of cluster "1af3ddf7-bc64-4eaa-96da-7ae509382d89"
I170628 18:19:26.956276 59748 server/node.go:401 [n1] node=1: started with [[]=] engine(s) and attributes []
I170628 18:19:26.962093 60036 storage/replica_command.go:2694 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2]
I170628 18:19:26.963625 59748 sql/executor.go:358 [n1] creating distSQLPlanner with address {tcp 127.0.0.1:60907}
E170628 18:19:27.022144 60037 storage/queue.go:658 [replicate,n1,s1,r1/1:/{Min-System/}] range requires a replication change, but lacks a quorum of live replicas (0/1)
I170628 18:19:27.032052 60036 storage/replica_command.go:2694 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/tsd [r3]
I170628 18:19:27.038615 59748 server/server.go:769 [n1] starting https server at 127.0.0.1:43705
I170628 18:19:27.038740 59748 server/server.go:770 [n1] starting grpc/postgres server at 127.0.0.1:60907
I170628 18:19:27.038784 59748 server/server.go:771 [n1] advertising CockroachDB node at 127.0.0.1:60907
E170628 18:19:27.082339 60037 storage/queue.go:658 [replicate,n1,s1,r1/1:/{Min-System/}] range requires a replication change, but lacks a quorum of live replicas (0/1)
E170628 18:19:27.290094 59984 storage/replica_proposal.go:470 [n1,s1,r2/1:/{System/-Max}] could not load SystemConfig span: must retry later due to intent on SystemConfigSpan
I170628 18:19:27.295656 60036 storage/replica_command.go:2694 [split,n1,s1,r3/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r4]
I170628 18:19:27.303122 59748 sql/event_log.go:101 [n1] Event: "alter_table", target: 12, info: {TableName:eventlog Statement:ALTER TABLE system.eventlog ALTER COLUMN uniqueid SET DEFAULT uuid_v4() User:node MutationID:0 CascadeDroppedViews:[]}
I170628 18:19:27.390008 59748 sql/lease.go:372 [n1] publish: descID=12 (eventlog) version=2 mtime=2017-06-28 18:19:27.38991686 +0000 UTC
I170628 18:19:27.445220 60036 storage/replica_command.go:2694 [split,n1,s1,r4/1:/{System/tse-Max}] initiating a split of this range at key /Table/0 [r5]
I170628 18:19:27.568315 60036 storage/replica_command.go:2694 [split,n1,s1,r5/1:/{Table/0-Max}] initiating a split of this range at key /Table/11 [r6]
I170628 18:19:27.639950 59748 server/server.go:906 [n1] done ensuring all necessary migrations have run
I170628 18:19:27.640079 59748 server/server.go:908 [n1] serving sql connections
I170628 18:19:27.674349 60036 storage/replica_command.go:2694 [split,n1,s1,r6/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r7]
I170628 18:19:27.718370 60206 sql/event_log.go:101 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:60907} Attrs: Locality:} ClusterID:1af3ddf7-bc64-4eaa-96da-7ae509382d89 StartedAt:1498673966956040633 LastUp:1498673966956040633}
I170628 18:19:27.772910 60036 storage/replica_command.go:2694 [split,n1,s1,r7/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r8]
I170628 18:19:27.868291 60036 storage/replica_command.go:2694 [split,n1,s1,r8/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r9]
I170628 18:19:27.980417 60036 storage/replica_command.go:2694 [split,n1,s1,r9/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r10]
W170628 18:19:28.177356 59748 server/status/runtime.go:111 Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
W170628 18:19:28.187826 59748 gossip/gossip.go:1196 [n?] no incoming or outgoing connections
I170628 18:19:28.222038 59748 server/config.go:479 [n?] 1 storage engine initialized
I170628 18:19:28.222136 59748 server/config.go:481 [n?] RocksDB cache size: 512 MiB
I170628 18:19:28.222172 59748 server/config.go:481 [n?] store 0: in-memory, size 100 MiB
I170628 18:19:28.223344 59748 server/node.go:450 [n?] store [n0,s0] not bootstrapped
I170628 18:19:28.223445 59748 storage/stores.go:296 [n?] read 0 node addresses from persistent storage
I170628 18:19:28.223540 59748 server/node.go:604 [n?] connecting to gossip network to verify cluster ID...
I170628 18:19:28.266534 60424 gossip/client.go:131 [n?] started gossip client to 127.0.0.1:60907
I170628 18:19:28.274630 60444 gossip/server.go:234 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:56027}
I170628 18:19:28.278409 59748 server/node.go:629 [n?] node connected via gossip and verified as part of cluster "1af3ddf7-bc64-4eaa-96da-7ae509382d89"
I170628 18:19:28.283062 60516 storage/stores.go:312 [n?] wrote 1 node addresses to persistent storage
I170628 18:19:28.291610 59748 kv/dist_sender.go:370 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I170628 18:19:28.296676 59748 server/node.go:340 [n?] new node allocated ID 2
I170628 18:19:28.297684 59748 gossip/gossip.go:297 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:56027" > attrs:<> locality:<>
I170628 18:19:28.298477 59748 server/node.go:401 [n2] node=2: started with [[]=] engine(s) and attributes []
I170628 18:19:28.306128 60500 storage/stores.go:312 [n1] wrote 1 node addresses to persistent storage
I170628 18:19:28.314202 59748 sql/executor.go:358 [n2] creating distSQLPlanner with address {tcp 127.0.0.1:56027}
I170628 18:19:28.326292 60530 server/node.go:585 [n2] bootstrapped store [n2,s2]
I170628 18:19:28.334735 59748 server/server.go:769 [n2] starting https server at 127.0.0.1:35587
I170628 18:19:28.335700 59748 server/server.go:770 [n2] starting grpc/postgres server at 127.0.0.1:56027
I170628 18:19:28.335764 59748 server/server.go:771 [n2] advertising CockroachDB node at 127.0.0.1:56027
I170628 18:19:28.349165 59748 server/server.go:906 [n2] done ensuring all necessary migrations have run
I170628 18:19:28.349309 59748 server/server.go:908 [n2] serving sql connections
W170628 18:19:28.472859 59748 server/status/runtime.go:111 Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
W170628 18:19:28.486721 59748 gossip/gossip.go:1196 [n?] no incoming or outgoing connections
I170628 18:19:28.487189 60485 sql/event_log.go:101 [n2] Event: "node_join", target: 2, info: {Descriptor:{NodeID:2 Address:{NetworkField:tcp AddressField:127.0.0.1:56027} Attrs: Locality:} ClusterID:1af3ddf7-bc64-4eaa-96da-7ae509382d89 StartedAt:1498673968298206396 LastUp:1498673968298206396}
I170628 18:19:28.495009 59748 server/config.go:479 [n?] 1 storage engine initialized
I170628 18:19:28.495172 59748 server/config.go:481 [n?] RocksDB cache size: 512 MiB
I170628 18:19:28.495235 59748 server/config.go:481 [n?] store 0: in-memory, size 100 MiB
I170628 18:19:28.496808 59748 server/node.go:450 [n?] store [n0,s0] not bootstrapped
I170628 18:19:28.496967 59748 storage/stores.go:296 [n?] read 0 node addresses from persistent storage
I170628 18:19:28.497131 59748 server/node.go:604 [n?] connecting to gossip network to verify cluster ID...
I170628 18:19:28.569793 60660 gossip/client.go:131 [n?] started gossip client to 127.0.0.1:60907
I170628 18:19:28.573156 60688 gossip/server.go:234 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:51287}
I170628 18:19:28.579376 60732 storage/stores.go:312 [n?] wrote 1 node addresses to persistent storage
I170628 18:19:28.579857 60732 storage/stores.go:312 [n?] wrote 2 node addresses to persistent storage
I170628 18:19:28.580963 59748 server/node.go:629 [n?] node connected via gossip and verified as part of cluster "1af3ddf7-bc64-4eaa-96da-7ae509382d89"
I170628 18:19:28.588662 59748 kv/dist_sender.go:370 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I170628 18:19:28.595038 59748 server/node.go:340 [n?] new node allocated ID 3
I170628 18:19:28.595312 59748 gossip/gossip.go:297 [n3] NodeDescriptor set to node_id:3 address:<network_field:"tcp" address_field:"127.0.0.1:51287" > attrs:<> locality:<>
I170628 18:19:28.596090 59748 server/node.go:401 [n3] node=3: started with [[]=] engine(s) and attributes []
I170628 18:19:28.598922 60734 storage/stores.go:312 [n1] wrote 2 node addresses to persistent storage
I170628 18:19:28.601072 60735 storage/stores.go:312 [n2] wrote 2 node addresses to persistent storage
I170628 18:19:28.606873 59748 sql/executor.go:358 [n3] creating distSQLPlanner with address {tcp 127.0.0.1:51287}
I170628 18:19:28.705041 59748 server/server.go:769 [n3] starting https server at 127.0.0.1:35566
I170628 18:19:28.709481 59748 server/server.go:770 [n3] starting grpc/postgres server at 127.0.0.1:51287
I170628 18:19:28.709815 59748 server/server.go:771 [n3] advertising CockroachDB node at 127.0.0.1:51287
I170628 18:19:28.719717 60458 server/node.go:585 [n3] bootstrapped store [n3,s3]
I170628 18:19:28.728480 59748 server/server.go:906 [n3] done ensuring all necessary migrations have run
I170628 18:19:28.728625 59748 server/server.go:908 [n3] serving sql connections
I170628 18:19:28.730292 59761 storage/replica_raftstorage.go:442 [replicate,n1,s1,r4/1:/{System/tse-Table/0}] generated preemptive snapshot 834298d3 at index 18
I170628 18:19:28.786057 60856 sql/event_log.go:101 [n3] Event: "node_join", target: 3, info: {Descriptor:{NodeID:3 Address:{NetworkField:tcp AddressField:127.0.0.1:51287} Attrs: Locality:} ClusterID:1af3ddf7-bc64-4eaa-96da-7ae509382d89 StartedAt:1498673968595824323 LastUp:1498673968595824323}
I170628 18:19:28.870534 59761 storage/store.go:3416 [replicate,n1,s1,r4/1:/{System/tse-Table/0}] streamed snapshot to (n2,s2):?: kv pairs: 10, log entries: 8, rate-limit: 8.0 MiB/sec, 3ms
I170628 18:19:28.871641 60885 storage/replica_raftstorage.go:639 [n2,s2,r4/?:{-}] applying preemptive snapshot at index 18 (id=834298d3, encoded size=7150, 1 rocksdb batches, 8 log entries)
I170628 18:19:28.873387 60885 storage/replica_raftstorage.go:647 [n2,s2,r4/?:/{System/tse-Table/0}] applied preemptive snapshot in 2ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:19:28.878068 59761 storage/replica_command.go:3615 [replicate,n1,s1,r4/1:/{System/tse-Table/0}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r4:/{System/tse-Table/0} [(n1,s1):1, next=2]
W170628 18:19:28.893016 59748 server/status/runtime.go:111 Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
W170628 18:19:28.943107 59748 gossip/gossip.go:1196 [n?] no incoming or outgoing connections
I170628 18:19:28.944462 59748 server/config.go:479 [n?] 1 storage engine initialized
I170628 18:19:28.944561 59748 server/config.go:481 [n?] RocksDB cache size: 512 MiB
I170628 18:19:28.944591 59748 server/config.go:481 [n?] store 0: in-memory, size 100 MiB
I170628 18:19:28.945863 59748 server/node.go:450 [n?] store [n0,s0] not bootstrapped
I170628 18:19:28.946430 59748 storage/stores.go:296 [n?] read 0 node addresses from persistent storage
I170628 18:19:28.946582 59748 server/node.go:604 [n?] connecting to gossip network to verify cluster ID...
I170628 18:19:28.949510 60887 storage/replica.go:2887 [n1,s1,r4/1:/{System/tse-Table/0}] proposing ADD_REPLICA (n2,s2):2: [(n1,s1):1 (n2,s2):2]
I170628 18:19:28.956131 59761 storage/replica_raftstorage.go:442 [replicate,n1,s1,r9/1:/Table/1{4-5}] generated preemptive snapshot 2b81cd71 at index 19
I170628 18:19:28.966638 60973 storage/raft_transport.go:456 [n2] raft transport stream to node 1 established
I170628 18:19:29.024770 60938 gossip/client.go:131 [n?] started gossip client to 127.0.0.1:60907
I170628 18:19:29.025389 61026 gossip/server.go:234 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:55080}
I170628 18:19:29.029447 61019 storage/stores.go:312 [n?] wrote 1 node addresses to persistent storage
I170628 18:19:29.029671 61019 storage/stores.go:312 [n?] wrote 2 node addresses to persistent storage
I170628 18:19:29.030017 61019 storage/stores.go:312 [n?] wrote 3 node addresses to persistent storage
I170628 18:19:29.032725 59748 server/node.go:629 [n?] node connected via gossip and verified as part of cluster "1af3ddf7-bc64-4eaa-96da-7ae509382d89"
I170628 18:19:29.039740 59748 kv/dist_sender.go:370 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I170628 18:19:29.044997 59748 server/node.go:340 [n?] new node allocated ID 4
I170628 18:19:29.045263 59748 gossip/gossip.go:297 [n4] NodeDescriptor set to node_id:4 address:<network_field:"tcp" address_field:"127.0.0.1:55080" > attrs:<> locality:<>
I170628 18:19:29.046079 59748 server/node.go:401 [n4] node=4: started with [[]=] engine(s) and attributes []
I170628 18:19:29.050814 59748 sql/executor.go:358 [n4] creating distSQLPlanner with address {tcp 127.0.0.1:55080}
I170628 18:19:29.052304 60881 storage/stores.go:312 [n1] wrote 3 node addresses to persistent storage
I170628 18:19:29.054674 61044 storage/stores.go:312 [n3] wrote 3 node addresses to persistent storage
I170628 18:19:29.067522 61045 storage/stores.go:312 [n2] wrote 3 node addresses to persistent storage
I170628 18:19:29.087904 61029 server/node.go:585 [n4] bootstrapped store [n4,s4]
I170628 18:19:29.124306 59748 server/server.go:769 [n4] starting https server at 127.0.0.1:46733
I170628 18:19:29.127303 59748 server/server.go:770 [n4] starting grpc/postgres server at 127.0.0.1:55080
I170628 18:19:29.127363 59748 server/server.go:771 [n4] advertising CockroachDB node at 127.0.0.1:55080
I170628 18:19:29.135969 59748 server/server.go:906 [n4] done ensuring all necessary migrations have run
I170628 18:19:29.136123 59748 server/server.go:908 [n4] serving sql connections
I170628 18:19:29.158397 59761 storage/store.go:3416 [replicate,n1,s1,r9/1:/Table/1{4-5}] streamed snapshot to (n3,s3):?: kv pairs: 10, log entries: 9, rate-limit: 8.0 MiB/sec, 12ms
I170628 18:19:29.160364 61220 storage/replica_raftstorage.go:639 [n3,s3,r9/?:{-}] applying preemptive snapshot at index 19 (id=2b81cd71, encoded size=5888, 1 rocksdb batches, 9 log entries)
I170628 18:19:29.165350 61220 storage/replica_raftstorage.go:647 [n3,s3,r9/?:/Table/1{4-5}] applied preemptive snapshot in 4ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:19:29.182713 59761 storage/replica_command.go:3615 [replicate,n1,s1,r9/1:/Table/1{4-5}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r9:/Table/1{4-5} [(n1,s1):1, next=2]
W170628 18:19:29.256663 59748 server/status/runtime.go:111 Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I170628 18:19:29.269370 59761 storage/replica_command.go:3615 [replicate,n1,s1,r9/1:/Table/1{4-5}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r9:/Table/1{4-5} [(n1,s1):1, next=2]
I170628 18:19:29.274478 61203 sql/event_log.go:101 [n4] Event: "node_join", target: 4, info: {Descriptor:{NodeID:4 Address:{NetworkField:tcp AddressField:127.0.0.1:55080} Attrs: Locality:} ClusterID:1af3ddf7-bc64-4eaa-96da-7ae509382d89 StartedAt:1498673969045805303 LastUp:1498673969045805303}
W170628 18:19:29.278205 59748 gossip/gossip.go:1196 [n?] no incoming or outgoing connections
I170628 18:19:29.286150 59748 server/config.go:479 [n?] 1 storage engine initialized
I170628 18:19:29.286256 59748 server/config.go:481 [n?] RocksDB cache size: 512 MiB
I170628 18:19:29.286303 59748 server/config.go:481 [n?] store 0: in-memory, size 100 MiB
I170628 18:19:29.287507 59748 server/node.go:450 [n?] store [n0,s0] not bootstrapped
I170628 18:19:29.287623 59748 storage/stores.go:296 [n?] read 0 node addresses from persistent storage
I170628 18:19:29.287726 59748 server/node.go:604 [n?] connecting to gossip network to verify cluster ID...
I170628 18:19:29.322028 61284 storage/replica.go:2887 [n1,s1,r9/1:/Table/1{4-5}] proposing ADD_REPLICA (n3,s3):2: [(n1,s1):1 (n3,s3):2]
I170628 18:19:29.341042 59761 storage/replica_raftstorage.go:442 [replicate,n1,s1,r8/1:/Table/1{3-4}] generated preemptive snapshot f3451e2a at index 29
I170628 18:19:29.347402 59761 storage/store.go:3416 [replicate,n1,s1,r8/1:/Table/1{3-4}] streamed snapshot to (n2,s2):?: kv pairs: 74, log entries: 19, rate-limit: 8.0 MiB/sec, 6ms
I170628 18:19:29.350003 61269 storage/replica_raftstorage.go:639 [n2,s2,r8/?:{-}] applying preemptive snapshot at index 29 (id=f3451e2a, encoded size=25874, 1 rocksdb batches, 19 log entries)
I170628 18:19:29.352485 61272 storage/raft_transport.go:456 [n3] raft transport stream to node 1 established
I170628 18:19:29.352809 61269 storage/replica_raftstorage.go:647 [n2,s2,r8/?:/Table/1{3-4}] applied preemptive snapshot in 3ms [clear=0ms batch=0ms entries=2ms commit=0ms]
I170628 18:19:29.358503 59761 storage/replica_command.go:3615 [replicate,n1,s1,r8/1:/Table/1{3-4}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r8:/Table/1{3-4} [(n1,s1):1, next=2]
I170628 18:19:29.388466 61278 storage/replica.go:2887 [n1,s1,r8/1:/Table/1{3-4}] proposing ADD_REPLICA (n2,s2):2: [(n1,s1):1 (n2,s2):2]
I170628 18:19:29.392519 61242 gossip/client.go:131 [n?] started gossip client to 127.0.0.1:60907
I170628 18:19:29.394936 61196 gossip/server.go:234 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:52141}
I170628 18:19:29.398870 59748 server/node.go:629 [n?] node connected via gossip and verified as part of cluster "1af3ddf7-bc64-4eaa-96da-7ae509382d89"
I170628 18:19:29.400196 61197 storage/stores.go:312 [n?] wrote 1 node addresses to persistent storage
I170628 18:19:29.403140 61198 storage/stores.go:312 [n?] wrote 2 node addresses to persistent storage
I170628 18:19:29.404525 61198 storage/stores.go:312 [n?] wrote 3 node addresses to persistent storage
I170628 18:19:29.404779 61198 storage/stores.go:312 [n?] wrote 4 node addresses to persistent storage
I170628 18:19:29.407898 59761 storage/replica_raftstorage.go:442 [replicate,n1,s1,r2/1:/System/{-tsd}] generated preemptive snapshot 58ce5105 at index 47
I170628 18:19:29.422232 59748 kv/dist_sender.go:370 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I170628 18:19:29.433788 59761 storage/store.go:3416 [replicate,n1,s1,r2/1:/System/{-tsd}] streamed snapshot to (n3,s3):?: kv pairs: 35, log entries: 4, rate-limit: 8.0 MiB/sec, 24ms
I170628 18:19:29.435133 59748 server/node.go:340 [n?] new node allocated ID 5
I170628 18:19:29.435431 59748 gossip/gossip.go:297 [n5] NodeDescriptor set to node_id:5 address:<network_field:"tcp" address_field:"127.0.0.1:52141" > attrs:<> locality:<>
I170628 18:19:29.435627 61334 storage/replica_raftstorage.go:639 [n3,s3,r2/?:{-}] applying preemptive snapshot at index 47 (id=58ce5105, encoded size=89880, 1 rocksdb batches, 4 log entries)
I170628 18:19:29.436311 59748 server/node.go:401 [n5] node=5: started with [[]=] engine(s) and attributes []
I170628 18:19:29.437666 61334 storage/replica_raftstorage.go:647 [n3,s3,r2/?:/System/{-tsd}] applied preemptive snapshot in 2ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:19:29.438808 59748 sql/executor.go:358 [n5] creating distSQLPlanner with address {tcp 127.0.0.1:52141}
I170628 18:19:29.448889 61196 gossip/server.go:285 [n1] refusing gossip from node 5 (max 3 conns); forwarding to 2 ({tcp 127.0.0.1:56027})
I170628 18:19:29.449025 59761 storage/replica_command.go:3615 [replicate,n1,s1,r2/1:/System/{-tsd}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r2:/System/{-tsd} [(n1,s1):1, next=2]
I170628 18:19:29.450520 61242 gossip/client.go:136 [n5] closing client to node 1 (127.0.0.1:60907): received forward from node 1 to 2 (127.0.0.1:56027)
I170628 18:19:29.451179 61254 gossip/gossip.go:1210 [n5] node has connected to cluster via gossip
I170628 18:19:29.451618 61254 storage/stores.go:312 [n5] wrote 4 node addresses to persistent storage
I170628 18:19:29.457497 61336 storage/stores.go:312 [n1] wrote 4 node addresses to persistent storage
I170628 18:19:29.462760 61301 storage/stores.go:312 [n3] wrote 4 node addresses to persistent storage
I170628 18:19:29.464380 61292 storage/stores.go:312 [n2] wrote 4 node addresses to persistent storage
I170628 18:19:29.464665 61302 storage/stores.go:312 [n4] wrote 4 node addresses to persistent storage
I170628 18:19:29.515681 61365 server/node.go:585 [n5] bootstrapped store [n5,s5]
I170628 18:19:29.518391 60696 storage/store.go:2165 [replicaGC,n3,s3,r2/?:/System/{-tsd}] removing replica
I170628 18:19:29.519153 59748 server/server.go:769 [n5] starting https server at 127.0.0.1:46391
I170628 18:19:29.519286 59748 server/server.go:770 [n5] starting grpc/postgres server at 127.0.0.1:52141
I170628 18:19:29.519329 59748 server/server.go:771 [n5] advertising CockroachDB node at 127.0.0.1:52141
I170628 18:19:29.520349 60696 storage/replica.go:721 [replicaGC,n3,s3,r2/?:/System/{-tsd}] removed 26 (17+9) keys in 2ms [clear=0ms commit=1ms]
I170628 18:19:29.579236 59748 server/server.go:906 [n5] done ensuring all necessary migrations have run
I170628 18:19:29.579364 59748 server/server.go:908 [n5] serving sql connections
I170628 18:19:29.586001 61296 storage/replica.go:2887 [n1,s1,r2/1:/System/{-tsd}] proposing ADD_REPLICA (n3,s3):2: [(n1,s1):1 (n3,s3):2]
I170628 18:19:29.612717 59761 storage/replica_raftstorage.go:442 [replicate,n1,s1,r1/1:/{Min-System/}] generated preemptive snapshot 8d917606 at index 57
I170628 18:19:29.617307 60040 storage/replica_raftstorage.go:442 [raftsnapshot,n1,s1,r2/1:/System/{-tsd}] generated Raft snapshot 944abd67 at index 52
I170628 18:19:29.621174 61291 gossip/client.go:131 [n5] started gossip client to 127.0.0.1:56027
I170628 18:19:29.644015 60040 storage/store.go:3416 [raftsnapshot,n1,s1,r2/1:/System/{-tsd}] streamed snapshot to (n3,s3):2: kv pairs: 40, log entries: 9, rate-limit: 8.0 MiB/sec, 15ms
I170628 18:19:29.646455 61556 storage/replica_raftstorage.go:639 [n3,s3,r2/2:{-}] applying Raft snapshot at index 52 (id=944abd67, encoded size=112186, 1 rocksdb batches, 9 log entries)
I170628 18:19:29.649500 61556 storage/replica_raftstorage.go:647 [n3,s3,r2/2:/System/{-tsd}] applied Raft snapshot in 3ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:19:29.698268 61312 sql/event_log.go:101 [n5] Event: "node_join", target: 5, info: {Descriptor:{NodeID:5 Address:{NetworkField:tcp AddressField:127.0.0.1:52141} Attrs: Locality:} ClusterID:1af3ddf7-bc64-4eaa-96da-7ae509382d89 StartedAt:1498673969436006758 LastUp:1498673969436006758}
I170628 18:19:29.734279 59761 storage/store.go:3416 [replicate,n1,s1,r1/1:/{Min-System/}] streamed snapshot to (n4,s4):?: kv pairs: 37, log entries: 47, rate-limit: 8.0 MiB/sec, 3ms
I170628 18:19:29.736262 61592 storage/replica_raftstorage.go:639 [n4,s4,r1/?:{-}] applying preemptive snapshot at index 57 (id=8d917606, encoded size=26870, 1 rocksdb batches, 47 log entries)
I170628 18:19:29.740770 61592 storage/replica_raftstorage.go:647 [n4,s4,r1/?:/{Min-System/}] applied preemptive snapshot in 4ms [clear=0ms batch=0ms entries=3ms commit=0ms]
I170628 18:19:29.746051 59761 storage/replica_command.go:3615 [replicate,n1,s1,r1/1:/{Min-System/}] change replicas (ADD_REPLICA (n4,s4):2): read existing descriptor r1:/{Min-System/} [(n1,s1):1, next=2]
I170628 18:19:29.789160 61530 storage/replica.go:2887 [n1,s1,r1/1:/{Min-System/}] proposing ADD_REPLICA (n4,s4):2: [(n1,s1):1 (n4,s4):2]
I170628 18:19:29.795848 59761 storage/replica_raftstorage.go:442 [replicate,n1,s1,r5/1:/Table/{0-11}] generated preemptive snapshot 6f476a10 at index 23
I170628 18:19:29.835863 61535 storage/raft_transport.go:456 [n4] raft transport stream to node 1 established
I170628 18:19:29.899633 59761 storage/store.go:3416 [replicate,n1,s1,r5/1:/Table/{0-11}] streamed snapshot to (n5,s5):?: kv pairs: 39, log entries: 13, rate-limit: 8.0 MiB/sec, 6ms
I170628 18:19:29.901866 61671 storage/replica_raftstorage.go:639 [n5,s5,r5/?:{-}] applying preemptive snapshot at index 23 (id=6f476a10, encoded size=16399, 1 rocksdb batches, 13 log entries)
I170628 18:19:29.903920 61671 storage/replica_raftstorage.go:647 [n5,s5,r5/?:/Table/{0-11}] applied preemptive snapshot in 2ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:19:29.909102 59761 storage/replica_command.go:3615 [replicate,n1,s1,r5/1:/Table/{0-11}] change replicas (ADD_REPLICA (n5,s5):2): read existing descriptor r5:/Table/{0-11} [(n1,s1):1, next=2]
I170628 18:19:29.945750 61706 storage/replica.go:2887 [n1,s1,r5/1:/Table/{0-11}] proposing ADD_REPLICA (n5,s5):2: [(n1,s1):1 (n5,s5):2]
I170628 18:19:29.950290 59761 storage/replica_raftstorage.go:442 [replicate,n1,s1,r6/1:/Table/1{1-2}] generated preemptive snapshot a017ec3c at index 19
I170628 18:19:29.964296 59761 storage/store.go:3416 [replicate,n1,s1,r6/1:/Table/1{1-2}] streamed snapshot to (n3,s3):?: kv pairs: 10, log entries: 9, rate-limit: 8.0 MiB/sec, 13ms
I170628 18:19:29.966145 61677 storage/replica_raftstorage.go:639 [n3,s3,r6/?:{-}] applying preemptive snapshot at index 19 (id=a017ec3c, encoded size=8472, 1 rocksdb batches, 9 log entries)
I170628 18:19:29.968087 61677 storage/replica_raftstorage.go:647 [n3,s3,r6/?:/Table/1{1-2}] applied preemptive snapshot in 2ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:19:30.025825 59761 storage/replica_command.go:3615 [replicate,n1,s1,r6/1:/Table/1{1-2}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r6:/Table/1{1-2} [(n1,s1):1, next=2]
I170628 18:19:30.036406 61764 storage/raft_transport.go:456 [n5] raft transport stream to node 1 established
I170628 18:19:30.102937 60696 storage/store.go:2165 [replicaGC,n3,s3,r6/?:/Table/1{1-2}] removing replica
I170628 18:19:30.104466 60696 storage/replica.go:721 [replicaGC,n3,s3,r6/?:/Table/1{1-2}] removed 9 (0+9) keys in 1ms [clear=1ms commit=0ms]
I170628 18:19:30.128148 61665 storage/replica.go:2887 [n1,s1,r6/1:/Table/1{1-2}] proposing ADD_REPLICA (n3,s3):2: [(n1,s1):1 (n3,s3):2]
I170628 18:19:30.137554 59761 storage/replica_raftstorage.go:442 [replicate,n1,s1,r10/1:/{Table/15-Max}] generated preemptive snapshot 4dccf8fa at index 11
I170628 18:19:30.143860 59761 storage/store.go:3416 [replicate,n1,s1,r10/1:/{Table/15-Max}] streamed snapshot to (n4,s4):?: kv pairs: 9, log entries: 1, rate-limit: 8.0 MiB/sec, 4ms
I170628 18:19:30.147099 61696 storage/replica_raftstorage.go:639 [n4,s4,r10/?:{-}] applying preemptive snapshot at index 11 (id=4dccf8fa, encoded size=548, 1 rocksdb batches, 1 log entries)
I170628 18:19:30.148382 61696 storage/replica_raftstorage.go:647 [n4,s4,r10/?:/{Table/15-Max}] applied preemptive snapshot in 1ms [clear=0ms batch=0ms entries=0ms commit=0ms]
I170628 18:19:30.152879 59761 storage/replica_command.go:3615 [replicate,n1,s1,r10/1:/{Table/15-Max}] change replicas (ADD_REPLICA (n4,s4):2): read existing descriptor r10:/{Table/15-Max} [(n1,s1):1, next=2]
I170628 18:19:30.159491 60040 storage/replica_raftstorage.go:442 [raftsnapshot,n1,s1,r6/1:/Table/1{1-2}] generated Raft snapshot 6f92afff at index 21
I170628 18:19:30.164192 60040 storage/store.go:3416 [raftsnapshot,n1,s1,r6/1:/Table/1{1-2}] streamed snapshot to (n3,s3):2: kv pairs: 12, log entries: 11, rate-limit: 8.0 MiB/sec, 4ms
I170628 18:19:30.166648 61844 storage/replica_raftstorage.go:639 [n3,s3,r6/2:{-}] applying Raft snapshot at index 21 (id=6f92afff, encoded size=10427, 1 rocksdb batches, 11 log entries)
I170628 18:19:30.171939 61844 storage/replica_raftstorage.go:647 [n3,s3,r6/2:/Table/1{1-2}] applied Raft snapshot in 5ms [clear=0ms batch=0ms entries=2ms commit=0ms]
I170628 18:19:30.214291 61805 storage/replica.go:2887 [n1,s1,r10/1:/{Table/15-Max}] proposing ADD_REPLICA (n4,s4):2: [(n1,s1):1 (n4,s4):2]
I170628 18:19:30.221353 59761 storage/replica_raftstorage.go:442 [replicate,n1,s1,r3/1:/System/ts{d-e}] generated preemptive snapshot aa13cdf1 at index 29
I170628 18:19:30.275186 59761 storage/store.go:3416 [replicate,n1,s1,r3/1:/System/ts{d-e}] streamed snapshot to (n5,s5):?: kv pairs: 1404, log entries: 2, rate-limit: 8.0 MiB/sec, 52ms
I170628 18:19:30.283673 61729 storage/replica_raftstorage.go:639 [n5,s5,r3/?:{-}] applying preemptive snapshot at index 29 (id=aa13cdf1, encoded size=170084, 1 rocksdb batches, 2 log entries)
I170628 18:19:30.288120 61729 storage/replica_raftstorage.go:647 [n5,s5,r3/?:/System/ts{d-e}] applied preemptive snapshot in 4ms [clear=0ms batch=0ms entries=0ms commit=1ms]
I170628 18:19:30.293251 59761 storage/replica_command.go:3615 [replicate,n1,s1,r3/1:/System/ts{d-e}] change replicas (ADD_REPLICA (n5,s5):2): read existing descriptor r3:/System/ts{d-e} [(n1,s1):1, next=2]
I170628 18:19:30.340202 61897 storage/replica.go:2887 [n1,s1,r3/1:/System/ts{d-e}] proposing ADD_REPLICA (n5,s5):2: [(n1,s1):1 (n5,s5):2]
I170628 18:19:30.347308 59761 storage/replica_raftstorage.go:442 [replicate,n1,s1,r7/1:/Table/1{2-3}] generated preemptive snapshot 9da60953 at index 29
I170628 18:19:30.367821 59761 storage/store.go:3416 [replicate,n1,s1,r7/1:/Table/1{2-3}] streamed snapshot to (n4,s4):?: kv pairs: 40, log entries: 19, rate-limit: 8.0 MiB/sec, 7ms
I170628 18:19:30.371450 61834 storage/replica_raftstorage.go:639 [n4,s4,r7/?:{-}] applying preemptive snapshot at index 29 (id=9da60953, encoded size=22998, 1 rocksdb batches, 19 log entries)
I170628 18:19:30.384499 61834 storage/replica_raftstorage.go:647 [n4,s4,r7/?:/Table/1{2-3}] applied preemptive snapshot in 13ms [clear=0ms batch=0ms entries=1ms commit=2ms]
I170628 18:19:30.390455 59761 storage/replica_command.go:3615 [replicate,n1,s1,r7/1:/Table/1{2-3}] change replicas (ADD_REPLICA (n4,s4):2): read existing descriptor r7:/Table/1{2-3} [(n1,s1):1, next=2]
I170628 18:19:30.430196 61882 storage/replica.go:2887 [n1,s1,r7/1:/Table/1{2-3}] proposing ADD_REPLICA (n4,s4):2: [(n1,s1):1 (n4,s4):2]
I170628 18:19:30.447304 59761 storage/queue.go:725 [n1,replicate] purgatory is now empty
I170628 18:19:30.449102 60037 storage/replica_raftstorage.go:442 [replicate,n1,s1,r8/1:/Table/1{3-4}] generated preemptive snapshot d3cbfc51 at index 46
I170628 18:19:30.462464 60037 storage/store.go:3416 [replicate,n1,s1,r8/1:/Table/1{3-4}] streamed snapshot to (n5,s5):?: kv pairs: 120, log entries: 36, rate-limit: 8.0 MiB/sec, 8ms
I170628 18:19:30.464628 61867 storage/replica_raftstorage.go:639 [n5,s5,r8/?:{-}] applying preemptive snapshot at index 46 (id=d3cbfc51, encoded size=47495, 1 rocksdb batches, 36 log entries)
I170628 18:19:30.470105 61867 storage/replica_raftstorage.go:647 [n5,s5,r8/?:/Table/1{3-4}] applied preemptive snapshot in 5ms [clear=0ms batch=0ms entries=3ms commit=1ms]
I170628 18:19:30.477830 60037 storage/replica_command.go:3615 [replicate,n1,s1,r8/1:/Table/1{3-4}] change replicas (ADD_REPLICA (n5,s5):3): read existing descriptor r8:/Table/1{3-4} [(n1,s1):1, (n2,s2):2, next=3]
I170628 18:19:30.541988 61496 storage/store.go:2165 [replicaGC,n5,s5,r8/?:/Table/1{3-4}] removing replica
I170628 18:19:30.542818 61496 storage/replica.go:721 [replicaGC,n5,s5,r8/?:/Table/1{3-4}] removed 113 (104+9) keys in 0ms [clear=0ms commit=0ms]
I170628 18:19:30.545050 61825 storage/replica.go:2887 [n1,s1,r8/1:/Table/1{3-4}] proposing ADD_REPLICA (n5,s5):3: [(n1,s1):1 (n2,s2):2 (n5,s5):3]
I170628 18:19:30.555962 60037 storage/replica_raftstorage.go:442 [replicate,n1,s1,r4/1:/{System/tse-Table/0}] generated preemptive snapshot c1dc6eca at index 22
I170628 18:19:30.560213 60037 storage/store.go:3416 [replicate,n1,s1,r4/1:/{System/tse-Table/0}] streamed snapshot to (n5,s5):?: kv pairs: 11, log entries: 12, rate-limit: 8.0 MiB/sec, 4ms
I170628 18:19:30.563486 62003 storage/replica_raftstorage.go:639 [n5,s5,r4/?:{-}] applying preemptive snapshot at index 22 (id=c1dc6eca, encoded size=9647, 1 rocksdb batches, 12 log entries)
I170628 18:19:30.565824 62003 storage/replica_raftstorage.go:647 [n5,s5,r4/?:/{System/tse-Table/0}] applied preemptive snapshot in 2ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:19:30.570757 60037 storage/replica_command.go:3615 [replicate,n1,s1,r4/1:/{System/tse-Table/0}] change replicas (ADD_REPLICA (n5,s5):3): read existing descriptor r4:/{System/tse-Table/0} [(n1,s1):1, (n2,s2):2, next=3]
I170628 18:19:30.596395 60040 storage/replica_raftstorage.go:442 [raftsnapshot,n1,s1,r8/1:/Table/1{3-4}] generated Raft snapshot e879a0a4 at index 50
I170628 18:19:30.606719 61913 storage/replica_raftstorage.go:639 [n5,s5,r8/3:{-}] applying Raft snapshot at index 50 (id=e879a0a4, encoded size=51409, 1 rocksdb batches, 40 log entries)
I170628 18:19:30.617347 61991 storage/replica.go:2887 [n1,s1,r4/1:/{System/tse-Table/0}] proposing ADD_REPLICA (n5,s5):3: [(n1,s1):1 (n2,s2):2 (n5,s5):3]
I170628 18:19:30.619964 60040 storage/store.go:3416 [raftsnapshot,n1,s1,r8/1:/Table/1{3-4}] streamed snapshot to (n5,s5):3: kv pairs: 122, log entries: 40, rate-limit: 8.0 MiB/sec, 22ms
I170628 18:19:30.630824 60037 storage/replica_raftstorage.go:442 [replicate,n1,s1,r3/1:/System/ts{d-e}] generated preemptive snapshot b20eaa8f at index 33
I170628 18:19:30.643256 61913 storage/replica_raftstorage.go:647 [n5,s5,r8/3:/Table/1{3-4}] applied Raft snapshot in 36ms [clear=3ms batch=0ms entries=32ms commit=1ms]
I170628 18:19:30.697769 60037 storage/store.go:3416 [replicate,n1,s1,r3/1:/System/ts{d-e}] streamed snapshot to (n3,s3):?: kv pairs: 1405, log entries: 6, rate-limit: 8.0 MiB/sec, 60ms
I170628 18:19:30.703182 62013 storage/replica_raftstorage.go:639 [n3,s3,r3/?:{-}] applying preemptive snapshot at index 33 (id=b20eaa8f, encoded size=172593, 1 rocksdb batches, 6 log entries)
I170628 18:19:30.705901 62013 storage/replica_raftstorage.go:647 [n3,s3,r3/?:/System/ts{d-e}] applied preemptive snapshot in 3ms [clear=0ms batch=0ms entries=1ms commit=1ms]
I170628 18:19:30.711595 60037 storage/replica_command.go:3615 [replicate,n1,s1,r3/1:/System/ts{d-e}] change replicas (ADD_REPLICA (n3,s3):3): read existing descriptor r3:/System/ts{d-e} [(n1,s1):1, (n5,s5):2, next=3]
I170628 18:19:30.763362 61920 storage/replica.go:2887 [n1,s1,r3/1:/System/ts{d-e}] proposing ADD_REPLICA (n3,s3):3: [(n1,s1):1 (n5,s5):2 (n3,s3):3]
I170628 18:19:30.804030 60037 storage/replica_raftstorage.go:442 [replicate,n1,s1,r2/1:/System/{-tsd}] generated preemptive snapshot 2dd698b4 at index 57
I170628 18:19:30.839688 60037 storage/store.go:3416 [replicate,n1,s1,r2/1:/System/{-tsd}] streamed snapshot to (n2,s2):?: kv pairs: 40, log entries: 14, rate-limit: 8.0 MiB/sec, 35ms
I170628 18:19:30.841510 62025 storage/replica_raftstorage.go:639 [n2,s2,r2/?:{-}] applying preemptive snapshot at index 57 (id=2dd698b4, encoded size=135175, 1 rocksdb batches, 14 log entries)
I170628 18:19:30.844632 62025 storage/replica_raftstorage.go:647 [n2,s2,r2/?:/System/{-tsd}] applied preemptive snapshot in 3ms [clear=0ms batch=0ms entries=2ms commit=1ms]
I170628 18:19:30.851728 60037 storage/replica_command.go:3615 [replicate,n1,s1,r2/1:/System/{-tsd}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r2:/System/{-tsd} [(n1,s1):1, (n3,s3):2, next=3]
I170628 18:19:30.918014 60536 storage/store.go:2165 [replicaGC,n2,s2,r2/?:/System/{-tsd}] removing replica
I170628 18:19:30.918848 60536 storage/replica.go:721 [replicaGC,n2,s2,r2/?:/System/{-tsd}] removed 28 (19+9) keys in 0ms [clear=0ms commit=0ms]
I170628 18:19:30.925877 62045 storage/replica.go:2887 [n1,s1,r2/1:/System/{-tsd}] proposing ADD_REPLICA (n2,s2):3: [(n1,s1):1 (n3,s3):2 (n2,s2):3]
I170628 18:19:30.958636 60037 storage/replica_raftstorage.go:442 [replicate,n1,s1,r7/1:/Table/1{2-3}] generated preemptive snapshot cbbbdfc4 at index 34
I170628 18:19:30.964708 60037 storage/store.go:3416 [replicate,n1,s1,r7/1:/Table/1{2-3}] streamed snapshot to (n2,s2):?: kv pairs: 41, log entries: 24, rate-limit: 8.0 MiB/sec, 5ms
I170628 18:19:30.966865 62089 storage/replica_raftstorage.go:639 [n2,s2,r7/?:{-}] applying preemptive snapshot at index 34 (id=cbbbdfc4, encoded size=25826, 1 rocksdb batches, 24 log entries)
I170628 18:19:30.969905 62089 storage/replica_raftstorage.go:647 [n2,s2,r7/?:/Table/1{2-3}] applied preemptive snapshot in 3ms [clear=0ms batch=0ms entries=2ms commit=0ms]
I170628 18:19:30.975758 60037 storage/replica_command.go:3615 [replicate,n1,s1,r7/1:/Table/1{2-3}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r7:/Table/1{2-3} [(n1,s1):1, (n4,s4):2, next=3]
I170628 18:19:30.999837 60040 storage/replica_raftstorage.go:442 [raftsnapshot,n1,s1,r2/1:/System/{-tsd}] generated Raft snapshot 0e88df0d at index 60
I170628 18:19:31.010932 60040 storage/store.go:3416 [raftsnapshot,n1,s1,r2/1:/System/{-tsd}] streamed snapshot to (n2,s2):3: kv pairs: 41, log entries: 17, rate-limit: 8.0 MiB/sec, 10ms
I170628 18:19:31.023660 62132 storage/replica_raftstorage.go:639 [n2,s2,r2/3:{-}] applying Raft snapshot at index 60 (id=0e88df0d, encoded size=137259, 1 rocksdb batches, 17 log entries)
I170628 18:19:31.027405 62132 storage/replica_raftstorage.go:647 [n2,s2,r2/3:/System/{-tsd}] applied Raft snapshot in 4ms [clear=0ms batch=0ms entries=2ms commit=1ms]
I170628 18:19:31.051320 60536 storage/store.go:2165 [replicaGC,n2,s2,r7/?:/Table/1{2-3}] removing replica
I170628 18:19:31.067706 60536 storage/replica.go:721 [replicaGC,n2,s2,r7/?:/Table/1{2-3}] removed 39 (30+9) keys in 16ms [clear=0ms commit=16ms]
I170628 18:19:31.072259 62058 storage/replica.go:2887 [n1,s1,r7/1:/Table/1{2-3}] proposing ADD_REPLICA (n2,s2):3: [(n1,s1):1 (n4,s4):2 (n2,s2):3]
I170628 18:19:31.089686 60037 storage/replica_raftstorage.go:442 [replicate,n1,s1,r5/1:/Table/{0-11}] generated preemptive snapshot 59b945ac at index 28
I170628 18:19:31.096495 60037 storage/store.go:3416 [replicate,n1,s1,r5/1:/Table/{0-11}] streamed snapshot to (n4,s4):?: kv pairs: 40, log entries: 18, rate-limit: 8.0 MiB/sec, 6ms
I170628 18:19:31.098948 62097 storage/replica_raftstorage.go:639 [n4,s4,r5/?:{-}] applying preemptive snapshot at index 28 (id=59b945ac, encoded size=19227, 1 rocksdb batches, 18 log entries)
I170628 18:19:31.101251 62097 storage/replica_raftstorage.go:647 [n4,s4,r5/?:/Table/{0-11}] applied preemptive snapshot in 2ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:19:31.110786 60037 storage/replica_command.go:3615 [replicate,n1,s1,r5/1:/Table/{0-11}] change replicas (ADD_REPLICA (n4,s4):3): read existing descriptor r5:/Table/{0-11} [(n1,s1):1, (n5,s5):2, next=3]
I170628 18:19:31.150104 60040 storage/replica_raftstorage.go:442 [raftsnapshot,n1,s1,r7/1:/Table/1{2-3}] generated Raft snapshot 8ab855b2 at index 37
I170628 18:19:31.182644 60040 storage/store.go:3416 [raftsnapshot,n1,s1,r7/1:/Table/1{2-3}] streamed snapshot to (n2,s2):3: kv pairs: 42, log entries: 27, rate-limit: 8.0 MiB/sec, 9ms
I170628 18:19:31.183564 62063 storage/replica_raftstorage.go:639 [n2,s2,r7/3:{-}] applying Raft snapshot at index 37 (id=8ab855b2, encoded size=27898, 1 rocksdb batches, 27 log entries)
I170628 18:19:31.191245 62063 storage/replica_raftstorage.go:647 [n2,s2,r7/3:/Table/1{2-3}] applied Raft snapshot in 8ms [clear=0ms batch=0ms entries=5ms commit=0ms]
I170628 18:19:31.200817 62199 storage/replica.go:2887 [n1,s1,r5/1:/Table/{0-11}] proposing ADD_REPLICA (n4,s4):3: [(n1,s1):1 (n5,s5):2 (n4,s4):3]
I170628 18:19:31.230850 60037 storage/replica_raftstorage.go:442 [replicate,n1,s1,r1/1:/{Min-System/}] generated preemptive snapshot b85d771c at index 87
I170628 18:19:31.261757 60037 storage/store.go:3416 [replicate,n1,s1,r1/1:/{Min-System/}] streamed snapshot to (n5,s5):?: kv pairs: 51, log entries: 77, rate-limit: 8.0 MiB/sec, 29ms
I170628 18:19:31.269096 62212 storage/replica_raftstorage.go:639 [n5,s5,r1/?:{-}] applying preemptive snapshot at index 87 (id=b85d771c, encoded size=42192, 1 rocksdb batches, 77 log entries)
I170628 18:19:31.280988 62212 storage/replica_raftstorage.go:647 [n5,s5,r1/?:/{Min-System/}] applied preemptive snapshot in 11ms [clear=0ms batch=0ms entries=8ms commit=1ms]
I170628 18:19:31.331934 60037 storage/replica_command.go:3615 [replicate,n1,s1,r1/1:/{Min-System/}] change replicas (ADD_REPLICA (n5,s5):3): read existing descriptor r1:/{Min-System/} [(n1,s1):1, (n4,s4):2, next=3]
I170628 18:19:31.339130 61496 storage/store.go:2165 [replicaGC,n5,s5,r1/?:/{Min-System/}] removing replica
I170628 18:19:31.339889 61496 storage/replica.go:721 [replicaGC,n5,s5,r1/?:/{Min-System/}] removed 21 (11+10) keys in 0ms [clear=0ms commit=0ms]
I170628 18:19:31.402815 62169 storage/replica.go:2887 [n1,s1,r1/1:/{Min-System/}] proposing ADD_REPLICA (n5,s5):3: [(n1,s1):1 (n4,s4):2 (n5,s5):3]
I170628 18:19:31.428621 60037 storage/replica_raftstorage.go:442 [replicate,n1,s1,r9/1:/Table/1{4-5}] generated preemptive snapshot a1f46630 at index 25
I170628 18:19:31.432257 60037 storage/store.go:3416 [replicate,n1,s1,r9/1:/Table/1{4-5}] streamed snapshot to (n4,s4):?: kv pairs: 11, log entries: 15, rate-limit: 8.0 MiB/sec, 3ms
I170628 18:19:31.436971 62203 storage/replica_raftstorage.go:639 [n4,s4,r9/?:{-}] applying preemptive snapshot at index 25 (id=a1f46630, encoded size=9427, 1 rocksdb batches, 15 log entries)
I170628 18:19:31.442506 62203 storage/replica_raftstorage.go:647 [n4,s4,r9/?:/Table/1{4-5}] applied preemptive snapshot in 5ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:19:31.455773 60037 storage/replica_command.go:3615 [replicate,n1,s1,r9/1:/Table/1{4-5}] change replicas (ADD_REPLICA (n4,s4):3): read existing descriptor r9:/Table/1{4-5} [(n1,s1):1, (n3,s3):2, next=3]
I170628 18:19:31.480824 60040 storage/replica_raftstorage.go:442 [raftsnapshot,n1,s1,r1/1:/{Min-System/}] generated Raft snapshot 42ddeed5 at index 93
I170628 18:19:31.511401 61049 storage/store.go:2165 [replicaGC,n4,s4,r9/?:/Table/1{4-5}] removing replica
I170628 18:19:31.516246 61049 storage/replica.go:721 [replicaGC,n4,s4,r9/?:/Table/1{4-5}] removed 9 (0+9) keys in 4ms [clear=4ms commit=0ms]
I170628 18:19:31.516943 60040 storage/store.go:3416 [raftsnapshot,n1,s1,r1/1:/{Min-System/}] streamed snapshot to (n5,s5):3: kv pairs: 53, log entries: 83, rate-limit: 8.0 MiB/sec, 29ms
I170628 18:19:31.518817 62276 storage/replica_raftstorage.go:639 [n5,s5,r1/3:{-}] applying Raft snapshot at index 93 (id=42ddeed5, encoded size=45885, 1 rocksdb batches, 83 log entries)
I170628 18:19:31.531955 62276 storage/replica_raftstorage.go:647 [n5,s5,r1/3:/{Min-System/}] applied Raft snapshot in 13ms [clear=0ms batch=0ms entries=11ms commit=0ms]
I170628 18:19:31.574326 62322 storage/replica.go:2887 [n1,s1,r9/1:/Table/1{4-5}] proposing ADD_REPLICA (n4,s4):3: [(n1,s1):1 (n3,s3):2 (n4,s4):3]
W170628 18:19:31.585532 62176 storage/replica.go:2528 [n1,s1,r2/1:/System/{-tsd}] con
```
Please assign, take a look and update the issue accordingly. | 1.0 | teamcity: failed tests on master: testrace/TestStoreRangeCorruptionChangeReplicas, testrace/TestSystemZoneConfigs - The following tests appear to have failed:
[#282905](https://teamcity.cockroachdb.com/viewLog.html?buildId=282905):
```
--- FAIL: testrace/TestStoreRangeCorruptionChangeReplicas (0.900s)
client_raft_test.go:1210: [NotLeaseHolderError] r1: replica (n1,s1):1 not lease holder; current lease is repl=(n4,s4):4 start=0.000000123,427 exp=0.900000123,427 pro=0.000000123,428
------- Stdout: -------
I170628 18:18:42.039506 32093 gossip/gossip.go:297 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:54812" > attrs:<> locality:<>
W170628 18:18:42.051538 32093 gossip/gossip.go:1196 [n2] no incoming or outgoing connections
I170628 18:18:42.055676 32365 gossip/client.go:131 [n2] started gossip client to 127.0.0.1:54812
I170628 18:18:42.069941 32093 gossip/gossip.go:297 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:48975" > attrs:<> locality:<>
W170628 18:18:42.083864 32093 gossip/gossip.go:1196 [n3] no incoming or outgoing connections
I170628 18:18:42.093278 32428 gossip/client.go:131 [n3] started gossip client to 127.0.0.1:54812
I170628 18:18:42.103542 32093 storage/store.go:1265 [n3,s3]: failed initial metrics computation: [n3,s3]: system config not yet available
I170628 18:18:42.103846 32093 gossip/gossip.go:297 [n3] NodeDescriptor set to node_id:3 address:<network_field:"tcp" address_field:"127.0.0.1:34226" > attrs:<> locality:<>
W170628 18:18:42.127052 32093 gossip/gossip.go:1196 [n4] no incoming or outgoing connections
I170628 18:18:42.131932 32633 gossip/client.go:131 [n4] started gossip client to 127.0.0.1:54812
I170628 18:18:42.142942 32093 gossip/gossip.go:297 [n4] NodeDescriptor set to node_id:4 address:<network_field:"tcp" address_field:"127.0.0.1:40101" > attrs:<> locality:<>
W170628 18:18:42.182637 32093 gossip/gossip.go:1196 [n5] no incoming or outgoing connections
I170628 18:18:42.193550 32093 storage/store.go:1265 [n5,s5]: failed initial metrics computation: [n5,s5]: system config not yet available
I170628 18:18:42.193833 32093 gossip/gossip.go:297 [n5] NodeDescriptor set to node_id:5 address:<network_field:"tcp" address_field:"127.0.0.1:34547" > attrs:<> locality:<>
I170628 18:18:42.197140 32617 gossip/client.go:131 [n5] started gossip client to 127.0.0.1:54812
I170628 18:18:42.198251 32749 gossip/server.go:285 [n1] refusing gossip from node 5 (max 3 conns); forwarding to 3 ({tcp 127.0.0.1:34226})
I170628 18:18:42.206388 32617 gossip/client.go:136 [n5] closing client to node 1 (127.0.0.1:54812): received forward from node 1 to 3 (127.0.0.1:34226)
I170628 18:18:42.206818 32771 gossip/gossip.go:1210 [n5] node has connected to cluster via gossip
I170628 18:18:42.207910 32584 gossip/client.go:131 [n5] started gossip client to 127.0.0.1:34226
W170628 18:18:42.286554 32093 gossip/gossip.go:1196 [n6] no incoming or outgoing connections
I170628 18:18:42.288066 32190 gossip/client.go:131 [n6] started gossip client to 127.0.0.1:54812
I170628 18:18:42.289202 32696 gossip/server.go:285 [n1] refusing gossip from node 6 (max 3 conns); forwarding to 2 ({tcp 127.0.0.1:48975})
I170628 18:18:42.292831 32190 gossip/client.go:136 [n6] closing client to node 1 (127.0.0.1:54812): received forward from node 1 to 2 (127.0.0.1:48975)
I170628 18:18:42.294080 32916 gossip/gossip.go:1210 [n6] node has connected to cluster via gossip
I170628 18:18:42.295214 32855 gossip/client.go:131 [n6] started gossip client to 127.0.0.1:48975
I170628 18:18:42.313508 32093 gossip/gossip.go:297 [n6] NodeDescriptor set to node_id:6 address:<network_field:"tcp" address_field:"127.0.0.1:34349" > attrs:<> locality:<>
I170628 18:18:42.373654 32093 storage/client_test.go:410 gossip network initialized
I170628 18:18:42.374929 32093 storage/replica_raftstorage.go:442 [replicate,s1,r1/1:/M{in-ax}] generated preemptive snapshot ec6a1960 at index 18
I170628 18:18:42.389357 32093 storage/store.go:3416 [replicate,s1,r1/1:/M{in-ax}] streamed snapshot to (n6,s6):?: kv pairs: 36, log entries: 8, rate-limit: 8.0 MiB/sec, 14ms
I170628 18:18:42.390259 32938 storage/replica_raftstorage.go:639 [s6,r1/?:{-}] applying preemptive snapshot at index 18 (id=ec6a1960, encoded size=7114, 1 rocksdb batches, 8 log entries)
I170628 18:18:42.391805 32938 storage/replica_raftstorage.go:647 [s6,r1/?:/M{in-ax}] applied preemptive snapshot in 1ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:18:42.393956 32093 storage/replica_command.go:3615 [replicate,s1,r1/1:/M{in-ax}] change replicas (ADD_REPLICA (n6,s6):2): read existing descriptor r1:/M{in-ax} [(n1,s1):1, next=2]
I170628 18:18:42.398815 33092 storage/replica.go:2887 [replicate,s1,r1/1:/M{in-ax}] proposing ADD_REPLICA (n6,s6):2: [(n1,s1):1 (n6,s6):2]
I170628 18:18:42.402861 32093 storage/replica_raftstorage.go:442 [replicate,s1,r1/1:/M{in-ax}] generated preemptive snapshot 30e6642c at index 20
I170628 18:18:42.408045 32093 storage/store.go:3416 [replicate,s1,r1/1:/M{in-ax}] streamed snapshot to (n3,s3):?: kv pairs: 39, log entries: 10, rate-limit: 8.0 MiB/sec, 3ms
I170628 18:18:42.412534 33052 storage/replica_raftstorage.go:639 [s3,r1/?:{-}] applying preemptive snapshot at index 20 (id=30e6642c, encoded size=8680, 1 rocksdb batches, 10 log entries)
I170628 18:18:42.414655 33052 storage/replica_raftstorage.go:647 [s3,r1/?:/M{in-ax}] applied preemptive snapshot in 2ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:18:42.436549 32093 storage/replica_command.go:3615 [replicate,s1,r1/1:/M{in-ax}] change replicas (ADD_REPLICA (n3,s3):3): read existing descriptor r1:/M{in-ax} [(n1,s1):1, (n6,s6):2, next=3]
I170628 18:18:42.459365 33124 storage/raft_transport.go:456 raft transport stream to node 1 established
I170628 18:18:42.468188 33037 storage/replica.go:2887 [replicate,s1,r1/1:/M{in-ax}] proposing ADD_REPLICA (n3,s3):3: [(n1,s1):1 (n6,s6):2 (n3,s3):3]
E170628 18:18:42.491831 32925 storage/replica.go:4924 [s6,r1/2:/M{in-ax}] stalling replica due to: boom
I170628 18:18:42.498379 32093 storage/replica_command.go:3615 [replicate,s1,r1/1:/M{in-ax}] change replicas (REMOVE_REPLICA (n6,s6):2): read existing descriptor r1:/M{in-ax} [(n1,s1):1, (n6,s6):2, (n3,s3):3, next=4]
W170628 18:18:42.515704 33123 storage/store.go:3231 [s1] got error from r1, replica (n6,s6):2: replica corruption (processed=true): boom
W170628 18:18:42.519010 33123 storage/store.go:3231 [s1] got error from r1, replica (n6,s6):2: replica corruption (processed=true): boom
W170628 18:18:42.519165 33123 storage/store.go:3231 [s1] got error from r1, replica (n6,s6):2: replica corruption (processed=true): boom
I170628 18:18:42.523686 33073 storage/replica.go:2887 [replicate,s1,r1/1:/M{in-ax}] proposing REMOVE_REPLICA (n6,s6):2: [(n1,s1):1 (n3,s3):3]
W170628 18:18:42.530618 33123 storage/store.go:3231 [s1] got error from r1, replica (n6,s6):2: replica corruption (processed=true): boom
W170628 18:18:42.536108 33123 storage/store.go:3231 [s1] got error from r1, replica (n6,s6):2: replica corruption (processed=true): boom
I170628 18:18:42.540300 32093 storage/replica_raftstorage.go:442 [replicate,s1,r1/1:/M{in-ax}] generated preemptive snapshot a570ed02 at index 25
I170628 18:18:42.554709 32093 storage/store.go:3416 [replicate,s1,r1/1:/M{in-ax}] streamed snapshot to (n4,s4):?: kv pairs: 46, log entries: 15, rate-limit: 8.0 MiB/sec, 13ms
I170628 18:18:42.555791 33119 storage/replica_raftstorage.go:639 [s4,r1/?:{-}] applying preemptive snapshot at index 25 (id=a570ed02, encoded size=12373, 1 rocksdb batches, 15 log entries)
I170628 18:18:42.562230 33119 storage/replica_raftstorage.go:647 [s4,r1/?:/M{in-ax}] applied preemptive snapshot in 6ms [clear=0ms batch=0ms entries=1ms commit=4ms]
I170628 18:18:42.579036 32093 storage/replica_command.go:3615 [replicate,s1,r1/1:/M{in-ax}] change replicas (ADD_REPLICA (n4,s4):4): read existing descriptor r1:/M{in-ax} [(n1,s1):1, (n3,s3):3, next=4]
I170628 18:18:42.588160 33088 storage/replica.go:2887 [replicate,s1,r1/1:/M{in-ax}] proposing ADD_REPLICA (n4,s4):4: [(n1,s1):1 (n3,s3):3 (n4,s4):4]
E170628 18:18:42.600565 32518 storage/replica.go:4924 [s3,r1/3:/M{in-ax}] stalling replica due to: boom
I170628 18:18:42.615557 32093 storage/replica_command.go:3615 [replicate,s1,r1/1:/M{in-ax}] change replicas (REMOVE_REPLICA (n3,s3):3): read existing descriptor r1:/M{in-ax} [(n1,s1):1, (n3,s3):3, (n4,s4):4, next=5]
W170628 18:18:42.619227 33099 storage/store.go:3231 [s1] got error from r1, replica (n3,s3):3: replica corruption (processed=true): boom
I170628 18:18:42.625580 33235 storage/replica.go:2887 [replicate,s1,r1/1:/M{in-ax}] proposing REMOVE_REPLICA (n3,s3):3: [(n1,s1):1 (n4,s4):4]
W170628 18:18:42.628331 33099 storage/store.go:3231 [s1] got error from r1, replica (n3,s3):3: replica corruption (processed=true): boom
W170628 18:18:42.628508 33099 storage/store.go:3231 [s1] got error from r1, replica (n3,s3):3: replica corruption (processed=true): boom
W170628 18:18:42.631838 33099 storage/store.go:3231 [s1] got error from r1, replica (n3,s3):3: replica corruption (processed=true): boom
I170628 18:18:42.634528 32096 storage/replica_raftstorage.go:442 [replicate,s1,r1/1:/M{in-ax}] generated preemptive snapshot 3f71ce49 at index 30
I170628 18:18:42.639944 32096 storage/store.go:3416 [replicate,s1,r1/1:/M{in-ax}] streamed snapshot to (n2,s2):?: kv pairs: 53, log entries: 20, rate-limit: 8.0 MiB/sec, 3ms
I170628 18:18:42.643043 33224 storage/replica_raftstorage.go:639 [s2,r1/?:{-}] applying preemptive snapshot at index 30 (id=3f71ce49, encoded size=16066, 1 rocksdb batches, 20 log entries)
I170628 18:18:42.651484 33224 storage/replica_raftstorage.go:647 [s2,r1/?:/M{in-ax}] applied preemptive snapshot in 8ms [clear=0ms batch=0ms entries=7ms commit=0ms]
I170628 18:18:42.655067 32096 storage/replica_command.go:3615 [replicate,s1,r1/1:/M{in-ax}] change replicas (ADD_REPLICA (n2,s2):5): read existing descriptor r1:/M{in-ax} [(n1,s1):1, (n4,s4):4, next=5]
I170628 18:18:42.673879 33191 storage/replica.go:2887 [replicate,s1,r1/1:/M{in-ax}] proposing ADD_REPLICA (n2,s2):5: [(n1,s1):1 (n4,s4):4 (n2,s2):5]
I170628 18:18:42.687305 32093 storage/replica_raftstorage.go:442 [replicate,s1,r1/1:/M{in-ax}] generated preemptive snapshot f9c76c99 at index 32
I170628 18:18:42.691517 32093 storage/store.go:3416 [replicate,s1,r1/1:/M{in-ax}] streamed snapshot to (n5,s5):?: kv pairs: 56, log entries: 22, rate-limit: 2.0 MiB/sec, 4ms
I170628 18:18:42.692640 33165 storage/replica_raftstorage.go:639 [s5,r1/?:{-}] applying preemptive snapshot at index 32 (id=f9c76c99, encoded size=17696, 1 rocksdb batches, 22 log entries)
I170628 18:18:42.695242 33165 storage/replica_raftstorage.go:647 [s5,r1/?:/M{in-ax}] applied preemptive snapshot in 2ms [clear=0ms batch=0ms entries=2ms commit=0ms]
I170628 18:18:42.698120 32093 storage/replica_command.go:3615 [replicate,s1,r1/1:/M{in-ax}] change replicas (ADD_REPLICA (n5,s5):6): read existing descriptor r1:/M{in-ax} [(n1,s1):1, (n4,s4):4, (n2,s2):5, next=6]
I170628 18:18:42.709954 33259 storage/replica.go:2887 [replicate,s1,r1/1:/M{in-ax}] proposing ADD_REPLICA (n5,s5):6: [(n1,s1):1 (n4,s4):4 (n2,s2):5 (n5,s5):6]
E170628 18:18:42.770293 32785 storage/replica.go:4924 [s5,r1/?:/M{in-ax}] stalling replica due to: boom
E170628 18:18:42.771138 32785 storage/replica.go:4924 [s5,r1/?:/M{in-ax}] stalling replica due to: boom
E170628 18:18:42.772612 32785 storage/replica.go:4924 [s5,r1/6:/M{in-ax}] stalling replica due to: boom
W170628 18:18:42.792463 32342 storage/node_liveness.go:318 [hb] failed node liveness heartbeat: node unavailable; try another peer
W170628 18:18:42.795256 33139 storage/raft_transport.go:272 unable to accept Raft message from (n5,s5):6: no handler registered for (n1,s1):1
W170628 18:18:42.795945 33264 storage/store.go:3231 [s4] got error from r1, replica (n5,s5):6: replica corruption (processed=true): boom
W170628 18:18:42.796387 33126 storage/store.go:3227 [s5] raft error: node 1 claims to not contain store 1 for replica (n1,s1):1: store 1 was not found
W170628 18:18:42.796559 33139 storage/raft_transport.go:272 unable to accept Raft message from (n4,s4):4: no handler registered for (n1,s1):1
W170628 18:18:42.796716 33124 storage/raft_transport.go:462 raft transport stream to node 1 failed: store 1 was not found
W170628 18:18:42.800187 33264 storage/store.go:3231 [s4] got error from r1, replica (n5,s5):6: replica corruption (processed=true): boom
W170628 18:18:42.800756 33244 storage/raft_transport.go:272 unable to accept Raft message from (n4,s4):4: no handler registered for (n2,s2):5
W170628 18:18:42.801735 33268 storage/store.go:3227 [s4] raft error: node 2 claims to not contain store 2 for replica (n2,s2):5: store 2 was not found
W170628 18:18:42.801965 33266 storage/raft_transport.go:462 raft transport stream to node 2 failed: store 2 was not found
W170628 18:18:42.884557 32916 gossip/gossip.go:1196 [n6] no incoming or outgoing connections
W170628 18:18:42.885603 33057 storage/raft_transport.go:462 raft transport stream to node 6 failed: rpc error: code = Internal desc = transport is closing
I170628 18:18:42.886036 32614 vendor/google.golang.org/grpc/transport/http2_server.go:399 transport: http2Server.HandleStreams failed to read frame: read tcp 127.0.0.1:34547->127.0.0.1:60106: use of closed network connection
I170628 18:18:42.887097 32425 vendor/google.golang.org/grpc/transport/http2_server.go:399 transport: http2Server.HandleStreams failed to read frame: read tcp 127.0.0.1:34226->127.0.0.1:34086: use of closed network connection
I170628 18:18:42.887865 32370 vendor/google.golang.org/grpc/transport/http2_server.go:399 transport: http2Server.HandleStreams failed to read frame: read tcp 127.0.0.1:48975->127.0.0.1:47604: use of closed network connection
I170628 18:18:42.888284 32160 vendor/google.golang.org/grpc/transport/http2_server.go:399 transport: http2Server.HandleStreams failed to read frame: read tcp 127.0.0.1:54812->127.0.0.1:41161: use of closed network connection
I170628 18:18:42.888693 32377 vendor/google.golang.org/grpc/transport/http2_server.go:399 transport: http2Server.HandleStreams failed to read frame: read tcp 127.0.0.1:40101->127.0.0.1:57971: use of closed network connection
W170628 18:18:42.889788 33262 storage/raft_transport.go:462 raft transport stream to node 5 failed: rpc error: code = Internal desc = transport is closing
--- FAIL: testrace/TestSystemZoneConfigs (54.980s)
<autogenerated>:14: storage/client_replica_test.go:1575, condition failed to evaluate within 45s: got 41 replicas, want 39; details: map[5:3 12:3 1:3 8:3 13:3 3:3 4:4 6:3 7:4 9:3 10:3 11:3 2:3]
------- Stdout: -------
W170628 18:19:26.872997 59748 server/status/runtime.go:111 Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I170628 18:19:26.890981 59748 server/config.go:479 [n?] 1 storage engine initialized
I170628 18:19:26.891063 59748 server/config.go:481 [n?] RocksDB cache size: 512 MiB
I170628 18:19:26.891106 59748 server/config.go:481 [n?] store 0: in-memory, size 100 MiB
I170628 18:19:26.892471 59748 server/node.go:450 [n?] store [n0,s0] not bootstrapped
I170628 18:19:26.918588 59748 server/node.go:385 [n?] **** cluster 1af3ddf7-bc64-4eaa-96da-7ae509382d89 has been created
I170628 18:19:26.918683 59748 server/node.go:386 [n?] **** add additional nodes by specifying --join=127.0.0.1:60907
I170628 18:19:26.954622 59748 server/node.go:463 [n1] initialized store [n1,s1]: {Capacity:536870912 Available:536870912 RangeCount:1 LeaseCount:1 WritesPerSecond:58.1946861908287}
I170628 18:19:26.954878 59748 server/node.go:347 [n1] node ID 1 initialized
I170628 18:19:26.955096 59748 gossip/gossip.go:297 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:60907" > attrs:<> locality:<>
I170628 18:19:26.955642 59748 storage/stores.go:296 [n1] read 0 node addresses from persistent storage
I170628 18:19:26.955896 59748 server/node.go:604 [n1] connecting to gossip network to verify cluster ID...
I170628 18:19:26.955998 59748 server/node.go:629 [n1] node connected via gossip and verified as part of cluster "1af3ddf7-bc64-4eaa-96da-7ae509382d89"
I170628 18:19:26.956276 59748 server/node.go:401 [n1] node=1: started with [[]=] engine(s) and attributes []
I170628 18:19:26.962093 60036 storage/replica_command.go:2694 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2]
I170628 18:19:26.963625 59748 sql/executor.go:358 [n1] creating distSQLPlanner with address {tcp 127.0.0.1:60907}
E170628 18:19:27.022144 60037 storage/queue.go:658 [replicate,n1,s1,r1/1:/{Min-System/}] range requires a replication change, but lacks a quorum of live replicas (0/1)
I170628 18:19:27.032052 60036 storage/replica_command.go:2694 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/tsd [r3]
I170628 18:19:27.038615 59748 server/server.go:769 [n1] starting https server at 127.0.0.1:43705
I170628 18:19:27.038740 59748 server/server.go:770 [n1] starting grpc/postgres server at 127.0.0.1:60907
I170628 18:19:27.038784 59748 server/server.go:771 [n1] advertising CockroachDB node at 127.0.0.1:60907
E170628 18:19:27.082339 60037 storage/queue.go:658 [replicate,n1,s1,r1/1:/{Min-System/}] range requires a replication change, but lacks a quorum of live replicas (0/1)
E170628 18:19:27.290094 59984 storage/replica_proposal.go:470 [n1,s1,r2/1:/{System/-Max}] could not load SystemConfig span: must retry later due to intent on SystemConfigSpan
I170628 18:19:27.295656 60036 storage/replica_command.go:2694 [split,n1,s1,r3/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r4]
I170628 18:19:27.303122 59748 sql/event_log.go:101 [n1] Event: "alter_table", target: 12, info: {TableName:eventlog Statement:ALTER TABLE system.eventlog ALTER COLUMN uniqueid SET DEFAULT uuid_v4() User:node MutationID:0 CascadeDroppedViews:[]}
I170628 18:19:27.390008 59748 sql/lease.go:372 [n1] publish: descID=12 (eventlog) version=2 mtime=2017-06-28 18:19:27.38991686 +0000 UTC
I170628 18:19:27.445220 60036 storage/replica_command.go:2694 [split,n1,s1,r4/1:/{System/tse-Max}] initiating a split of this range at key /Table/0 [r5]
I170628 18:19:27.568315 60036 storage/replica_command.go:2694 [split,n1,s1,r5/1:/{Table/0-Max}] initiating a split of this range at key /Table/11 [r6]
I170628 18:19:27.639950 59748 server/server.go:906 [n1] done ensuring all necessary migrations have run
I170628 18:19:27.640079 59748 server/server.go:908 [n1] serving sql connections
I170628 18:19:27.674349 60036 storage/replica_command.go:2694 [split,n1,s1,r6/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r7]
I170628 18:19:27.718370 60206 sql/event_log.go:101 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:60907} Attrs: Locality:} ClusterID:1af3ddf7-bc64-4eaa-96da-7ae509382d89 StartedAt:1498673966956040633 LastUp:1498673966956040633}
I170628 18:19:27.772910 60036 storage/replica_command.go:2694 [split,n1,s1,r7/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r8]
I170628 18:19:27.868291 60036 storage/replica_command.go:2694 [split,n1,s1,r8/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r9]
I170628 18:19:27.980417 60036 storage/replica_command.go:2694 [split,n1,s1,r9/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r10]
W170628 18:19:28.177356 59748 server/status/runtime.go:111 Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
W170628 18:19:28.187826 59748 gossip/gossip.go:1196 [n?] no incoming or outgoing connections
I170628 18:19:28.222038 59748 server/config.go:479 [n?] 1 storage engine initialized
I170628 18:19:28.222136 59748 server/config.go:481 [n?] RocksDB cache size: 512 MiB
I170628 18:19:28.222172 59748 server/config.go:481 [n?] store 0: in-memory, size 100 MiB
I170628 18:19:28.223344 59748 server/node.go:450 [n?] store [n0,s0] not bootstrapped
I170628 18:19:28.223445 59748 storage/stores.go:296 [n?] read 0 node addresses from persistent storage
I170628 18:19:28.223540 59748 server/node.go:604 [n?] connecting to gossip network to verify cluster ID...
I170628 18:19:28.266534 60424 gossip/client.go:131 [n?] started gossip client to 127.0.0.1:60907
I170628 18:19:28.274630 60444 gossip/server.go:234 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:56027}
I170628 18:19:28.278409 59748 server/node.go:629 [n?] node connected via gossip and verified as part of cluster "1af3ddf7-bc64-4eaa-96da-7ae509382d89"
I170628 18:19:28.283062 60516 storage/stores.go:312 [n?] wrote 1 node addresses to persistent storage
I170628 18:19:28.291610 59748 kv/dist_sender.go:370 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I170628 18:19:28.296676 59748 server/node.go:340 [n?] new node allocated ID 2
I170628 18:19:28.297684 59748 gossip/gossip.go:297 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:56027" > attrs:<> locality:<>
I170628 18:19:28.298477 59748 server/node.go:401 [n2] node=2: started with [[]=] engine(s) and attributes []
I170628 18:19:28.306128 60500 storage/stores.go:312 [n1] wrote 1 node addresses to persistent storage
I170628 18:19:28.314202 59748 sql/executor.go:358 [n2] creating distSQLPlanner with address {tcp 127.0.0.1:56027}
I170628 18:19:28.326292 60530 server/node.go:585 [n2] bootstrapped store [n2,s2]
I170628 18:19:28.334735 59748 server/server.go:769 [n2] starting https server at 127.0.0.1:35587
I170628 18:19:28.335700 59748 server/server.go:770 [n2] starting grpc/postgres server at 127.0.0.1:56027
I170628 18:19:28.335764 59748 server/server.go:771 [n2] advertising CockroachDB node at 127.0.0.1:56027
I170628 18:19:28.349165 59748 server/server.go:906 [n2] done ensuring all necessary migrations have run
I170628 18:19:28.349309 59748 server/server.go:908 [n2] serving sql connections
W170628 18:19:28.472859 59748 server/status/runtime.go:111 Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
W170628 18:19:28.486721 59748 gossip/gossip.go:1196 [n?] no incoming or outgoing connections
I170628 18:19:28.487189 60485 sql/event_log.go:101 [n2] Event: "node_join", target: 2, info: {Descriptor:{NodeID:2 Address:{NetworkField:tcp AddressField:127.0.0.1:56027} Attrs: Locality:} ClusterID:1af3ddf7-bc64-4eaa-96da-7ae509382d89 StartedAt:1498673968298206396 LastUp:1498673968298206396}
I170628 18:19:28.495009 59748 server/config.go:479 [n?] 1 storage engine initialized
I170628 18:19:28.495172 59748 server/config.go:481 [n?] RocksDB cache size: 512 MiB
I170628 18:19:28.495235 59748 server/config.go:481 [n?] store 0: in-memory, size 100 MiB
I170628 18:19:28.496808 59748 server/node.go:450 [n?] store [n0,s0] not bootstrapped
I170628 18:19:28.496967 59748 storage/stores.go:296 [n?] read 0 node addresses from persistent storage
I170628 18:19:28.497131 59748 server/node.go:604 [n?] connecting to gossip network to verify cluster ID...
I170628 18:19:28.569793 60660 gossip/client.go:131 [n?] started gossip client to 127.0.0.1:60907
I170628 18:19:28.573156 60688 gossip/server.go:234 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:51287}
I170628 18:19:28.579376 60732 storage/stores.go:312 [n?] wrote 1 node addresses to persistent storage
I170628 18:19:28.579857 60732 storage/stores.go:312 [n?] wrote 2 node addresses to persistent storage
I170628 18:19:28.580963 59748 server/node.go:629 [n?] node connected via gossip and verified as part of cluster "1af3ddf7-bc64-4eaa-96da-7ae509382d89"
I170628 18:19:28.588662 59748 kv/dist_sender.go:370 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I170628 18:19:28.595038 59748 server/node.go:340 [n?] new node allocated ID 3
I170628 18:19:28.595312 59748 gossip/gossip.go:297 [n3] NodeDescriptor set to node_id:3 address:<network_field:"tcp" address_field:"127.0.0.1:51287" > attrs:<> locality:<>
I170628 18:19:28.596090 59748 server/node.go:401 [n3] node=3: started with [[]=] engine(s) and attributes []
I170628 18:19:28.598922 60734 storage/stores.go:312 [n1] wrote 2 node addresses to persistent storage
I170628 18:19:28.601072 60735 storage/stores.go:312 [n2] wrote 2 node addresses to persistent storage
I170628 18:19:28.606873 59748 sql/executor.go:358 [n3] creating distSQLPlanner with address {tcp 127.0.0.1:51287}
I170628 18:19:28.705041 59748 server/server.go:769 [n3] starting https server at 127.0.0.1:35566
I170628 18:19:28.709481 59748 server/server.go:770 [n3] starting grpc/postgres server at 127.0.0.1:51287
I170628 18:19:28.709815 59748 server/server.go:771 [n3] advertising CockroachDB node at 127.0.0.1:51287
I170628 18:19:28.719717 60458 server/node.go:585 [n3] bootstrapped store [n3,s3]
I170628 18:19:28.728480 59748 server/server.go:906 [n3] done ensuring all necessary migrations have run
I170628 18:19:28.728625 59748 server/server.go:908 [n3] serving sql connections
I170628 18:19:28.730292 59761 storage/replica_raftstorage.go:442 [replicate,n1,s1,r4/1:/{System/tse-Table/0}] generated preemptive snapshot 834298d3 at index 18
I170628 18:19:28.786057 60856 sql/event_log.go:101 [n3] Event: "node_join", target: 3, info: {Descriptor:{NodeID:3 Address:{NetworkField:tcp AddressField:127.0.0.1:51287} Attrs: Locality:} ClusterID:1af3ddf7-bc64-4eaa-96da-7ae509382d89 StartedAt:1498673968595824323 LastUp:1498673968595824323}
I170628 18:19:28.870534 59761 storage/store.go:3416 [replicate,n1,s1,r4/1:/{System/tse-Table/0}] streamed snapshot to (n2,s2):?: kv pairs: 10, log entries: 8, rate-limit: 8.0 MiB/sec, 3ms
I170628 18:19:28.871641 60885 storage/replica_raftstorage.go:639 [n2,s2,r4/?:{-}] applying preemptive snapshot at index 18 (id=834298d3, encoded size=7150, 1 rocksdb batches, 8 log entries)
I170628 18:19:28.873387 60885 storage/replica_raftstorage.go:647 [n2,s2,r4/?:/{System/tse-Table/0}] applied preemptive snapshot in 2ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:19:28.878068 59761 storage/replica_command.go:3615 [replicate,n1,s1,r4/1:/{System/tse-Table/0}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r4:/{System/tse-Table/0} [(n1,s1):1, next=2]
W170628 18:19:28.893016 59748 server/status/runtime.go:111 Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
W170628 18:19:28.943107 59748 gossip/gossip.go:1196 [n?] no incoming or outgoing connections
I170628 18:19:28.944462 59748 server/config.go:479 [n?] 1 storage engine initialized
I170628 18:19:28.944561 59748 server/config.go:481 [n?] RocksDB cache size: 512 MiB
I170628 18:19:28.944591 59748 server/config.go:481 [n?] store 0: in-memory, size 100 MiB
I170628 18:19:28.945863 59748 server/node.go:450 [n?] store [n0,s0] not bootstrapped
I170628 18:19:28.946430 59748 storage/stores.go:296 [n?] read 0 node addresses from persistent storage
I170628 18:19:28.946582 59748 server/node.go:604 [n?] connecting to gossip network to verify cluster ID...
I170628 18:19:28.949510 60887 storage/replica.go:2887 [n1,s1,r4/1:/{System/tse-Table/0}] proposing ADD_REPLICA (n2,s2):2: [(n1,s1):1 (n2,s2):2]
I170628 18:19:28.956131 59761 storage/replica_raftstorage.go:442 [replicate,n1,s1,r9/1:/Table/1{4-5}] generated preemptive snapshot 2b81cd71 at index 19
I170628 18:19:28.966638 60973 storage/raft_transport.go:456 [n2] raft transport stream to node 1 established
I170628 18:19:29.024770 60938 gossip/client.go:131 [n?] started gossip client to 127.0.0.1:60907
I170628 18:19:29.025389 61026 gossip/server.go:234 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:55080}
I170628 18:19:29.029447 61019 storage/stores.go:312 [n?] wrote 1 node addresses to persistent storage
I170628 18:19:29.029671 61019 storage/stores.go:312 [n?] wrote 2 node addresses to persistent storage
I170628 18:19:29.030017 61019 storage/stores.go:312 [n?] wrote 3 node addresses to persistent storage
I170628 18:19:29.032725 59748 server/node.go:629 [n?] node connected via gossip and verified as part of cluster "1af3ddf7-bc64-4eaa-96da-7ae509382d89"
I170628 18:19:29.039740 59748 kv/dist_sender.go:370 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I170628 18:19:29.044997 59748 server/node.go:340 [n?] new node allocated ID 4
I170628 18:19:29.045263 59748 gossip/gossip.go:297 [n4] NodeDescriptor set to node_id:4 address:<network_field:"tcp" address_field:"127.0.0.1:55080" > attrs:<> locality:<>
I170628 18:19:29.046079 59748 server/node.go:401 [n4] node=4: started with [[]=] engine(s) and attributes []
I170628 18:19:29.050814 59748 sql/executor.go:358 [n4] creating distSQLPlanner with address {tcp 127.0.0.1:55080}
I170628 18:19:29.052304 60881 storage/stores.go:312 [n1] wrote 3 node addresses to persistent storage
I170628 18:19:29.054674 61044 storage/stores.go:312 [n3] wrote 3 node addresses to persistent storage
I170628 18:19:29.067522 61045 storage/stores.go:312 [n2] wrote 3 node addresses to persistent storage
I170628 18:19:29.087904 61029 server/node.go:585 [n4] bootstrapped store [n4,s4]
I170628 18:19:29.124306 59748 server/server.go:769 [n4] starting https server at 127.0.0.1:46733
I170628 18:19:29.127303 59748 server/server.go:770 [n4] starting grpc/postgres server at 127.0.0.1:55080
I170628 18:19:29.127363 59748 server/server.go:771 [n4] advertising CockroachDB node at 127.0.0.1:55080
I170628 18:19:29.135969 59748 server/server.go:906 [n4] done ensuring all necessary migrations have run
I170628 18:19:29.136123 59748 server/server.go:908 [n4] serving sql connections
I170628 18:19:29.158397 59761 storage/store.go:3416 [replicate,n1,s1,r9/1:/Table/1{4-5}] streamed snapshot to (n3,s3):?: kv pairs: 10, log entries: 9, rate-limit: 8.0 MiB/sec, 12ms
I170628 18:19:29.160364 61220 storage/replica_raftstorage.go:639 [n3,s3,r9/?:{-}] applying preemptive snapshot at index 19 (id=2b81cd71, encoded size=5888, 1 rocksdb batches, 9 log entries)
I170628 18:19:29.165350 61220 storage/replica_raftstorage.go:647 [n3,s3,r9/?:/Table/1{4-5}] applied preemptive snapshot in 4ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:19:29.182713 59761 storage/replica_command.go:3615 [replicate,n1,s1,r9/1:/Table/1{4-5}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r9:/Table/1{4-5} [(n1,s1):1, next=2]
W170628 18:19:29.256663 59748 server/status/runtime.go:111 Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I170628 18:19:29.269370 59761 storage/replica_command.go:3615 [replicate,n1,s1,r9/1:/Table/1{4-5}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r9:/Table/1{4-5} [(n1,s1):1, next=2]
I170628 18:19:29.274478 61203 sql/event_log.go:101 [n4] Event: "node_join", target: 4, info: {Descriptor:{NodeID:4 Address:{NetworkField:tcp AddressField:127.0.0.1:55080} Attrs: Locality:} ClusterID:1af3ddf7-bc64-4eaa-96da-7ae509382d89 StartedAt:1498673969045805303 LastUp:1498673969045805303}
W170628 18:19:29.278205 59748 gossip/gossip.go:1196 [n?] no incoming or outgoing connections
I170628 18:19:29.286150 59748 server/config.go:479 [n?] 1 storage engine initialized
I170628 18:19:29.286256 59748 server/config.go:481 [n?] RocksDB cache size: 512 MiB
I170628 18:19:29.286303 59748 server/config.go:481 [n?] store 0: in-memory, size 100 MiB
I170628 18:19:29.287507 59748 server/node.go:450 [n?] store [n0,s0] not bootstrapped
I170628 18:19:29.287623 59748 storage/stores.go:296 [n?] read 0 node addresses from persistent storage
I170628 18:19:29.287726 59748 server/node.go:604 [n?] connecting to gossip network to verify cluster ID...
I170628 18:19:29.322028 61284 storage/replica.go:2887 [n1,s1,r9/1:/Table/1{4-5}] proposing ADD_REPLICA (n3,s3):2: [(n1,s1):1 (n3,s3):2]
I170628 18:19:29.341042 59761 storage/replica_raftstorage.go:442 [replicate,n1,s1,r8/1:/Table/1{3-4}] generated preemptive snapshot f3451e2a at index 29
I170628 18:19:29.347402 59761 storage/store.go:3416 [replicate,n1,s1,r8/1:/Table/1{3-4}] streamed snapshot to (n2,s2):?: kv pairs: 74, log entries: 19, rate-limit: 8.0 MiB/sec, 6ms
I170628 18:19:29.350003 61269 storage/replica_raftstorage.go:639 [n2,s2,r8/?:{-}] applying preemptive snapshot at index 29 (id=f3451e2a, encoded size=25874, 1 rocksdb batches, 19 log entries)
I170628 18:19:29.352485 61272 storage/raft_transport.go:456 [n3] raft transport stream to node 1 established
I170628 18:19:29.352809 61269 storage/replica_raftstorage.go:647 [n2,s2,r8/?:/Table/1{3-4}] applied preemptive snapshot in 3ms [clear=0ms batch=0ms entries=2ms commit=0ms]
I170628 18:19:29.358503 59761 storage/replica_command.go:3615 [replicate,n1,s1,r8/1:/Table/1{3-4}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r8:/Table/1{3-4} [(n1,s1):1, next=2]
I170628 18:19:29.388466 61278 storage/replica.go:2887 [n1,s1,r8/1:/Table/1{3-4}] proposing ADD_REPLICA (n2,s2):2: [(n1,s1):1 (n2,s2):2]
I170628 18:19:29.392519 61242 gossip/client.go:131 [n?] started gossip client to 127.0.0.1:60907
I170628 18:19:29.394936 61196 gossip/server.go:234 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:52141}
I170628 18:19:29.398870 59748 server/node.go:629 [n?] node connected via gossip and verified as part of cluster "1af3ddf7-bc64-4eaa-96da-7ae509382d89"
I170628 18:19:29.400196 61197 storage/stores.go:312 [n?] wrote 1 node addresses to persistent storage
I170628 18:19:29.403140 61198 storage/stores.go:312 [n?] wrote 2 node addresses to persistent storage
I170628 18:19:29.404525 61198 storage/stores.go:312 [n?] wrote 3 node addresses to persistent storage
I170628 18:19:29.404779 61198 storage/stores.go:312 [n?] wrote 4 node addresses to persistent storage
I170628 18:19:29.407898 59761 storage/replica_raftstorage.go:442 [replicate,n1,s1,r2/1:/System/{-tsd}] generated preemptive snapshot 58ce5105 at index 47
I170628 18:19:29.422232 59748 kv/dist_sender.go:370 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping
I170628 18:19:29.433788 59761 storage/store.go:3416 [replicate,n1,s1,r2/1:/System/{-tsd}] streamed snapshot to (n3,s3):?: kv pairs: 35, log entries: 4, rate-limit: 8.0 MiB/sec, 24ms
I170628 18:19:29.435133 59748 server/node.go:340 [n?] new node allocated ID 5
I170628 18:19:29.435431 59748 gossip/gossip.go:297 [n5] NodeDescriptor set to node_id:5 address:<network_field:"tcp" address_field:"127.0.0.1:52141" > attrs:<> locality:<>
I170628 18:19:29.435627 61334 storage/replica_raftstorage.go:639 [n3,s3,r2/?:{-}] applying preemptive snapshot at index 47 (id=58ce5105, encoded size=89880, 1 rocksdb batches, 4 log entries)
I170628 18:19:29.436311 59748 server/node.go:401 [n5] node=5: started with [[]=] engine(s) and attributes []
I170628 18:19:29.437666 61334 storage/replica_raftstorage.go:647 [n3,s3,r2/?:/System/{-tsd}] applied preemptive snapshot in 2ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:19:29.438808 59748 sql/executor.go:358 [n5] creating distSQLPlanner with address {tcp 127.0.0.1:52141}
I170628 18:19:29.448889 61196 gossip/server.go:285 [n1] refusing gossip from node 5 (max 3 conns); forwarding to 2 ({tcp 127.0.0.1:56027})
I170628 18:19:29.449025 59761 storage/replica_command.go:3615 [replicate,n1,s1,r2/1:/System/{-tsd}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r2:/System/{-tsd} [(n1,s1):1, next=2]
I170628 18:19:29.450520 61242 gossip/client.go:136 [n5] closing client to node 1 (127.0.0.1:60907): received forward from node 1 to 2 (127.0.0.1:56027)
I170628 18:19:29.451179 61254 gossip/gossip.go:1210 [n5] node has connected to cluster via gossip
I170628 18:19:29.451618 61254 storage/stores.go:312 [n5] wrote 4 node addresses to persistent storage
I170628 18:19:29.457497 61336 storage/stores.go:312 [n1] wrote 4 node addresses to persistent storage
I170628 18:19:29.462760 61301 storage/stores.go:312 [n3] wrote 4 node addresses to persistent storage
I170628 18:19:29.464380 61292 storage/stores.go:312 [n2] wrote 4 node addresses to persistent storage
I170628 18:19:29.464665 61302 storage/stores.go:312 [n4] wrote 4 node addresses to persistent storage
I170628 18:19:29.515681 61365 server/node.go:585 [n5] bootstrapped store [n5,s5]
I170628 18:19:29.518391 60696 storage/store.go:2165 [replicaGC,n3,s3,r2/?:/System/{-tsd}] removing replica
I170628 18:19:29.519153 59748 server/server.go:769 [n5] starting https server at 127.0.0.1:46391
I170628 18:19:29.519286 59748 server/server.go:770 [n5] starting grpc/postgres server at 127.0.0.1:52141
I170628 18:19:29.519329 59748 server/server.go:771 [n5] advertising CockroachDB node at 127.0.0.1:52141
I170628 18:19:29.520349 60696 storage/replica.go:721 [replicaGC,n3,s3,r2/?:/System/{-tsd}] removed 26 (17+9) keys in 2ms [clear=0ms commit=1ms]
I170628 18:19:29.579236 59748 server/server.go:906 [n5] done ensuring all necessary migrations have run
I170628 18:19:29.579364 59748 server/server.go:908 [n5] serving sql connections
I170628 18:19:29.586001 61296 storage/replica.go:2887 [n1,s1,r2/1:/System/{-tsd}] proposing ADD_REPLICA (n3,s3):2: [(n1,s1):1 (n3,s3):2]
I170628 18:19:29.612717 59761 storage/replica_raftstorage.go:442 [replicate,n1,s1,r1/1:/{Min-System/}] generated preemptive snapshot 8d917606 at index 57
I170628 18:19:29.617307 60040 storage/replica_raftstorage.go:442 [raftsnapshot,n1,s1,r2/1:/System/{-tsd}] generated Raft snapshot 944abd67 at index 52
I170628 18:19:29.621174 61291 gossip/client.go:131 [n5] started gossip client to 127.0.0.1:56027
I170628 18:19:29.644015 60040 storage/store.go:3416 [raftsnapshot,n1,s1,r2/1:/System/{-tsd}] streamed snapshot to (n3,s3):2: kv pairs: 40, log entries: 9, rate-limit: 8.0 MiB/sec, 15ms
I170628 18:19:29.646455 61556 storage/replica_raftstorage.go:639 [n3,s3,r2/2:{-}] applying Raft snapshot at index 52 (id=944abd67, encoded size=112186, 1 rocksdb batches, 9 log entries)
I170628 18:19:29.649500 61556 storage/replica_raftstorage.go:647 [n3,s3,r2/2:/System/{-tsd}] applied Raft snapshot in 3ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:19:29.698268 61312 sql/event_log.go:101 [n5] Event: "node_join", target: 5, info: {Descriptor:{NodeID:5 Address:{NetworkField:tcp AddressField:127.0.0.1:52141} Attrs: Locality:} ClusterID:1af3ddf7-bc64-4eaa-96da-7ae509382d89 StartedAt:1498673969436006758 LastUp:1498673969436006758}
I170628 18:19:29.734279 59761 storage/store.go:3416 [replicate,n1,s1,r1/1:/{Min-System/}] streamed snapshot to (n4,s4):?: kv pairs: 37, log entries: 47, rate-limit: 8.0 MiB/sec, 3ms
I170628 18:19:29.736262 61592 storage/replica_raftstorage.go:639 [n4,s4,r1/?:{-}] applying preemptive snapshot at index 57 (id=8d917606, encoded size=26870, 1 rocksdb batches, 47 log entries)
I170628 18:19:29.740770 61592 storage/replica_raftstorage.go:647 [n4,s4,r1/?:/{Min-System/}] applied preemptive snapshot in 4ms [clear=0ms batch=0ms entries=3ms commit=0ms]
I170628 18:19:29.746051 59761 storage/replica_command.go:3615 [replicate,n1,s1,r1/1:/{Min-System/}] change replicas (ADD_REPLICA (n4,s4):2): read existing descriptor r1:/{Min-System/} [(n1,s1):1, next=2]
I170628 18:19:29.789160 61530 storage/replica.go:2887 [n1,s1,r1/1:/{Min-System/}] proposing ADD_REPLICA (n4,s4):2: [(n1,s1):1 (n4,s4):2]
I170628 18:19:29.795848 59761 storage/replica_raftstorage.go:442 [replicate,n1,s1,r5/1:/Table/{0-11}] generated preemptive snapshot 6f476a10 at index 23
I170628 18:19:29.835863 61535 storage/raft_transport.go:456 [n4] raft transport stream to node 1 established
I170628 18:19:29.899633 59761 storage/store.go:3416 [replicate,n1,s1,r5/1:/Table/{0-11}] streamed snapshot to (n5,s5):?: kv pairs: 39, log entries: 13, rate-limit: 8.0 MiB/sec, 6ms
I170628 18:19:29.901866 61671 storage/replica_raftstorage.go:639 [n5,s5,r5/?:{-}] applying preemptive snapshot at index 23 (id=6f476a10, encoded size=16399, 1 rocksdb batches, 13 log entries)
I170628 18:19:29.903920 61671 storage/replica_raftstorage.go:647 [n5,s5,r5/?:/Table/{0-11}] applied preemptive snapshot in 2ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:19:29.909102 59761 storage/replica_command.go:3615 [replicate,n1,s1,r5/1:/Table/{0-11}] change replicas (ADD_REPLICA (n5,s5):2): read existing descriptor r5:/Table/{0-11} [(n1,s1):1, next=2]
I170628 18:19:29.945750 61706 storage/replica.go:2887 [n1,s1,r5/1:/Table/{0-11}] proposing ADD_REPLICA (n5,s5):2: [(n1,s1):1 (n5,s5):2]
I170628 18:19:29.950290 59761 storage/replica_raftstorage.go:442 [replicate,n1,s1,r6/1:/Table/1{1-2}] generated preemptive snapshot a017ec3c at index 19
I170628 18:19:29.964296 59761 storage/store.go:3416 [replicate,n1,s1,r6/1:/Table/1{1-2}] streamed snapshot to (n3,s3):?: kv pairs: 10, log entries: 9, rate-limit: 8.0 MiB/sec, 13ms
I170628 18:19:29.966145 61677 storage/replica_raftstorage.go:639 [n3,s3,r6/?:{-}] applying preemptive snapshot at index 19 (id=a017ec3c, encoded size=8472, 1 rocksdb batches, 9 log entries)
I170628 18:19:29.968087 61677 storage/replica_raftstorage.go:647 [n3,s3,r6/?:/Table/1{1-2}] applied preemptive snapshot in 2ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:19:30.025825 59761 storage/replica_command.go:3615 [replicate,n1,s1,r6/1:/Table/1{1-2}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r6:/Table/1{1-2} [(n1,s1):1, next=2]
I170628 18:19:30.036406 61764 storage/raft_transport.go:456 [n5] raft transport stream to node 1 established
I170628 18:19:30.102937 60696 storage/store.go:2165 [replicaGC,n3,s3,r6/?:/Table/1{1-2}] removing replica
I170628 18:19:30.104466 60696 storage/replica.go:721 [replicaGC,n3,s3,r6/?:/Table/1{1-2}] removed 9 (0+9) keys in 1ms [clear=1ms commit=0ms]
I170628 18:19:30.128148 61665 storage/replica.go:2887 [n1,s1,r6/1:/Table/1{1-2}] proposing ADD_REPLICA (n3,s3):2: [(n1,s1):1 (n3,s3):2]
I170628 18:19:30.137554 59761 storage/replica_raftstorage.go:442 [replicate,n1,s1,r10/1:/{Table/15-Max}] generated preemptive snapshot 4dccf8fa at index 11
I170628 18:19:30.143860 59761 storage/store.go:3416 [replicate,n1,s1,r10/1:/{Table/15-Max}] streamed snapshot to (n4,s4):?: kv pairs: 9, log entries: 1, rate-limit: 8.0 MiB/sec, 4ms
I170628 18:19:30.147099 61696 storage/replica_raftstorage.go:639 [n4,s4,r10/?:{-}] applying preemptive snapshot at index 11 (id=4dccf8fa, encoded size=548, 1 rocksdb batches, 1 log entries)
I170628 18:19:30.148382 61696 storage/replica_raftstorage.go:647 [n4,s4,r10/?:/{Table/15-Max}] applied preemptive snapshot in 1ms [clear=0ms batch=0ms entries=0ms commit=0ms]
I170628 18:19:30.152879 59761 storage/replica_command.go:3615 [replicate,n1,s1,r10/1:/{Table/15-Max}] change replicas (ADD_REPLICA (n4,s4):2): read existing descriptor r10:/{Table/15-Max} [(n1,s1):1, next=2]
I170628 18:19:30.159491 60040 storage/replica_raftstorage.go:442 [raftsnapshot,n1,s1,r6/1:/Table/1{1-2}] generated Raft snapshot 6f92afff at index 21
I170628 18:19:30.164192 60040 storage/store.go:3416 [raftsnapshot,n1,s1,r6/1:/Table/1{1-2}] streamed snapshot to (n3,s3):2: kv pairs: 12, log entries: 11, rate-limit: 8.0 MiB/sec, 4ms
I170628 18:19:30.166648 61844 storage/replica_raftstorage.go:639 [n3,s3,r6/2:{-}] applying Raft snapshot at index 21 (id=6f92afff, encoded size=10427, 1 rocksdb batches, 11 log entries)
I170628 18:19:30.171939 61844 storage/replica_raftstorage.go:647 [n3,s3,r6/2:/Table/1{1-2}] applied Raft snapshot in 5ms [clear=0ms batch=0ms entries=2ms commit=0ms]
I170628 18:19:30.214291 61805 storage/replica.go:2887 [n1,s1,r10/1:/{Table/15-Max}] proposing ADD_REPLICA (n4,s4):2: [(n1,s1):1 (n4,s4):2]
I170628 18:19:30.221353 59761 storage/replica_raftstorage.go:442 [replicate,n1,s1,r3/1:/System/ts{d-e}] generated preemptive snapshot aa13cdf1 at index 29
I170628 18:19:30.275186 59761 storage/store.go:3416 [replicate,n1,s1,r3/1:/System/ts{d-e}] streamed snapshot to (n5,s5):?: kv pairs: 1404, log entries: 2, rate-limit: 8.0 MiB/sec, 52ms
I170628 18:19:30.283673 61729 storage/replica_raftstorage.go:639 [n5,s5,r3/?:{-}] applying preemptive snapshot at index 29 (id=aa13cdf1, encoded size=170084, 1 rocksdb batches, 2 log entries)
I170628 18:19:30.288120 61729 storage/replica_raftstorage.go:647 [n5,s5,r3/?:/System/ts{d-e}] applied preemptive snapshot in 4ms [clear=0ms batch=0ms entries=0ms commit=1ms]
I170628 18:19:30.293251 59761 storage/replica_command.go:3615 [replicate,n1,s1,r3/1:/System/ts{d-e}] change replicas (ADD_REPLICA (n5,s5):2): read existing descriptor r3:/System/ts{d-e} [(n1,s1):1, next=2]
I170628 18:19:30.340202 61897 storage/replica.go:2887 [n1,s1,r3/1:/System/ts{d-e}] proposing ADD_REPLICA (n5,s5):2: [(n1,s1):1 (n5,s5):2]
I170628 18:19:30.347308 59761 storage/replica_raftstorage.go:442 [replicate,n1,s1,r7/1:/Table/1{2-3}] generated preemptive snapshot 9da60953 at index 29
I170628 18:19:30.367821 59761 storage/store.go:3416 [replicate,n1,s1,r7/1:/Table/1{2-3}] streamed snapshot to (n4,s4):?: kv pairs: 40, log entries: 19, rate-limit: 8.0 MiB/sec, 7ms
I170628 18:19:30.371450 61834 storage/replica_raftstorage.go:639 [n4,s4,r7/?:{-}] applying preemptive snapshot at index 29 (id=9da60953, encoded size=22998, 1 rocksdb batches, 19 log entries)
I170628 18:19:30.384499 61834 storage/replica_raftstorage.go:647 [n4,s4,r7/?:/Table/1{2-3}] applied preemptive snapshot in 13ms [clear=0ms batch=0ms entries=1ms commit=2ms]
I170628 18:19:30.390455 59761 storage/replica_command.go:3615 [replicate,n1,s1,r7/1:/Table/1{2-3}] change replicas (ADD_REPLICA (n4,s4):2): read existing descriptor r7:/Table/1{2-3} [(n1,s1):1, next=2]
I170628 18:19:30.430196 61882 storage/replica.go:2887 [n1,s1,r7/1:/Table/1{2-3}] proposing ADD_REPLICA (n4,s4):2: [(n1,s1):1 (n4,s4):2]
I170628 18:19:30.447304 59761 storage/queue.go:725 [n1,replicate] purgatory is now empty
I170628 18:19:30.449102 60037 storage/replica_raftstorage.go:442 [replicate,n1,s1,r8/1:/Table/1{3-4}] generated preemptive snapshot d3cbfc51 at index 46
I170628 18:19:30.462464 60037 storage/store.go:3416 [replicate,n1,s1,r8/1:/Table/1{3-4}] streamed snapshot to (n5,s5):?: kv pairs: 120, log entries: 36, rate-limit: 8.0 MiB/sec, 8ms
I170628 18:19:30.464628 61867 storage/replica_raftstorage.go:639 [n5,s5,r8/?:{-}] applying preemptive snapshot at index 46 (id=d3cbfc51, encoded size=47495, 1 rocksdb batches, 36 log entries)
I170628 18:19:30.470105 61867 storage/replica_raftstorage.go:647 [n5,s5,r8/?:/Table/1{3-4}] applied preemptive snapshot in 5ms [clear=0ms batch=0ms entries=3ms commit=1ms]
I170628 18:19:30.477830 60037 storage/replica_command.go:3615 [replicate,n1,s1,r8/1:/Table/1{3-4}] change replicas (ADD_REPLICA (n5,s5):3): read existing descriptor r8:/Table/1{3-4} [(n1,s1):1, (n2,s2):2, next=3]
I170628 18:19:30.541988 61496 storage/store.go:2165 [replicaGC,n5,s5,r8/?:/Table/1{3-4}] removing replica
I170628 18:19:30.542818 61496 storage/replica.go:721 [replicaGC,n5,s5,r8/?:/Table/1{3-4}] removed 113 (104+9) keys in 0ms [clear=0ms commit=0ms]
I170628 18:19:30.545050 61825 storage/replica.go:2887 [n1,s1,r8/1:/Table/1{3-4}] proposing ADD_REPLICA (n5,s5):3: [(n1,s1):1 (n2,s2):2 (n5,s5):3]
I170628 18:19:30.555962 60037 storage/replica_raftstorage.go:442 [replicate,n1,s1,r4/1:/{System/tse-Table/0}] generated preemptive snapshot c1dc6eca at index 22
I170628 18:19:30.560213 60037 storage/store.go:3416 [replicate,n1,s1,r4/1:/{System/tse-Table/0}] streamed snapshot to (n5,s5):?: kv pairs: 11, log entries: 12, rate-limit: 8.0 MiB/sec, 4ms
I170628 18:19:30.563486 62003 storage/replica_raftstorage.go:639 [n5,s5,r4/?:{-}] applying preemptive snapshot at index 22 (id=c1dc6eca, encoded size=9647, 1 rocksdb batches, 12 log entries)
I170628 18:19:30.565824 62003 storage/replica_raftstorage.go:647 [n5,s5,r4/?:/{System/tse-Table/0}] applied preemptive snapshot in 2ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:19:30.570757 60037 storage/replica_command.go:3615 [replicate,n1,s1,r4/1:/{System/tse-Table/0}] change replicas (ADD_REPLICA (n5,s5):3): read existing descriptor r4:/{System/tse-Table/0} [(n1,s1):1, (n2,s2):2, next=3]
I170628 18:19:30.596395 60040 storage/replica_raftstorage.go:442 [raftsnapshot,n1,s1,r8/1:/Table/1{3-4}] generated Raft snapshot e879a0a4 at index 50
I170628 18:19:30.606719 61913 storage/replica_raftstorage.go:639 [n5,s5,r8/3:{-}] applying Raft snapshot at index 50 (id=e879a0a4, encoded size=51409, 1 rocksdb batches, 40 log entries)
I170628 18:19:30.617347 61991 storage/replica.go:2887 [n1,s1,r4/1:/{System/tse-Table/0}] proposing ADD_REPLICA (n5,s5):3: [(n1,s1):1 (n2,s2):2 (n5,s5):3]
I170628 18:19:30.619964 60040 storage/store.go:3416 [raftsnapshot,n1,s1,r8/1:/Table/1{3-4}] streamed snapshot to (n5,s5):3: kv pairs: 122, log entries: 40, rate-limit: 8.0 MiB/sec, 22ms
I170628 18:19:30.630824 60037 storage/replica_raftstorage.go:442 [replicate,n1,s1,r3/1:/System/ts{d-e}] generated preemptive snapshot b20eaa8f at index 33
I170628 18:19:30.643256 61913 storage/replica_raftstorage.go:647 [n5,s5,r8/3:/Table/1{3-4}] applied Raft snapshot in 36ms [clear=3ms batch=0ms entries=32ms commit=1ms]
I170628 18:19:30.697769 60037 storage/store.go:3416 [replicate,n1,s1,r3/1:/System/ts{d-e}] streamed snapshot to (n3,s3):?: kv pairs: 1405, log entries: 6, rate-limit: 8.0 MiB/sec, 60ms
I170628 18:19:30.703182 62013 storage/replica_raftstorage.go:639 [n3,s3,r3/?:{-}] applying preemptive snapshot at index 33 (id=b20eaa8f, encoded size=172593, 1 rocksdb batches, 6 log entries)
I170628 18:19:30.705901 62013 storage/replica_raftstorage.go:647 [n3,s3,r3/?:/System/ts{d-e}] applied preemptive snapshot in 3ms [clear=0ms batch=0ms entries=1ms commit=1ms]
I170628 18:19:30.711595 60037 storage/replica_command.go:3615 [replicate,n1,s1,r3/1:/System/ts{d-e}] change replicas (ADD_REPLICA (n3,s3):3): read existing descriptor r3:/System/ts{d-e} [(n1,s1):1, (n5,s5):2, next=3]
I170628 18:19:30.763362 61920 storage/replica.go:2887 [n1,s1,r3/1:/System/ts{d-e}] proposing ADD_REPLICA (n3,s3):3: [(n1,s1):1 (n5,s5):2 (n3,s3):3]
I170628 18:19:30.804030 60037 storage/replica_raftstorage.go:442 [replicate,n1,s1,r2/1:/System/{-tsd}] generated preemptive snapshot 2dd698b4 at index 57
I170628 18:19:30.839688 60037 storage/store.go:3416 [replicate,n1,s1,r2/1:/System/{-tsd}] streamed snapshot to (n2,s2):?: kv pairs: 40, log entries: 14, rate-limit: 8.0 MiB/sec, 35ms
I170628 18:19:30.841510 62025 storage/replica_raftstorage.go:639 [n2,s2,r2/?:{-}] applying preemptive snapshot at index 57 (id=2dd698b4, encoded size=135175, 1 rocksdb batches, 14 log entries)
I170628 18:19:30.844632 62025 storage/replica_raftstorage.go:647 [n2,s2,r2/?:/System/{-tsd}] applied preemptive snapshot in 3ms [clear=0ms batch=0ms entries=2ms commit=1ms]
I170628 18:19:30.851728 60037 storage/replica_command.go:3615 [replicate,n1,s1,r2/1:/System/{-tsd}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r2:/System/{-tsd} [(n1,s1):1, (n3,s3):2, next=3]
I170628 18:19:30.918014 60536 storage/store.go:2165 [replicaGC,n2,s2,r2/?:/System/{-tsd}] removing replica
I170628 18:19:30.918848 60536 storage/replica.go:721 [replicaGC,n2,s2,r2/?:/System/{-tsd}] removed 28 (19+9) keys in 0ms [clear=0ms commit=0ms]
I170628 18:19:30.925877 62045 storage/replica.go:2887 [n1,s1,r2/1:/System/{-tsd}] proposing ADD_REPLICA (n2,s2):3: [(n1,s1):1 (n3,s3):2 (n2,s2):3]
I170628 18:19:30.958636 60037 storage/replica_raftstorage.go:442 [replicate,n1,s1,r7/1:/Table/1{2-3}] generated preemptive snapshot cbbbdfc4 at index 34
I170628 18:19:30.964708 60037 storage/store.go:3416 [replicate,n1,s1,r7/1:/Table/1{2-3}] streamed snapshot to (n2,s2):?: kv pairs: 41, log entries: 24, rate-limit: 8.0 MiB/sec, 5ms
I170628 18:19:30.966865 62089 storage/replica_raftstorage.go:639 [n2,s2,r7/?:{-}] applying preemptive snapshot at index 34 (id=cbbbdfc4, encoded size=25826, 1 rocksdb batches, 24 log entries)
I170628 18:19:30.969905 62089 storage/replica_raftstorage.go:647 [n2,s2,r7/?:/Table/1{2-3}] applied preemptive snapshot in 3ms [clear=0ms batch=0ms entries=2ms commit=0ms]
I170628 18:19:30.975758 60037 storage/replica_command.go:3615 [replicate,n1,s1,r7/1:/Table/1{2-3}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r7:/Table/1{2-3} [(n1,s1):1, (n4,s4):2, next=3]
I170628 18:19:30.999837 60040 storage/replica_raftstorage.go:442 [raftsnapshot,n1,s1,r2/1:/System/{-tsd}] generated Raft snapshot 0e88df0d at index 60
I170628 18:19:31.010932 60040 storage/store.go:3416 [raftsnapshot,n1,s1,r2/1:/System/{-tsd}] streamed snapshot to (n2,s2):3: kv pairs: 41, log entries: 17, rate-limit: 8.0 MiB/sec, 10ms
I170628 18:19:31.023660 62132 storage/replica_raftstorage.go:639 [n2,s2,r2/3:{-}] applying Raft snapshot at index 60 (id=0e88df0d, encoded size=137259, 1 rocksdb batches, 17 log entries)
I170628 18:19:31.027405 62132 storage/replica_raftstorage.go:647 [n2,s2,r2/3:/System/{-tsd}] applied Raft snapshot in 4ms [clear=0ms batch=0ms entries=2ms commit=1ms]
I170628 18:19:31.051320 60536 storage/store.go:2165 [replicaGC,n2,s2,r7/?:/Table/1{2-3}] removing replica
I170628 18:19:31.067706 60536 storage/replica.go:721 [replicaGC,n2,s2,r7/?:/Table/1{2-3}] removed 39 (30+9) keys in 16ms [clear=0ms commit=16ms]
I170628 18:19:31.072259 62058 storage/replica.go:2887 [n1,s1,r7/1:/Table/1{2-3}] proposing ADD_REPLICA (n2,s2):3: [(n1,s1):1 (n4,s4):2 (n2,s2):3]
I170628 18:19:31.089686 60037 storage/replica_raftstorage.go:442 [replicate,n1,s1,r5/1:/Table/{0-11}] generated preemptive snapshot 59b945ac at index 28
I170628 18:19:31.096495 60037 storage/store.go:3416 [replicate,n1,s1,r5/1:/Table/{0-11}] streamed snapshot to (n4,s4):?: kv pairs: 40, log entries: 18, rate-limit: 8.0 MiB/sec, 6ms
I170628 18:19:31.098948 62097 storage/replica_raftstorage.go:639 [n4,s4,r5/?:{-}] applying preemptive snapshot at index 28 (id=59b945ac, encoded size=19227, 1 rocksdb batches, 18 log entries)
I170628 18:19:31.101251 62097 storage/replica_raftstorage.go:647 [n4,s4,r5/?:/Table/{0-11}] applied preemptive snapshot in 2ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:19:31.110786 60037 storage/replica_command.go:3615 [replicate,n1,s1,r5/1:/Table/{0-11}] change replicas (ADD_REPLICA (n4,s4):3): read existing descriptor r5:/Table/{0-11} [(n1,s1):1, (n5,s5):2, next=3]
I170628 18:19:31.150104 60040 storage/replica_raftstorage.go:442 [raftsnapshot,n1,s1,r7/1:/Table/1{2-3}] generated Raft snapshot 8ab855b2 at index 37
I170628 18:19:31.182644 60040 storage/store.go:3416 [raftsnapshot,n1,s1,r7/1:/Table/1{2-3}] streamed snapshot to (n2,s2):3: kv pairs: 42, log entries: 27, rate-limit: 8.0 MiB/sec, 9ms
I170628 18:19:31.183564 62063 storage/replica_raftstorage.go:639 [n2,s2,r7/3:{-}] applying Raft snapshot at index 37 (id=8ab855b2, encoded size=27898, 1 rocksdb batches, 27 log entries)
I170628 18:19:31.191245 62063 storage/replica_raftstorage.go:647 [n2,s2,r7/3:/Table/1{2-3}] applied Raft snapshot in 8ms [clear=0ms batch=0ms entries=5ms commit=0ms]
I170628 18:19:31.200817 62199 storage/replica.go:2887 [n1,s1,r5/1:/Table/{0-11}] proposing ADD_REPLICA (n4,s4):3: [(n1,s1):1 (n5,s5):2 (n4,s4):3]
I170628 18:19:31.230850 60037 storage/replica_raftstorage.go:442 [replicate,n1,s1,r1/1:/{Min-System/}] generated preemptive snapshot b85d771c at index 87
I170628 18:19:31.261757 60037 storage/store.go:3416 [replicate,n1,s1,r1/1:/{Min-System/}] streamed snapshot to (n5,s5):?: kv pairs: 51, log entries: 77, rate-limit: 8.0 MiB/sec, 29ms
I170628 18:19:31.269096 62212 storage/replica_raftstorage.go:639 [n5,s5,r1/?:{-}] applying preemptive snapshot at index 87 (id=b85d771c, encoded size=42192, 1 rocksdb batches, 77 log entries)
I170628 18:19:31.280988 62212 storage/replica_raftstorage.go:647 [n5,s5,r1/?:/{Min-System/}] applied preemptive snapshot in 11ms [clear=0ms batch=0ms entries=8ms commit=1ms]
I170628 18:19:31.331934 60037 storage/replica_command.go:3615 [replicate,n1,s1,r1/1:/{Min-System/}] change replicas (ADD_REPLICA (n5,s5):3): read existing descriptor r1:/{Min-System/} [(n1,s1):1, (n4,s4):2, next=3]
I170628 18:19:31.339130 61496 storage/store.go:2165 [replicaGC,n5,s5,r1/?:/{Min-System/}] removing replica
I170628 18:19:31.339889 61496 storage/replica.go:721 [replicaGC,n5,s5,r1/?:/{Min-System/}] removed 21 (11+10) keys in 0ms [clear=0ms commit=0ms]
I170628 18:19:31.402815 62169 storage/replica.go:2887 [n1,s1,r1/1:/{Min-System/}] proposing ADD_REPLICA (n5,s5):3: [(n1,s1):1 (n4,s4):2 (n5,s5):3]
I170628 18:19:31.428621 60037 storage/replica_raftstorage.go:442 [replicate,n1,s1,r9/1:/Table/1{4-5}] generated preemptive snapshot a1f46630 at index 25
I170628 18:19:31.432257 60037 storage/store.go:3416 [replicate,n1,s1,r9/1:/Table/1{4-5}] streamed snapshot to (n4,s4):?: kv pairs: 11, log entries: 15, rate-limit: 8.0 MiB/sec, 3ms
I170628 18:19:31.436971 62203 storage/replica_raftstorage.go:639 [n4,s4,r9/?:{-}] applying preemptive snapshot at index 25 (id=a1f46630, encoded size=9427, 1 rocksdb batches, 15 log entries)
I170628 18:19:31.442506 62203 storage/replica_raftstorage.go:647 [n4,s4,r9/?:/Table/1{4-5}] applied preemptive snapshot in 5ms [clear=0ms batch=0ms entries=1ms commit=0ms]
I170628 18:19:31.455773 60037 storage/replica_command.go:3615 [replicate,n1,s1,r9/1:/Table/1{4-5}] change replicas (ADD_REPLICA (n4,s4):3): read existing descriptor r9:/Table/1{4-5} [(n1,s1):1, (n3,s3):2, next=3]
I170628 18:19:31.480824 60040 storage/replica_raftstorage.go:442 [raftsnapshot,n1,s1,r1/1:/{Min-System/}] generated Raft snapshot 42ddeed5 at index 93
I170628 18:19:31.511401 61049 storage/store.go:2165 [replicaGC,n4,s4,r9/?:/Table/1{4-5}] removing replica
I170628 18:19:31.516246 61049 storage/replica.go:721 [replicaGC,n4,s4,r9/?:/Table/1{4-5}] removed 9 (0+9) keys in 4ms [clear=4ms commit=0ms]
I170628 18:19:31.516943 60040 storage/store.go:3416 [raftsnapshot,n1,s1,r1/1:/{Min-System/}] streamed snapshot to (n5,s5):3: kv pairs: 53, log entries: 83, rate-limit: 8.0 MiB/sec, 29ms
I170628 18:19:31.518817 62276 storage/replica_raftstorage.go:639 [n5,s5,r1/3:{-}] applying Raft snapshot at index 93 (id=42ddeed5, encoded size=45885, 1 rocksdb batches, 83 log entries)
I170628 18:19:31.531955 62276 storage/replica_raftstorage.go:647 [n5,s5,r1/3:/{Min-System/}] applied Raft snapshot in 13ms [clear=0ms batch=0ms entries=11ms commit=0ms]
I170628 18:19:31.574326 62322 storage/replica.go:2887 [n1,s1,r9/1:/Table/1{4-5}] proposing ADD_REPLICA (n4,s4):3: [(n1,s1):1 (n3,s3):2 (n4,s4):3]
W170628 18:19:31.585532 62176 storage/replica.go:2528 [n1,s1,r2/1:/System/{-tsd}] con
```
Please assign, take a look and update the issue accordingly. | test | teamcity failed tests on master testrace teststorerangecorruptionchangereplicas testrace testsystemzoneconfigs the following tests appear to have failed fail testrace teststorerangecorruptionchangereplicas client raft test go replica not lease holder current lease is repl start exp pro stdout gossip gossip go nodedescriptor set to node id address attrs locality gossip gossip go no incoming or outgoing connections gossip client go started gossip client to gossip gossip go nodedescriptor set to node id address attrs locality gossip gossip go no incoming or outgoing connections gossip client go started gossip client to storage store go failed initial metrics computation system config not yet available gossip gossip go nodedescriptor set to node id address attrs locality gossip gossip go no incoming or outgoing connections gossip client go started gossip client to gossip gossip go nodedescriptor set to node id address attrs locality gossip gossip go no incoming or outgoing connections storage store go failed initial metrics computation system config not yet available gossip gossip go nodedescriptor set to node id address attrs locality gossip client go started gossip client to gossip server go refusing gossip from node max conns forwarding to tcp gossip client go closing client to node received forward from node to gossip gossip go node has connected to cluster via gossip gossip client go started gossip client to gossip gossip go no incoming or outgoing connections gossip client go started gossip client to gossip server go refusing gossip from node max conns forwarding to tcp gossip client go closing client to node received forward from node to gossip gossip go node has connected to cluster via gossip gossip client go started gossip client to gossip gossip go nodedescriptor set to node id address attrs locality storage client test go gossip network initialized storage replica raftstorage go generated preemptive snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor m in ax storage replica go proposing add replica storage replica raftstorage go generated preemptive snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor m in ax storage raft transport go raft transport stream to node established storage replica go proposing add replica storage replica go stalling replica due to boom storage replica command go change replicas remove replica read existing descriptor m in ax storage store go got error from replica replica corruption processed true boom storage store go got error from replica replica corruption processed true boom storage store go got error from replica replica corruption processed true boom storage replica go proposing remove replica storage store go got error from replica replica corruption processed true boom storage store go got error from replica replica corruption processed true boom storage replica raftstorage go generated preemptive snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor m in ax storage replica go proposing add replica storage replica go stalling replica due to boom storage replica command go change replicas remove replica read existing descriptor m in ax storage store go got error from replica replica corruption processed true boom storage replica go proposing remove replica storage store go got error from replica replica corruption processed true boom storage store go got error from replica replica corruption processed true boom storage store go got error from replica replica corruption processed true boom storage replica raftstorage go generated preemptive snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor m in ax storage replica go proposing add replica storage replica raftstorage go generated preemptive snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor m in ax storage replica go proposing add replica storage replica go stalling replica due to boom storage replica go stalling replica due to boom storage replica go stalling replica due to boom storage node liveness go failed node liveness heartbeat node unavailable try another peer storage raft transport go unable to accept raft message from no handler registered for storage store go got error from replica replica corruption processed true boom storage store go raft error node claims to not contain store for replica store was not found storage raft transport go unable to accept raft message from no handler registered for storage raft transport go raft transport stream to node failed store was not found storage store go got error from replica replica corruption processed true boom storage raft transport go unable to accept raft message from no handler registered for storage store go raft error node claims to not contain store for replica store was not found storage raft transport go raft transport stream to node failed store was not found gossip gossip go no incoming or outgoing connections storage raft transport go raft transport stream to node failed rpc error code internal desc transport is closing vendor google golang org grpc transport server go transport handlestreams failed to read frame read tcp use of closed network connection vendor google golang org grpc transport server go transport handlestreams failed to read frame read tcp use of closed network connection vendor google golang org grpc transport server go transport handlestreams failed to read frame read tcp use of closed network connection vendor google golang org grpc transport server go transport handlestreams failed to read frame read tcp use of closed network connection vendor google golang org grpc transport server go transport handlestreams failed to read frame read tcp use of closed network connection storage raft transport go raft transport stream to node failed rpc error code internal desc transport is closing fail testrace testsystemzoneconfigs storage client replica test go condition failed to evaluate within got replicas want details map stdout server status runtime go could not parse build timestamp parsing time as cannot parse as server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size mib server node go store not bootstrapped server node go cluster has been created server node go add additional nodes by specifying join server node go initialized store capacity available rangecount leasecount writespersecond server node go node id initialized gossip gossip go nodedescriptor set to node id address attrs locality storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster server node go node started with engine s and attributes storage replica command go initiating a split of this range at key system sql executor go creating distsqlplanner with address tcp storage queue go range requires a replication change but lacks a quorum of live replicas storage replica command go initiating a split of this range at key system tsd server server go starting https server at server server go starting grpc postgres server at server server go advertising cockroachdb node at storage queue go range requires a replication change but lacks a quorum of live replicas storage replica proposal go could not load systemconfig span must retry later due to intent on systemconfigspan storage replica command go initiating a split of this range at key system tse sql event log go event alter table target info tablename eventlog statement alter table system eventlog alter column uniqueid set default uuid user node mutationid cascadedroppedviews sql lease go publish descid eventlog version mtime utc storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table server server go done ensuring all necessary migrations have run server server go serving sql connections storage replica command go initiating a split of this range at key table sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality clusterid startedat lastup storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table server status runtime go could not parse build timestamp parsing time as cannot parse as gossip gossip go no incoming or outgoing connections server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size mib server node go store not bootstrapped storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id gossip client go started gossip client to gossip server go received initial cluster verification connection from tcp server node go node connected via gossip and verified as part of cluster storage stores go wrote node addresses to persistent storage kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping server node go new node allocated id gossip gossip go nodedescriptor set to node id address attrs locality server node go node started with engine s and attributes storage stores go wrote node addresses to persistent storage sql executor go creating distsqlplanner with address tcp server node go bootstrapped store server server go starting https server at server server go starting grpc postgres server at server server go advertising cockroachdb node at server server go done ensuring all necessary migrations have run server server go serving sql connections server status runtime go could not parse build timestamp parsing time as cannot parse as gossip gossip go no incoming or outgoing connections sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality clusterid startedat lastup server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size mib server node go store not bootstrapped storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id gossip client go started gossip client to gossip server go received initial cluster verification connection from tcp storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage server node go node connected via gossip and verified as part of cluster kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping server node go new node allocated id gossip gossip go nodedescriptor set to node id address attrs locality server node go node started with engine s and attributes storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage sql executor go creating distsqlplanner with address tcp server server go starting https server at server server go starting grpc postgres server at server server go advertising cockroachdb node at server node go bootstrapped store server server go done ensuring all necessary migrations have run server server go serving sql connections storage replica raftstorage go generated preemptive snapshot at index sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality clusterid startedat lastup storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor system tse table server status runtime go could not parse build timestamp parsing time as cannot parse as gossip gossip go no incoming or outgoing connections server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size mib server node go store not bootstrapped storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id storage replica go proposing add replica storage replica raftstorage go generated preemptive snapshot at index storage raft transport go raft transport stream to node established gossip client go started gossip client to gossip server go received initial cluster verification connection from tcp storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage server node go node connected via gossip and verified as part of cluster kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping server node go new node allocated id gossip gossip go nodedescriptor set to node id address attrs locality server node go node started with engine s and attributes sql executor go creating distsqlplanner with address tcp storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage server node go bootstrapped store server server go starting https server at server server go starting grpc postgres server at server server go advertising cockroachdb node at server server go done ensuring all necessary migrations have run server server go serving sql connections storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table server status runtime go could not parse build timestamp parsing time as cannot parse as storage replica command go change replicas add replica read existing descriptor table sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality clusterid startedat lastup gossip gossip go no incoming or outgoing connections server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size mib server node go store not bootstrapped storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id storage replica go proposing add replica storage replica raftstorage go generated preemptive snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage raft transport go raft transport stream to node established storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica gossip client go started gossip client to gossip server go received initial cluster verification connection from tcp server node go node connected via gossip and verified as part of cluster storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage storage replica raftstorage go generated preemptive snapshot at index kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping storage store go streamed snapshot to kv pairs log entries rate limit mib sec server node go new node allocated id gossip gossip go nodedescriptor set to node id address attrs locality storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries server node go node started with engine s and attributes storage replica raftstorage go applied preemptive snapshot in sql executor go creating distsqlplanner with address tcp gossip server go refusing gossip from node max conns forwarding to tcp storage replica command go change replicas add replica read existing descriptor system tsd gossip client go closing client to node received forward from node to gossip gossip go node has connected to cluster via gossip storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage server node go bootstrapped store storage store go removing replica server server go starting https server at server server go starting grpc postgres server at server server go advertising cockroachdb node at storage replica go removed keys in server server go done ensuring all necessary migrations have run server server go serving sql connections storage replica go proposing add replica storage replica raftstorage go generated preemptive snapshot at index storage replica raftstorage go generated raft snapshot at index gossip client go started gossip client to storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying raft snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied raft snapshot in sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality clusterid startedat lastup storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor min system storage replica go proposing add replica storage replica raftstorage go generated preemptive snapshot at index storage raft transport go raft transport stream to node established storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica storage replica raftstorage go generated preemptive snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage raft transport go raft transport stream to node established storage store go removing replica storage replica go removed keys in storage replica go proposing add replica storage replica raftstorage go generated preemptive snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table max storage replica raftstorage go generated raft snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying raft snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied raft snapshot in storage replica go proposing add replica storage replica raftstorage go generated preemptive snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor system ts d e storage replica go proposing add replica storage replica raftstorage go generated preemptive snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica storage queue go purgatory is now empty storage replica raftstorage go generated preemptive snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage store go removing replica storage replica go removed keys in storage replica go proposing add replica storage replica raftstorage go generated preemptive snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor system tse table storage replica raftstorage go generated raft snapshot at index storage replica raftstorage go applying raft snapshot at index id encoded size rocksdb batches log entries storage replica go proposing add replica storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go generated preemptive snapshot at index storage replica raftstorage go applied raft snapshot in storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor system ts d e storage replica go proposing add replica storage replica raftstorage go generated preemptive snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor system tsd storage store go removing replica storage replica go removed keys in storage replica go proposing add replica storage replica raftstorage go generated preemptive snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica raftstorage go generated raft snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying raft snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied raft snapshot in storage store go removing replica storage replica go removed keys in storage replica go proposing add replica storage replica raftstorage go generated preemptive snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica raftstorage go generated raft snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying raft snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied raft snapshot in storage replica go proposing add replica storage replica raftstorage go generated preemptive snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor min system storage store go removing replica storage replica go removed keys in storage replica go proposing add replica storage replica raftstorage go generated preemptive snapshot at index storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica raftstorage go generated raft snapshot at index storage store go removing replica storage replica go removed keys in storage store go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying raft snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied raft snapshot in storage replica go proposing add replica storage replica go con please assign take a look and update the issue accordingly | 1 |
557,206 | 16,503,611,021 | IssuesEvent | 2021-05-25 16:36:50 | magento/magento2 | https://api.github.com/repos/magento/magento2 | opened | [Issue] Code cleanup | Component: MediaGalleryIntegration Priority: P4 | This issue is automatically created based on existing pull request: magento/magento2#31773: Code cleanup
---------
### Description (*)
- Reduce early return usage
- Remove unused code
- Improve condition: replace substr/strlen by strpos
| 1.0 | [Issue] Code cleanup - This issue is automatically created based on existing pull request: magento/magento2#31773: Code cleanup
---------
### Description (*)
- Reduce early return usage
- Remove unused code
- Improve condition: replace substr/strlen by strpos
| non_test | code cleanup this issue is automatically created based on existing pull request magento code cleanup description reduce early return usage remove unused code improve condition replace substr strlen by strpos | 0 |
57,580 | 6,551,048,178 | IssuesEvent | 2017-09-05 13:31:04 | openbmc/openbmc-test-automation | https://api.github.com/repos/openbmc/openbmc-test-automation | closed | Clean up BMC dump in code update | Test | Needed to start a clean code update, dump cleanup needed.
- [x] BMC code update
- [x] Host code update | 1.0 | Clean up BMC dump in code update - Needed to start a clean code update, dump cleanup needed.
- [x] BMC code update
- [x] Host code update | test | clean up bmc dump in code update needed to start a clean code update dump cleanup needed bmc code update host code update | 1 |
399,903 | 11,763,252,166 | IssuesEvent | 2020-03-14 06:06:13 | TheLX5/Powerups | https://api.github.com/repos/TheLX5/Powerups | opened | Custom layer interaction is not properly implemented | bug help wanted low priority | Custom layer interaction is kind of broken since it really doesn't completely setup everything to make it stable.
It currently makes mini Mario act really weird when hitting ceilings, it appears to process other parts of Mario's body and proceed to push downwards the player.
The whole original code is a mess and I'm starting to question myself if there's any fix for this. | 1.0 | Custom layer interaction is not properly implemented - Custom layer interaction is kind of broken since it really doesn't completely setup everything to make it stable.
It currently makes mini Mario act really weird when hitting ceilings, it appears to process other parts of Mario's body and proceed to push downwards the player.
The whole original code is a mess and I'm starting to question myself if there's any fix for this. | non_test | custom layer interaction is not properly implemented custom layer interaction is kind of broken since it really doesn t completely setup everything to make it stable it currently makes mini mario act really weird when hitting ceilings it appears to process other parts of mario s body and proceed to push downwards the player the whole original code is a mess and i m starting to question myself if there s any fix for this | 0 |
76,778 | 26,594,155,580 | IssuesEvent | 2023-01-23 11:06:22 | vector-im/element-ios | https://api.github.com/repos/vector-im/element-ios | closed | Element immediately crashes | T-Defect crash A-SDK A-Startup S-Critical O-Uncommon | ### Steps to reproduce
The first time it crashed I was simply navigating between rooms. I don’t remember exactly what I tapped, but I’d been reading messages in one room and navigated back to the All chats screen. Now it crashes immediately:
1. At the iOS home screen, tap on Element.
### Outcome
#### What did you expect?
No crash
#### What happened instead?
Element briefly displays its All chats screen with up-to-date content. After approximately 0.5–5 seconds, Element disappears and the iOS home screen is displayed.
### Your phone model
iPhone15,2
### Operating system version
iOS 16.1.2
### Application version
1.9.13
### Homeserver
Synapse 1.72.0
### Will you send logs?
[Element-2022-12-07-171137.ips.log](https://github.com/vector-im/element-ios/files/10181503/Element-2022-12-07-171137.ips.log) | 1.0 | Element immediately crashes - ### Steps to reproduce
The first time it crashed I was simply navigating between rooms. I don’t remember exactly what I tapped, but I’d been reading messages in one room and navigated back to the All chats screen. Now it crashes immediately:
1. At the iOS home screen, tap on Element.
### Outcome
#### What did you expect?
No crash
#### What happened instead?
Element briefly displays its All chats screen with up-to-date content. After approximately 0.5–5 seconds, Element disappears and the iOS home screen is displayed.
### Your phone model
iPhone15,2
### Operating system version
iOS 16.1.2
### Application version
1.9.13
### Homeserver
Synapse 1.72.0
### Will you send logs?
[Element-2022-12-07-171137.ips.log](https://github.com/vector-im/element-ios/files/10181503/Element-2022-12-07-171137.ips.log) | non_test | element immediately crashes steps to reproduce the first time it crashed i was simply navigating between rooms i don’t remember exactly what i tapped but i’d been reading messages in one room and navigated back to the all chats screen now it crashes immediately at the ios home screen tap on element outcome what did you expect no crash what happened instead element briefly displays its all chats screen with up to date content after approximately – seconds element disappears and the ios home screen is displayed your phone model operating system version ios application version homeserver synapse will you send logs | 0 |
326,653 | 9,958,890,531 | IssuesEvent | 2019-07-06 00:18:27 | ChasakisD/Xamarin.Forms.BackgroundKit | https://api.github.com/repos/ChasakisD/Xamarin.Forms.BackgroundKit | opened | [Android] ClipPathManager should clip only when corner radius is not uniform | android bug performance priority-high up-for-grabs | When corner radius is uniform(the same on each corner) the clipping is done by the ViewOutlineProvider.
When the corner radius is not uniform, ViewOutlineProvider only draws the shadow and the clipping is done by the ClipPathManager.
Right now, the manager clips the child view in any case. It should be more effective, if the manager clips the child views only when the corner radius is not uniform. | 1.0 | [Android] ClipPathManager should clip only when corner radius is not uniform - When corner radius is uniform(the same on each corner) the clipping is done by the ViewOutlineProvider.
When the corner radius is not uniform, ViewOutlineProvider only draws the shadow and the clipping is done by the ClipPathManager.
Right now, the manager clips the child view in any case. It should be more effective, if the manager clips the child views only when the corner radius is not uniform. | non_test | clippathmanager should clip only when corner radius is not uniform when corner radius is uniform the same on each corner the clipping is done by the viewoutlineprovider when the corner radius is not uniform viewoutlineprovider only draws the shadow and the clipping is done by the clippathmanager right now the manager clips the child view in any case it should be more effective if the manager clips the child views only when the corner radius is not uniform | 0 |
353,877 | 25,139,399,814 | IssuesEvent | 2022-11-09 21:37:47 | capsule-corp-ternoa/ternoa-node | https://api.github.com/repos/capsule-corp-ternoa/ternoa-node | closed | DevOps: Redeploy Alphanet | documentation deployment | Tasks todo:
- Generate new Chain Specification for Alphanet
- Configure VMs
- Release a new version of Ternoa Node
We need to make sure that what ever we do here that we record and/or live stream it. Currently there is no documentation on how to fully deploy a chain and we need to fix that.
People who are responsible for this task: @soufiane-bouchaara @Leouarz @markopoloparadox | 1.0 | DevOps: Redeploy Alphanet - Tasks todo:
- Generate new Chain Specification for Alphanet
- Configure VMs
- Release a new version of Ternoa Node
We need to make sure that what ever we do here that we record and/or live stream it. Currently there is no documentation on how to fully deploy a chain and we need to fix that.
People who are responsible for this task: @soufiane-bouchaara @Leouarz @markopoloparadox | non_test | devops redeploy alphanet tasks todo generate new chain specification for alphanet configure vms release a new version of ternoa node we need to make sure that what ever we do here that we record and or live stream it currently there is no documentation on how to fully deploy a chain and we need to fix that people who are responsible for this task soufiane bouchaara leouarz markopoloparadox | 0 |
870 | 4,489,287,569 | IssuesEvent | 2016-08-30 10:25:38 | lukashaertel/TSS | https://api.github.com/repos/lukashaertel/TSS | opened | SA6 | Software Architecture | The project team MUST prefix all global names with its team name. Such names are e.g. project name, context path (root URL), database name, persistence units, server resources, etc.
This is to avoid naming conflicts, since the customer has to install the systems of many teams side-by-side. | 1.0 | SA6 - The project team MUST prefix all global names with its team name. Such names are e.g. project name, context path (root URL), database name, persistence units, server resources, etc.
This is to avoid naming conflicts, since the customer has to install the systems of many teams side-by-side. | non_test | the project team must prefix all global names with its team name such names are e g project name context path root url database name persistence units server resources etc this is to avoid naming conflicts since the customer has to install the systems of many teams side by side | 0 |
208,663 | 15,897,746,574 | IssuesEvent | 2021-04-11 22:22:18 | blynkkk/beta | https://api.github.com/repos/blynkkk/beta | closed | On Android, in SuperChart some colors of Datastreams does not show up | Android bug ready to test | These two colors does not visible before the text. İts just empty.
Android App: 1.0.8
S10e Android 11

<img width="598" alt="Screen Shot 2021-03-30 at 09 19 48" src="https://user-images.githubusercontent.com/81516108/112942759-4adb9500-9139-11eb-9b31-3f7f8f78ec86.png">
| 1.0 | On Android, in SuperChart some colors of Datastreams does not show up - These two colors does not visible before the text. İts just empty.
Android App: 1.0.8
S10e Android 11

<img width="598" alt="Screen Shot 2021-03-30 at 09 19 48" src="https://user-images.githubusercontent.com/81516108/112942759-4adb9500-9139-11eb-9b31-3f7f8f78ec86.png">
| test | on android in superchart some colors of datastreams does not show up these two colors does not visible before the text i̇ts just empty android app android img width alt screen shot at src | 1 |
281,542 | 21,315,413,151 | IssuesEvent | 2022-04-16 07:22:14 | Kidsnd274/pe | https://api.github.com/repos/Kidsnd274/pe | opened | Header indentation issues in developer guide | severity.VeryLow type.DocumentationBug | The sort feature in the developer guide under the implementation section has some indentation issues.
The photo below shows the issue:

This photo shows what the indentation is supposed to be:

<!--session: 1650088649060-79fd5dc5-68c0-4a9a-8cfa-a788c21533ff-->
<!--Version: Web v3.4.2--> | 1.0 | Header indentation issues in developer guide - The sort feature in the developer guide under the implementation section has some indentation issues.
The photo below shows the issue:

This photo shows what the indentation is supposed to be:

<!--session: 1650088649060-79fd5dc5-68c0-4a9a-8cfa-a788c21533ff-->
<!--Version: Web v3.4.2--> | non_test | header indentation issues in developer guide the sort feature in the developer guide under the implementation section has some indentation issues the photo below shows the issue this photo shows what the indentation is supposed to be | 0 |
99,905 | 21,056,486,552 | IssuesEvent | 2022-04-01 04:13:01 | virnect-corp/BookClub | https://api.github.com/repos/virnect-corp/BookClub | closed | The Clean Coder, Chapter 4 5 6, 2022-04-01 | 2022 The Clean Coder | ## 22 Mar
Chapter 4,5,6을 읽고 아래와 같이 진행
- (Required) chapter 당 한 개 이상 논의 내용 작성, 총 3개의 논의 내용 필요
- (Optional) 책 내용 정리 외에 추가 논의, 질문, 공유 정보 작성
- Reviewer는 자신을 제외한 멤버 모두 할당
## Before meeting
다른 멤버의 pull request review를 3일의 시간에 진행 해야 하며
pull request 당 적어도 1개 이상 코멘트 추가, 이후 approve를 통해 리뷰 했다는 걸 인증
## 25 Mar
회사 3-3 회의실에서 점심 + 리뷰 미팅 진행
발표 순서는 pull request 나중에 올린 순서 | 1.0 | The Clean Coder, Chapter 4 5 6, 2022-04-01 - ## 22 Mar
Chapter 4,5,6을 읽고 아래와 같이 진행
- (Required) chapter 당 한 개 이상 논의 내용 작성, 총 3개의 논의 내용 필요
- (Optional) 책 내용 정리 외에 추가 논의, 질문, 공유 정보 작성
- Reviewer는 자신을 제외한 멤버 모두 할당
## Before meeting
다른 멤버의 pull request review를 3일의 시간에 진행 해야 하며
pull request 당 적어도 1개 이상 코멘트 추가, 이후 approve를 통해 리뷰 했다는 걸 인증
## 25 Mar
회사 3-3 회의실에서 점심 + 리뷰 미팅 진행
발표 순서는 pull request 나중에 올린 순서 | non_test | the clean coder chapter mar chapter 읽고 아래와 같이 진행 required chapter 당 한 개 이상 논의 내용 작성 총 논의 내용 필요 optional 책 내용 정리 외에 추가 논의 질문 공유 정보 작성 reviewer는 자신을 제외한 멤버 모두 할당 before meeting 다른 멤버의 pull request review를 시간에 진행 해야 하며 pull request 당 적어도 이상 코멘트 추가 이후 approve를 통해 리뷰 했다는 걸 인증 mar 회사 회의실에서 점심 리뷰 미팅 진행 발표 순서는 pull request 나중에 올린 순서 | 0 |
70,729 | 13,529,048,669 | IssuesEvent | 2020-09-15 17:40:01 | KwanLab/Autometa | https://api.github.com/repos/KwanLab/Autometa | closed | [dev branch] Issues with autometa/config/project.py | code review | This file handles configuration handling for the Autometa user project. It contains the definition of the `Project` class. Here are the issues I found:
1. [x] The docstring for the `Project` class is not very descriptive and doesn't go into detail about the attributes/properties etc.
2. [x] There is no docstring for the `n_metagenomes` property in the `Project` class.
3. [x] There is no docstring for the `save` method in the `Property` class.
4. [x] In the `add` method of the `Project` class, it basically takes a path to a metagenome config file (supplied through the `fpath` argument) and is supposed to set up a new directory under the project. However, the first few things that it does is to set up directories and paths that do not use the `fpath` argument. If the config path it wants to create doesn't exist but the directory that it should be in does:
https://github.com/KwanLab/Autometa/blob/3e8c7718dc63e5a06654a899506535ae8bd6219e/autometa/config/project.py#L117-L125
...then we raise a FileNotFoundError and give a message. However, the message might be confusing to users if the config file they supplied (which has not been used yet), does exist. Perhaps it should be changed to something like "It appears that there is already a metagenome directory as <path> without a config file in the project hierarchy. Please check the integrity of the project file tree." | 1.0 | [dev branch] Issues with autometa/config/project.py - This file handles configuration handling for the Autometa user project. It contains the definition of the `Project` class. Here are the issues I found:
1. [x] The docstring for the `Project` class is not very descriptive and doesn't go into detail about the attributes/properties etc.
2. [x] There is no docstring for the `n_metagenomes` property in the `Project` class.
3. [x] There is no docstring for the `save` method in the `Property` class.
4. [x] In the `add` method of the `Project` class, it basically takes a path to a metagenome config file (supplied through the `fpath` argument) and is supposed to set up a new directory under the project. However, the first few things that it does is to set up directories and paths that do not use the `fpath` argument. If the config path it wants to create doesn't exist but the directory that it should be in does:
https://github.com/KwanLab/Autometa/blob/3e8c7718dc63e5a06654a899506535ae8bd6219e/autometa/config/project.py#L117-L125
...then we raise a FileNotFoundError and give a message. However, the message might be confusing to users if the config file they supplied (which has not been used yet), does exist. Perhaps it should be changed to something like "It appears that there is already a metagenome directory as <path> without a config file in the project hierarchy. Please check the integrity of the project file tree." | non_test | issues with autometa config project py this file handles configuration handling for the autometa user project it contains the definition of the project class here are the issues i found the docstring for the project class is not very descriptive and doesn t go into detail about the attributes properties etc there is no docstring for the n metagenomes property in the project class there is no docstring for the save method in the property class in the add method of the project class it basically takes a path to a metagenome config file supplied through the fpath argument and is supposed to set up a new directory under the project however the first few things that it does is to set up directories and paths that do not use the fpath argument if the config path it wants to create doesn t exist but the directory that it should be in does then we raise a filenotfounderror and give a message however the message might be confusing to users if the config file they supplied which has not been used yet does exist perhaps it should be changed to something like it appears that there is already a metagenome directory as without a config file in the project hierarchy please check the integrity of the project file tree | 0 |
187,023 | 14,426,916,083 | IssuesEvent | 2020-12-06 00:45:42 | kalexmills/github-vet-tests-dec2020 | https://api.github.com/repos/kalexmills/github-vet-tests-dec2020 | closed | futurewei-cloud/global-scheduler: pkg/controller/nodelifecycle/node_lifecycle_controller_test.go; 3 LoC | fresh test tiny |
Found a possible issue in [futurewei-cloud/global-scheduler](https://www.github.com/futurewei-cloud/global-scheduler) at [pkg/controller/nodelifecycle/node_lifecycle_controller_test.go](https://github.com/futurewei-cloud/global-scheduler/blob/b9329a9fbcd5ca571c8acc498b9d42f5e8b9c19e/pkg/controller/nodelifecycle/node_lifecycle_controller_test.go#L649-L651)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to ds at line 650 may start a goroutine
[Click here to see the code in its original context.](https://github.com/futurewei-cloud/global-scheduler/blob/b9329a9fbcd5ca571c8acc498b9d42f5e8b9c19e/pkg/controller/nodelifecycle/node_lifecycle_controller_test.go#L649-L651)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, ds := range item.daemonSets {
nodeController.daemonSetInformer.Informer().GetStore().Add(&ds)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: b9329a9fbcd5ca571c8acc498b9d42f5e8b9c19e
| 1.0 | futurewei-cloud/global-scheduler: pkg/controller/nodelifecycle/node_lifecycle_controller_test.go; 3 LoC -
Found a possible issue in [futurewei-cloud/global-scheduler](https://www.github.com/futurewei-cloud/global-scheduler) at [pkg/controller/nodelifecycle/node_lifecycle_controller_test.go](https://github.com/futurewei-cloud/global-scheduler/blob/b9329a9fbcd5ca571c8acc498b9d42f5e8b9c19e/pkg/controller/nodelifecycle/node_lifecycle_controller_test.go#L649-L651)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to ds at line 650 may start a goroutine
[Click here to see the code in its original context.](https://github.com/futurewei-cloud/global-scheduler/blob/b9329a9fbcd5ca571c8acc498b9d42f5e8b9c19e/pkg/controller/nodelifecycle/node_lifecycle_controller_test.go#L649-L651)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, ds := range item.daemonSets {
nodeController.daemonSetInformer.Informer().GetStore().Add(&ds)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: b9329a9fbcd5ca571c8acc498b9d42f5e8b9c19e
| test | futurewei cloud global scheduler pkg controller nodelifecycle node lifecycle controller test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to ds at line may start a goroutine click here to show the line s of go which triggered the analyzer go for ds range item daemonsets nodecontroller daemonsetinformer informer getstore add ds leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id | 1 |
271,709 | 8,488,748,538 | IssuesEvent | 2018-10-26 17:38:58 | Sage-Bionetworks/Agora | https://api.github.com/repos/Sage-Bionetworks/Agora | opened | Nominated Targets Page: List only shows 1-3 lines at a time | bug high priority | Scroll seems to be sticking to the top of the paragraph description. Update so the page scrolls until you get to the table's filter bar/dark blue bar, then allow scroll.
<img width="1242" alt="screen shot 2018-10-26 at 10 13 36 am" src="https://user-images.githubusercontent.com/38479289/47583123-517fd580-d90b-11e8-8411-a633536b5159.png">
| 1.0 | Nominated Targets Page: List only shows 1-3 lines at a time - Scroll seems to be sticking to the top of the paragraph description. Update so the page scrolls until you get to the table's filter bar/dark blue bar, then allow scroll.
<img width="1242" alt="screen shot 2018-10-26 at 10 13 36 am" src="https://user-images.githubusercontent.com/38479289/47583123-517fd580-d90b-11e8-8411-a633536b5159.png">
| non_test | nominated targets page list only shows lines at a time scroll seems to be sticking to the top of the paragraph description update so the page scrolls until you get to the table s filter bar dark blue bar then allow scroll img width alt screen shot at am src | 0 |
55,571 | 13,642,965,369 | IssuesEvent | 2020-09-25 16:19:49 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | closed | TensorFlow build is failing on Bazel CI (Release Bazel) for MacOS and Windows | subtype:bazel type:build/install | https://buildkite.com/bazel/tensorflow/builds/5772#d0cb252a-2fab-4c9c-aea3-a1b9e914ea15
Windows seems to have been failing since Sep 18 2020
MacOS seems to have been failing since Sep 19 2020
| 1.0 | TensorFlow build is failing on Bazel CI (Release Bazel) for MacOS and Windows - https://buildkite.com/bazel/tensorflow/builds/5772#d0cb252a-2fab-4c9c-aea3-a1b9e914ea15
Windows seems to have been failing since Sep 18 2020
MacOS seems to have been failing since Sep 19 2020
| non_test | tensorflow build is failing on bazel ci release bazel for macos and windows windows seems to have been failing since sep macos seems to have been failing since sep | 0 |
271,588 | 29,577,729,510 | IssuesEvent | 2023-06-07 01:17:45 | snykiotcubedev/arangodb-3.7.6 | https://api.github.com/repos/snykiotcubedev/arangodb-3.7.6 | closed | WS-2020-0208 (Medium) detected in highlight.js-9.18.1.tgz, highlight.js-9.15.6.tgz - autoclosed | Mend: dependency security vulnerability | ## WS-2020-0208 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>highlight.js-9.18.1.tgz</b>, <b>highlight.js-9.15.6.tgz</b></p></summary>
<p>
<details><summary><b>highlight.js-9.18.1.tgz</b></p></summary>
<p>Syntax highlighting with language autodetection.</p>
<p>Library home page: <a href="https://registry.npmjs.org/highlight.js/-/highlight.js-9.18.1.tgz">https://registry.npmjs.org/highlight.js/-/highlight.js-9.18.1.tgz</a></p>
<p>
Dependency Hierarchy:
- :x: **highlight.js-9.18.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>highlight.js-9.15.6.tgz</b></p></summary>
<p>Syntax highlighting with language autodetection.</p>
<p>Library home page: <a href="https://registry.npmjs.org/highlight.js/-/highlight.js-9.15.6.tgz">https://registry.npmjs.org/highlight.js/-/highlight.js-9.15.6.tgz</a></p>
<p>Path to dependency file: /js/node/package.json</p>
<p>Path to vulnerable library: /js/node/node_modules/highlight.js/package.json</p>
<p>
Dependency Hierarchy:
- :x: **highlight.js-9.15.6.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/snykiotcubedev/arangodb-3.7.6/commit/fce8f85f1c2f070c8e6a8e76d17210a2117d3833">fce8f85f1c2f070c8e6a8e76d17210a2117d3833</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
If are you are using Highlight.js to highlight user-provided data you are possibly vulnerable. On the client-side (in a browser or Electron environment) risks could include lengthy freezes or crashes... On the server-side infinite freezes could occur... effectively preventing users from accessing your app or service (ie, Denial of Service). This is an issue with grammars shipped with the parser (and potentially 3rd party grammars also), not the parser itself. If you are using Highlight.js with any of the following grammars you are vulnerable. If you are using highlightAuto to detect the language (and have any of these grammars registered) you are vulnerable.
<p>Publish Date: 2020-12-04
<p>URL: <a href=https://github.com/highlightjs/highlight.js/commit/373b9d862401162e832ce77305e49b859e110f9c>WS-2020-0208</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2020-12-04</p>
<p>Fix Resolution: 10.4.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2020-0208 (Medium) detected in highlight.js-9.18.1.tgz, highlight.js-9.15.6.tgz - autoclosed - ## WS-2020-0208 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>highlight.js-9.18.1.tgz</b>, <b>highlight.js-9.15.6.tgz</b></p></summary>
<p>
<details><summary><b>highlight.js-9.18.1.tgz</b></p></summary>
<p>Syntax highlighting with language autodetection.</p>
<p>Library home page: <a href="https://registry.npmjs.org/highlight.js/-/highlight.js-9.18.1.tgz">https://registry.npmjs.org/highlight.js/-/highlight.js-9.18.1.tgz</a></p>
<p>
Dependency Hierarchy:
- :x: **highlight.js-9.18.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>highlight.js-9.15.6.tgz</b></p></summary>
<p>Syntax highlighting with language autodetection.</p>
<p>Library home page: <a href="https://registry.npmjs.org/highlight.js/-/highlight.js-9.15.6.tgz">https://registry.npmjs.org/highlight.js/-/highlight.js-9.15.6.tgz</a></p>
<p>Path to dependency file: /js/node/package.json</p>
<p>Path to vulnerable library: /js/node/node_modules/highlight.js/package.json</p>
<p>
Dependency Hierarchy:
- :x: **highlight.js-9.15.6.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/snykiotcubedev/arangodb-3.7.6/commit/fce8f85f1c2f070c8e6a8e76d17210a2117d3833">fce8f85f1c2f070c8e6a8e76d17210a2117d3833</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
If are you are using Highlight.js to highlight user-provided data you are possibly vulnerable. On the client-side (in a browser or Electron environment) risks could include lengthy freezes or crashes... On the server-side infinite freezes could occur... effectively preventing users from accessing your app or service (ie, Denial of Service). This is an issue with grammars shipped with the parser (and potentially 3rd party grammars also), not the parser itself. If you are using Highlight.js with any of the following grammars you are vulnerable. If you are using highlightAuto to detect the language (and have any of these grammars registered) you are vulnerable.
<p>Publish Date: 2020-12-04
<p>URL: <a href=https://github.com/highlightjs/highlight.js/commit/373b9d862401162e832ce77305e49b859e110f9c>WS-2020-0208</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2020-12-04</p>
<p>Fix Resolution: 10.4.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | ws medium detected in highlight js tgz highlight js tgz autoclosed ws medium severity vulnerability vulnerable libraries highlight js tgz highlight js tgz highlight js tgz syntax highlighting with language autodetection library home page a href dependency hierarchy x highlight js tgz vulnerable library highlight js tgz syntax highlighting with language autodetection library home page a href path to dependency file js node package json path to vulnerable library js node node modules highlight js package json dependency hierarchy x highlight js tgz vulnerable library found in head commit a href found in base branch main vulnerability details if are you are using highlight js to highlight user provided data you are possibly vulnerable on the client side in a browser or electron environment risks could include lengthy freezes or crashes on the server side infinite freezes could occur effectively preventing users from accessing your app or service ie denial of service this is an issue with grammars shipped with the parser and potentially party grammars also not the parser itself if you are using highlight js with any of the following grammars you are vulnerable if you are using highlightauto to detect the language and have any of these grammars registered you are vulnerable publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution step up your open source security game with mend | 0 |
680,501 | 23,273,046,609 | IssuesEvent | 2022-08-05 02:44:05 | 4paradigm/OpenMLDB | https://api.github.com/repos/4paradigm/OpenMLDB | closed | Coredump or killed by OOM when enabling UnsafeRowOpt for Criteo dataset | bug high-priority batch-engine | Now we have some bugs which occurs coredump or killed by OOM when enabling UnsafeRowOpt for Criteo dataset. | 1.0 | Coredump or killed by OOM when enabling UnsafeRowOpt for Criteo dataset - Now we have some bugs which occurs coredump or killed by OOM when enabling UnsafeRowOpt for Criteo dataset. | non_test | coredump or killed by oom when enabling unsaferowopt for criteo dataset now we have some bugs which occurs coredump or killed by oom when enabling unsaferowopt for criteo dataset | 0 |
245,142 | 26,519,922,196 | IssuesEvent | 2023-01-19 01:03:50 | brogers588/netflix_conductor_fork | https://api.github.com/repos/brogers588/netflix_conductor_fork | opened | CVE-2022-25901 (Medium) detected in cookiejar-2.1.2.tgz | security vulnerability | ## CVE-2022-25901 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cookiejar-2.1.2.tgz</b></p></summary>
<p>simple persistent cookiejar system</p>
<p>Library home page: <a href="https://registry.npmjs.org/cookiejar/-/cookiejar-2.1.2.tgz">https://registry.npmjs.org/cookiejar/-/cookiejar-2.1.2.tgz</a></p>
<p>
Dependency Hierarchy:
- superagent-3.8.3.tgz (Root Library)
- :x: **cookiejar-2.1.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/brogers588/netflix_conductor_fork/commit/975ea99358eaa6f34b7c8c0c0dce2a0a92a39da5">975ea99358eaa6f34b7c8c0c0dce2a0a92a39da5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of the package cookiejar before 2.1.4 are vulnerable to Regular Expression Denial of Service (ReDoS) via the Cookie.parse function, which uses an insecure regular expression.
<p>Publish Date: 2023-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25901>CVE-2022-25901</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2023-01-18</p>
<p>Fix Resolution: cookiejar - 2.1.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-25901 (Medium) detected in cookiejar-2.1.2.tgz - ## CVE-2022-25901 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cookiejar-2.1.2.tgz</b></p></summary>
<p>simple persistent cookiejar system</p>
<p>Library home page: <a href="https://registry.npmjs.org/cookiejar/-/cookiejar-2.1.2.tgz">https://registry.npmjs.org/cookiejar/-/cookiejar-2.1.2.tgz</a></p>
<p>
Dependency Hierarchy:
- superagent-3.8.3.tgz (Root Library)
- :x: **cookiejar-2.1.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/brogers588/netflix_conductor_fork/commit/975ea99358eaa6f34b7c8c0c0dce2a0a92a39da5">975ea99358eaa6f34b7c8c0c0dce2a0a92a39da5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of the package cookiejar before 2.1.4 are vulnerable to Regular Expression Denial of Service (ReDoS) via the Cookie.parse function, which uses an insecure regular expression.
<p>Publish Date: 2023-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25901>CVE-2022-25901</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2023-01-18</p>
<p>Fix Resolution: cookiejar - 2.1.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve medium detected in cookiejar tgz cve medium severity vulnerability vulnerable library cookiejar tgz simple persistent cookiejar system library home page a href dependency hierarchy superagent tgz root library x cookiejar tgz vulnerable library found in head commit a href found in base branch master vulnerability details versions of the package cookiejar before are vulnerable to regular expression denial of service redos via the cookie parse function which uses an insecure regular expression publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version release date fix resolution cookiejar step up your open source security game with mend | 0 |
292,867 | 22,038,000,547 | IssuesEvent | 2022-05-28 22:44:33 | sdcervi/Waypoint | https://api.github.com/repos/sdcervi/Waypoint | closed | QA thoroughly | documentation | **General**
- [x] Alt text and aria labels
- [x] Responsive scaling across breakpoints
- [x] Links: no broken links or incorrect targets
- [x] No typos, filler content, or incorrect copy
- [x] All images and SVGs are correctly sized
- [x] Looks good on all device sizes, including older smaller devices (test on Pixel 3 and iPhone 6)
**Android**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Displays in full-screen mode when added to home screen
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**iPhone Chrome**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Displays in full-screen mode when added to home screen
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**iPhone Safari**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Displays in full-screen mode when added to home screen
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**iPad Chrome**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Displays in full-screen mode when added to home screen
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**iPad Safari**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Displays in full-screen mode when added to home screen
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**Windows Chrome**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**Windows Firefox**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**Windows IE/Edge**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**Mac Chrome**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**Mac Safari**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**Mac Firefox**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**Mac Edge**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues | 1.0 | QA thoroughly - **General**
- [x] Alt text and aria labels
- [x] Responsive scaling across breakpoints
- [x] Links: no broken links or incorrect targets
- [x] No typos, filler content, or incorrect copy
- [x] All images and SVGs are correctly sized
- [x] Looks good on all device sizes, including older smaller devices (test on Pixel 3 and iPhone 6)
**Android**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Displays in full-screen mode when added to home screen
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**iPhone Chrome**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Displays in full-screen mode when added to home screen
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**iPhone Safari**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Displays in full-screen mode when added to home screen
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**iPad Chrome**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Displays in full-screen mode when added to home screen
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**iPad Safari**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Displays in full-screen mode when added to home screen
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**Windows Chrome**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**Windows Firefox**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**Windows IE/Edge**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**Mac Chrome**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**Mac Safari**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**Mac Firefox**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues
**Mac Edge**
- [x] Add challenge
- [x] Load sample data on demo page
- [x] Delete challenge
- [x] Add progress
- [x] Edit progress
- [x] Milestones display correctly and only when needed
- [x] Profile page
- [x] Homepage
- [x] Dashboard
- [x] About
- [x] No other weird display and functionality issues | non_test | qa thoroughly general alt text and aria labels responsive scaling across breakpoints links no broken links or incorrect targets no typos filler content or incorrect copy all images and svgs are correctly sized looks good on all device sizes including older smaller devices test on pixel and iphone android add challenge load sample data on demo page delete challenge add progress edit progress milestones display correctly and only when needed displays in full screen mode when added to home screen profile page homepage dashboard about no other weird display and functionality issues iphone chrome add challenge load sample data on demo page delete challenge add progress edit progress milestones display correctly and only when needed displays in full screen mode when added to home screen profile page homepage dashboard about no other weird display and functionality issues iphone safari add challenge load sample data on demo page delete challenge add progress edit progress milestones display correctly and only when needed displays in full screen mode when added to home screen profile page homepage dashboard about no other weird display and functionality issues ipad chrome add challenge load sample data on demo page delete challenge add progress edit progress milestones display correctly and only when needed displays in full screen mode when added to home screen profile page homepage dashboard about no other weird display and functionality issues ipad safari add challenge load sample data on demo page delete challenge add progress edit progress milestones display correctly and only when needed displays in full screen mode when added to home screen profile page homepage dashboard about no other weird display and functionality issues windows chrome add challenge load sample data on demo page delete challenge add progress edit progress milestones display correctly and only when needed profile page homepage dashboard about no other weird display and functionality issues windows firefox add challenge load sample data on demo page delete challenge add progress edit progress milestones display correctly and only when needed profile page homepage dashboard about no other weird display and functionality issues windows ie edge add challenge load sample data on demo page delete challenge add progress edit progress milestones display correctly and only when needed profile page homepage dashboard about no other weird display and functionality issues mac chrome add challenge load sample data on demo page delete challenge add progress edit progress milestones display correctly and only when needed profile page homepage dashboard about no other weird display and functionality issues mac safari add challenge load sample data on demo page delete challenge add progress edit progress milestones display correctly and only when needed profile page homepage dashboard about no other weird display and functionality issues mac firefox add challenge load sample data on demo page delete challenge add progress edit progress milestones display correctly and only when needed profile page homepage dashboard about no other weird display and functionality issues mac edge add challenge load sample data on demo page delete challenge add progress edit progress milestones display correctly and only when needed profile page homepage dashboard about no other weird display and functionality issues | 0 |
136,719 | 11,081,416,572 | IssuesEvent | 2019-12-13 09:45:16 | dzhw/SLC-IntEr | https://api.github.com/repos/dzhw/SLC-IntEr | closed | Layout Episodenmodul | Modul: Episoden layout mittlere Priorität testing | 
- [x] Können wir auf diesen Seiten jeweils den Umbruch bereits nach "deutlich" und "genau" haben? Dann wäre in der zweiten Zeile immer ein Wort und die jeweilige Zahl. Das sähe dann einheitlicher aus. Betrifft die Seiten emaa43, emba43, emca43, emda43, emea43, emfa43 und em0043

- [x] Hier bitte die Antwortkategorien noch etwas entzerren. Das ist leider nur schwer lesbar. Betrifft die Seiten staa24, stab24, stac24, stad24, stae24, stba24, stbb24, stbc24, stbd, stca24, stcb24, stcc24, stda24, stdb24, stea24, st0025 | 1.0 | Layout Episodenmodul - 
- [x] Können wir auf diesen Seiten jeweils den Umbruch bereits nach "deutlich" und "genau" haben? Dann wäre in der zweiten Zeile immer ein Wort und die jeweilige Zahl. Das sähe dann einheitlicher aus. Betrifft die Seiten emaa43, emba43, emca43, emda43, emea43, emfa43 und em0043

- [x] Hier bitte die Antwortkategorien noch etwas entzerren. Das ist leider nur schwer lesbar. Betrifft die Seiten staa24, stab24, stac24, stad24, stae24, stba24, stbb24, stbc24, stbd, stca24, stcb24, stcc24, stda24, stdb24, stea24, st0025 | test | layout episodenmodul können wir auf diesen seiten jeweils den umbruch bereits nach deutlich und genau haben dann wäre in der zweiten zeile immer ein wort und die jeweilige zahl das sähe dann einheitlicher aus betrifft die seiten und hier bitte die antwortkategorien noch etwas entzerren das ist leider nur schwer lesbar betrifft die seiten stbd | 1 |
275,929 | 23,954,419,040 | IssuesEvent | 2022-09-12 13:59:09 | samtools/bcftools | https://api.github.com/repos/samtools/bcftools | closed | -i and -e are not complement when I use query subcommand | requires-test-case | I noticed the result of
`bcftools query -i 'FMT/DP>=20' -f '[%CHROM %POS %SAMPLE\n]' input.vcf`
and
`bcftools query -e 'FMT/DP<20' -f '[%CHROM %POS %SAMPLE\n]' input.vcf`
are not complement.
This occur if only I specified %SAMPLE but %FORMAT tags, so I can avoid it by specifying other %FORMAT tags but it is a little inconvenient and confusing.
I think this is because the process is skipped by the code shown below when I use `-e` and the condition is met in some samples.
If only %SAMPLE is specified, not %FMT/DP, etc., the flag BCF_UN_FMT is not set. When I use `-i`, this checking and skipping do not occur.
https://github.com/samtools/bcftools/blob/12088c0887d2c680ae55240d806d089659d61dca/vcfquery.c#L146
The simple solutions I think is to omit the above code or to change the code below to set the flag.
https://github.com/samtools/bcftools/blob/12088c0887d2c680ae55240d806d089659d61dca/convert.c#L1304
However, I'm not sure how this change will affect the rest of the code.
Thank you, | 1.0 | -i and -e are not complement when I use query subcommand - I noticed the result of
`bcftools query -i 'FMT/DP>=20' -f '[%CHROM %POS %SAMPLE\n]' input.vcf`
and
`bcftools query -e 'FMT/DP<20' -f '[%CHROM %POS %SAMPLE\n]' input.vcf`
are not complement.
This occur if only I specified %SAMPLE but %FORMAT tags, so I can avoid it by specifying other %FORMAT tags but it is a little inconvenient and confusing.
I think this is because the process is skipped by the code shown below when I use `-e` and the condition is met in some samples.
If only %SAMPLE is specified, not %FMT/DP, etc., the flag BCF_UN_FMT is not set. When I use `-i`, this checking and skipping do not occur.
https://github.com/samtools/bcftools/blob/12088c0887d2c680ae55240d806d089659d61dca/vcfquery.c#L146
The simple solutions I think is to omit the above code or to change the code below to set the flag.
https://github.com/samtools/bcftools/blob/12088c0887d2c680ae55240d806d089659d61dca/convert.c#L1304
However, I'm not sure how this change will affect the rest of the code.
Thank you, | test | i and e are not complement when i use query subcommand i noticed the result of bcftools query i fmt dp f input vcf and bcftools query e fmt dp f input vcf are not complement this occur if only i specified sample but format tags so i can avoid it by specifying other format tags but it is a little inconvenient and confusing i think this is because the process is skipped by the code shown below when i use e and the condition is met in some samples if only sample is specified not fmt dp etc the flag bcf un fmt is not set when i use i this checking and skipping do not occur the simple solutions i think is to omit the above code or to change the code below to set the flag however i m not sure how this change will affect the rest of the code thank you | 1 |
136,882 | 20,011,648,035 | IssuesEvent | 2022-02-01 07:28:58 | RRZE-Webteam/FAU-Einrichtungen | https://api.github.com/repos/RRZE-Webteam/FAU-Einrichtungen | opened | Mouseover Effekt funktioniert nicht bei Bildern die eine Bildunterschrift enthalten (Grid-Gallery) | Design Bug (Optik) | - [ ] Gallery Grid Darstellung: Der Mouseover Effekt (Zoom und SW-Darstellung) funktioniert nicht bei Bildern die eine Bildunterschrift enthalten.
- [ ] Gallery mit Slider: Die Vorschaubilder im Slider werden nicht zentriert dargestellt das sieht bei Gallerien die Bilder mit verschiedenen Formaten enthalten teilweise sehr unschön aus.
( 🎫 http://go.fau.de/9s-ck | Aus Helpdesk ) | 1.0 | Mouseover Effekt funktioniert nicht bei Bildern die eine Bildunterschrift enthalten (Grid-Gallery) - - [ ] Gallery Grid Darstellung: Der Mouseover Effekt (Zoom und SW-Darstellung) funktioniert nicht bei Bildern die eine Bildunterschrift enthalten.
- [ ] Gallery mit Slider: Die Vorschaubilder im Slider werden nicht zentriert dargestellt das sieht bei Gallerien die Bilder mit verschiedenen Formaten enthalten teilweise sehr unschön aus.
( 🎫 http://go.fau.de/9s-ck | Aus Helpdesk ) | non_test | mouseover effekt funktioniert nicht bei bildern die eine bildunterschrift enthalten grid gallery gallery grid darstellung der mouseover effekt zoom und sw darstellung funktioniert nicht bei bildern die eine bildunterschrift enthalten gallery mit slider die vorschaubilder im slider werden nicht zentriert dargestellt das sieht bei gallerien die bilder mit verschiedenen formaten enthalten teilweise sehr unschön aus 🎫 aus helpdesk | 0 |
15,408 | 3,462,138,752 | IssuesEvent | 2015-12-20 17:45:33 | odds-lang/odds | https://api.github.com/repos/odds-lang/odds | closed | String operations and standard library | compiler tests | - [ ] add `str_index` builtin function @lillyfwang
- [ ] add basic string standard library | 1.0 | String operations and standard library - - [ ] add `str_index` builtin function @lillyfwang
- [ ] add basic string standard library | test | string operations and standard library add str index builtin function lillyfwang add basic string standard library | 1 |
56,277 | 15,015,397,916 | IssuesEvent | 2021-02-01 08:14:54 | gy0512/comments.github.io | https://api.github.com/repos/gy0512/comments.github.io | opened | How do you determine priority and severity of a defect? - TestBirds | /post/determine-defect-the-priority-and-severity/ Gitalk | https://gy0512.github.io/post/determine-defect-the-priority-and-severity/
Basic Test Knowledge. | 1.0 | How do you determine priority and severity of a defect? - TestBirds - https://gy0512.github.io/post/determine-defect-the-priority-and-severity/
Basic Test Knowledge. | non_test | how do you determine priority and severity of a defect testbirds basic test knowledge | 0 |
12,553 | 3,621,295,309 | IssuesEvent | 2016-02-08 23:25:47 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | closed | RFE: Review Workflow Diagram | kind/documentation priority/P1 team/test-infra team/ux | Use Case:
Developer has some cool new feature that they wish to add to kubernetes. This diagram walks them through the process that the developer will follow to have their PR-reviewed in a timely fashion. It will also set the expectations for all parties involved in order to remove ambiguity.
@bgrant0607
/cc @childsb @kubernetes/rh-cluster-infra | 1.0 | RFE: Review Workflow Diagram - Use Case:
Developer has some cool new feature that they wish to add to kubernetes. This diagram walks them through the process that the developer will follow to have their PR-reviewed in a timely fashion. It will also set the expectations for all parties involved in order to remove ambiguity.
@bgrant0607
/cc @childsb @kubernetes/rh-cluster-infra | non_test | rfe review workflow diagram use case developer has some cool new feature that they wish to add to kubernetes this diagram walks them through the process that the developer will follow to have their pr reviewed in a timely fashion it will also set the expectations for all parties involved in order to remove ambiguity cc childsb kubernetes rh cluster infra | 0 |
47,756 | 13,066,198,589 | IssuesEvent | 2020-07-30 21:11:45 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | closed | cmake - I3_USE_ROOT macro set w/o root installed (Trac #1131) | Migrated from Trac cmake defect | if USE_ROOT is given to cmake, the C macro I3_USE_ROOT is set even if root isn't found.
see #1115
Migrated from https://code.icecube.wisc.edu/ticket/1131
```json
{
"status": "closed",
"changetime": "2019-01-11T23:59:15",
"description": "if USE_ROOT is given to cmake, the C macro I3_USE_ROOT is set even if root isn't found.\n\nsee #1121",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"_ts": "1547251155701656",
"component": "cmake",
"summary": "cmake - I3_USE_ROOT macro set w/o root installed",
"priority": "blocker",
"keywords": "",
"time": "2015-08-17T18:43:53",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
| 1.0 | cmake - I3_USE_ROOT macro set w/o root installed (Trac #1131) - if USE_ROOT is given to cmake, the C macro I3_USE_ROOT is set even if root isn't found.
see #1115
Migrated from https://code.icecube.wisc.edu/ticket/1131
```json
{
"status": "closed",
"changetime": "2019-01-11T23:59:15",
"description": "if USE_ROOT is given to cmake, the C macro I3_USE_ROOT is set even if root isn't found.\n\nsee #1121",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"_ts": "1547251155701656",
"component": "cmake",
"summary": "cmake - I3_USE_ROOT macro set w/o root installed",
"priority": "blocker",
"keywords": "",
"time": "2015-08-17T18:43:53",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
| non_test | cmake use root macro set w o root installed trac if use root is given to cmake the c macro use root is set even if root isn t found see migrated from json status closed changetime description if use root is given to cmake the c macro use root is set even if root isn t found n nsee reporter nega cc resolution fixed ts component cmake summary cmake use root macro set w o root installed priority blocker keywords time milestone owner nega type defect | 0 |
160,683 | 20,116,576,354 | IssuesEvent | 2022-02-07 20:12:27 | snowdensb/nifi | https://api.github.com/repos/snowdensb/nifi | opened | CVE-2022-21724 (High) detected in postgresql-42.2.19.jar | security vulnerability | ## CVE-2022-21724 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postgresql-42.2.19.jar</b></p></summary>
<p>PostgreSQL JDBC Driver Postgresql</p>
<p>Library home page: <a href="https://jdbc.postgresql.org">https://jdbc.postgresql.org</a></p>
<p>Path to dependency file: /nifi-registry/nifi-registry-core/nifi-registry-test/pom.xml</p>
<p>Path to vulnerable library: /r/.m2/repository/org/postgresql/postgresql/42.2.19/postgresql-42.2.19.jar</p>
<p>
Dependency Hierarchy:
- :x: **postgresql-42.2.19.jar** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
pgjdbc is the offical PostgreSQL JDBC Driver. A security hole was found in the jdbc driver for postgresql database while doing security research. The system using the postgresql library will be attacked when attacker control the jdbc url or properties. pgjdbc instantiates plugin instances based on class names provided via `authenticationPluginClassName`, `sslhostnameverifier`, `socketFactory`, `sslfactory`, `sslpasswordcallback` connection properties. However, the driver did not verify if the class implements the expected interface before instantiating the class. This can lead to code execution loaded via arbitrary classes. Users using plugins are advised to upgrade. There are no known workarounds for this issue.
<p>Publish Date: 2022-02-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21724>CVE-2022-21724</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-v7wg-cpwc-24m4">https://github.com/advisories/GHSA-v7wg-cpwc-24m4</a></p>
<p>Release Date: 2022-02-02</p>
<p>Fix Resolution: org.postgresql:postgresql:42.2.25,42.3.2</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.postgresql","packageName":"postgresql","packageVersion":"42.2.19","packageFilePaths":["/nifi-registry/nifi-registry-core/nifi-registry-test/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.postgresql:postgresql:42.2.19","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.postgresql:postgresql:42.2.25,42.3.2","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2022-21724","vulnerabilityDetails":"pgjdbc is the offical PostgreSQL JDBC Driver. A security hole was found in the jdbc driver for postgresql database while doing security research. The system using the postgresql library will be attacked when attacker control the jdbc url or properties. pgjdbc instantiates plugin instances based on class names provided via `authenticationPluginClassName`, `sslhostnameverifier`, `socketFactory`, `sslfactory`, `sslpasswordcallback` connection properties. However, the driver did not verify if the class implements the expected interface before instantiating the class. This can lead to code execution loaded via arbitrary classes. Users using plugins are advised to upgrade. There are no known workarounds for this issue.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21724","cvss3Severity":"high","cvss3Score":"7.0","cvss3Metrics":{"A":"High","AC":"High","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Local","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2022-21724 (High) detected in postgresql-42.2.19.jar - ## CVE-2022-21724 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postgresql-42.2.19.jar</b></p></summary>
<p>PostgreSQL JDBC Driver Postgresql</p>
<p>Library home page: <a href="https://jdbc.postgresql.org">https://jdbc.postgresql.org</a></p>
<p>Path to dependency file: /nifi-registry/nifi-registry-core/nifi-registry-test/pom.xml</p>
<p>Path to vulnerable library: /r/.m2/repository/org/postgresql/postgresql/42.2.19/postgresql-42.2.19.jar</p>
<p>
Dependency Hierarchy:
- :x: **postgresql-42.2.19.jar** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
pgjdbc is the offical PostgreSQL JDBC Driver. A security hole was found in the jdbc driver for postgresql database while doing security research. The system using the postgresql library will be attacked when attacker control the jdbc url or properties. pgjdbc instantiates plugin instances based on class names provided via `authenticationPluginClassName`, `sslhostnameverifier`, `socketFactory`, `sslfactory`, `sslpasswordcallback` connection properties. However, the driver did not verify if the class implements the expected interface before instantiating the class. This can lead to code execution loaded via arbitrary classes. Users using plugins are advised to upgrade. There are no known workarounds for this issue.
<p>Publish Date: 2022-02-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21724>CVE-2022-21724</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-v7wg-cpwc-24m4">https://github.com/advisories/GHSA-v7wg-cpwc-24m4</a></p>
<p>Release Date: 2022-02-02</p>
<p>Fix Resolution: org.postgresql:postgresql:42.2.25,42.3.2</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.postgresql","packageName":"postgresql","packageVersion":"42.2.19","packageFilePaths":["/nifi-registry/nifi-registry-core/nifi-registry-test/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.postgresql:postgresql:42.2.19","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.postgresql:postgresql:42.2.25,42.3.2","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2022-21724","vulnerabilityDetails":"pgjdbc is the offical PostgreSQL JDBC Driver. A security hole was found in the jdbc driver for postgresql database while doing security research. The system using the postgresql library will be attacked when attacker control the jdbc url or properties. pgjdbc instantiates plugin instances based on class names provided via `authenticationPluginClassName`, `sslhostnameverifier`, `socketFactory`, `sslfactory`, `sslpasswordcallback` connection properties. However, the driver did not verify if the class implements the expected interface before instantiating the class. This can lead to code execution loaded via arbitrary classes. Users using plugins are advised to upgrade. There are no known workarounds for this issue.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21724","cvss3Severity":"high","cvss3Score":"7.0","cvss3Metrics":{"A":"High","AC":"High","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Local","I":"High"},"extraData":{}}</REMEDIATE> --> | non_test | cve high detected in postgresql jar cve high severity vulnerability vulnerable library postgresql jar postgresql jdbc driver postgresql library home page a href path to dependency file nifi registry nifi registry core nifi registry test pom xml path to vulnerable library r repository org postgresql postgresql postgresql jar dependency hierarchy x postgresql jar vulnerable library found in base branch main vulnerability details pgjdbc is the offical postgresql jdbc driver a security hole was found in the jdbc driver for postgresql database while doing security research the system using the postgresql library will be attacked when attacker control the jdbc url or properties pgjdbc instantiates plugin instances based on class names provided via authenticationpluginclassname sslhostnameverifier socketfactory sslfactory sslpasswordcallback connection properties however the driver did not verify if the class implements the expected interface before instantiating the class this can lead to code execution loaded via arbitrary classes users using plugins are advised to upgrade there are no known workarounds for this issue publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org postgresql postgresql rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree org postgresql postgresql isminimumfixversionavailable true minimumfixversion org postgresql postgresql isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails pgjdbc is the offical postgresql jdbc driver a security hole was found in the jdbc driver for postgresql database while doing security research the system using the postgresql library will be attacked when attacker control the jdbc url or properties pgjdbc instantiates plugin instances based on class names provided via authenticationpluginclassname sslhostnameverifier socketfactory sslfactory sslpasswordcallback connection properties however the driver did not verify if the class implements the expected interface before instantiating the class this can lead to code execution loaded via arbitrary classes users using plugins are advised to upgrade there are no known workarounds for this issue vulnerabilityurl | 0 |
397,192 | 11,725,040,403 | IssuesEvent | 2020-03-10 12:11:13 | cliqz/user-agent-ios | https://api.github.com/repos/cliqz/user-agent-ios | closed | Rename bookmarks | Feature/Browser Priority/1 Type/Enhancement | Long press on a bookmark on bookmark list should open offer an option to open a UI that will allow user to change bookmark title and url. There should be such code in Firefox codebase already. | 1.0 | Rename bookmarks - Long press on a bookmark on bookmark list should open offer an option to open a UI that will allow user to change bookmark title and url. There should be such code in Firefox codebase already. | non_test | rename bookmarks long press on a bookmark on bookmark list should open offer an option to open a ui that will allow user to change bookmark title and url there should be such code in firefox codebase already | 0 |
52,782 | 13,051,121,254 | IssuesEvent | 2020-07-29 16:33:00 | vmware/clarity | https://api.github.com/repos/vmware/clarity | closed | Update CLI installer to support Angular 10, tslib 2.0 and @clr/ui 4.0 | @clr/angular flag: has workaround type: build | ## Describe the bug
The Clarity packages have not been updated to support Angular 10 and tslib 2.0.
## How to reproduce
I have Angular 10 installed, and tslib 2.0. When trying to install Clarity for the first time using the Angular CLI option 1 (ng add @clr/angular), it did not install @clr/core and gave me a warning about unmet peer dependencies. It did install the other Clarity elements.
When I installed @clr/core separately, it gave more peer dependency messages, but did install:
<img width="1093" alt="Screen Shot 2020-07-18 at 10 43 50 AM" src="https://user-images.githubusercontent.com/62292109/87855162-6329bd80-c8e4-11ea-9bb4-a1127486da4c.png">
Note that the Clarity ui @clr/ui is at version 4.0.0 and also triggers a peer dependency warning.
Steps to reproduce the behavior:
1. Go to Angular app root directory in terminal window
2. Enter "ng add @clr/core"
3. See npm warnings
## Expected behavior
CLI installation of Clarity should proceed without npm warnings.
## Versions
**App**
- Angular: 10
- Node: 13.11.0
- Clarity: 3.1.4
**Device:**
- Type: MacBook
- OS: iOS
- Browser None
- Version n/a
## Additional notes
_Add any other notes about the problem here._
| 1.0 | Update CLI installer to support Angular 10, tslib 2.0 and @clr/ui 4.0 - ## Describe the bug
The Clarity packages have not been updated to support Angular 10 and tslib 2.0.
## How to reproduce
I have Angular 10 installed, and tslib 2.0. When trying to install Clarity for the first time using the Angular CLI option 1 (ng add @clr/angular), it did not install @clr/core and gave me a warning about unmet peer dependencies. It did install the other Clarity elements.
When I installed @clr/core separately, it gave more peer dependency messages, but did install:
<img width="1093" alt="Screen Shot 2020-07-18 at 10 43 50 AM" src="https://user-images.githubusercontent.com/62292109/87855162-6329bd80-c8e4-11ea-9bb4-a1127486da4c.png">
Note that the Clarity ui @clr/ui is at version 4.0.0 and also triggers a peer dependency warning.
Steps to reproduce the behavior:
1. Go to Angular app root directory in terminal window
2. Enter "ng add @clr/core"
3. See npm warnings
## Expected behavior
CLI installation of Clarity should proceed without npm warnings.
## Versions
**App**
- Angular: 10
- Node: 13.11.0
- Clarity: 3.1.4
**Device:**
- Type: MacBook
- OS: iOS
- Browser None
- Version n/a
## Additional notes
_Add any other notes about the problem here._
| non_test | update cli installer to support angular tslib and clr ui describe the bug the clarity packages have not been updated to support angular and tslib how to reproduce i have angular installed and tslib when trying to install clarity for the first time using the angular cli option ng add clr angular it did not install clr core and gave me a warning about unmet peer dependencies it did install the other clarity elements when i installed clr core separately it gave more peer dependency messages but did install img width alt screen shot at am src note that the clarity ui clr ui is at version and also triggers a peer dependency warning steps to reproduce the behavior go to angular app root directory in terminal window enter ng add clr core see npm warnings expected behavior cli installation of clarity should proceed without npm warnings versions app angular node clarity device type macbook os ios browser none version n a additional notes add any other notes about the problem here | 0 |
337,352 | 30,247,551,012 | IssuesEvent | 2023-07-06 17:42:28 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | reopened | Fix matrix_and_vector_products.test_numpy_outer | NumPy Frontend Sub Task Failing Test | | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5478319447"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5478319447"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5423118933"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5478319447"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5478319447"><img src=https://img.shields.io/badge/-success-success></a>
| 1.0 | Fix matrix_and_vector_products.test_numpy_outer - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5478319447"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5478319447"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5423118933"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5478319447"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5478319447"><img src=https://img.shields.io/badge/-success-success></a>
| test | fix matrix and vector products test numpy outer tensorflow a href src jax a href src numpy a href src torch a href src paddle a href src | 1 |
177,401 | 6,583,299,393 | IssuesEvent | 2017-09-13 04:41:52 | ankidroid/Anki-Android | https://api.github.com/repos/ankidroid/Anki-Android | closed | Notification only shown when app is opened | advanced-reminders enhancement Priority-Low Widget | ###### Research
*Enter an [ x ] character to confirm the points below:*
[x] I have read the [support page](https://ankidroid.org/docs/help.html) and am reporting a bug or enhancement request specific to AnkiDroid
[x] I have checked the [manual](https://ankidroid.org/docs/manual.html) and the [FAQ](https://github.com/ankidroid/Anki-Android/wiki/FAQ) and could not find a solution to my issue
[x] I have searched for similar existing issues here and on the user forum
###### Reproduction Steps
1. Enable notifications on more than 10 due cards.
2. Enter some Cards
3. Wait until more than 10 cards are due.
###### Expected Result
A notification should be shown without opening the app.
###### Actual Result
I have to open the app, then a notification is shown. But this does not make sense as I can see the due cards in the allready opened app.
###### Debug info
AnkiDroid Version = 2.8.2
Android Version = 6.0.1
ACRA UUID = 78549792-7916-4041-a6b2-b58d85214139
| 1.0 | Notification only shown when app is opened - ###### Research
*Enter an [ x ] character to confirm the points below:*
[x] I have read the [support page](https://ankidroid.org/docs/help.html) and am reporting a bug or enhancement request specific to AnkiDroid
[x] I have checked the [manual](https://ankidroid.org/docs/manual.html) and the [FAQ](https://github.com/ankidroid/Anki-Android/wiki/FAQ) and could not find a solution to my issue
[x] I have searched for similar existing issues here and on the user forum
###### Reproduction Steps
1. Enable notifications on more than 10 due cards.
2. Enter some Cards
3. Wait until more than 10 cards are due.
###### Expected Result
A notification should be shown without opening the app.
###### Actual Result
I have to open the app, then a notification is shown. But this does not make sense as I can see the due cards in the allready opened app.
###### Debug info
AnkiDroid Version = 2.8.2
Android Version = 6.0.1
ACRA UUID = 78549792-7916-4041-a6b2-b58d85214139
| non_test | notification only shown when app is opened research enter an character to confirm the points below i have read the and am reporting a bug or enhancement request specific to ankidroid i have checked the and the and could not find a solution to my issue i have searched for similar existing issues here and on the user forum reproduction steps enable notifications on more than due cards enter some cards wait until more than cards are due expected result a notification should be shown without opening the app actual result i have to open the app then a notification is shown but this does not make sense as i can see the due cards in the allready opened app debug info ankidroid version android version acra uuid | 0 |
92,513 | 11,661,691,327 | IssuesEvent | 2020-03-03 07:27:43 | cagov/UX | https://api.github.com/repos/cagov/UX | closed | UX prototype: Apply online for LifeLine | Design Tech UX | **Parent User Story:** https://github.com/cagov/UX/issues/64
**content prototypes:**
#106
**Prototype Link(s):**
-- https://cagov.github.io/UX/apply-online-discounted-phone-service
- [ ] HTML page(s) created and linked in UX repo
- [ ] Meta description added
- [ ] Page title updated
- [ ] Content integrated according to design system and standards
- [ ] Automated accessibility checks performed (score of 100%)
- [ ] Manual accessibility checks performed ([See checklist](https://drive.google.com/open?id=0B72FqFUONAggazJwQURQeTBPMEZGRjJXSFNGcDJrMF9ENERR))
- [ ] HTML structure verified
- [ ] Design issue template # (Closed)
- [ ] Migrated to Alpha repo
**11y Status:** Issues and resolution status if applicable
**Design Issue Link:** Link to Design issue if applicable#
| 1.0 | UX prototype: Apply online for LifeLine - **Parent User Story:** https://github.com/cagov/UX/issues/64
**content prototypes:**
#106
**Prototype Link(s):**
-- https://cagov.github.io/UX/apply-online-discounted-phone-service
- [ ] HTML page(s) created and linked in UX repo
- [ ] Meta description added
- [ ] Page title updated
- [ ] Content integrated according to design system and standards
- [ ] Automated accessibility checks performed (score of 100%)
- [ ] Manual accessibility checks performed ([See checklist](https://drive.google.com/open?id=0B72FqFUONAggazJwQURQeTBPMEZGRjJXSFNGcDJrMF9ENERR))
- [ ] HTML structure verified
- [ ] Design issue template # (Closed)
- [ ] Migrated to Alpha repo
**11y Status:** Issues and resolution status if applicable
**Design Issue Link:** Link to Design issue if applicable#
| non_test | ux prototype apply online for lifeline parent user story content prototypes prototype link s html page s created and linked in ux repo meta description added page title updated content integrated according to design system and standards automated accessibility checks performed score of manual accessibility checks performed html structure verified design issue template closed migrated to alpha repo status issues and resolution status if applicable design issue link link to design issue if applicable | 0 |
204,524 | 15,500,234,521 | IssuesEvent | 2021-03-11 09:03:32 | openmsupply/mobile | https://api.github.com/repos/openmsupply/mobile | closed | TESTING: Supplier Requisitions Page | Needs testing Test Plan Template | # TESTING
This is an issue for testing `SupplierRequisitionsPage`.
Please copy and paste this list of tests and post your results in a comment in this issue.
#### TESTER NOTES:
- Ensure one of the EXTERNAL SUPPLIERS (`Name` record which is a facility, NOT a store, and is a supplier) has a very long name
- If testing both `SupplierRequisitionPage` and `SupplierRequisitionsPage`, will be faster to do them both without programs first and then with programs after
## Supplier Requisitions Page
This page should be tested WITH NO PROGRAMS and WITH PROGRAMS. Alterations to the behaviour of this page are listed at the BOTTOM of this list.
### NAVIGATION
- [ ] Can navigate to this page through the Menu Page, `"Supplier Requisition"` button
- [ ] Pressing either the top left back button, or hardware back button navigates to the Menu Page
- [ ] Navigating to this page from a supplier requisition shows the correctly updated details (finalise or add items to a supplier requisition, navigating back shows that supplier requisition as finalised or the increase in items)
### SYNC
- [ ] Supplier requisitions are synced to desktop
- [ ] Supplier requisitions are synced from desktop
### Table
- [ ] List all supplier requisitions for this store (matches internal orders list in desktop)
- [ ] Initial sort is by requisition number in descending order
#### Row Press
- [ ] Pressing a row (except for the `"DELETE"` column), navigates to the `SupplierRequisitionPage`
- [ ] The supplier requisition navigated to is the row that was pressed
#### Filtering
- [ ] Supplier requisitions are filtered by requisition number when a search term is entered
- [ ] Filtering is cleared when the search term is cleared [by backspacing or pressing the X].
#### Columns
- [ ] **REQUISITION NUMBER**
- [ ] Sortable (and sorts correctly)
- [ ] New requisitions are correctly incremented (customer and supplier requisitions share numbers)
- [ ] Sorting icon updates correctly to indicate ascending/descending and no sorting
- [ ] Cells are NOT editable
- [ ] Left aligned
- [ ] **SUPPLIER**
- [ ] Sortable (and sorts correctly)
- [ ] Supplier names are correctly listed for every supplier requisition
- [ ] Sorting icon updates correctly to indicate ascending/descending and no sorting.
- [ ] Cells NOT editable
- [ ] Text overflow handled (create a Supplier name that is very long)
- [ ] Left aligned
- [ ] **PROGRAM** (NOTE: only displays if at least one requisition is associated with a program)
- [ ] Does not display if no requisitions are associated with a program (includes current AND past requisitions)
- [ ] Displays if at least one requisition is associated with a program (includes current AND past requisitions)
- [ ] Sortable (and sorts correctly)
- [ ] Programs are listed correctly for every customer requisition (`"N/A"` if requisition has no program)
- [ ] Sorting icon updates correctly to indicate ascending/descending and no sorting
- [ ] Cells are NOT editable
- [ ] Text overflow handled (create a program name that is very long)
- [ ] Left aligned
- [ ] **ITEMS**
- [ ] Sortable (and sorts correctly)
- [ ] Shows the correct number of items in the supplier requisition
- [ ] Sorting icon updates correctly to indicate ascending/descending and no sorting.
- [ ] Right aligned
- [ ] Number of items is correctly updated (when navigating back from a supplier requisition)
- [ ] **ENTERED DATE**
- [ ] Sortable (and sorts correctly): this might be difficult if the data file is new or there are no old records, so clicking the sort button and having the icon change is sufficient
- [ ] Shows the correct date an invoice was created
- [ ] Dates are in the form: `WeekDay Month Year`, e.g. `17 May 2008`
- [ ] Sorting icon updates correctly to indicate ascending/descending and no sorting
- [ ] Left aligned
- [ ] Cells are NOT editable
- [ ] **STATUS**
- [ ] NOT sortable
- [ ] Shows the correct status for an invoice
- [ ] Status is one of: `"In Progress"` or `"Finalised"` (or localised equivalents)
- [ ] Left aligned
- [ ] Cells are NOT editable
- [ ] **Delete**
- [ ] NOT a sortable column
- [ ] Cells NOT editable
- [ ] Finalised invoices are not selectable
- [ ] Finalised rows indicate they are not selectable
- [ ] Rows not finalised are selectable
- [ ] Multiple rows are selectable
- [ ] Selecting a row opens a confirmation dialog at the bottom of the page
- [ ] Unselecting all rows closes the confirmation dialog at the bottom of the page
- [ ] Confirming the confirmation dialog deletes the supplier requisitions
- [ ] Cancelling the confirmation dialog removes the current selection and closes the confirmation dialog
- [ ] Finalised invoices are not selectable
### Buttons
- [ ] `"New Requisition"`
- [ ] Button is displayed.
- [ ] On press, button opens `"Select a supplier"` modal
### Modals
- [ ] **New Requisition Modal**
- [ ] Title is `"Select a supplier"`
- [ ] Options to select are names records that are: VISIBLE, a SUPPLIER, not THIS store and a STORE type
- [ ] Search Bar is auto focused on opening
- [ ] Closing the modal, closes the modal
- [ ] Filtering the list, filters by supplier name.
- [ ] Clearing the search term clears filtering
- [ ] Clicking a customer creates a new supplier requisition and navigates to a `SupplierRequisitionPage` with the newly created supplier requisition open
#### WITH PROGRAMS
*Note: Having a program: Having a master list which is a program related to the store with a store tag matching a tag in this stores tag*
### Buttons
- [ ] `"New Requisition"`
- [ ] Button is displayed
- [ ] On press, button opens `ByProgramModal`
### Modals
- [ ] **ByProgramModal**
- [ ] Must select steps in order: program, supplier, order type, period
- [ ] Can back pedal steps, where the progress is adjusted correctly
- [ ] Only order types which are defined for a program are displayed
- [ ] Only periods for the order type defined are displayed
- [ ] Order types, when being selected show the correct details (MOS etc.)
- [ ] Periods, when being selected show the correct details (number of requisitions etc.)
- [ ] Switching between general and program orders adjusts the steps correctly
- [ ] OK is enabled when all steps are complete
- [ ] closing the modal, closes the modal
- [ ] programs items are added to the requisition
| 2.0 | TESTING: Supplier Requisitions Page - # TESTING
This is an issue for testing `SupplierRequisitionsPage`.
Please copy and paste this list of tests and post your results in a comment in this issue.
#### TESTER NOTES:
- Ensure one of the EXTERNAL SUPPLIERS (`Name` record which is a facility, NOT a store, and is a supplier) has a very long name
- If testing both `SupplierRequisitionPage` and `SupplierRequisitionsPage`, will be faster to do them both without programs first and then with programs after
## Supplier Requisitions Page
This page should be tested WITH NO PROGRAMS and WITH PROGRAMS. Alterations to the behaviour of this page are listed at the BOTTOM of this list.
### NAVIGATION
- [ ] Can navigate to this page through the Menu Page, `"Supplier Requisition"` button
- [ ] Pressing either the top left back button, or hardware back button navigates to the Menu Page
- [ ] Navigating to this page from a supplier requisition shows the correctly updated details (finalise or add items to a supplier requisition, navigating back shows that supplier requisition as finalised or the increase in items)
### SYNC
- [ ] Supplier requisitions are synced to desktop
- [ ] Supplier requisitions are synced from desktop
### Table
- [ ] List all supplier requisitions for this store (matches internal orders list in desktop)
- [ ] Initial sort is by requisition number in descending order
#### Row Press
- [ ] Pressing a row (except for the `"DELETE"` column), navigates to the `SupplierRequisitionPage`
- [ ] The supplier requisition navigated to is the row that was pressed
#### Filtering
- [ ] Supplier requisitions are filtered by requisition number when a search term is entered
- [ ] Filtering is cleared when the search term is cleared [by backspacing or pressing the X].
#### Columns
- [ ] **REQUISITION NUMBER**
- [ ] Sortable (and sorts correctly)
- [ ] New requisitions are correctly incremented (customer and supplier requisitions share numbers)
- [ ] Sorting icon updates correctly to indicate ascending/descending and no sorting
- [ ] Cells are NOT editable
- [ ] Left aligned
- [ ] **SUPPLIER**
- [ ] Sortable (and sorts correctly)
- [ ] Supplier names are correctly listed for every supplier requisition
- [ ] Sorting icon updates correctly to indicate ascending/descending and no sorting.
- [ ] Cells NOT editable
- [ ] Text overflow handled (create a Supplier name that is very long)
- [ ] Left aligned
- [ ] **PROGRAM** (NOTE: only displays if at least one requisition is associated with a program)
- [ ] Does not display if no requisitions are associated with a program (includes current AND past requisitions)
- [ ] Displays if at least one requisition is associated with a program (includes current AND past requisitions)
- [ ] Sortable (and sorts correctly)
- [ ] Programs are listed correctly for every customer requisition (`"N/A"` if requisition has no program)
- [ ] Sorting icon updates correctly to indicate ascending/descending and no sorting
- [ ] Cells are NOT editable
- [ ] Text overflow handled (create a program name that is very long)
- [ ] Left aligned
- [ ] **ITEMS**
- [ ] Sortable (and sorts correctly)
- [ ] Shows the correct number of items in the supplier requisition
- [ ] Sorting icon updates correctly to indicate ascending/descending and no sorting.
- [ ] Right aligned
- [ ] Number of items is correctly updated (when navigating back from a supplier requisition)
- [ ] **ENTERED DATE**
- [ ] Sortable (and sorts correctly): this might be difficult if the data file is new or there are no old records, so clicking the sort button and having the icon change is sufficient
- [ ] Shows the correct date an invoice was created
- [ ] Dates are in the form: `WeekDay Month Year`, e.g. `17 May 2008`
- [ ] Sorting icon updates correctly to indicate ascending/descending and no sorting
- [ ] Left aligned
- [ ] Cells are NOT editable
- [ ] **STATUS**
- [ ] NOT sortable
- [ ] Shows the correct status for an invoice
- [ ] Status is one of: `"In Progress"` or `"Finalised"` (or localised equivalents)
- [ ] Left aligned
- [ ] Cells are NOT editable
- [ ] **Delete**
- [ ] NOT a sortable column
- [ ] Cells NOT editable
- [ ] Finalised invoices are not selectable
- [ ] Finalised rows indicate they are not selectable
- [ ] Rows not finalised are selectable
- [ ] Multiple rows are selectable
- [ ] Selecting a row opens a confirmation dialog at the bottom of the page
- [ ] Unselecting all rows closes the confirmation dialog at the bottom of the page
- [ ] Confirming the confirmation dialog deletes the supplier requisitions
- [ ] Cancelling the confirmation dialog removes the current selection and closes the confirmation dialog
- [ ] Finalised invoices are not selectable
### Buttons
- [ ] `"New Requisition"`
- [ ] Button is displayed.
- [ ] On press, button opens `"Select a supplier"` modal
### Modals
- [ ] **New Requisition Modal**
- [ ] Title is `"Select a supplier"`
- [ ] Options to select are names records that are: VISIBLE, a SUPPLIER, not THIS store and a STORE type
- [ ] Search Bar is auto focused on opening
- [ ] Closing the modal, closes the modal
- [ ] Filtering the list, filters by supplier name.
- [ ] Clearing the search term clears filtering
- [ ] Clicking a customer creates a new supplier requisition and navigates to a `SupplierRequisitionPage` with the newly created supplier requisition open
#### WITH PROGRAMS
*Note: Having a program: Having a master list which is a program related to the store with a store tag matching a tag in this stores tag*
### Buttons
- [ ] `"New Requisition"`
- [ ] Button is displayed
- [ ] On press, button opens `ByProgramModal`
### Modals
- [ ] **ByProgramModal**
- [ ] Must select steps in order: program, supplier, order type, period
- [ ] Can back pedal steps, where the progress is adjusted correctly
- [ ] Only order types which are defined for a program are displayed
- [ ] Only periods for the order type defined are displayed
- [ ] Order types, when being selected show the correct details (MOS etc.)
- [ ] Periods, when being selected show the correct details (number of requisitions etc.)
- [ ] Switching between general and program orders adjusts the steps correctly
- [ ] OK is enabled when all steps are complete
- [ ] closing the modal, closes the modal
- [ ] programs items are added to the requisition
| test | testing supplier requisitions page testing this is an issue for testing supplierrequisitionspage please copy and paste this list of tests and post your results in a comment in this issue tester notes ensure one of the external suppliers name record which is a facility not a store and is a supplier has a very long name if testing both supplierrequisitionpage and supplierrequisitionspage will be faster to do them both without programs first and then with programs after supplier requisitions page this page should be tested with no programs and with programs alterations to the behaviour of this page are listed at the bottom of this list navigation can navigate to this page through the menu page supplier requisition button pressing either the top left back button or hardware back button navigates to the menu page navigating to this page from a supplier requisition shows the correctly updated details finalise or add items to a supplier requisition navigating back shows that supplier requisition as finalised or the increase in items sync supplier requisitions are synced to desktop supplier requisitions are synced from desktop table list all supplier requisitions for this store matches internal orders list in desktop initial sort is by requisition number in descending order row press pressing a row except for the delete column navigates to the supplierrequisitionpage the supplier requisition navigated to is the row that was pressed filtering supplier requisitions are filtered by requisition number when a search term is entered filtering is cleared when the search term is cleared columns requisition number sortable and sorts correctly new requisitions are correctly incremented customer and supplier requisitions share numbers sorting icon updates correctly to indicate ascending descending and no sorting cells are not editable left aligned supplier sortable and sorts correctly supplier names are correctly listed for every supplier requisition sorting icon updates correctly to indicate ascending descending and no sorting cells not editable text overflow handled create a supplier name that is very long left aligned program note only displays if at least one requisition is associated with a program does not display if no requisitions are associated with a program includes current and past requisitions displays if at least one requisition is associated with a program includes current and past requisitions sortable and sorts correctly programs are listed correctly for every customer requisition n a if requisition has no program sorting icon updates correctly to indicate ascending descending and no sorting cells are not editable text overflow handled create a program name that is very long left aligned items sortable and sorts correctly shows the correct number of items in the supplier requisition sorting icon updates correctly to indicate ascending descending and no sorting right aligned number of items is correctly updated when navigating back from a supplier requisition entered date sortable and sorts correctly this might be difficult if the data file is new or there are no old records so clicking the sort button and having the icon change is sufficient shows the correct date an invoice was created dates are in the form weekday month year e g may sorting icon updates correctly to indicate ascending descending and no sorting left aligned cells are not editable status not sortable shows the correct status for an invoice status is one of in progress or finalised or localised equivalents left aligned cells are not editable delete not a sortable column cells not editable finalised invoices are not selectable finalised rows indicate they are not selectable rows not finalised are selectable multiple rows are selectable selecting a row opens a confirmation dialog at the bottom of the page unselecting all rows closes the confirmation dialog at the bottom of the page confirming the confirmation dialog deletes the supplier requisitions cancelling the confirmation dialog removes the current selection and closes the confirmation dialog finalised invoices are not selectable buttons new requisition button is displayed on press button opens select a supplier modal modals new requisition modal title is select a supplier options to select are names records that are visible a supplier not this store and a store type search bar is auto focused on opening closing the modal closes the modal filtering the list filters by supplier name clearing the search term clears filtering clicking a customer creates a new supplier requisition and navigates to a supplierrequisitionpage with the newly created supplier requisition open with programs note having a program having a master list which is a program related to the store with a store tag matching a tag in this stores tag buttons new requisition button is displayed on press button opens byprogrammodal modals byprogrammodal must select steps in order program supplier order type period can back pedal steps where the progress is adjusted correctly only order types which are defined for a program are displayed only periods for the order type defined are displayed order types when being selected show the correct details mos etc periods when being selected show the correct details number of requisitions etc switching between general and program orders adjusts the steps correctly ok is enabled when all steps are complete closing the modal closes the modal programs items are added to the requisition | 1 |
89,902 | 11,301,627,090 | IssuesEvent | 2020-01-17 15:57:41 | YardSale2020/yard-sale-web-app | https://api.github.com/repos/YardSale2020/yard-sale-web-app | closed | Create Paper Prototype UX | design | Create a paper prototype UX for the home page - @Miriampion
- [x] loading / home screen
- [x] location / maximum distance selection
- [x] map screen | 1.0 | Create Paper Prototype UX - Create a paper prototype UX for the home page - @Miriampion
- [x] loading / home screen
- [x] location / maximum distance selection
- [x] map screen | non_test | create paper prototype ux create a paper prototype ux for the home page miriampion loading home screen location maximum distance selection map screen | 0 |
180,097 | 13,920,004,174 | IssuesEvent | 2020-10-21 09:50:02 | AnssiR66/AlanStdLib | https://api.github.com/repos/AnssiR66/AlanStdLib | closed | Test Suite Scripts: Delete *.log Files at Each Execution | :star: test suite :warning: USEFUL | In the test suite scripts:
- [ ] before each execution of the tests, delete all `*.log` files.
This ensure that the local folder is kept free from left-over clutter created by failed compilations during sources editing.
The `.log` files generated by the ALAN compiler are ignored by Git but, when things go _horribly wrong_ while editing adventure sources, hundreds of these temporary logs file can easily pile up locally. Deleting them is the best option.
----
**NOTE** — The main script `RUNTESTS.bat` already does delete all `*.log` files; it's only the individual folder-scripts that don't!
| 1.0 | Test Suite Scripts: Delete *.log Files at Each Execution - In the test suite scripts:
- [ ] before each execution of the tests, delete all `*.log` files.
This ensure that the local folder is kept free from left-over clutter created by failed compilations during sources editing.
The `.log` files generated by the ALAN compiler are ignored by Git but, when things go _horribly wrong_ while editing adventure sources, hundreds of these temporary logs file can easily pile up locally. Deleting them is the best option.
----
**NOTE** — The main script `RUNTESTS.bat` already does delete all `*.log` files; it's only the individual folder-scripts that don't!
| test | test suite scripts delete log files at each execution in the test suite scripts before each execution of the tests delete all log files this ensure that the local folder is kept free from left over clutter created by failed compilations during sources editing the log files generated by the alan compiler are ignored by git but when things go horribly wrong while editing adventure sources hundreds of these temporary logs file can easily pile up locally deleting them is the best option note — the main script runtests bat already does delete all log files it s only the individual folder scripts that don t | 1 |
761,241 | 26,672,730,563 | IssuesEvent | 2023-01-26 11:46:39 | frequenz-floss/frequenz-sdk-python | https://api.github.com/repos/frequenz-floss/frequenz-sdk-python | opened | There is a lack of clear standards and documentation on actor development | priority:❓ type:bug | ### What happened?
I wanted to add nox checks to my FCR actor and I combined the nox file from peak shaving actor and isort feature from EVC actor. And this failed because the isort rule is inconsistent with the black rules.
- black wants:
```
from frequenz.actor.fcr.charge_discharge_management import (
get_charge_discharge_management,
)
```
- isort wants:
```
from frequenz.actor.fcr.charge_discharge_management import \
get_charge_discharge_management
```
### What did you expect instead?
Instead I would expect there is a well-documented standard for actors including the nox configuration that for any actor can be just copied and trusted and it will work straight away.
### Affected version(s)
All versions
### Affected part(s)
Documentation (part:docs), Build script, CI, dependencies, etc. (part:tooling)
### Extra information
We have to automate this and it should be part of every release. Otherwise we cannot expect that the actors would follow good standard if any of the following is true:
- the standards or documentation on them does not exist,
- there is no mechanism or process to keep the standards and their documentation updated,
- the standards work only partially or for some actors only.
I think the problem is **having good standards** seems not a requirement, but this is in contradiction with requiring that **actors follow good standards**. | 1.0 | There is a lack of clear standards and documentation on actor development - ### What happened?
I wanted to add nox checks to my FCR actor and I combined the nox file from peak shaving actor and isort feature from EVC actor. And this failed because the isort rule is inconsistent with the black rules.
- black wants:
```
from frequenz.actor.fcr.charge_discharge_management import (
get_charge_discharge_management,
)
```
- isort wants:
```
from frequenz.actor.fcr.charge_discharge_management import \
get_charge_discharge_management
```
### What did you expect instead?
Instead I would expect there is a well-documented standard for actors including the nox configuration that for any actor can be just copied and trusted and it will work straight away.
### Affected version(s)
All versions
### Affected part(s)
Documentation (part:docs), Build script, CI, dependencies, etc. (part:tooling)
### Extra information
We have to automate this and it should be part of every release. Otherwise we cannot expect that the actors would follow good standard if any of the following is true:
- the standards or documentation on them does not exist,
- there is no mechanism or process to keep the standards and their documentation updated,
- the standards work only partially or for some actors only.
I think the problem is **having good standards** seems not a requirement, but this is in contradiction with requiring that **actors follow good standards**. | non_test | there is a lack of clear standards and documentation on actor development what happened i wanted to add nox checks to my fcr actor and i combined the nox file from peak shaving actor and isort feature from evc actor and this failed because the isort rule is inconsistent with the black rules black wants from frequenz actor fcr charge discharge management import get charge discharge management isort wants from frequenz actor fcr charge discharge management import get charge discharge management what did you expect instead instead i would expect there is a well documented standard for actors including the nox configuration that for any actor can be just copied and trusted and it will work straight away affected version s all versions affected part s documentation part docs build script ci dependencies etc part tooling extra information we have to automate this and it should be part of every release otherwise we cannot expect that the actors would follow good standard if any of the following is true the standards or documentation on them does not exist there is no mechanism or process to keep the standards and their documentation updated the standards work only partially or for some actors only i think the problem is having good standards seems not a requirement but this is in contradiction with requiring that actors follow good standards | 0 |
308,213 | 26,588,600,843 | IssuesEvent | 2023-01-23 05:47:30 | rizinorg/rz-ghidra | https://api.github.com/repos/rizinorg/rz-ghidra | closed | Strings Truncated in Decompiler Output | bug test-attached | **Environment information**
* Operating System: Fedora 31 (x86_64)
* Cutter version: Appimage 1.10
* Obtained from:
- [ ] Built from source
- [x] Downloaded from release
- [ ] Distribution repository
* File format: EXE
**Describe the bug**
It appears that some strings are truncated in the decompiled output
The asm:

The decompiled output (which is showing only the first letter of the lpName)

**To Reproduce**
Open the following file and open in Cutter. (password: infected)
[mal.zip](https://github.com/rizinorg/cutter/files/4028768/mal.zip)
Call to CreateMutex is at `0x1400018b8`
**Expected behavior**
Cutter to behave same as Ghidra

I searched through back issues and didn't see this one listed, apologies if this is already known. | 1.0 | Strings Truncated in Decompiler Output - **Environment information**
* Operating System: Fedora 31 (x86_64)
* Cutter version: Appimage 1.10
* Obtained from:
- [ ] Built from source
- [x] Downloaded from release
- [ ] Distribution repository
* File format: EXE
**Describe the bug**
It appears that some strings are truncated in the decompiled output
The asm:

The decompiled output (which is showing only the first letter of the lpName)

**To Reproduce**
Open the following file and open in Cutter. (password: infected)
[mal.zip](https://github.com/rizinorg/cutter/files/4028768/mal.zip)
Call to CreateMutex is at `0x1400018b8`
**Expected behavior**
Cutter to behave same as Ghidra

I searched through back issues and didn't see this one listed, apologies if this is already known. | test | strings truncated in decompiler output environment information operating system fedora cutter version appimage obtained from built from source downloaded from release distribution repository file format exe describe the bug it appears that some strings are truncated in the decompiled output the asm the decompiled output which is showing only the first letter of the lpname to reproduce open the following file and open in cutter password infected call to createmutex is at expected behavior cutter to behave same as ghidra i searched through back issues and didn t see this one listed apologies if this is already known | 1 |
40,539 | 12,799,560,275 | IssuesEvent | 2020-07-02 15:35:04 | TreyM-WSS/WhiteSource-Demo | https://api.github.com/repos/TreyM-WSS/WhiteSource-Demo | opened | CVE-2017-7525 (High) detected in jackson-databind-2.8.1.jar | security vulnerability | ## CVE-2017-7525 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-scm/WhiteSource-Demo/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.4.0.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/TreyM-WSS/WhiteSource-Demo/commits/75659f691fb82d67ecd666ba6076394defeb92d0">75659f691fb82d67ecd666ba6076394defeb92d0</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A deserialization flaw was discovered in the jackson-databind, versions before 2.6.7.1, 2.7.9.1 and 2.8.9, which could allow an unauthenticated user to perform code execution by sending the maliciously crafted input to the readValue method of the ObjectMapper.
<p>Publish Date: 2018-02-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-7525>CVE-2017-7525</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-7525">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-7525</a></p>
<p>Release Date: 2018-02-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.1,2.7.9.1,2.8.9</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.1","isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.4.0.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.8.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.6.7.1,2.7.9.1,2.8.9"}],"vulnerabilityIdentifier":"CVE-2017-7525","vulnerabilityDetails":"A deserialization flaw was discovered in the jackson-databind, versions before 2.6.7.1, 2.7.9.1 and 2.8.9, which could allow an unauthenticated user to perform code execution by sending the maliciously crafted input to the readValue method of the ObjectMapper.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-7525","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2017-7525 (High) detected in jackson-databind-2.8.1.jar - ## CVE-2017-7525 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-scm/WhiteSource-Demo/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.4.0.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/TreyM-WSS/WhiteSource-Demo/commits/75659f691fb82d67ecd666ba6076394defeb92d0">75659f691fb82d67ecd666ba6076394defeb92d0</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A deserialization flaw was discovered in the jackson-databind, versions before 2.6.7.1, 2.7.9.1 and 2.8.9, which could allow an unauthenticated user to perform code execution by sending the maliciously crafted input to the readValue method of the ObjectMapper.
<p>Publish Date: 2018-02-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-7525>CVE-2017-7525</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-7525">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-7525</a></p>
<p>Release Date: 2018-02-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.1,2.7.9.1,2.8.9</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.1","isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.4.0.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.8.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.6.7.1,2.7.9.1,2.8.9"}],"vulnerabilityIdentifier":"CVE-2017-7525","vulnerabilityDetails":"A deserialization flaw was discovered in the jackson-databind, versions before 2.6.7.1, 2.7.9.1 and 2.8.9, which could allow an unauthenticated user to perform code execution by sending the maliciously crafted input to the readValue method of the ObjectMapper.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-7525","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_test | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws scm whitesource demo pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library x jackson databind jar vulnerable library found in head commit a href vulnerability details a deserialization flaw was discovered in the jackson databind versions before and which could allow an unauthenticated user to perform code execution by sending the maliciously crafted input to the readvalue method of the objectmapper publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails a deserialization flaw was discovered in the jackson databind versions before and which could allow an unauthenticated user to perform code execution by sending the maliciously crafted input to the readvalue method of the objectmapper vulnerabilityurl | 0 |
818,638 | 30,697,497,073 | IssuesEvent | 2023-07-26 19:51:01 | sequencefilm/docs-and-issues | https://api.github.com/repos/sequencefilm/docs-and-issues | closed | [SEQ-1857] Timeline rename gets funky behavior after deleting a timeline | bug High priority | Seems like some keys get screwed up and the index can become off by one, resulting in a context menu action getting applied to the wrong timeline
* rename also not properly working
<sub>From [SyncLinear.com](https://synclinear.com) | [SEQ-1857](https://linear.app/sequencefilm/issue/SEQ-1857/timeline-rename-gets-funky-behavior-after-deleting-a-timeline)</sub> | 1.0 | [SEQ-1857] Timeline rename gets funky behavior after deleting a timeline - Seems like some keys get screwed up and the index can become off by one, resulting in a context menu action getting applied to the wrong timeline
* rename also not properly working
<sub>From [SyncLinear.com](https://synclinear.com) | [SEQ-1857](https://linear.app/sequencefilm/issue/SEQ-1857/timeline-rename-gets-funky-behavior-after-deleting-a-timeline)</sub> | non_test | timeline rename gets funky behavior after deleting a timeline seems like some keys get screwed up and the index can become off by one resulting in a context menu action getting applied to the wrong timeline rename also not properly working from | 0 |
52,562 | 10,881,583,048 | IssuesEvent | 2019-11-17 18:44:55 | sourcegraph/sourcegraph | https://api.github.com/repos/sourcegraph/sourcegraph | opened | Add way to quickly/easily identify when LSIF is working in the browser | code-intel | Follow up from https://github.com/sourcegraph/sourcegraph/pull/6642#discussion_r347016777
From Felix:
> Maybe it would be easier if the LSIF handler in the extension logged a statement "Using LSIF code intelligence" on init
| 1.0 | Add way to quickly/easily identify when LSIF is working in the browser - Follow up from https://github.com/sourcegraph/sourcegraph/pull/6642#discussion_r347016777
From Felix:
> Maybe it would be easier if the LSIF handler in the extension logged a statement "Using LSIF code intelligence" on init
| non_test | add way to quickly easily identify when lsif is working in the browser follow up from from felix maybe it would be easier if the lsif handler in the extension logged a statement using lsif code intelligence on init | 0 |
248,188 | 21,001,774,550 | IssuesEvent | 2022-03-29 18:11:05 | TriBITSPub/TriBITS | https://api.github.com/repos/TriBITSPub/TriBITS | closed | Set up for automerge of TriBITS PR | type: enhancement component: testing | GitHub now supports automerging of PRs as described [here](https://docs.github.com/en/pull-requests/collaborating-with-pull-requests/incorporating-changes-from-a-pull-request/automatically-merging-a-pull-request).
To require that all of the tests pass and that reviews are done, you need to:
* Enable auto-merge
* Add GitHub Actions tests (or other tool)
* Require all changes to the repo through PRs (so you can require a review)
For the final details, see [below](https://github.com/TriBITSPub/TriBITS/issues/473#issuecomment-1082192597).
| 1.0 | Set up for automerge of TriBITS PR - GitHub now supports automerging of PRs as described [here](https://docs.github.com/en/pull-requests/collaborating-with-pull-requests/incorporating-changes-from-a-pull-request/automatically-merging-a-pull-request).
To require that all of the tests pass and that reviews are done, you need to:
* Enable auto-merge
* Add GitHub Actions tests (or other tool)
* Require all changes to the repo through PRs (so you can require a review)
For the final details, see [below](https://github.com/TriBITSPub/TriBITS/issues/473#issuecomment-1082192597).
| test | set up for automerge of tribits pr github now supports automerging of prs as described to require that all of the tests pass and that reviews are done you need to enable auto merge add github actions tests or other tool require all changes to the repo through prs so you can require a review for the final details see | 1 |
236,432 | 19,541,260,713 | IssuesEvent | 2022-01-01 00:17:37 | kubernetes/minikube | https://api.github.com/repos/kubernetes/minikube | closed | Frequent test failures of `TestFunctional/parallel/MountCmd/specific-port` | priority/backlog kind/failing-test | This test has high flake rates for the following environments:
|Environment|Flake Rate (%)|
|---|---|
|[KVM_Linux_containerd](https://storage.googleapis.com/minikube-flake-rate/flake_chart.html?env=KVM_Linux_containerd&test=TestFunctional/parallel/MountCmd/specific-port)|20.00| | 1.0 | Frequent test failures of `TestFunctional/parallel/MountCmd/specific-port` - This test has high flake rates for the following environments:
|Environment|Flake Rate (%)|
|---|---|
|[KVM_Linux_containerd](https://storage.googleapis.com/minikube-flake-rate/flake_chart.html?env=KVM_Linux_containerd&test=TestFunctional/parallel/MountCmd/specific-port)|20.00| | test | frequent test failures of testfunctional parallel mountcmd specific port this test has high flake rates for the following environments environment flake rate | 1 |
81,403 | 3,590,536,314 | IssuesEvent | 2016-02-01 06:41:24 | jamesmontemagno/Xamarin.Plugins | https://api.github.com/repos/jamesmontemagno/Xamarin.Plugins | closed | CameraDevice.Front doesn't seem to work on Android | bug Media priority-low | I have specified CameraDevice.Front in the options, which works in iOS. However, it seems that it is not acknowledged in Android. Am I missing a step, or is this functionality not yet implemented on Android? | 1.0 | CameraDevice.Front doesn't seem to work on Android - I have specified CameraDevice.Front in the options, which works in iOS. However, it seems that it is not acknowledged in Android. Am I missing a step, or is this functionality not yet implemented on Android? | non_test | cameradevice front doesn t seem to work on android i have specified cameradevice front in the options which works in ios however it seems that it is not acknowledged in android am i missing a step or is this functionality not yet implemented on android | 0 |
450,817 | 31,994,079,569 | IssuesEvent | 2023-09-21 08:06:23 | cubefs/compass | https://api.github.com/repos/cubefs/compass | closed | Failed to run `com.oppo.cloud.parser.spark.eventlog.ReplayEventLogsTest` | documentation enhancement | Since we set the mysql dependency to test in https://github.com/cubefs/compass/commit/06109bc259dfc39026123ab63ad31fcc54109e46, the mysql dependency will not be passed to subsequent modules.
Error:
```
2023-09-20 17:25:50,270 ERROR 14036 [main] [] : [com.alibaba.druid.pool.DruidDataSource:960] {dataSource-1} init error
java.sql.SQLException: com.mysql.jdbc.Driver
at com.alibaba.druid.util.JdbcUtils.createDriver(JdbcUtils.java:688)
at com.alibaba.druid.pool.DruidDataSource.resolveDriver(DruidDataSource.java:1248)
at com.alibaba.druid.pool.DruidDataSource.init(DruidDataSource.java:876)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.invokeCustomInitMethod(AbstractAutowireCapableBeanFactory.java:1930)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.invokeInitMethods(AbstractAutowireCapableBeanFactory.java:1872)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.initializeBean(AbstractAutowireCapableBeanFactory.java:1800)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:620)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:542)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:335)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:234)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:333)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:208)
at org.springframework.beans.factory.config.DependencyDescriptor.resolveCandidate(DependencyDescriptor.java:276)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.addCandidateEntry(DefaultListableBeanFactory.java:1609)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.findAutowireCandidates(DefaultListableBeanFactory.java:1573)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.resolveMultipleBeans(DefaultListableBeanFactory.java:1492)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.doResolveDependency(DefaultListableBeanFactory.java:1349)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.resolveDependency(DefaultListableBeanFactory.java:1311)
at org.springframework.beans.factory.support.ConstructorResolver.resolveAutowiredArgument(ConstructorResolver.java:887)
at org.springframework.beans.factory.support.ConstructorResolver.createArgumentArray(ConstructorResolver.java:791)
at org.springframework.beans.factory.support.ConstructorResolver.instantiateUsingFactoryMethod(ConstructorResolver.java:541)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.instantiateUsingFactoryMethod(AbstractAutowireCapableBeanFactory.java:1352)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:1195)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:582)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:542)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:335)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:234)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:333)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:208)
at org.springframework.beans.factory.config.DependencyDescriptor.resolveCandidate(DependencyDescriptor.java:276)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.addCandidateEntry(DefaultListableBeanFactory.java:1616)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.findAutowireCandidates(DefaultListableBeanFactory.java:1573)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.resolveMultipleBeans(DefaultListableBeanFactory.java:1417)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.doResolveDependency(DefaultListableBeanFactory.java:1349)
at org.springframework.beans.factory.support.DefaultListableBeanFactory$DependencyObjectProvider.resolveStream(DefaultListableBeanFactory.java:2119)
at org.springframework.beans.factory.support.DefaultListableBeanFactory$DependencyObjectProvider.orderedStream(DefaultListableBeanFactory.java:2113)
at org.springframework.boot.actuate.autoconfigure.metrics.MeterRegistryConfigurer.addBinders(MeterRegistryConfigurer.java:86)
at org.springframework.boot.actuate.autoconfigure.metrics.MeterRegistryConfigurer.configure(MeterRegistryConfigurer.java:68)
at org.springframework.boot.actuate.autoconfigure.metrics.MeterRegistryPostProcessor.postProcessAfterInitialization(MeterRegistryPostProcessor.java:64)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.applyBeanPostProcessorsAfterInitialization(AbstractAutowireCapableBeanFactory.java:455)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.initializeBean(AbstractAutowireCapableBeanFactory.java:1808)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:620)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:542)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:335)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:234)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:333)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:208)
at org.springframework.beans.factory.config.DependencyDescriptor.resolveCandidate(DependencyDescriptor.java:276)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.doResolveDependency(DefaultListableBeanFactory.java:1391)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.resolveDependency(DefaultListableBeanFactory.java:1311)
at org.springframework.beans.factory.support.ConstructorResolver.resolveAutowiredArgument(ConstructorResolver.java:887)
at org.springframework.beans.factory.support.ConstructorResolver.createArgumentArray(ConstructorResolver.java:791)
at org.springframework.beans.factory.support.ConstructorResolver.instantiateUsingFactoryMethod(ConstructorResolver.java:541)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.instantiateUsingFactoryMethod(AbstractAutowireCapableBeanFactory.java:1352)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:1195)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:582)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:542)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:335)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:234)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:333)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:208)
at org.springframework.beans.factory.config.DependencyDescriptor.resolveCandidate(DependencyDescriptor.java:276)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.addCandidateEntry(DefaultListableBeanFactory.java:1616)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.findAutowireCandidates(DefaultListableBeanFactory.java:1573)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.resolveMultipleBeans(DefaultListableBeanFactory.java:1417)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.doResolveDependency(DefaultListableBeanFactory.java:1349)
at org.springframework.beans.factory.support.DefaultListableBeanFactory$DependencyObjectProvider.resolveStream(DefaultListableBeanFactory.java:2119)
at org.springframework.beans.factory.support.DefaultListableBeanFactory$DependencyObjectProvider.orderedStream(DefaultListableBeanFactory.java:2113)
at org.springframework.boot.autoconfigure.data.redis.LettuceConnectionConfiguration.lettuceClientResources(LettuceConnectionConfiguration.java:72)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at org.springframework.beans.factory.support.SimpleInstantiationStrategy.instantiate(SimpleInstantiationStrategy.java:154)
at org.springframework.beans.factory.support.ConstructorResolver.instantiate(ConstructorResolver.java:653)
at org.springframework.beans.factory.support.ConstructorResolver.instantiateUsingFactoryMethod(ConstructorResolver.java:638)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.instantiateUsingFactoryMethod(AbstractAutowireCapableBeanFactory.java:1352)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:1195)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:582)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:542)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:335)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:234)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:333)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:208)
at org.springframework.beans.factory.config.DependencyDescriptor.resolveCandidate(DependencyDescriptor.java:276)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.doResolveDependency(DefaultListableBeanFactory.java:1391)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.resolveDependency(DefaultListableBeanFactory.java:1311)
at org.springframework.beans.factory.support.ConstructorResolver.resolveAutowiredArgument(ConstructorResolver.java:887)
at org.springframework.beans.factory.support.ConstructorResolver.createArgumentArray(ConstructorResolver.java:791)
at org.springframework.beans.factory.support.ConstructorResolver.instantiateUsingFactoryMethod(ConstructorResolver.java:541)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.instantiateUsingFactoryMethod(AbstractAutowireCapableBeanFactory.java:1352)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:1195)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:582)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:542)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:335)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:234)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:333)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:208)
at org.springframework.beans.factory.config.DependencyDescriptor.resolveCandidate(DependencyDescriptor.java:276)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.doResolveDependency(DefaultListableBeanFactory.java:1391)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.resolveDependency(DefaultListableBeanFactory.java:1311)
at org.springframework.beans.factory.support.ConstructorResolver.resolveAutowiredArgument(ConstructorResolver.java:887)
at org.springframework.beans.factory.support.ConstructorResolver.createArgumentArray(ConstructorResolver.java:791)
at org.springframework.beans.factory.support.ConstructorResolver.instantiateUsingFactoryMethod(ConstructorResolver.java:541)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.instantiateUsingFactoryMethod(AbstractAutowireCapableBeanFactory.java:1352)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:1195)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:582)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:542)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:335)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:234)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:333)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:208)
at org.springframework.beans.factory.config.DependencyDescriptor.resolveCandidate(DependencyDescriptor.java:276)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.doResolveDependency(DefaultListableBeanFactory.java:1391)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.resolveDependency(DefaultListableBeanFactory.java:1311)
at org.springframework.beans.factory.annotation.AutowiredAnnotationBeanPostProcessor$AutowiredFieldElement.resolveFieldValue(AutowiredAnnotationBeanPostProcessor.java:657)
at org.springframework.beans.factory.annotation.AutowiredAnnotationBeanPostProcessor$AutowiredFieldElement.inject(AutowiredAnnotationBeanPostProcessor.java:640)
at org.springframework.beans.factory.annotation.InjectionMetadata.inject(InjectionMetadata.java:119)
at org.springframework.beans.factory.annotation.AutowiredAnnotationBeanPostProcessor.postProcessProperties(AutowiredAnnotationBeanPostProcessor.java:399)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.populateBean(AbstractAutowireCapableBeanFactory.java:1431)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:619)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:542)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:335)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:234)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:333)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:213)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.resolveBeanByName(AbstractAutowireCapableBeanFactory.java:479)
at org.springframework.context.annotation.CommonAnnotationBeanPostProcessor.autowireResource(CommonAnnotationBeanPostProcessor.java:550)
at org.springframework.context.annotation.CommonAnnotationBeanPostProcessor.getResource(CommonAnnotationBeanPostProcessor.java:520)
at org.springframework.context.annotation.CommonAnnotationBeanPostProcessor$ResourceElement.getResourceToInject(CommonAnnotationBeanPostProcessor.java:673)
at org.springframework.beans.factory.annotation.InjectionMetadata$InjectedElement.inject(InjectionMetadata.java:228)
at org.springframework.beans.factory.annotation.InjectionMetadata.inject(InjectionMetadata.java:119)
at org.springframework.context.annotation.CommonAnnotationBeanPostProcessor.postProcessProperties(CommonAnnotationBeanPostProcessor.java:329)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.populateBean(AbstractAutowireCapableBeanFactory.java:1431)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:619)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:542)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:335)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:234)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:333)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:208)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.preInstantiateSingletons(DefaultListableBeanFactory.java:955)
at org.springframework.context.support.AbstractApplicationContext.finishBeanFactoryInitialization(AbstractApplicationContext.java:918)
at org.springframework.context.support.AbstractApplicationContext.refresh(AbstractApplicationContext.java:583)
at org.springframework.boot.SpringApplication.refresh(SpringApplication.java:731)
at org.springframework.boot.SpringApplication.refreshContext(SpringApplication.java:408)
at org.springframework.boot.SpringApplication.run(SpringApplication.java:307)
at org.springframework.boot.test.context.SpringBootContextLoader.loadContext(SpringBootContextLoader.java:136)
at org.springframework.test.context.cache.DefaultCacheAwareContextLoaderDelegate.loadContextInternal(DefaultCacheAwareContextLoaderDelegate.java:141)
at org.springframework.test.context.cache.DefaultCacheAwareContextLoaderDelegate.loadContext(DefaultCacheAwareContextLoaderDelegate.java:90)
at org.springframework.test.context.support.DefaultTestContext.getApplicationContext(DefaultTestContext.java:124)
at org.springframework.test.context.web.ServletTestExecutionListener.setUpRequestContextIfNecessary(ServletTestExecutionListener.java:190)
at org.springframework.test.context.web.ServletTestExecutionListener.prepareTestInstance(ServletTestExecutionListener.java:132)
at org.springframework.test.context.TestContextManager.prepareTestInstance(TestContextManager.java:248)
at org.springframework.test.context.junit.jupiter.SpringExtension.postProcessTestInstance(SpringExtension.java:138)
at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.lambda$invokeTestInstancePostProcessors$8(ClassBasedTestDescriptor.java:363)
at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.executeAndMaskThrowable(ClassBasedTestDescriptor.java:368)
at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.lambda$invokeTestInstancePostProcessors$9(ClassBasedTestDescriptor.java:363)
at java.base/java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:195)
at java.base/java.util.stream.ReferencePipeline$2$1.accept(ReferencePipeline.java:177)
at java.base/java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1655)
at java.base/java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:484)
at java.base/java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:474)
at java.base/java.util.stream.StreamSpliterators$WrappingSpliterator.forEachRemaining(StreamSpliterators.java:312)
at java.base/java.util.stream.Streams$ConcatSpliterator.forEachRemaining(Streams.java:735)
at java.base/java.util.stream.Streams$ConcatSpliterator.forEachRemaining(Streams.java:734)
at java.base/java.util.stream.ReferencePipeline$Head.forEach(ReferencePipeline.java:658)
at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.invokeTestInstancePostProcessors(ClassBasedTestDescriptor.java:362)
at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.lambda$instantiateAndPostProcessTestInstance$6(ClassBasedTestDescriptor.java:283)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.instantiateAndPostProcessTestInstance(ClassBasedTestDescriptor.java:282)
at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.lambda$testInstancesProvider$4(ClassBasedTestDescriptor.java:272)
at java.base/java.util.Optional.orElseGet(Optional.java:369)
at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.lambda$testInstancesProvider$5(ClassBasedTestDescriptor.java:271)
at org.junit.jupiter.engine.execution.TestInstancesProvider.getTestInstances(TestInstancesProvider.java:31)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$prepare$0(TestMethodTestDescriptor.java:102)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.prepare(TestMethodTestDescriptor.java:101)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.prepare(TestMethodTestDescriptor.java:66)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$prepare$2(NodeTestTask.java:123)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.prepare(NodeTestTask.java:123)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:90)
at java.base/java.util.ArrayList.forEach(ArrayList.java:1541)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:41)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:155)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95)
at java.base/java.util.ArrayList.forEach(ArrayList.java:1541)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:41)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:155)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.submit(SameThreadHierarchicalTestExecutorService.java:35)
at org.junit.platform.engine.support.hierarchical.HierarchicalTestExecutor.execute(HierarchicalTestExecutor.java:57)
at org.junit.platform.engine.support.hierarchical.HierarchicalTestEngine.execute(HierarchicalTestEngine.java:54)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:107)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:88)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.lambda$execute$0(EngineExecutionOrchestrator.java:54)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.withInterceptedStreams(EngineExecutionOrchestrator.java:67)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:52)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:114)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:86)
at org.junit.platform.launcher.core.DefaultLauncherSession$DelegatingLauncher.execute(DefaultLauncherSession.java:86)
at org.junit.platform.launcher.core.SessionPerRequestLauncher.execute(SessionPerRequestLauncher.java:53)
at com.intellij.junit5.JUnit5IdeaTestRunner.startRunnerWithArgs(JUnit5IdeaTestRunner.java:71)
at com.intellij.rt.junit.IdeaTestRunner$Repeater$1.execute(IdeaTestRunner.java:38)
at com.intellij.rt.execution.junit.TestsRepeater.repeat(TestsRepeater.java:11)
at com.intellij.rt.junit.IdeaTestRunner$Repeater.startRunnerWithArgs(IdeaTestRunner.java:35)
at com.intellij.rt.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:235)
at com.intellij.rt.junit.JUnitStarter.main(JUnitStarter.java:54)
Caused by: java.lang.ClassNotFoundException: com.mysql.jdbc.Driver
at java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:581)
at java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:178)
at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:521)
at java.base/java.lang.Class.forName0(Native Method)
at java.base/java.lang.Class.forName(Class.java:315)
at com.alibaba.druid.util.JdbcUtils.createDriver(JdbcUtils.java:686)
... 223 common frames omitted
```
| 1.0 | Failed to run `com.oppo.cloud.parser.spark.eventlog.ReplayEventLogsTest` - Since we set the mysql dependency to test in https://github.com/cubefs/compass/commit/06109bc259dfc39026123ab63ad31fcc54109e46, the mysql dependency will not be passed to subsequent modules.
Error:
```
2023-09-20 17:25:50,270 ERROR 14036 [main] [] : [com.alibaba.druid.pool.DruidDataSource:960] {dataSource-1} init error
java.sql.SQLException: com.mysql.jdbc.Driver
at com.alibaba.druid.util.JdbcUtils.createDriver(JdbcUtils.java:688)
at com.alibaba.druid.pool.DruidDataSource.resolveDriver(DruidDataSource.java:1248)
at com.alibaba.druid.pool.DruidDataSource.init(DruidDataSource.java:876)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.invokeCustomInitMethod(AbstractAutowireCapableBeanFactory.java:1930)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.invokeInitMethods(AbstractAutowireCapableBeanFactory.java:1872)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.initializeBean(AbstractAutowireCapableBeanFactory.java:1800)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:620)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:542)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:335)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:234)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:333)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:208)
at org.springframework.beans.factory.config.DependencyDescriptor.resolveCandidate(DependencyDescriptor.java:276)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.addCandidateEntry(DefaultListableBeanFactory.java:1609)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.findAutowireCandidates(DefaultListableBeanFactory.java:1573)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.resolveMultipleBeans(DefaultListableBeanFactory.java:1492)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.doResolveDependency(DefaultListableBeanFactory.java:1349)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.resolveDependency(DefaultListableBeanFactory.java:1311)
at org.springframework.beans.factory.support.ConstructorResolver.resolveAutowiredArgument(ConstructorResolver.java:887)
at org.springframework.beans.factory.support.ConstructorResolver.createArgumentArray(ConstructorResolver.java:791)
at org.springframework.beans.factory.support.ConstructorResolver.instantiateUsingFactoryMethod(ConstructorResolver.java:541)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.instantiateUsingFactoryMethod(AbstractAutowireCapableBeanFactory.java:1352)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:1195)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:582)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:542)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:335)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:234)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:333)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:208)
at org.springframework.beans.factory.config.DependencyDescriptor.resolveCandidate(DependencyDescriptor.java:276)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.addCandidateEntry(DefaultListableBeanFactory.java:1616)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.findAutowireCandidates(DefaultListableBeanFactory.java:1573)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.resolveMultipleBeans(DefaultListableBeanFactory.java:1417)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.doResolveDependency(DefaultListableBeanFactory.java:1349)
at org.springframework.beans.factory.support.DefaultListableBeanFactory$DependencyObjectProvider.resolveStream(DefaultListableBeanFactory.java:2119)
at org.springframework.beans.factory.support.DefaultListableBeanFactory$DependencyObjectProvider.orderedStream(DefaultListableBeanFactory.java:2113)
at org.springframework.boot.actuate.autoconfigure.metrics.MeterRegistryConfigurer.addBinders(MeterRegistryConfigurer.java:86)
at org.springframework.boot.actuate.autoconfigure.metrics.MeterRegistryConfigurer.configure(MeterRegistryConfigurer.java:68)
at org.springframework.boot.actuate.autoconfigure.metrics.MeterRegistryPostProcessor.postProcessAfterInitialization(MeterRegistryPostProcessor.java:64)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.applyBeanPostProcessorsAfterInitialization(AbstractAutowireCapableBeanFactory.java:455)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.initializeBean(AbstractAutowireCapableBeanFactory.java:1808)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:620)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:542)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:335)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:234)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:333)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:208)
at org.springframework.beans.factory.config.DependencyDescriptor.resolveCandidate(DependencyDescriptor.java:276)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.doResolveDependency(DefaultListableBeanFactory.java:1391)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.resolveDependency(DefaultListableBeanFactory.java:1311)
at org.springframework.beans.factory.support.ConstructorResolver.resolveAutowiredArgument(ConstructorResolver.java:887)
at org.springframework.beans.factory.support.ConstructorResolver.createArgumentArray(ConstructorResolver.java:791)
at org.springframework.beans.factory.support.ConstructorResolver.instantiateUsingFactoryMethod(ConstructorResolver.java:541)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.instantiateUsingFactoryMethod(AbstractAutowireCapableBeanFactory.java:1352)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:1195)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:582)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:542)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:335)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:234)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:333)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:208)
at org.springframework.beans.factory.config.DependencyDescriptor.resolveCandidate(DependencyDescriptor.java:276)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.addCandidateEntry(DefaultListableBeanFactory.java:1616)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.findAutowireCandidates(DefaultListableBeanFactory.java:1573)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.resolveMultipleBeans(DefaultListableBeanFactory.java:1417)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.doResolveDependency(DefaultListableBeanFactory.java:1349)
at org.springframework.beans.factory.support.DefaultListableBeanFactory$DependencyObjectProvider.resolveStream(DefaultListableBeanFactory.java:2119)
at org.springframework.beans.factory.support.DefaultListableBeanFactory$DependencyObjectProvider.orderedStream(DefaultListableBeanFactory.java:2113)
at org.springframework.boot.autoconfigure.data.redis.LettuceConnectionConfiguration.lettuceClientResources(LettuceConnectionConfiguration.java:72)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at org.springframework.beans.factory.support.SimpleInstantiationStrategy.instantiate(SimpleInstantiationStrategy.java:154)
at org.springframework.beans.factory.support.ConstructorResolver.instantiate(ConstructorResolver.java:653)
at org.springframework.beans.factory.support.ConstructorResolver.instantiateUsingFactoryMethod(ConstructorResolver.java:638)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.instantiateUsingFactoryMethod(AbstractAutowireCapableBeanFactory.java:1352)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:1195)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:582)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:542)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:335)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:234)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:333)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:208)
at org.springframework.beans.factory.config.DependencyDescriptor.resolveCandidate(DependencyDescriptor.java:276)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.doResolveDependency(DefaultListableBeanFactory.java:1391)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.resolveDependency(DefaultListableBeanFactory.java:1311)
at org.springframework.beans.factory.support.ConstructorResolver.resolveAutowiredArgument(ConstructorResolver.java:887)
at org.springframework.beans.factory.support.ConstructorResolver.createArgumentArray(ConstructorResolver.java:791)
at org.springframework.beans.factory.support.ConstructorResolver.instantiateUsingFactoryMethod(ConstructorResolver.java:541)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.instantiateUsingFactoryMethod(AbstractAutowireCapableBeanFactory.java:1352)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:1195)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:582)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:542)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:335)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:234)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:333)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:208)
at org.springframework.beans.factory.config.DependencyDescriptor.resolveCandidate(DependencyDescriptor.java:276)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.doResolveDependency(DefaultListableBeanFactory.java:1391)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.resolveDependency(DefaultListableBeanFactory.java:1311)
at org.springframework.beans.factory.support.ConstructorResolver.resolveAutowiredArgument(ConstructorResolver.java:887)
at org.springframework.beans.factory.support.ConstructorResolver.createArgumentArray(ConstructorResolver.java:791)
at org.springframework.beans.factory.support.ConstructorResolver.instantiateUsingFactoryMethod(ConstructorResolver.java:541)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.instantiateUsingFactoryMethod(AbstractAutowireCapableBeanFactory.java:1352)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:1195)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:582)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:542)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:335)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:234)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:333)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:208)
at org.springframework.beans.factory.config.DependencyDescriptor.resolveCandidate(DependencyDescriptor.java:276)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.doResolveDependency(DefaultListableBeanFactory.java:1391)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.resolveDependency(DefaultListableBeanFactory.java:1311)
at org.springframework.beans.factory.annotation.AutowiredAnnotationBeanPostProcessor$AutowiredFieldElement.resolveFieldValue(AutowiredAnnotationBeanPostProcessor.java:657)
at org.springframework.beans.factory.annotation.AutowiredAnnotationBeanPostProcessor$AutowiredFieldElement.inject(AutowiredAnnotationBeanPostProcessor.java:640)
at org.springframework.beans.factory.annotation.InjectionMetadata.inject(InjectionMetadata.java:119)
at org.springframework.beans.factory.annotation.AutowiredAnnotationBeanPostProcessor.postProcessProperties(AutowiredAnnotationBeanPostProcessor.java:399)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.populateBean(AbstractAutowireCapableBeanFactory.java:1431)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:619)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:542)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:335)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:234)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:333)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:213)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.resolveBeanByName(AbstractAutowireCapableBeanFactory.java:479)
at org.springframework.context.annotation.CommonAnnotationBeanPostProcessor.autowireResource(CommonAnnotationBeanPostProcessor.java:550)
at org.springframework.context.annotation.CommonAnnotationBeanPostProcessor.getResource(CommonAnnotationBeanPostProcessor.java:520)
at org.springframework.context.annotation.CommonAnnotationBeanPostProcessor$ResourceElement.getResourceToInject(CommonAnnotationBeanPostProcessor.java:673)
at org.springframework.beans.factory.annotation.InjectionMetadata$InjectedElement.inject(InjectionMetadata.java:228)
at org.springframework.beans.factory.annotation.InjectionMetadata.inject(InjectionMetadata.java:119)
at org.springframework.context.annotation.CommonAnnotationBeanPostProcessor.postProcessProperties(CommonAnnotationBeanPostProcessor.java:329)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.populateBean(AbstractAutowireCapableBeanFactory.java:1431)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:619)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:542)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:335)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:234)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:333)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:208)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.preInstantiateSingletons(DefaultListableBeanFactory.java:955)
at org.springframework.context.support.AbstractApplicationContext.finishBeanFactoryInitialization(AbstractApplicationContext.java:918)
at org.springframework.context.support.AbstractApplicationContext.refresh(AbstractApplicationContext.java:583)
at org.springframework.boot.SpringApplication.refresh(SpringApplication.java:731)
at org.springframework.boot.SpringApplication.refreshContext(SpringApplication.java:408)
at org.springframework.boot.SpringApplication.run(SpringApplication.java:307)
at org.springframework.boot.test.context.SpringBootContextLoader.loadContext(SpringBootContextLoader.java:136)
at org.springframework.test.context.cache.DefaultCacheAwareContextLoaderDelegate.loadContextInternal(DefaultCacheAwareContextLoaderDelegate.java:141)
at org.springframework.test.context.cache.DefaultCacheAwareContextLoaderDelegate.loadContext(DefaultCacheAwareContextLoaderDelegate.java:90)
at org.springframework.test.context.support.DefaultTestContext.getApplicationContext(DefaultTestContext.java:124)
at org.springframework.test.context.web.ServletTestExecutionListener.setUpRequestContextIfNecessary(ServletTestExecutionListener.java:190)
at org.springframework.test.context.web.ServletTestExecutionListener.prepareTestInstance(ServletTestExecutionListener.java:132)
at org.springframework.test.context.TestContextManager.prepareTestInstance(TestContextManager.java:248)
at org.springframework.test.context.junit.jupiter.SpringExtension.postProcessTestInstance(SpringExtension.java:138)
at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.lambda$invokeTestInstancePostProcessors$8(ClassBasedTestDescriptor.java:363)
at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.executeAndMaskThrowable(ClassBasedTestDescriptor.java:368)
at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.lambda$invokeTestInstancePostProcessors$9(ClassBasedTestDescriptor.java:363)
at java.base/java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:195)
at java.base/java.util.stream.ReferencePipeline$2$1.accept(ReferencePipeline.java:177)
at java.base/java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1655)
at java.base/java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:484)
at java.base/java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:474)
at java.base/java.util.stream.StreamSpliterators$WrappingSpliterator.forEachRemaining(StreamSpliterators.java:312)
at java.base/java.util.stream.Streams$ConcatSpliterator.forEachRemaining(Streams.java:735)
at java.base/java.util.stream.Streams$ConcatSpliterator.forEachRemaining(Streams.java:734)
at java.base/java.util.stream.ReferencePipeline$Head.forEach(ReferencePipeline.java:658)
at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.invokeTestInstancePostProcessors(ClassBasedTestDescriptor.java:362)
at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.lambda$instantiateAndPostProcessTestInstance$6(ClassBasedTestDescriptor.java:283)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.instantiateAndPostProcessTestInstance(ClassBasedTestDescriptor.java:282)
at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.lambda$testInstancesProvider$4(ClassBasedTestDescriptor.java:272)
at java.base/java.util.Optional.orElseGet(Optional.java:369)
at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.lambda$testInstancesProvider$5(ClassBasedTestDescriptor.java:271)
at org.junit.jupiter.engine.execution.TestInstancesProvider.getTestInstances(TestInstancesProvider.java:31)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$prepare$0(TestMethodTestDescriptor.java:102)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.prepare(TestMethodTestDescriptor.java:101)
at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.prepare(TestMethodTestDescriptor.java:66)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$prepare$2(NodeTestTask.java:123)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.prepare(NodeTestTask.java:123)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:90)
at java.base/java.util.ArrayList.forEach(ArrayList.java:1541)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:41)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:155)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95)
at java.base/java.util.ArrayList.forEach(ArrayList.java:1541)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:41)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:155)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138)
at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95)
at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.submit(SameThreadHierarchicalTestExecutorService.java:35)
at org.junit.platform.engine.support.hierarchical.HierarchicalTestExecutor.execute(HierarchicalTestExecutor.java:57)
at org.junit.platform.engine.support.hierarchical.HierarchicalTestEngine.execute(HierarchicalTestEngine.java:54)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:107)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:88)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.lambda$execute$0(EngineExecutionOrchestrator.java:54)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.withInterceptedStreams(EngineExecutionOrchestrator.java:67)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:52)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:114)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:86)
at org.junit.platform.launcher.core.DefaultLauncherSession$DelegatingLauncher.execute(DefaultLauncherSession.java:86)
at org.junit.platform.launcher.core.SessionPerRequestLauncher.execute(SessionPerRequestLauncher.java:53)
at com.intellij.junit5.JUnit5IdeaTestRunner.startRunnerWithArgs(JUnit5IdeaTestRunner.java:71)
at com.intellij.rt.junit.IdeaTestRunner$Repeater$1.execute(IdeaTestRunner.java:38)
at com.intellij.rt.execution.junit.TestsRepeater.repeat(TestsRepeater.java:11)
at com.intellij.rt.junit.IdeaTestRunner$Repeater.startRunnerWithArgs(IdeaTestRunner.java:35)
at com.intellij.rt.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:235)
at com.intellij.rt.junit.JUnitStarter.main(JUnitStarter.java:54)
Caused by: java.lang.ClassNotFoundException: com.mysql.jdbc.Driver
at java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:581)
at java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:178)
at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:521)
at java.base/java.lang.Class.forName0(Native Method)
at java.base/java.lang.Class.forName(Class.java:315)
at com.alibaba.druid.util.JdbcUtils.createDriver(JdbcUtils.java:686)
... 223 common frames omitted
```
| non_test | failed to run com oppo cloud parser spark eventlog replayeventlogstest since we set the mysql dependency to test in the mysql dependency will not be passed to subsequent modules error error datasource init error java sql sqlexception com mysql jdbc driver at com alibaba druid util jdbcutils createdriver jdbcutils java at com alibaba druid pool druiddatasource resolvedriver druiddatasource java at com alibaba druid pool druiddatasource init druiddatasource java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org springframework beans factory support abstractautowirecapablebeanfactory invokecustominitmethod abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory invokeinitmethods abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory initializebean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory docreatebean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory createbean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractbeanfactory lambda dogetbean abstractbeanfactory java at org springframework beans factory support defaultsingletonbeanregistry getsingleton defaultsingletonbeanregistry java at org springframework beans factory support abstractbeanfactory dogetbean abstractbeanfactory java at org springframework beans factory support abstractbeanfactory getbean abstractbeanfactory java at org springframework beans factory config dependencydescriptor resolvecandidate dependencydescriptor java at org springframework beans factory support defaultlistablebeanfactory addcandidateentry defaultlistablebeanfactory java at org springframework beans factory support defaultlistablebeanfactory findautowirecandidates defaultlistablebeanfactory java at org springframework beans factory support defaultlistablebeanfactory resolvemultiplebeans defaultlistablebeanfactory java at org springframework beans factory support defaultlistablebeanfactory doresolvedependency defaultlistablebeanfactory java at org springframework beans factory support defaultlistablebeanfactory resolvedependency defaultlistablebeanfactory java at org springframework beans factory support constructorresolver resolveautowiredargument constructorresolver java at org springframework beans factory support constructorresolver createargumentarray constructorresolver java at org springframework beans factory support constructorresolver instantiateusingfactorymethod constructorresolver java at org springframework beans factory support abstractautowirecapablebeanfactory instantiateusingfactorymethod abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory createbeaninstance abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory docreatebean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory createbean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractbeanfactory lambda dogetbean abstractbeanfactory java at org springframework beans factory support defaultsingletonbeanregistry getsingleton defaultsingletonbeanregistry java at org springframework beans factory support abstractbeanfactory dogetbean abstractbeanfactory java at org springframework beans factory support abstractbeanfactory getbean abstractbeanfactory java at org springframework beans factory config dependencydescriptor resolvecandidate dependencydescriptor java at org springframework beans factory support defaultlistablebeanfactory addcandidateentry defaultlistablebeanfactory java at org springframework beans factory support defaultlistablebeanfactory findautowirecandidates defaultlistablebeanfactory java at org springframework beans factory support defaultlistablebeanfactory resolvemultiplebeans defaultlistablebeanfactory java at org springframework beans factory support defaultlistablebeanfactory doresolvedependency defaultlistablebeanfactory java at org springframework beans factory support defaultlistablebeanfactory dependencyobjectprovider resolvestream defaultlistablebeanfactory java at org springframework beans factory support defaultlistablebeanfactory dependencyobjectprovider orderedstream defaultlistablebeanfactory java at org springframework boot actuate autoconfigure metrics meterregistryconfigurer addbinders meterregistryconfigurer java at org springframework boot actuate autoconfigure metrics meterregistryconfigurer configure meterregistryconfigurer java at org springframework boot actuate autoconfigure metrics meterregistrypostprocessor postprocessafterinitialization meterregistrypostprocessor java at org springframework beans factory support abstractautowirecapablebeanfactory applybeanpostprocessorsafterinitialization abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory initializebean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory docreatebean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory createbean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractbeanfactory lambda dogetbean abstractbeanfactory java at org springframework beans factory support defaultsingletonbeanregistry getsingleton defaultsingletonbeanregistry java at org springframework beans factory support abstractbeanfactory dogetbean abstractbeanfactory java at org springframework beans factory support abstractbeanfactory getbean abstractbeanfactory java at org springframework beans factory config dependencydescriptor resolvecandidate dependencydescriptor java at org springframework beans factory support defaultlistablebeanfactory doresolvedependency defaultlistablebeanfactory java at org springframework beans factory support defaultlistablebeanfactory resolvedependency defaultlistablebeanfactory java at org springframework beans factory support constructorresolver resolveautowiredargument constructorresolver java at org springframework beans factory support constructorresolver createargumentarray constructorresolver java at org springframework beans factory support constructorresolver instantiateusingfactorymethod constructorresolver java at org springframework beans factory support abstractautowirecapablebeanfactory instantiateusingfactorymethod abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory createbeaninstance abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory docreatebean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory createbean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractbeanfactory lambda dogetbean abstractbeanfactory java at org springframework beans factory support defaultsingletonbeanregistry getsingleton defaultsingletonbeanregistry java at org springframework beans factory support abstractbeanfactory dogetbean abstractbeanfactory java at org springframework beans factory support abstractbeanfactory getbean abstractbeanfactory java at org springframework beans factory config dependencydescriptor resolvecandidate dependencydescriptor java at org springframework beans factory support defaultlistablebeanfactory addcandidateentry defaultlistablebeanfactory java at org springframework beans factory support defaultlistablebeanfactory findautowirecandidates defaultlistablebeanfactory java at org springframework beans factory support defaultlistablebeanfactory resolvemultiplebeans defaultlistablebeanfactory java at org springframework beans factory support defaultlistablebeanfactory doresolvedependency defaultlistablebeanfactory java at org springframework beans factory support defaultlistablebeanfactory dependencyobjectprovider resolvestream defaultlistablebeanfactory java at org springframework beans factory support defaultlistablebeanfactory dependencyobjectprovider orderedstream defaultlistablebeanfactory java at org springframework boot autoconfigure data redis lettuceconnectionconfiguration lettuceclientresources lettuceconnectionconfiguration java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org springframework beans factory support simpleinstantiationstrategy instantiate simpleinstantiationstrategy java at org springframework beans factory support constructorresolver instantiate constructorresolver java at org springframework beans factory support constructorresolver instantiateusingfactorymethod constructorresolver java at org springframework beans factory support abstractautowirecapablebeanfactory instantiateusingfactorymethod abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory createbeaninstance abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory docreatebean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory createbean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractbeanfactory lambda dogetbean abstractbeanfactory java at org springframework beans factory support defaultsingletonbeanregistry getsingleton defaultsingletonbeanregistry java at org springframework beans factory support abstractbeanfactory dogetbean abstractbeanfactory java at org springframework beans factory support abstractbeanfactory getbean abstractbeanfactory java at org springframework beans factory config dependencydescriptor resolvecandidate dependencydescriptor java at org springframework beans factory support defaultlistablebeanfactory doresolvedependency defaultlistablebeanfactory java at org springframework beans factory support defaultlistablebeanfactory resolvedependency defaultlistablebeanfactory java at org springframework beans factory support constructorresolver resolveautowiredargument constructorresolver java at org springframework beans factory support constructorresolver createargumentarray constructorresolver java at org springframework beans factory support constructorresolver instantiateusingfactorymethod constructorresolver java at org springframework beans factory support abstractautowirecapablebeanfactory instantiateusingfactorymethod abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory createbeaninstance abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory docreatebean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory createbean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractbeanfactory lambda dogetbean abstractbeanfactory java at org springframework beans factory support defaultsingletonbeanregistry getsingleton defaultsingletonbeanregistry java at org springframework beans factory support abstractbeanfactory dogetbean abstractbeanfactory java at org springframework beans factory support abstractbeanfactory getbean abstractbeanfactory java at org springframework beans factory config dependencydescriptor resolvecandidate dependencydescriptor java at org springframework beans factory support defaultlistablebeanfactory doresolvedependency defaultlistablebeanfactory java at org springframework beans factory support defaultlistablebeanfactory resolvedependency defaultlistablebeanfactory java at org springframework beans factory support constructorresolver resolveautowiredargument constructorresolver java at org springframework beans factory support constructorresolver createargumentarray constructorresolver java at org springframework beans factory support constructorresolver instantiateusingfactorymethod constructorresolver java at org springframework beans factory support abstractautowirecapablebeanfactory instantiateusingfactorymethod abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory createbeaninstance abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory docreatebean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory createbean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractbeanfactory lambda dogetbean abstractbeanfactory java at org springframework beans factory support defaultsingletonbeanregistry getsingleton defaultsingletonbeanregistry java at org springframework beans factory support abstractbeanfactory dogetbean abstractbeanfactory java at org springframework beans factory support abstractbeanfactory getbean abstractbeanfactory java at org springframework beans factory config dependencydescriptor resolvecandidate dependencydescriptor java at org springframework beans factory support defaultlistablebeanfactory doresolvedependency defaultlistablebeanfactory java at org springframework beans factory support defaultlistablebeanfactory resolvedependency defaultlistablebeanfactory java at org springframework beans factory annotation autowiredannotationbeanpostprocessor autowiredfieldelement resolvefieldvalue autowiredannotationbeanpostprocessor java at org springframework beans factory annotation autowiredannotationbeanpostprocessor autowiredfieldelement inject autowiredannotationbeanpostprocessor java at org springframework beans factory annotation injectionmetadata inject injectionmetadata java at org springframework beans factory annotation autowiredannotationbeanpostprocessor postprocessproperties autowiredannotationbeanpostprocessor java at org springframework beans factory support abstractautowirecapablebeanfactory populatebean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory docreatebean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory createbean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractbeanfactory lambda dogetbean abstractbeanfactory java at org springframework beans factory support defaultsingletonbeanregistry getsingleton defaultsingletonbeanregistry java at org springframework beans factory support abstractbeanfactory dogetbean abstractbeanfactory java at org springframework beans factory support abstractbeanfactory getbean abstractbeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory resolvebeanbyname abstractautowirecapablebeanfactory java at org springframework context annotation commonannotationbeanpostprocessor autowireresource commonannotationbeanpostprocessor java at org springframework context annotation commonannotationbeanpostprocessor getresource commonannotationbeanpostprocessor java at org springframework context annotation commonannotationbeanpostprocessor resourceelement getresourcetoinject commonannotationbeanpostprocessor java at org springframework beans factory annotation injectionmetadata injectedelement inject injectionmetadata java at org springframework beans factory annotation injectionmetadata inject injectionmetadata java at org springframework context annotation commonannotationbeanpostprocessor postprocessproperties commonannotationbeanpostprocessor java at org springframework beans factory support abstractautowirecapablebeanfactory populatebean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory docreatebean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory createbean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractbeanfactory lambda dogetbean abstractbeanfactory java at org springframework beans factory support defaultsingletonbeanregistry getsingleton defaultsingletonbeanregistry java at org springframework beans factory support abstractbeanfactory dogetbean abstractbeanfactory java at org springframework beans factory support abstractbeanfactory getbean abstractbeanfactory java at org springframework beans factory support defaultlistablebeanfactory preinstantiatesingletons defaultlistablebeanfactory java at org springframework context support abstractapplicationcontext finishbeanfactoryinitialization abstractapplicationcontext java at org springframework context support abstractapplicationcontext refresh abstractapplicationcontext java at org springframework boot springapplication refresh springapplication java at org springframework boot springapplication refreshcontext springapplication java at org springframework boot springapplication run springapplication java at org springframework boot test context springbootcontextloader loadcontext springbootcontextloader java at org springframework test context cache defaultcacheawarecontextloaderdelegate loadcontextinternal defaultcacheawarecontextloaderdelegate java at org springframework test context cache defaultcacheawarecontextloaderdelegate loadcontext defaultcacheawarecontextloaderdelegate java at org springframework test context support defaulttestcontext getapplicationcontext defaulttestcontext java at org springframework test context web servlettestexecutionlistener setuprequestcontextifnecessary servlettestexecutionlistener java at org springframework test context web servlettestexecutionlistener preparetestinstance servlettestexecutionlistener java at org springframework test context testcontextmanager preparetestinstance testcontextmanager java at org springframework test context junit jupiter springextension postprocesstestinstance springextension java at org junit jupiter engine descriptor classbasedtestdescriptor lambda invoketestinstancepostprocessors classbasedtestdescriptor java at org junit jupiter engine descriptor classbasedtestdescriptor executeandmaskthrowable classbasedtestdescriptor java at org junit jupiter engine descriptor classbasedtestdescriptor lambda invoketestinstancepostprocessors classbasedtestdescriptor java at java base java util stream referencepipeline accept referencepipeline java at java base java util stream referencepipeline accept referencepipeline java at java base java util arraylist arraylistspliterator foreachremaining arraylist java at java base java util stream abstractpipeline copyinto abstractpipeline java at java base java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java base java util stream streamspliterators wrappingspliterator foreachremaining streamspliterators java at java base java util stream streams concatspliterator foreachremaining streams java at java base java util stream streams concatspliterator foreachremaining streams java at java base java util stream referencepipeline head foreach referencepipeline java at org junit jupiter engine descriptor classbasedtestdescriptor invoketestinstancepostprocessors classbasedtestdescriptor java at org junit jupiter engine descriptor classbasedtestdescriptor lambda instantiateandpostprocesstestinstance classbasedtestdescriptor java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit jupiter engine descriptor classbasedtestdescriptor instantiateandpostprocesstestinstance classbasedtestdescriptor java at org junit jupiter engine descriptor classbasedtestdescriptor lambda testinstancesprovider classbasedtestdescriptor java at java base java util optional orelseget optional java at org junit jupiter engine descriptor classbasedtestdescriptor lambda testinstancesprovider classbasedtestdescriptor java at org junit jupiter engine execution testinstancesprovider gettestinstances testinstancesprovider java at org junit jupiter engine descriptor testmethodtestdescriptor lambda prepare testmethodtestdescriptor java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit jupiter engine descriptor testmethodtestdescriptor prepare testmethodtestdescriptor java at org junit jupiter engine descriptor testmethodtestdescriptor prepare testmethodtestdescriptor java at org junit platform engine support hierarchical nodetesttask lambda prepare nodetesttask java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit platform engine support hierarchical nodetesttask prepare nodetesttask java at org junit platform engine support hierarchical nodetesttask execute nodetesttask java at java base java util arraylist foreach arraylist java at org junit platform engine support hierarchical samethreadhierarchicaltestexecutorservice invokeall samethreadhierarchicaltestexecutorservice java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical node around node java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit platform engine support hierarchical nodetesttask executerecursively nodetesttask java at org junit platform engine support hierarchical nodetesttask execute nodetesttask java at java base java util arraylist foreach arraylist java at org junit platform engine support hierarchical samethreadhierarchicaltestexecutorservice invokeall samethreadhierarchicaltestexecutorservice java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical node around node java at org junit platform engine support hierarchical nodetesttask lambda executerecursively nodetesttask java at org junit platform engine support hierarchical throwablecollector execute throwablecollector java at org junit platform engine support hierarchical nodetesttask executerecursively nodetesttask java at org junit platform engine support hierarchical nodetesttask execute nodetesttask java at org junit platform engine support hierarchical samethreadhierarchicaltestexecutorservice submit samethreadhierarchicaltestexecutorservice java at org junit platform engine support hierarchical hierarchicaltestexecutor execute hierarchicaltestexecutor java at org junit platform engine support hierarchical hierarchicaltestengine execute hierarchicaltestengine java at org junit platform launcher core engineexecutionorchestrator execute engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator execute engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator lambda execute engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator withinterceptedstreams engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator execute engineexecutionorchestrator java at org junit platform launcher core defaultlauncher execute defaultlauncher java at org junit platform launcher core defaultlauncher execute defaultlauncher java at org junit platform launcher core defaultlaunchersession delegatinglauncher execute defaultlaunchersession java at org junit platform launcher core sessionperrequestlauncher execute sessionperrequestlauncher java at com intellij startrunnerwithargs java at com intellij rt junit ideatestrunner repeater execute ideatestrunner java at com intellij rt execution junit testsrepeater repeat testsrepeater java at com intellij rt junit ideatestrunner repeater startrunnerwithargs ideatestrunner java at com intellij rt junit junitstarter preparestreamsandstart junitstarter java at com intellij rt junit junitstarter main junitstarter java caused by java lang classnotfoundexception com mysql jdbc driver at java base jdk internal loader builtinclassloader loadclass builtinclassloader java at java base jdk internal loader classloaders appclassloader loadclass classloaders java at java base java lang classloader loadclass classloader java at java base java lang class native method at java base java lang class forname class java at com alibaba druid util jdbcutils createdriver jdbcutils java common frames omitted | 0 |
598,236 | 18,240,868,069 | IssuesEvent | 2021-10-01 12:48:47 | bexis/Module_ResearchAreaManagement | https://api.github.com/repos/bexis/Module_ResearchAreaManagement | closed | Update of subplots of many plots via file upload needed | Type: enhancement Priority: High | Plots admins wants to do subplot updates / additions on a file on their computer. See #70
This file needs to be uploaded into the system and the changes needs to be done on the database, including the correct use of history tables.
There is already some kind of upload function. | 1.0 | Update of subplots of many plots via file upload needed - Plots admins wants to do subplot updates / additions on a file on their computer. See #70
This file needs to be uploaded into the system and the changes needs to be done on the database, including the correct use of history tables.
There is already some kind of upload function. | non_test | update of subplots of many plots via file upload needed plots admins wants to do subplot updates additions on a file on their computer see this file needs to be uploaded into the system and the changes needs to be done on the database including the correct use of history tables there is already some kind of upload function | 0 |
213,394 | 16,516,532,816 | IssuesEvent | 2021-05-26 10:17:56 | rancher/dashboard | https://api.github.com/repos/rancher/dashboard | closed | Add support for custom colors | [zube]: To Test | Allow the user to override UI colors via a setting page.
Store colors in a JSON setting: ui-colors.
Detail TBD | 1.0 | Add support for custom colors - Allow the user to override UI colors via a setting page.
Store colors in a JSON setting: ui-colors.
Detail TBD | test | add support for custom colors allow the user to override ui colors via a setting page store colors in a json setting ui colors detail tbd | 1 |
72,747 | 8,774,525,621 | IssuesEvent | 2018-12-18 20:05:02 | brave/brave-ios | https://api.github.com/repos/brave/brave-ios | opened | redesign the error page | needs design | ### Description:
The error page in V1.7 is not clear and needs design work
<img width="437" alt="screen shot 2018-12-18 at 11 15 19 am" src="https://user-images.githubusercontent.com/7606853/50178478-78c6a380-02b9-11e9-9db2-360a875cf0e4.png">
| 1.0 | redesign the error page - ### Description:
The error page in V1.7 is not clear and needs design work
<img width="437" alt="screen shot 2018-12-18 at 11 15 19 am" src="https://user-images.githubusercontent.com/7606853/50178478-78c6a380-02b9-11e9-9db2-360a875cf0e4.png">
| non_test | redesign the error page description the error page in is not clear and needs design work img width alt screen shot at am src | 0 |
13,372 | 3,330,177,048 | IssuesEvent | 2015-11-11 08:52:39 | Piiit/tpg | https://api.github.com/repos/Piiit/tpg | opened | SET OPERATIONS: function calls in target list do not work | 3-IN DEVELOPMENT bug test case | The following statement fails at the last SELECT statement.
```
create table e
(
a int,
b char,
t int4range
);
insert into e values
(1,'A','[2,5)'),
(2,'A','[3,4)'),
(3,'A','[7,9)');
create table f
(
a int,
b char,
t int4range
);
insert into f values
(1,'A','[3,4)');
select a,b,lower(t) ts, upper(t) te from e
UNION PERIOD WITH (ts,te,ts,te)
select * from f
order by 1,2,3,4;
``` | 1.0 | SET OPERATIONS: function calls in target list do not work - The following statement fails at the last SELECT statement.
```
create table e
(
a int,
b char,
t int4range
);
insert into e values
(1,'A','[2,5)'),
(2,'A','[3,4)'),
(3,'A','[7,9)');
create table f
(
a int,
b char,
t int4range
);
insert into f values
(1,'A','[3,4)');
select a,b,lower(t) ts, upper(t) te from e
UNION PERIOD WITH (ts,te,ts,te)
select * from f
order by 1,2,3,4;
``` | test | set operations function calls in target list do not work the following statement fails at the last select statement create table e a int b char t insert into e values a a a create table f a int b char t insert into f values a select a b lower t ts upper t te from e union period with ts te ts te select from f order by | 1 |
236,772 | 19,573,633,626 | IssuesEvent | 2022-01-04 13:03:35 | The-Microservice-Dungeon/generic-player | https://api.github.com/repos/The-Microservice-Dungeon/generic-player | closed | "Integration Testing Mode" for the Generic Player | testing | Ability to start the service with a multitude of player (e.g. 20 players), each performing random walks and commands, so that each service can watch if there are any problems. | 1.0 | "Integration Testing Mode" for the Generic Player - Ability to start the service with a multitude of player (e.g. 20 players), each performing random walks and commands, so that each service can watch if there are any problems. | test | integration testing mode for the generic player ability to start the service with a multitude of player e g players each performing random walks and commands so that each service can watch if there are any problems | 1 |
543,166 | 15,878,736,331 | IssuesEvent | 2021-04-09 11:27:45 | OpenSRP/opensrp-client-reveal | https://api.github.com/repos/OpenSRP/opensrp-client-reveal | closed | Change Edit Date Format Under the Household Task Tab | Priority: High Thailand | On the task tab in the household view in the Thailand APK, the task edit date format is given as mm/dd. The client would like this changed to the format dd/mm/yy. See sample present screenshot below.

| 1.0 | Change Edit Date Format Under the Household Task Tab - On the task tab in the household view in the Thailand APK, the task edit date format is given as mm/dd. The client would like this changed to the format dd/mm/yy. See sample present screenshot below.

| non_test | change edit date format under the household task tab on the task tab in the household view in the thailand apk the task edit date format is given as mm dd the client would like this changed to the format dd mm yy see sample present screenshot below | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.