Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
12,723
9,935,783,452
IssuesEvent
2019-07-02 17:25:31
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
dart2js buildbots are slow
area-infrastructure closed-obsolete type-bug
The build cycle time on a dart2js bot appears to be more than one hour. Perhaps we should add more shards?
1.0
dart2js buildbots are slow - The build cycle time on a dart2js bot appears to be more than one hour. Perhaps we should add more shards?
non_process
buildbots are slow the build cycle time on a bot appears to be more than one hour perhaps we should add more shards
0
21,248
28,371,216,914
IssuesEvent
2023-04-12 17:07:36
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
closed
Allow additional testkube testsuite executionRequest variables
enhancement test process
### Problem For different environments, we want to set specific variables in testkube testsuite's `.spec.executionRequest.variables` ### Solution Support `.Values.testkube.test.extraExecutionRequestVariables` config in `hedera-mirror-common` chart, and customize it for different envs if necessary ### Alternatives _No response_
1.0
Allow additional testkube testsuite executionRequest variables - ### Problem For different environments, we want to set specific variables in testkube testsuite's `.spec.executionRequest.variables` ### Solution Support `.Values.testkube.test.extraExecutionRequestVariables` config in `hedera-mirror-common` chart, and customize it for different envs if necessary ### Alternatives _No response_
process
allow additional testkube testsuite executionrequest variables problem for different environments we want to set specific variables in testkube testsuite s spec executionrequest variables solution support values testkube test extraexecutionrequestvariables config in hedera mirror common chart and customize it for different envs if necessary alternatives no response
1
8,907
12,013,911,666
IssuesEvent
2020-04-10 10:01:02
topcoder-platform/community-app
https://api.github.com/repos/topcoder-platform/community-app
opened
Submit button: "You are not authorized to access this page" error
Dev Env submissions processor
Register to a challenge and then click on submit. You get the following error You are not authorized to access this page. Please see attached video. [submit.mov.zip](https://github.com/topcoder-platform/community-app/files/4461017/submit.mov.zip)
1.0
Submit button: "You are not authorized to access this page" error - Register to a challenge and then click on submit. You get the following error You are not authorized to access this page. Please see attached video. [submit.mov.zip](https://github.com/topcoder-platform/community-app/files/4461017/submit.mov.zip)
process
submit button you are not authorized to access this page error register to a challenge and then click on submit you get the following error you are not authorized to access this page please see attached video
1
19,403
25,543,991,012
IssuesEvent
2022-11-29 17:15:08
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
Support log sampling
enhancement processor/probabilisticsampler
**Is your feature request related to a problem? Please describe.** Similar to what is offered with Probabilistic Sampling Processor, offer a log sampler. **Describe the solution you'd like** Ideally, the exact same config as the trace equivalent, with the addition of being able to sample logs according to their severity. **Describe alternatives you've considered** None come to mind. **Additional context** N/A
1.0
Support log sampling - **Is your feature request related to a problem? Please describe.** Similar to what is offered with Probabilistic Sampling Processor, offer a log sampler. **Describe the solution you'd like** Ideally, the exact same config as the trace equivalent, with the addition of being able to sample logs according to their severity. **Describe alternatives you've considered** None come to mind. **Additional context** N/A
process
support log sampling is your feature request related to a problem please describe similar to what is offered with probabilistic sampling processor offer a log sampler describe the solution you d like ideally the exact same config as the trace equivalent with the addition of being able to sample logs according to their severity describe alternatives you ve considered none come to mind additional context n a
1
181,974
14,086,030,439
IssuesEvent
2020-11-05 02:35:14
CSU-Booking-Platform/application
https://api.github.com/repos/CSU-Booking-Platform/application
opened
Acceptance Tests for #43 Delete Roles
acceptance-test
### User story #43 ### Acceptance criteria checklist - [ ] Administrators can delete roles from the system - [ ] Non-administrators cannot delete roles from the system - [ ] When a role is deleted, it is disassociated from all of its linked users.
1.0
Acceptance Tests for #43 Delete Roles - ### User story #43 ### Acceptance criteria checklist - [ ] Administrators can delete roles from the system - [ ] Non-administrators cannot delete roles from the system - [ ] When a role is deleted, it is disassociated from all of its linked users.
non_process
acceptance tests for delete roles user story acceptance criteria checklist administrators can delete roles from the system non administrators cannot delete roles from the system when a role is deleted it is disassociated from all of its linked users
0
15,139
18,892,586,409
IssuesEvent
2021-11-15 14:44:04
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
Fix flaky `system-tests-firefox` job
stage: needs review process: flaky test
`system-tests-firefox` seems to fail frequently on snapshot matches when it takes longer than normal to launch Firefox: https://app.circleci.com/pipelines/github/cypress-io/cypress/25418/workflows/c6f39a26-4ae6-4c34-a562-730b7d5e9001/jobs/953049 Two ways to fix it: 1. Support some env var like `CYPRESS_CDP_CONNECTION_TIMEOUT` in `protocol.ts` and pass it in `system-tests.ts` 2. Filter out the noisy stdout in `system-tests.ts` like we do for video snapshot errors in FF: https://github.com/cypress-io/cypress/blob/65458d6ac86ac8e4ba84844203c9177cee9904d8/system-tests/lib/system-tests.ts#L954-L968
1.0
Fix flaky `system-tests-firefox` job - `system-tests-firefox` seems to fail frequently on snapshot matches when it takes longer than normal to launch Firefox: https://app.circleci.com/pipelines/github/cypress-io/cypress/25418/workflows/c6f39a26-4ae6-4c34-a562-730b7d5e9001/jobs/953049 Two ways to fix it: 1. Support some env var like `CYPRESS_CDP_CONNECTION_TIMEOUT` in `protocol.ts` and pass it in `system-tests.ts` 2. Filter out the noisy stdout in `system-tests.ts` like we do for video snapshot errors in FF: https://github.com/cypress-io/cypress/blob/65458d6ac86ac8e4ba84844203c9177cee9904d8/system-tests/lib/system-tests.ts#L954-L968
process
fix flaky system tests firefox job system tests firefox seems to fail frequently on snapshot matches when it takes longer than normal to launch firefox two ways to fix it support some env var like cypress cdp connection timeout in protocol ts and pass it in system tests ts filter out the noisy stdout in system tests ts like we do for video snapshot errors in ff
1
19,352
25,482,326,798
IssuesEvent
2022-11-26 00:10:00
bitfocus/companion-module-requests
https://api.github.com/repos/bitfocus/companion-module-requests
opened
Broadlink RM Mini 3
NOT YET PROCESSED
- [X] **I have researched the list of existing Companion modules and requests and have determined this has not yet been requested** The name of the device, hardware, or software you would like to control: Broadlink RM Mini 3 A small, cheap IR blaster that attaches over WiFi What you would like to be able to make it do from Companion: Learn IR commands, Send IR Commands Direct links or attachments to the ethernet control protocol or API: https://github.com/mjg59/python-broadlink https://github.com/lbschenkel/broadlink-bridge https://github.com/401Unauthorized/broadlink-bridge
1.0
Broadlink RM Mini 3 - - [X] **I have researched the list of existing Companion modules and requests and have determined this has not yet been requested** The name of the device, hardware, or software you would like to control: Broadlink RM Mini 3 A small, cheap IR blaster that attaches over WiFi What you would like to be able to make it do from Companion: Learn IR commands, Send IR Commands Direct links or attachments to the ethernet control protocol or API: https://github.com/mjg59/python-broadlink https://github.com/lbschenkel/broadlink-bridge https://github.com/401Unauthorized/broadlink-bridge
process
broadlink rm mini i have researched the list of existing companion modules and requests and have determined this has not yet been requested the name of the device hardware or software you would like to control broadlink rm mini a small cheap ir blaster that attaches over wifi what you would like to be able to make it do from companion learn ir commands send ir commands direct links or attachments to the ethernet control protocol or api
1
132,395
28,138,162,115
IssuesEvent
2023-04-01 16:26:37
Howdju/howdju
https://api.github.com/repos/Howdju/howdju
opened
Allow EditorFields components to omit editorDispatch
clean code
Not all `EditorFields` components need to dispatch actions to the edit entity. Those that don't should be able to omit the `editorDispatch` property. Doing so simplifies reusing them outside of withEditor. E.g. `PropositionEditorFields` in `CreatePropositionPage`. But since `EntityEditorFieldsProps` determines both what `EditorFields` require and accept, marking `editorDispatch` optional or required won't work for the `EditorFields` that either require or omit `editorDispatch`, respectively. Add typing to `withEditor` to allow `EditorFields` to either require or omit `editorDispatch`. An idea would be to accept a config object to `withEditor` that determines the type of the EditorFields (with `editorDispatch` or not.)
1.0
Allow EditorFields components to omit editorDispatch - Not all `EditorFields` components need to dispatch actions to the edit entity. Those that don't should be able to omit the `editorDispatch` property. Doing so simplifies reusing them outside of withEditor. E.g. `PropositionEditorFields` in `CreatePropositionPage`. But since `EntityEditorFieldsProps` determines both what `EditorFields` require and accept, marking `editorDispatch` optional or required won't work for the `EditorFields` that either require or omit `editorDispatch`, respectively. Add typing to `withEditor` to allow `EditorFields` to either require or omit `editorDispatch`. An idea would be to accept a config object to `withEditor` that determines the type of the EditorFields (with `editorDispatch` or not.)
non_process
allow editorfields components to omit editordispatch not all editorfields components need to dispatch actions to the edit entity those that don t should be able to omit the editordispatch property doing so simplifies reusing them outside of witheditor e g propositioneditorfields in createpropositionpage but since entityeditorfieldsprops determines both what editorfields require and accept marking editordispatch optional or required won t work for the editorfields that either require or omit editordispatch respectively add typing to witheditor to allow editorfields to either require or omit editordispatch an idea would be to accept a config object to witheditor that determines the type of the editorfields with editordispatch or not
0
5,852
8,679,071,084
IssuesEvent
2018-11-30 22:10:46
habitat-sh/habitat
https://api.github.com/repos/habitat-sh/habitat
closed
Feature request: provide a hook for service member change
A-process-management A-supervisor C-feature V-sup
Need a hook for when the service group members change - even if there isn't a config change.
1.0
Feature request: provide a hook for service member change - Need a hook for when the service group members change - even if there isn't a config change.
process
feature request provide a hook for service member change need a hook for when the service group members change even if there isn t a config change
1
36,163
5,037,594,782
IssuesEvent
2016-12-17 19:18:16
jantman/awslimitchecker
https://api.github.com/repos/jantman/awslimitchecker
opened
pytest yield tests
testing
Pytest is now throwing a bunch of warnings about yield tests scheduled to be deprecated (see below). Figure out what's involved in converting these to new-style tests, if that's even possible. If it is, do it. If it isn't, and there's no clear alternative, pin pytest to <4.0 and suppress the warning :( ``` WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 ```
1.0
pytest yield tests - Pytest is now throwing a bunch of warnings about yield tests scheduled to be deprecated (see below). Figure out what's involved in converting these to new-style tests, if that's even possible. If it is, do it. If it isn't, and there's no clear alternative, pin pytest to <4.0 and suppress the warning :( ``` WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/test_integration.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 WC1 /home/jantman/GIT/awslimitchecker/awslimitchecker/tests/services/test_base.py yield tests are deprecated, and scheduled to be removed in pytest 4.0 ```
non_process
pytest yield tests pytest is now throwing a bunch of warnings about yield tests scheduled to be deprecated see below figure out what s involved in converting these to new style tests if that s even possible if it is do it if it isn t and there s no clear alternative pin pytest to and suppress the warning home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests test integration py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests services test base py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests services test base py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests services test base py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests services test base py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests services test base py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests services test base py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests services test base py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests services test base py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests services test base py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests services test base py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests services test base py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests services test base py yield tests are deprecated and scheduled to be removed in pytest home jantman git awslimitchecker awslimitchecker tests services test base py yield tests are deprecated and scheduled to be removed in pytest
0
21,007
27,883,337,274
IssuesEvent
2023-03-21 21:14:25
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Poor examples which compare boolean parameter and 'true'
doc-enhancement Pri1 azure-devops-pipelines/svc azure-devops-pipelines-process/subsvc
There are examples in the code which show a comparison of a boolean parameter with a string, for example here: https://learn.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops ``` parameters: - name: doThing default: true type: boolean steps: - script: echo I did a thing condition: and(succeeded(), eq('${{ parameters.doThing }}', 'true')) ``` This works as expected but only because `'true'` resolves to a truthy value. The following line would have the same effect: `condition: and(succeeded(), eq('${{ parameters.doThing }}', 'false'))` I don't think this would be expected by someone reading the docs. A better example can be found on this page: https://learn.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops ``` #azure-pipeline.yml parameters: - name: experimentalTemplate displayName: 'Use experimental build process?' type: boolean default: false steps: - ${{ if eq(parameters.experimentalTemplate, true) }}: - template: experimental.yml - ${{ if not(eq(parameters.experimentalTemplate, true)) }}: - template: stable.yml ``` Note how `true` is not enclosed in quotes. --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 21e5cee4-eaae-3a96-db91-540ac759e83a * Version Independent ID: 9bdc837c-ffe0-d999-f922-f3a5debc7f92 * Content: [Conditions - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/conditions?view=azure-devops&tabs=yaml%2Cstages) * Content Source: [docs/pipelines/process/conditions.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/conditions.md) * Service: **azure-devops-pipelines** * Sub-service: **azure-devops-pipelines-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Poor examples which compare boolean parameter and 'true' - There are examples in the code which show a comparison of a boolean parameter with a string, for example here: https://learn.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops ``` parameters: - name: doThing default: true type: boolean steps: - script: echo I did a thing condition: and(succeeded(), eq('${{ parameters.doThing }}', 'true')) ``` This works as expected but only because `'true'` resolves to a truthy value. The following line would have the same effect: `condition: and(succeeded(), eq('${{ parameters.doThing }}', 'false'))` I don't think this would be expected by someone reading the docs. A better example can be found on this page: https://learn.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops ``` #azure-pipeline.yml parameters: - name: experimentalTemplate displayName: 'Use experimental build process?' type: boolean default: false steps: - ${{ if eq(parameters.experimentalTemplate, true) }}: - template: experimental.yml - ${{ if not(eq(parameters.experimentalTemplate, true)) }}: - template: stable.yml ``` Note how `true` is not enclosed in quotes. --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 21e5cee4-eaae-3a96-db91-540ac759e83a * Version Independent ID: 9bdc837c-ffe0-d999-f922-f3a5debc7f92 * Content: [Conditions - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/conditions?view=azure-devops&tabs=yaml%2Cstages) * Content Source: [docs/pipelines/process/conditions.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/conditions.md) * Service: **azure-devops-pipelines** * Sub-service: **azure-devops-pipelines-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
poor examples which compare boolean parameter and true there are examples in the code which show a comparison of a boolean parameter with a string for example here parameters name dothing default true type boolean steps script echo i did a thing condition and succeeded eq parameters dothing true this works as expected but only because true resolves to a truthy value the following line would have the same effect condition and succeeded eq parameters dothing false i don t think this would be expected by someone reading the docs a better example can be found on this page azure pipeline yml parameters name experimentaltemplate displayname use experimental build process type boolean default false steps if eq parameters experimentaltemplate true template experimental yml if not eq parameters experimentaltemplate true template stable yml note how true is not enclosed in quotes document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id eaae version independent id content content source service azure devops pipelines sub service azure devops pipelines process github login juliakm microsoft alias jukullam
1
17,637
23,458,562,034
IssuesEvent
2022-08-16 11:07:14
anitsh/til
https://api.github.com/repos/anitsh/til
opened
Objectives and Key Results OKR
process
Begin with the end in mind - Steven Covey, The Seven Habits of Highly Successful People When you want to achieve something, anything, your desires are based on a preferred outcome. It’s a very simple, basic human concept developed at an early age. One way of putting it into a sentence is: I will ________ as measured by ____________. This is how John Doerr explained objectives and key results (OKR) in his book Measure what Matters. While it’s a simple concept, the business world is a complex place with many moving parts. Still, by adhering to simple principles, big results can manifest. A common mistake with OKRs is confusing desired outcomes with actions used to reach objectives. A popular management strategy that defines objectives and tracks results. Helps create alignment and engagement around measurable goals. "When employees perceive their jobs as high in task significance, they display higher job performance." -- “[The Significance of Task Significance: Job Performance Effects, Relational Mechanisms, and Boundary Conditions](http://www.management.wharton.upenn.edu/grant/Grant_JAP2008b_TaskSignificance.pdf),” in the Journal of Applied Psychology. More useful to think about the “OKR process” than it is to just consider the “objective” and the “key result”. Static goals that aren’t regularly reviewed and revised as the operating environment changes quickly become stale and meaningless. Combining clear objectives with a small set of specific, measurable results and a regular process of reviewing progress toward those measures is what makes OKRs truly useful. Objective: Where do you want to go? Key results: How do you know you're getting there? How do you measure that progress? OKRs move teams away from an “output” mindset (“What projects or tasks did we do?”) to an “outcome” mindset (“What was the business result of the project or task?”). You can set OKRs at the beginning of a key time period for your business (like a quarter or a fiscal year) and then check in regularly with those OKRs to keep your team focused on the importance of their work rather than its volume. "Objectives: are memorable, qualitative descriptions of what you want to achieve. Objectives should be short, inspirational, and engaging. An objective should motivate and challenge the team. Key results: are a set of metrics that measure your progress towards the objective. For each objective, you should have a set of two to five key results. More than that and no one will remember them." -- Felipe Castro, an OKR coach, author, speaker, and evangelist. Objective should be concise and engaging, so a team can easily remember it. Next, there should be a small number of metrics to track the key results. These metrics should be something you can measure on a timely basis. If you can only see the results after two years, it’s not possible to review your progress on a quarterly basis. It’s important to find key results that are leading indicators of your objective, as opposed to lagging indicators. You want results that can be measured regularly and frequently, and lead to the overall objective. Some high-level objectives are: Improve customer satisfaction Increase recurring revenue Scale system performance Increase the number of customers served Reduce the number of data errors in the system Objectives: Reduce the number of data errors in the system Key results: as measured by the number of data quality errors reported to the support desk as measured by the number of orders that can’t be filled automatically as measured by the order errors reported by customers Measure OKRs and communicate progress. OKRs and agile development Agile development and OKRs go together perfectly. It's been said that agile development decreases the costs of changing your mind. That’s because agile planning calls for a regular, planned cadence of checking the progress of development, evaluating the context of the enterprise (the competition, customer requirements, regulatory requirements, market conditions), and making adjustments based on this information. Having clearly communicated objectives and key results provides transparency. A quarterly review can shift from “did we deliver the things we said we’d deliver?” to “did we achieve the objectives we wanted?” Combining those two questions helps an organization decide whether to continue on its path or modify it. Align top-line OKRs with organization’s mission, vision, and “north star” values. Where OKRs help define the "what" and the "how," vision should enforce the "why." ![image](https://user-images.githubusercontent.com/414141/184859353-ac8cd48e-299e-474a-8a43-d3ad4ffe7e65.png) Objectives should support organizations mission and values, and they should be tightly aligned to long-term and short-term goals. ## Principles Three key principles to provide effective OKR management are: - Connect work and outcomes: Clearly connecting each team's daily work with business goals - Align individuals and impact: Creating a strong culture of employee engagement and growth - Orchestrate the OKR process: Setting up and running personalized OKR programs to meets each team's specific needs ## How OKRs add value to an organization - Gives a strategic advantage to organization of any size and can be adopted by individuals in any role. OKRs add value to your organization in the following ways: - Helps organization navigate rapid change, reduce risk, and identify waste. The OKR framework gives a level of visibility into work done across your organization that allows you to proactively manage risk. It helps you identify themes, trends, and changes to diverse data sets that can go unnoticed without a unifying system. Your organization can shift focus from activity to outcomes, making the right changes at the right time to stay competitive.- - Motivates and fosters involvement. When done well and made the driving force behind your organization's strategic rhythm, OKRs can keep every employee connected to and invested in your broader mission. - Creates cross-functional cooperation that unifies initiatives and improves collaboration. When each individual, team, and department goal is guided by the organization's broader strategy, teams can clearly prioritize each project, task, and outcome. - Offers clear, contextual communication with continual progress reviews. The OKR framework doesn’t stop when your goals are written. Instead, it creates a strategic rhythm around the most important outcomes that your organization is pursuing. The OKR framework should influence every interaction and foster continuous improvement in every review process. - Creates clarity that is documented, measurable, and owned. The OKR framework provides accountability that is tracked and measured. This accountability works in both directions: each individual is responsible for their OKRs, and your organization's leadership is responsible for enabling their work. ## Examples Objective 1: Be the top US provider of learning platforms to schools. Key results: 45 percent of K-12 schools using our platform A 12 percent increase in student engagement, as measured through internal systems A 95 percent satisfaction rate from quarterly parent surveys Objective 2: Build a technology platform that supports every person in our business to innovate and create. Key results: Five new applications developed and adopted across the organization At least two members that use the Microsoft Power Platform on each team New cloud technologies like data analytics and machine learning introduced in all customer-facing applications Objective 3: Transform our approach from sales-driven to data-driven. Key results: Increasing pipeline coverage from 50 percent to 200 percent Increasing closing rates for sales engagements by 5 percent Reducing the time to close deals by 8 percent - https://www.atlassian.com/agile/agile-at-scale/okr - https://docs.microsoft.com/en-us/azure/cloud-adoption-framework/strategy/business-outcomes/okr #92
1.0
Objectives and Key Results OKR - Begin with the end in mind - Steven Covey, The Seven Habits of Highly Successful People When you want to achieve something, anything, your desires are based on a preferred outcome. It’s a very simple, basic human concept developed at an early age. One way of putting it into a sentence is: I will ________ as measured by ____________. This is how John Doerr explained objectives and key results (OKR) in his book Measure what Matters. While it’s a simple concept, the business world is a complex place with many moving parts. Still, by adhering to simple principles, big results can manifest. A common mistake with OKRs is confusing desired outcomes with actions used to reach objectives. A popular management strategy that defines objectives and tracks results. Helps create alignment and engagement around measurable goals. "When employees perceive their jobs as high in task significance, they display higher job performance." -- “[The Significance of Task Significance: Job Performance Effects, Relational Mechanisms, and Boundary Conditions](http://www.management.wharton.upenn.edu/grant/Grant_JAP2008b_TaskSignificance.pdf),” in the Journal of Applied Psychology. More useful to think about the “OKR process” than it is to just consider the “objective” and the “key result”. Static goals that aren’t regularly reviewed and revised as the operating environment changes quickly become stale and meaningless. Combining clear objectives with a small set of specific, measurable results and a regular process of reviewing progress toward those measures is what makes OKRs truly useful. Objective: Where do you want to go? Key results: How do you know you're getting there? How do you measure that progress? OKRs move teams away from an “output” mindset (“What projects or tasks did we do?”) to an “outcome” mindset (“What was the business result of the project or task?”). You can set OKRs at the beginning of a key time period for your business (like a quarter or a fiscal year) and then check in regularly with those OKRs to keep your team focused on the importance of their work rather than its volume. "Objectives: are memorable, qualitative descriptions of what you want to achieve. Objectives should be short, inspirational, and engaging. An objective should motivate and challenge the team. Key results: are a set of metrics that measure your progress towards the objective. For each objective, you should have a set of two to five key results. More than that and no one will remember them." -- Felipe Castro, an OKR coach, author, speaker, and evangelist. Objective should be concise and engaging, so a team can easily remember it. Next, there should be a small number of metrics to track the key results. These metrics should be something you can measure on a timely basis. If you can only see the results after two years, it’s not possible to review your progress on a quarterly basis. It’s important to find key results that are leading indicators of your objective, as opposed to lagging indicators. You want results that can be measured regularly and frequently, and lead to the overall objective. Some high-level objectives are: Improve customer satisfaction Increase recurring revenue Scale system performance Increase the number of customers served Reduce the number of data errors in the system Objectives: Reduce the number of data errors in the system Key results: as measured by the number of data quality errors reported to the support desk as measured by the number of orders that can’t be filled automatically as measured by the order errors reported by customers Measure OKRs and communicate progress. OKRs and agile development Agile development and OKRs go together perfectly. It's been said that agile development decreases the costs of changing your mind. That’s because agile planning calls for a regular, planned cadence of checking the progress of development, evaluating the context of the enterprise (the competition, customer requirements, regulatory requirements, market conditions), and making adjustments based on this information. Having clearly communicated objectives and key results provides transparency. A quarterly review can shift from “did we deliver the things we said we’d deliver?” to “did we achieve the objectives we wanted?” Combining those two questions helps an organization decide whether to continue on its path or modify it. Align top-line OKRs with organization’s mission, vision, and “north star” values. Where OKRs help define the "what" and the "how," vision should enforce the "why." ![image](https://user-images.githubusercontent.com/414141/184859353-ac8cd48e-299e-474a-8a43-d3ad4ffe7e65.png) Objectives should support organizations mission and values, and they should be tightly aligned to long-term and short-term goals. ## Principles Three key principles to provide effective OKR management are: - Connect work and outcomes: Clearly connecting each team's daily work with business goals - Align individuals and impact: Creating a strong culture of employee engagement and growth - Orchestrate the OKR process: Setting up and running personalized OKR programs to meets each team's specific needs ## How OKRs add value to an organization - Gives a strategic advantage to organization of any size and can be adopted by individuals in any role. OKRs add value to your organization in the following ways: - Helps organization navigate rapid change, reduce risk, and identify waste. The OKR framework gives a level of visibility into work done across your organization that allows you to proactively manage risk. It helps you identify themes, trends, and changes to diverse data sets that can go unnoticed without a unifying system. Your organization can shift focus from activity to outcomes, making the right changes at the right time to stay competitive.- - Motivates and fosters involvement. When done well and made the driving force behind your organization's strategic rhythm, OKRs can keep every employee connected to and invested in your broader mission. - Creates cross-functional cooperation that unifies initiatives and improves collaboration. When each individual, team, and department goal is guided by the organization's broader strategy, teams can clearly prioritize each project, task, and outcome. - Offers clear, contextual communication with continual progress reviews. The OKR framework doesn’t stop when your goals are written. Instead, it creates a strategic rhythm around the most important outcomes that your organization is pursuing. The OKR framework should influence every interaction and foster continuous improvement in every review process. - Creates clarity that is documented, measurable, and owned. The OKR framework provides accountability that is tracked and measured. This accountability works in both directions: each individual is responsible for their OKRs, and your organization's leadership is responsible for enabling their work. ## Examples Objective 1: Be the top US provider of learning platforms to schools. Key results: 45 percent of K-12 schools using our platform A 12 percent increase in student engagement, as measured through internal systems A 95 percent satisfaction rate from quarterly parent surveys Objective 2: Build a technology platform that supports every person in our business to innovate and create. Key results: Five new applications developed and adopted across the organization At least two members that use the Microsoft Power Platform on each team New cloud technologies like data analytics and machine learning introduced in all customer-facing applications Objective 3: Transform our approach from sales-driven to data-driven. Key results: Increasing pipeline coverage from 50 percent to 200 percent Increasing closing rates for sales engagements by 5 percent Reducing the time to close deals by 8 percent - https://www.atlassian.com/agile/agile-at-scale/okr - https://docs.microsoft.com/en-us/azure/cloud-adoption-framework/strategy/business-outcomes/okr #92
process
objectives and key results okr begin with the end in mind steven covey the seven habits of highly successful people when you want to achieve something anything your desires are based on a preferred outcome it’s a very simple basic human concept developed at an early age one way of putting it into a sentence is i will as measured by this is how john doerr explained objectives and key results okr in his book measure what matters while it’s a simple concept the business world is a complex place with many moving parts still by adhering to simple principles big results can manifest a common mistake with okrs is confusing desired outcomes with actions used to reach objectives a popular management strategy that defines objectives and tracks results helps create alignment and engagement around measurable goals when employees perceive their jobs as high in task significance they display higher job performance “ in the journal of applied psychology more useful to think about the “okr process” than it is to just consider the “objective” and the “key result” static goals that aren’t regularly reviewed and revised as the operating environment changes quickly become stale and meaningless combining clear objectives with a small set of specific measurable results and a regular process of reviewing progress toward those measures is what makes okrs truly useful objective where do you want to go key results how do you know you re getting there how do you measure that progress okrs move teams away from an “output” mindset “what projects or tasks did we do ” to an “outcome” mindset “what was the business result of the project or task ” you can set okrs at the beginning of a key time period for your business like a quarter or a fiscal year and then check in regularly with those okrs to keep your team focused on the importance of their work rather than its volume objectives are memorable qualitative descriptions of what you want to achieve objectives should be short inspirational and engaging an objective should motivate and challenge the team key results are a set of metrics that measure your progress towards the objective for each objective you should have a set of two to five key results more than that and no one will remember them felipe castro an okr coach author speaker and evangelist objective should be concise and engaging so a team can easily remember it next there should be a small number of metrics to track the key results these metrics should be something you can measure on a timely basis if you can only see the results after two years it’s not possible to review your progress on a quarterly basis it’s important to find key results that are leading indicators of your objective as opposed to lagging indicators you want results that can be measured regularly and frequently and lead to the overall objective some high level objectives are improve customer satisfaction increase recurring revenue scale system performance increase the number of customers served reduce the number of data errors in the system objectives reduce the number of data errors in the system key results as measured by the number of data quality errors reported to the support desk as measured by the number of orders that can’t be filled automatically as measured by the order errors reported by customers measure okrs and communicate progress okrs and agile development agile development and okrs go together perfectly it s been said that agile development decreases the costs of changing your mind that’s because agile planning calls for a regular planned cadence of checking the progress of development evaluating the context of the enterprise the competition customer requirements regulatory requirements market conditions and making adjustments based on this information having clearly communicated objectives and key results provides transparency a quarterly review can shift from “did we deliver the things we said we’d deliver ” to “did we achieve the objectives we wanted ” combining those two questions helps an organization decide whether to continue on its path or modify it align top line okrs with organization’s mission vision and “north star” values where okrs help define the what and the how vision should enforce the why objectives should support organizations mission and values and they should be tightly aligned to long term and short term goals principles three key principles to provide effective okr management are connect work and outcomes clearly connecting each team s daily work with business goals align individuals and impact creating a strong culture of employee engagement and growth orchestrate the okr process setting up and running personalized okr programs to meets each team s specific needs how okrs add value to an organization gives a strategic advantage to organization of any size and can be adopted by individuals in any role okrs add value to your organization in the following ways helps organization navigate rapid change reduce risk and identify waste the okr framework gives a level of visibility into work done across your organization that allows you to proactively manage risk it helps you identify themes trends and changes to diverse data sets that can go unnoticed without a unifying system your organization can shift focus from activity to outcomes making the right changes at the right time to stay competitive motivates and fosters involvement when done well and made the driving force behind your organization s strategic rhythm okrs can keep every employee connected to and invested in your broader mission creates cross functional cooperation that unifies initiatives and improves collaboration when each individual team and department goal is guided by the organization s broader strategy teams can clearly prioritize each project task and outcome offers clear contextual communication with continual progress reviews the okr framework doesn’t stop when your goals are written instead it creates a strategic rhythm around the most important outcomes that your organization is pursuing the okr framework should influence every interaction and foster continuous improvement in every review process creates clarity that is documented measurable and owned the okr framework provides accountability that is tracked and measured this accountability works in both directions each individual is responsible for their okrs and your organization s leadership is responsible for enabling their work examples objective be the top us provider of learning platforms to schools key results percent of k schools using our platform a percent increase in student engagement as measured through internal systems a percent satisfaction rate from quarterly parent surveys objective build a technology platform that supports every person in our business to innovate and create key results five new applications developed and adopted across the organization at least two members that use the microsoft power platform on each team new cloud technologies like data analytics and machine learning introduced in all customer facing applications objective transform our approach from sales driven to data driven key results increasing pipeline coverage from percent to percent increasing closing rates for sales engagements by percent reducing the time to close deals by percent
1
223,989
17,147,768,785
IssuesEvent
2021-07-13 16:24:51
MDAnalysis/solvation-analysis
https://api.github.com/repos/MDAnalysis/solvation-analysis
closed
Add Jupyter notebook for validation of visual elements of minima finding.
documentation testing
The minima finding element of the tests requires manual visual validation of the minima identified algorithmically. This process should be easily repeatable and IMO should be put into a jupyter-notebook somewhere near the tests.
1.0
Add Jupyter notebook for validation of visual elements of minima finding. - The minima finding element of the tests requires manual visual validation of the minima identified algorithmically. This process should be easily repeatable and IMO should be put into a jupyter-notebook somewhere near the tests.
non_process
add jupyter notebook for validation of visual elements of minima finding the minima finding element of the tests requires manual visual validation of the minima identified algorithmically this process should be easily repeatable and imo should be put into a jupyter notebook somewhere near the tests
0
13,179
15,608,923,977
IssuesEvent
2021-03-19 11:13:29
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Add `timeout` and `killSignal` options to `child_process` `spawn()` and `fork()`
child_process feature request
**Is your feature request related to a problem? Please describe.** The `timeout` and `killSignal` options are available in `execFile[Sync]()`, `exec[Sync]()`, `spawnSync()` but not in `spawn()` nor `fork()`. Those methods could be useful for `spawn()` and `fork()` too. For example, projects like [`execa`](https://github.com/sindresorhus/execa) (by @sindresorhus) use userland code to bring those options to `spawn()`. **Describe the solution you'd like** Add those options to `spawn()` and `fork()`.
1.0
Add `timeout` and `killSignal` options to `child_process` `spawn()` and `fork()` - **Is your feature request related to a problem? Please describe.** The `timeout` and `killSignal` options are available in `execFile[Sync]()`, `exec[Sync]()`, `spawnSync()` but not in `spawn()` nor `fork()`. Those methods could be useful for `spawn()` and `fork()` too. For example, projects like [`execa`](https://github.com/sindresorhus/execa) (by @sindresorhus) use userland code to bring those options to `spawn()`. **Describe the solution you'd like** Add those options to `spawn()` and `fork()`.
process
add timeout and killsignal options to child process spawn and fork is your feature request related to a problem please describe the timeout and killsignal options are available in execfile exec spawnsync but not in spawn nor fork those methods could be useful for spawn and fork too for example projects like by sindresorhus use userland code to bring those options to spawn describe the solution you d like add those options to spawn and fork
1
3,153
2,607,985,844
IssuesEvent
2015-02-26 00:51:40
chrsmithdemos/zen-coding
https://api.github.com/repos/chrsmithdemos/zen-coding
closed
Expanding in Sublime Text 2 doesn't work immediately after another expand
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Example: create a ul (with no children) and expand it with TAB 2. in the created "<ul></ul>", create an li element like so "<ul>li</ul>" and hit TAB 3. The cursor moves to the end of the line without expanding 4. putting the cursor back and hitting TAB after that does expand, so it only happens right after another expand. What is the expected output? What do you see instead? Expected: expanding of the li element Result: cursor jumps to end-of-line without expanding What version of the product are you using? On what operating system? Sublime Text 2, build 2181 Tested on Windows Vista & Windows 7 Please provide any additional information below. I think no more info is needed at this time. ``` ----- Original issue reported on code.google.com by `rp.ka...@gmail.com` on 3 May 2012 at 2:52
1.0
Expanding in Sublime Text 2 doesn't work immediately after another expand - ``` What steps will reproduce the problem? 1. Example: create a ul (with no children) and expand it with TAB 2. in the created "<ul></ul>", create an li element like so "<ul>li</ul>" and hit TAB 3. The cursor moves to the end of the line without expanding 4. putting the cursor back and hitting TAB after that does expand, so it only happens right after another expand. What is the expected output? What do you see instead? Expected: expanding of the li element Result: cursor jumps to end-of-line without expanding What version of the product are you using? On what operating system? Sublime Text 2, build 2181 Tested on Windows Vista & Windows 7 Please provide any additional information below. I think no more info is needed at this time. ``` ----- Original issue reported on code.google.com by `rp.ka...@gmail.com` on 3 May 2012 at 2:52
non_process
expanding in sublime text doesn t work immediately after another expand what steps will reproduce the problem example create a ul with no children and expand it with tab in the created create an li element like so li and hit tab the cursor moves to the end of the line without expanding putting the cursor back and hitting tab after that does expand so it only happens right after another expand what is the expected output what do you see instead expected expanding of the li element result cursor jumps to end of line without expanding what version of the product are you using on what operating system sublime text build tested on windows vista windows please provide any additional information below i think no more info is needed at this time original issue reported on code google com by rp ka gmail com on may at
0
74,196
7,388,880,174
IssuesEvent
2018-03-16 05:44:12
pixelhumain/co2
https://api.github.com/repos/pixelhumain/co2
closed
Ntework - content répertoire list height wrong
to test
La height du container change quand on va sur un élément et qu'on revient à la liste ![capture d ecran de 2018-03-15 16-46-30](https://user-images.githubusercontent.com/6576514/37474418-8d39a46c-2870-11e8-8ca6-163fcd4164e5.png)
1.0
Ntework - content répertoire list height wrong - La height du container change quand on va sur un élément et qu'on revient à la liste ![capture d ecran de 2018-03-15 16-46-30](https://user-images.githubusercontent.com/6576514/37474418-8d39a46c-2870-11e8-8ca6-163fcd4164e5.png)
non_process
ntework content répertoire list height wrong la height du container change quand on va sur un élément et qu on revient à la liste
0
314,731
9,602,133,199
IssuesEvent
2019-05-10 13:55:10
ngageoint/hootenanny
https://api.github.com/repos/ngageoint/hootenanny
closed
Change layout of tag-dist output and add a conflation reviews summary example
Category: Core Priority: Medium Type: Feature
A helpful feature to have as I debug #3136. Would use ```tag-distribution``` to give you a summary and distribution of feature reviews...something like what's below. I think I would rather see the counts and percentages on the left hand side though....so modify ```tag-distribution```. Generated with: ```hoot tag-distribution input.osm hoot:review:note hoot::ReviewRelationCriterion``` ```very little building overlap building orientation not similar building edges not very close to each other : 20 (14.6%) very little building overlap similar building orientation building edges not very close to each other : 15 (10.95%) very little building overlap semisimilar building orientation building edges not very close to each other : 14 (10.22%) very little building overlap very similar building orientation building edges not very close to each other : 12 (8.759%) conflicting information multiple features have been matched to the same feature and require review : 8 (5.839%) small building overlap building orientation not similar building edges not very close to each other : 3 (2.19%) small building overlap similar building orientation building edges not very close to each other : 1 (0.7299%) small building overlap semisimilar building orientation building edges not very close to each other : 1 (0.7299%) medium building overlap building orientation not similar building edges not very close to each other : 1 (0.7299%)```
1.0
Change layout of tag-dist output and add a conflation reviews summary example - A helpful feature to have as I debug #3136. Would use ```tag-distribution``` to give you a summary and distribution of feature reviews...something like what's below. I think I would rather see the counts and percentages on the left hand side though....so modify ```tag-distribution```. Generated with: ```hoot tag-distribution input.osm hoot:review:note hoot::ReviewRelationCriterion``` ```very little building overlap building orientation not similar building edges not very close to each other : 20 (14.6%) very little building overlap similar building orientation building edges not very close to each other : 15 (10.95%) very little building overlap semisimilar building orientation building edges not very close to each other : 14 (10.22%) very little building overlap very similar building orientation building edges not very close to each other : 12 (8.759%) conflicting information multiple features have been matched to the same feature and require review : 8 (5.839%) small building overlap building orientation not similar building edges not very close to each other : 3 (2.19%) small building overlap similar building orientation building edges not very close to each other : 1 (0.7299%) small building overlap semisimilar building orientation building edges not very close to each other : 1 (0.7299%) medium building overlap building orientation not similar building edges not very close to each other : 1 (0.7299%)```
non_process
change layout of tag dist output and add a conflation reviews summary example a helpful feature to have as i debug would use tag distribution to give you a summary and distribution of feature reviews something like what s below i think i would rather see the counts and percentages on the left hand side though so modify tag distribution generated with hoot tag distribution input osm hoot review note hoot reviewrelationcriterion very little building overlap building orientation not similar building edges not very close to each other very little building overlap similar building orientation building edges not very close to each other very little building overlap semisimilar building orientation building edges not very close to each other very little building overlap very similar building orientation building edges not very close to each other conflicting information multiple features have been matched to the same feature and require review small building overlap building orientation not similar building edges not very close to each other small building overlap similar building orientation building edges not very close to each other small building overlap semisimilar building orientation building edges not very close to each other medium building overlap building orientation not similar building edges not very close to each other
0
349,157
10,459,628,535
IssuesEvent
2019-09-20 11:28:14
yalla-coop/rent-check
https://api.github.com/repos/yalla-coop/rent-check
closed
AS A developer I NEED TO understand the technologies we'll be using SO THAT I can work without blocks
priority-1
- [x] decide tech stack - [x] plan technical spikes - [x] demonstrate new tech
1.0
AS A developer I NEED TO understand the technologies we'll be using SO THAT I can work without blocks - - [x] decide tech stack - [x] plan technical spikes - [x] demonstrate new tech
non_process
as a developer i need to understand the technologies we ll be using so that i can work without blocks decide tech stack plan technical spikes demonstrate new tech
0
15,513
19,703,266,910
IssuesEvent
2022-01-12 18:52:19
googleapis/java-data-fusion
https://api.github.com/repos/googleapis/java-data-fusion
opened
Your .repo-metadata.json file has a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * release_level must be equal to one of the allowed values in .repo-metadata.json * api_shortname 'data-fusion' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * release_level must be equal to one of the allowed values in .repo-metadata.json * api_shortname 'data-fusion' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 release level must be equal to one of the allowed values in repo metadata json api shortname data fusion invalid in repo metadata json ☝️ once you correct these problems you can close this issue reach out to go github automation if you have any questions
1
8,971
3,013,876,967
IssuesEvent
2015-07-29 11:53:03
Semantic-Org/Semantic-UI
https://api.github.com/repos/Semantic-Org/Semantic-UI
closed
[button] .ui.animated.orange.basic.button bug
Needs Test Case Should Use StackOverflow / Gitter
Sorry, my English is not well. source-code: .ui.basic.orange.button:hover, .ui.basic.orange.buttons .button:hover { color: #F26202 !important; } Correct code: .ui.basic.orange.button:hover, .ui.basic.orange.button:hover > .content, .ui.basic.orange.buttons .button:hover, .ui.basic.orange.buttons .button:hover > .content { color: #F26202 !important; } I mean, when use the following code ,the icon's color is default. <div class="ui animated orange basic button"> <div class="visible content">下一步</div> <div class="hidden content"> <i class="right arrow icon"></i> </div> </div> Meybe my code is wrong, please come and fix it.
1.0
[button] .ui.animated.orange.basic.button bug - Sorry, my English is not well. source-code: .ui.basic.orange.button:hover, .ui.basic.orange.buttons .button:hover { color: #F26202 !important; } Correct code: .ui.basic.orange.button:hover, .ui.basic.orange.button:hover > .content, .ui.basic.orange.buttons .button:hover, .ui.basic.orange.buttons .button:hover > .content { color: #F26202 !important; } I mean, when use the following code ,the icon's color is default. <div class="ui animated orange basic button"> <div class="visible content">下一步</div> <div class="hidden content"> <i class="right arrow icon"></i> </div> </div> Meybe my code is wrong, please come and fix it.
non_process
ui animated orange basic button bug sorry my english is not well source code ui basic orange button hover ui basic orange buttons button hover color important correct code ui basic orange button hover ui basic orange button hover content ui basic orange buttons button hover ui basic orange buttons button hover content color important i mean when use the following code the icon s color is default 下一步 meybe my code is wrong please come and fix it
0
6,396
3,799,200,640
IssuesEvent
2016-03-23 15:13:30
globaleaks/GlobaLeaks
https://api.github.com/repos/globaleaks/GlobaLeaks
closed
Utility to Download, Update, Upload to a website Tor Browser Bundle in several languages
A: Amnesty C: Backend C: Build Scripts C: Client F: Security T: Enhancement U: Admin
This ticket is to introduce a functionality that has been required by Amnesty as a part of their deployment strategy, to enable distribution of Tor Browser Bundle from a private website. The use case is: Country where torproject.org is blocked Distribution of access URL is done privately End user is already known to local government to be an activist End user is seasoned old, non-technology savvy person This ticket is to implement a software, to be part of GlobaLeaks software suite but run independently, that will: * Download the latests version of Tor Browser Bundle for a defined set of supported languages (Urdu, Arabic, English) The OS to be supported are: * Windows * OSX The software will run under a period background job (let's say once per day) and, if a new version of TBB is available, will replace the previous version. The latest version must be checked with https://www.torproject.org/projects/torbrowser/RecommendedTBBVersions . Download should be done from https://dist.torproject.org/torbrowser/4.0.1/ . The software may be integrated within the GlobaLeaks software, available and configurable from within the GlobaLeaks interface itself. <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/6686713-utility-to-download-update-upload-to-a-website-tor-browser-bundle-in-several-languages?utm_campaign=plugin&utm_content=tracker%2F316378&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F316378&utm_medium=issues&utm_source=github). </bountysource-plugin>
1.0
Utility to Download, Update, Upload to a website Tor Browser Bundle in several languages - This ticket is to introduce a functionality that has been required by Amnesty as a part of their deployment strategy, to enable distribution of Tor Browser Bundle from a private website. The use case is: Country where torproject.org is blocked Distribution of access URL is done privately End user is already known to local government to be an activist End user is seasoned old, non-technology savvy person This ticket is to implement a software, to be part of GlobaLeaks software suite but run independently, that will: * Download the latests version of Tor Browser Bundle for a defined set of supported languages (Urdu, Arabic, English) The OS to be supported are: * Windows * OSX The software will run under a period background job (let's say once per day) and, if a new version of TBB is available, will replace the previous version. The latest version must be checked with https://www.torproject.org/projects/torbrowser/RecommendedTBBVersions . Download should be done from https://dist.torproject.org/torbrowser/4.0.1/ . The software may be integrated within the GlobaLeaks software, available and configurable from within the GlobaLeaks interface itself. <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/6686713-utility-to-download-update-upload-to-a-website-tor-browser-bundle-in-several-languages?utm_campaign=plugin&utm_content=tracker%2F316378&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F316378&utm_medium=issues&utm_source=github). </bountysource-plugin>
non_process
utility to download update upload to a website tor browser bundle in several languages this ticket is to introduce a functionality that has been required by amnesty as a part of their deployment strategy to enable distribution of tor browser bundle from a private website the use case is country where torproject org is blocked distribution of access url is done privately end user is already known to local government to be an activist end user is seasoned old non technology savvy person this ticket is to implement a software to be part of globaleaks software suite but run independently that will download the latests version of tor browser bundle for a defined set of supported languages urdu arabic english the os to be supported are windows osx the software will run under a period background job let s say once per day and if a new version of tbb is available will replace the previous version the latest version must be checked with download should be done from the software may be integrated within the globaleaks software available and configurable from within the globaleaks interface itself want to back this issue we accept bounties via
0
144,560
22,392,895,554
IssuesEvent
2022-06-17 09:27:28
ZigZagExchange/frontend
https://api.github.com/repos/ZigZagExchange/frontend
opened
fix: Disable "buy" / "sell" button without input / with 0 input
uiRedesign
Currently, I can click buy/sell when there is no input at all, which refreshes the page. I can also put `0` as input and I can still click buy/sell https://gyazo.com/208c7634a4c8fb3a542d31264d75376c Buy/sell button should be disabled when there's a bad input (no input, 0 as input)
1.0
fix: Disable "buy" / "sell" button without input / with 0 input - Currently, I can click buy/sell when there is no input at all, which refreshes the page. I can also put `0` as input and I can still click buy/sell https://gyazo.com/208c7634a4c8fb3a542d31264d75376c Buy/sell button should be disabled when there's a bad input (no input, 0 as input)
non_process
fix disable buy sell button without input with input currently i can click buy sell when there is no input at all which refreshes the page i can also put as input and i can still click buy sell buy sell button should be disabled when there s a bad input no input as input
0
323,239
27,707,412,815
IssuesEvent
2023-03-14 12:10:10
jhipster/generator-jhipster
https://api.github.com/repos/jhipster/generator-jhipster
closed
Drop protractor.
theme: tests area: cleanup v8
##### **Overview of the feature request** Protractor is deprecated and should be dropped. <!-- Explain the feature request --> ##### **Motivation for or Use Case** <!-- Explain why this new feature is important for you --> ##### **Related issues or PR** <!-- Has a similar feature request been asked for before? Please search both closed & open issues --> - [ ] Checking this box is mandatory (this is just to show you read everything) <!-- Love JHipster? Please consider supporting our collective: 👉 https://opencollective.com/generator-jhipster/donate -->
1.0
Drop protractor. - ##### **Overview of the feature request** Protractor is deprecated and should be dropped. <!-- Explain the feature request --> ##### **Motivation for or Use Case** <!-- Explain why this new feature is important for you --> ##### **Related issues or PR** <!-- Has a similar feature request been asked for before? Please search both closed & open issues --> - [ ] Checking this box is mandatory (this is just to show you read everything) <!-- Love JHipster? Please consider supporting our collective: 👉 https://opencollective.com/generator-jhipster/donate -->
non_process
drop protractor overview of the feature request protractor is deprecated and should be dropped motivation for or use case related issues or pr checking this box is mandatory this is just to show you read everything love jhipster please consider supporting our collective 👉
0
38,532
6,676,179,398
IssuesEvent
2017-10-05 03:34:18
flyve-mdm/flyve-mdm-blackberry-inventory-agent
https://api.github.com/repos/flyve-mdm/flyve-mdm-blackberry-inventory-agent
closed
Create project page
documentation
Hi, @rafaelje I will create gh-pages as requested in this issue: flyve-mdm/flyve-mdm.github.io#13 Best regards
1.0
Create project page - Hi, @rafaelje I will create gh-pages as requested in this issue: flyve-mdm/flyve-mdm.github.io#13 Best regards
non_process
create project page hi rafaelje i will create gh pages as requested in this issue flyve mdm flyve mdm github io best regards
0
6,472
9,547,726,998
IssuesEvent
2019-05-02 00:54:14
parcel-bundler/parcel
https://api.github.com/repos/parcel-bundler/parcel
closed
BUG: postcss module generates wrong hashes
:bug: Bug CSS Preprocessing
<!--- Thanks for filing an issue 😄 ! Before you submit, please read the following: Search open/closed issues before submitting since someone might have asked the same thing before! --> This a 🐛 bug report <!--- Provide a general summary of the issue in the title above --> ### 🎛 Configuration (.babelrc, package.json, cli command) <!--- If describing a bug, tell us what your babel configuration looks like --> postcss.config.js ```js module.exports = { "modules": true, "plugins": { "autoprefixer": { "grid": true, "browsers": [ ">1%", "last 2 versions", "not ie < 10" ], "flexbox": "no-2009" }, "postcss-modules": { "generateScopedName": "[local]__[hash:base64:4]" } } } ``` package.json ```js { "name": "react-project", "version": "1.0.0", "description": "Small test", "main": "index.js", "scripts": { "start": "parcel public/index.html -p 4000 --no-cache", "build": "parcel build public/index.html --out-dir build --no-cache --public-url ./", "test": "echo \"Error: no test specified\" && exit 1" }, "repository": {}, "author": "", "license": "ISC", "homepage": "", "dependencies": { "apollo-boost": "^0.1.2", "babel-eslint": "^7.2.3", "classnames": "^2.2.5", "graphql": "^0.13.1", "react": "^16.2.0", "react-apollo": "^2.0.4", "react-dom": "^16.2.0", "react-loadable": "^5.3.1", "react-router-dom": "^4.2.2" }, "devDependencies": { "autoprefixer": "^8.1.0", "babel-core": "^6.26.0", "babel-plugin-module-resolver": "^3.1.0", "babel-preset-react-app": "^3.1.1", "eslint": "^4.18.2", "eslint-config-react-app": "^2.1.0", "eslint-plugin-flowtype": "^2.46.1", "eslint-plugin-import": "^2.9.0", "eslint-plugin-jsx-a11y": "^5.1.1", "eslint-plugin-react": "^7.7.0", "node-sass": "^4.7.2", "postcss-modules": "^1.1.0", "postcss-scss": "^1.0.4" } } ``` .babelrc ```js { "plugins": [ ["module-resolver", { "root": [ "./src/**/components" ], "alias": {} }] ], "presets": [ "react-app" ] } ``` ### 🤔 Expected Behavior <!--- If you're describing a bug, tell us what should happen --> <!--- If you're suggesting a change/improvement, tell us how it should work --> Classnames in generated .css file should have hash which is 4 chars long and :global classname should not contain hash ### 😯 Current Behavior <!--- If describing a bug, tell us what happens instead of the expected behavior --> <!--- If you are seeing an error, please include the full error message and stack trace --> <!--- If suggesting a change/improvement, explain the difference from current behavior --> Seems working correctly after restarting parcel. Example: `.page-header {}` is converted to `.page-header__1x8y {}` and `:global .app-root` is converted to `.app-root {}` Classnames are generated incorrectly after changing and saving css file `.page-header {}` is converted to `._page-header__1x8y_1dxnl_1 {}` and `:global .app-root` is converted to `._app-root_1dxnl_12` When saving again then classnames are converted as shown below `.page-header {}` -> `._page-header__1x8y_15rqa_1 {}` `:global .app-root` -> `._app-root_15rqa_12` ### 💁 Possible Solution <!--- Not obligatory, but suggest a fix/reason for the bug, --> <!--- or ideas how to implement the addition or change --> ### 🔦 Context <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> This is blocker bug. ### 💻 Code Sample <!-- If you are seeing an error, please provide a code repository, gist or sample files to reproduce the issue --> ### 🌍 Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> | Software | Version(s) | | ---------------- | ---------- | | Parcel | 1.62 | Node | 8.1.3 | npm/Yarn | npm 5.7.1 | Operating System | macOS High Sierra <!-- Love parcel? Please consider supporting our collective: 👉 https://opencollective.com/parcel/donate -->
1.0
BUG: postcss module generates wrong hashes - <!--- Thanks for filing an issue 😄 ! Before you submit, please read the following: Search open/closed issues before submitting since someone might have asked the same thing before! --> This a 🐛 bug report <!--- Provide a general summary of the issue in the title above --> ### 🎛 Configuration (.babelrc, package.json, cli command) <!--- If describing a bug, tell us what your babel configuration looks like --> postcss.config.js ```js module.exports = { "modules": true, "plugins": { "autoprefixer": { "grid": true, "browsers": [ ">1%", "last 2 versions", "not ie < 10" ], "flexbox": "no-2009" }, "postcss-modules": { "generateScopedName": "[local]__[hash:base64:4]" } } } ``` package.json ```js { "name": "react-project", "version": "1.0.0", "description": "Small test", "main": "index.js", "scripts": { "start": "parcel public/index.html -p 4000 --no-cache", "build": "parcel build public/index.html --out-dir build --no-cache --public-url ./", "test": "echo \"Error: no test specified\" && exit 1" }, "repository": {}, "author": "", "license": "ISC", "homepage": "", "dependencies": { "apollo-boost": "^0.1.2", "babel-eslint": "^7.2.3", "classnames": "^2.2.5", "graphql": "^0.13.1", "react": "^16.2.0", "react-apollo": "^2.0.4", "react-dom": "^16.2.0", "react-loadable": "^5.3.1", "react-router-dom": "^4.2.2" }, "devDependencies": { "autoprefixer": "^8.1.0", "babel-core": "^6.26.0", "babel-plugin-module-resolver": "^3.1.0", "babel-preset-react-app": "^3.1.1", "eslint": "^4.18.2", "eslint-config-react-app": "^2.1.0", "eslint-plugin-flowtype": "^2.46.1", "eslint-plugin-import": "^2.9.0", "eslint-plugin-jsx-a11y": "^5.1.1", "eslint-plugin-react": "^7.7.0", "node-sass": "^4.7.2", "postcss-modules": "^1.1.0", "postcss-scss": "^1.0.4" } } ``` .babelrc ```js { "plugins": [ ["module-resolver", { "root": [ "./src/**/components" ], "alias": {} }] ], "presets": [ "react-app" ] } ``` ### 🤔 Expected Behavior <!--- If you're describing a bug, tell us what should happen --> <!--- If you're suggesting a change/improvement, tell us how it should work --> Classnames in generated .css file should have hash which is 4 chars long and :global classname should not contain hash ### 😯 Current Behavior <!--- If describing a bug, tell us what happens instead of the expected behavior --> <!--- If you are seeing an error, please include the full error message and stack trace --> <!--- If suggesting a change/improvement, explain the difference from current behavior --> Seems working correctly after restarting parcel. Example: `.page-header {}` is converted to `.page-header__1x8y {}` and `:global .app-root` is converted to `.app-root {}` Classnames are generated incorrectly after changing and saving css file `.page-header {}` is converted to `._page-header__1x8y_1dxnl_1 {}` and `:global .app-root` is converted to `._app-root_1dxnl_12` When saving again then classnames are converted as shown below `.page-header {}` -> `._page-header__1x8y_15rqa_1 {}` `:global .app-root` -> `._app-root_15rqa_12` ### 💁 Possible Solution <!--- Not obligatory, but suggest a fix/reason for the bug, --> <!--- or ideas how to implement the addition or change --> ### 🔦 Context <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> This is blocker bug. ### 💻 Code Sample <!-- If you are seeing an error, please provide a code repository, gist or sample files to reproduce the issue --> ### 🌍 Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> | Software | Version(s) | | ---------------- | ---------- | | Parcel | 1.62 | Node | 8.1.3 | npm/Yarn | npm 5.7.1 | Operating System | macOS High Sierra <!-- Love parcel? Please consider supporting our collective: 👉 https://opencollective.com/parcel/donate -->
process
bug postcss module generates wrong hashes thanks for filing an issue 😄 before you submit please read the following search open closed issues before submitting since someone might have asked the same thing before this a 🐛 bug report 🎛 configuration babelrc package json cli command postcss config js js module exports modules true plugins autoprefixer grid true browsers last versions not ie flexbox no postcss modules generatescopedname package json js name react project version description small test main index js scripts start parcel public index html p no cache build parcel build public index html out dir build no cache public url test echo error no test specified exit repository author license isc homepage dependencies apollo boost babel eslint classnames graphql react react apollo react dom react loadable react router dom devdependencies autoprefixer babel core babel plugin module resolver babel preset react app eslint eslint config react app eslint plugin flowtype eslint plugin import eslint plugin jsx eslint plugin react node sass postcss modules postcss scss babelrc js plugins module resolver root src components alias presets react app 🤔 expected behavior classnames in generated css file should have hash which is chars long and global classname should not contain hash 😯 current behavior seems working correctly after restarting parcel example page header is converted to page header and global app root is converted to app root classnames are generated incorrectly after changing and saving css file page header is converted to page header and global app root is converted to app root when saving again then classnames are converted as shown below page header page header global app root app root 💁 possible solution 🔦 context this is blocker bug 💻 code sample 🌍 your environment software version s parcel node npm yarn npm operating system macos high sierra love parcel please consider supporting our collective 👉
1
61,893
8,560,975,941
IssuesEvent
2018-11-09 04:07:17
mysteriumnetwork/docs
https://api.github.com/repos/mysteriumnetwork/docs
opened
Sync wiki to the node repo
documentation enhancement
Docs repo now synced to the docs repo wiki. We need to move this wiki to the node repo as a main repo for the product. > Make sure that all wiki pages from the node repo moved to the docs or some other places. > It will be overwritten by this documentation.
1.0
Sync wiki to the node repo - Docs repo now synced to the docs repo wiki. We need to move this wiki to the node repo as a main repo for the product. > Make sure that all wiki pages from the node repo moved to the docs or some other places. > It will be overwritten by this documentation.
non_process
sync wiki to the node repo docs repo now synced to the docs repo wiki we need to move this wiki to the node repo as a main repo for the product make sure that all wiki pages from the node repo moved to the docs or some other places it will be overwritten by this documentation
0
6,449
9,546,284,614
IssuesEvent
2019-05-01 19:29:21
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
closed
Transcript - should be required
Apply Process State Dept.
Who: Internship applicants What: required to provide a transcript Why: required by State Once I have uploaded a transcript, I still must go select the radio button to select this transcript. If I do not select the radio button, I can click Save and continue without having selected a transcript. Transcript should be required on Save and Continue. Use Case 1: I have not uploaded any transcripts and select save and continue Use Case 2: I have added a transcript but have not selected the radio button and select save and continue.
1.0
Transcript - should be required - Who: Internship applicants What: required to provide a transcript Why: required by State Once I have uploaded a transcript, I still must go select the radio button to select this transcript. If I do not select the radio button, I can click Save and continue without having selected a transcript. Transcript should be required on Save and Continue. Use Case 1: I have not uploaded any transcripts and select save and continue Use Case 2: I have added a transcript but have not selected the radio button and select save and continue.
process
transcript should be required who internship applicants what required to provide a transcript why required by state once i have uploaded a transcript i still must go select the radio button to select this transcript if i do not select the radio button i can click save and continue without having selected a transcript transcript should be required on save and continue use case i have not uploaded any transcripts and select save and continue use case i have added a transcript but have not selected the radio button and select save and continue
1
440,315
12,697,693,685
IssuesEvent
2020-06-22 12:16:17
bitgapp/eqMac
https://api.github.com/repos/bitgapp/eqMac
closed
Enhancement: Improve Volume handling
priority: high status: pending type: enhancement
**Disclaimer:** Please go through [existing feature requests](https://github.com/bitgapp/eqMac/issues?q=is%3Aopen+is%3Aissue+label%3A%22type%3A+enhancement%22) to avoid creating duplicates. If you found a similar issue please post a comment there. If you are certain your issue is not related to others then create the issue. Obvious ignorance of this rule might get you blocked from posting new issues. Please tick the box below to show that you have followed the guideline. - [x] I have checked for a similar issue and sure it hasn't been reported before. **Description:** Users are experiencing volume behaviour inconsistency as mentioned in #111 and #246 Need to stay as close to native volume behaviour of macOS
1.0
Enhancement: Improve Volume handling - **Disclaimer:** Please go through [existing feature requests](https://github.com/bitgapp/eqMac/issues?q=is%3Aopen+is%3Aissue+label%3A%22type%3A+enhancement%22) to avoid creating duplicates. If you found a similar issue please post a comment there. If you are certain your issue is not related to others then create the issue. Obvious ignorance of this rule might get you blocked from posting new issues. Please tick the box below to show that you have followed the guideline. - [x] I have checked for a similar issue and sure it hasn't been reported before. **Description:** Users are experiencing volume behaviour inconsistency as mentioned in #111 and #246 Need to stay as close to native volume behaviour of macOS
non_process
enhancement improve volume handling disclaimer please go through to avoid creating duplicates if you found a similar issue please post a comment there if you are certain your issue is not related to others then create the issue obvious ignorance of this rule might get you blocked from posting new issues please tick the box below to show that you have followed the guideline i have checked for a similar issue and sure it hasn t been reported before description users are experiencing volume behaviour inconsistency as mentioned in and need to stay as close to native volume behaviour of macos
0
17,092
3,354,409,097
IssuesEvent
2015-11-18 11:59:25
interactivethings/catalog
https://api.github.com/repos/interactivethings/catalog
closed
Allow for project-specific design
design enhancement
- Minimize eigenbranding - Icon, logo, or something comparable - Colors
1.0
Allow for project-specific design - - Minimize eigenbranding - Icon, logo, or something comparable - Colors
non_process
allow for project specific design minimize eigenbranding icon logo or something comparable colors
0
16,913
22,240,519,300
IssuesEvent
2022-06-09 04:27:04
camunda/feel-scala
https://api.github.com/repos/camunda/feel-scala
opened
Adopt cycle() function from Zeebe
type: enhancement scope: zeebe team/process-automation
**Is your feature request related to a problem? Please describe.** I can use the `cycle()` function in the FEEL engine. Currently, the function is implemented in Zeebe (see [here](https://github.com/camunda/zeebe/blob/main/feel/src/main/scala/io/camunda/zeebe/feel/impl/FeelFunctionProvider.scala#L23-L48)). It is used to create a string for repeating intervals. For maintainability, it makes sense to move the function in the FEEL engine. **Describe the solution you'd like** * expose the `cycle()` function * add documentation * extend the function for an optional start time * maybe, create a new type for intervals with properties * think about other useful extensions of this method **Related issues** * Camunda Autormation Platform 7: <!-- link the issue: https://jira.camunda.com/browse/CAM- --> * Zeebe broker: <!-- link the issue: https://github.com/zeebe-io/zeebe/issues# -->
1.0
Adopt cycle() function from Zeebe - **Is your feature request related to a problem? Please describe.** I can use the `cycle()` function in the FEEL engine. Currently, the function is implemented in Zeebe (see [here](https://github.com/camunda/zeebe/blob/main/feel/src/main/scala/io/camunda/zeebe/feel/impl/FeelFunctionProvider.scala#L23-L48)). It is used to create a string for repeating intervals. For maintainability, it makes sense to move the function in the FEEL engine. **Describe the solution you'd like** * expose the `cycle()` function * add documentation * extend the function for an optional start time * maybe, create a new type for intervals with properties * think about other useful extensions of this method **Related issues** * Camunda Autormation Platform 7: <!-- link the issue: https://jira.camunda.com/browse/CAM- --> * Zeebe broker: <!-- link the issue: https://github.com/zeebe-io/zeebe/issues# -->
process
adopt cycle function from zeebe is your feature request related to a problem please describe i can use the cycle function in the feel engine currently the function is implemented in zeebe see it is used to create a string for repeating intervals for maintainability it makes sense to move the function in the feel engine describe the solution you d like expose the cycle function add documentation extend the function for an optional start time maybe create a new type for intervals with properties think about other useful extensions of this method related issues camunda autormation platform zeebe broker
1
347,471
24,888,591,294
IssuesEvent
2022-10-28 09:54:25
isanidiot/ped
https://api.github.com/repos/isanidiot/ped
opened
Inaccurate examples in UG
type.DocumentationBug severity.Low
Some examples given in UG doesn't follow the correct format. ![image.png](https://raw.githubusercontent.com/isanidiot/ped/main/files/807cee21-26ee-487a-be06-5e935ec02efa.png) ![image.png](https://raw.githubusercontent.com/isanidiot/ped/main/files/04ed3bc9-cf74-43f1-9142-f6ad0262f3f2.png) <!--session: 1666945844337-67510e57-6d90-4772-9916-3787fce01316--> <!--Version: Web v3.4.4-->
1.0
Inaccurate examples in UG - Some examples given in UG doesn't follow the correct format. ![image.png](https://raw.githubusercontent.com/isanidiot/ped/main/files/807cee21-26ee-487a-be06-5e935ec02efa.png) ![image.png](https://raw.githubusercontent.com/isanidiot/ped/main/files/04ed3bc9-cf74-43f1-9142-f6ad0262f3f2.png) <!--session: 1666945844337-67510e57-6d90-4772-9916-3787fce01316--> <!--Version: Web v3.4.4-->
non_process
inaccurate examples in ug some examples given in ug doesn t follow the correct format
0
35,185
4,965,321,392
IssuesEvent
2016-12-04 07:40:18
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
github.com/cockroachdb/cockroach/pkg/internal/client: TestClientGetAndPutProto failed under stress
Robot test-failure
SHA: https://github.com/cockroachdb/cockroach/commits/53ae9bb78ba376b9ec314caa5081d0ade866f3cd Parameters: ``` COCKROACH_PROPOSER_EVALUATED_KV=true TAGS=stress GOFLAGS= ``` Stress build found a failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=79293&tab=buildLog ``` I161204 07:40:15.352822 793 storage/store.go:1223 [n1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available I161204 07:40:15.352842 793 server/node.go:432 [n1] initialized store [n1,s1]: {Capacity:536870912 Available:536870912 RangeCount:1 LeaseCount:1} I161204 07:40:15.352857 793 server/node.go:317 [n1] node ID 1 initialized I161204 07:40:15.356225 793 gossip/gossip.go:290 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:36350" > attrs:<> locality:<> I161204 07:40:15.356289 793 storage/stores.go:296 [n1] read 0 node addresses from persistent storage I161204 07:40:15.356335 793 server/node.go:562 [n1] connecting to gossip network to verify cluster ID... I161204 07:40:15.356355 793 server/node.go:582 [n1] node connected via gossip and verified as part of cluster "bf3738a3-d325-4568-ab46-754e93c99ba6" I161204 07:40:15.356380 793 server/node.go:367 [n1] node=1: started with [[]=] engine(s) and attributes [] I161204 07:40:15.356409 793 sql/executor.go:291 [n1] creating distSQLPlanner with address {tcp 127.0.0.1:36350} I161204 07:40:15.361206 869 sql/event_log.go:95 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:36350} Attrs: Locality:} ClusterID:bf3738a3-d325-4568-ab46-754e93c99ba6 StartedAt:1480837215356362559} I161204 07:40:15.361721 862 storage/split_queue.go:103 [n1,split,s1,r1/1:/M{in-ax},@c42018ad80] splitting at keys [/Table/11/0 /Table/12/0 /Table/13/0 /Table/14/0] I161204 07:40:15.364411 862 storage/replica_command.go:2369 [n1,split,s1,r1/1:/M{in-ax},@c42018ad80] initiating a split of this range at key /Table/11 [r2] W161204 07:40:15.116065 914 util/hlc/hlc.go:145 backward time jump detected (-0.249007 seconds) I161204 07:40:15.122162 793 server/server.go:633 [n1] starting https server at 127.0.0.1:46912 I161204 07:40:15.122178 793 server/server.go:634 [n1] starting grpc/postgres server at 127.0.0.1:36350 I161204 07:40:15.122185 793 server/server.go:635 [n1] advertising CockroachDB node at 127.0.0.1:36350 E161204 07:40:15.131826 862 storage/queue.go:598 [n1,split,s1,r1/1:/{Min-Table/11},@c42018ad80] unable to split [n1,s1,r1/1:/{Min-Table/11}] at key "/Table/12/0": key range /Table/12/0-/Table/12/0 outside of bounds of range /Min-/Max I161204 07:40:15.131944 862 storage/split_queue.go:103 [n1,split,s1,r2/1:/{Table/11-Max},@c4202c6900] splitting at keys [/Table/12/0 /Table/13/0 /Table/14/0] I161204 07:40:15.131997 862 storage/replica_command.go:2369 [n1,split,s1,r2/1:/{Table/11-Max},@c4202c6900] initiating a split of this range at key /Table/12 [r3] E161204 07:40:15.136980 862 storage/queue.go:598 [n1,split,s1,r2/1:/Table/1{1-2},@c4202c6900] unable to split [n1,s1,r2/1:/Table/1{1-2}] at key "/Table/13/0": key range /Table/13/0-/Table/13/0 outside of bounds of range /Table/11-/Max I161204 07:40:15.137729 862 storage/split_queue.go:103 [n1,split,s1,r3/1:/{Table/12-Max},@c420099200] splitting at keys [/Table/13/0 /Table/14/0] I161204 07:40:15.137766 862 storage/replica_command.go:2369 [n1,split,s1,r3/1:/{Table/12-Max},@c420099200] initiating a split of this range at key /Table/13 [r4] E161204 07:40:15.142334 862 storage/queue.go:598 [n1,split,s1,r3/1:/Table/1{2-3},@c420099200] unable to split [n1,s1,r3/1:/Table/1{2-3}] at key "/Table/14/0": key range /Table/14/0-/Table/14/0 outside of bounds of range /Table/12-/Max I161204 07:40:15.142446 862 storage/split_queue.go:103 [n1,split,s1,r4/1:/{Table/13-Max},@c420510000] splitting at keys [/Table/14/0] I161204 07:40:15.142476 862 storage/replica_command.go:2369 [n1,split,s1,r4/1:/{Table/13-Max},@c420510000] initiating a split of this range at key /Table/14 [r5] F161204 07:40:15.148452 890 server/server.go:160 [n1] fewer than half the known nodes are within the maximum offset of 250ms (0 of 1) goroutine 890 [running]: github.com/cockroachdb/cockroach/pkg/util/log.getStacks(0x301, 0x8d93531, 0x2872aa0, 0xc420ad7a88) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:849 +0xa5 github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).outputLogEntry(0x2874000, 0xc400000004, 0x20a38bd, 0x10, 0xa0, 0xc4202ef800, 0x54) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:714 +0x7b1 github.com/cockroachdb/cockroach/pkg/util/log.addStructured(0x2b53e1e56a78, 0xc420bce930, 0x4, 0x2, 0x0, 0x0, 0xc420ad7d18, 0x1, 0x1) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/structured.go:146 +0x296 github.com/cockroachdb/cockroach/pkg/util/log.logDepth(0x2b53e1e56a78, 0xc420bce930, 0x1, 0x4, 0x0, 0x0, 0xc420ad7d18, 0x1, 0x1) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/log.go:88 +0x8c github.com/cockroachdb/cockroach/pkg/util/log.Fatal(0x2b53e1e56a78, 0xc420bce930, 0xc420ad7d18, 0x1, 0x1) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/log.go:181 +0x76 github.com/cockroachdb/cockroach/pkg/server.NewServer.func1() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:160 +0xcc github.com/cockroachdb/cockroach/pkg/rpc.(*Context).runHeartbeat(0xc420926600, 0xc42000ac00, 0xc42088ed70, 0xf, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/rpc/context.go:323 +0x47f github.com/cockroachdb/cockroach/pkg/rpc.(*Context).GRPCDial.func1.1.1() /go/src/github.com/cockroachdb/cockroach/pkg/rpc/context.go:227 +0x69 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420b9f920) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 1 [chan receive]: testing.(*T).Run(0xc4203240c0, 0x188dc17, 0x18, 0x1a3ea58, 0xc42027bc01) /usr/local/go/src/testing/testing.go:647 +0x316 testing.RunTests.func1(0xc4203240c0) /usr/local/go/src/testing/testing.go:793 +0x6d testing.tRunner(0xc4203240c0, 0xc420045e00) /usr/local/go/src/testing/testing.go:610 +0x81 testing.RunTests(0x1a41d40, 0x2395ea0, 0x31, 0x31, 0x28af6d0) /usr/local/go/src/testing/testing.go:799 +0x2f5 testing.(*M).Run(0xc420045ef8, 0x28af200) /usr/local/go/src/testing/testing.go:743 +0x85 github.com/cockroachdb/cockroach/pkg/internal/client_test.TestMain(0xc420045ef8) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/main_test.go:35 +0x76 main.main() github.com/cockroachdb/cockroach/pkg/internal/client/_test/_testmain.go:150 +0xc6 goroutine 17 [syscall, locked to thread]: runtime.goexit() /usr/local/go/src/runtime/asm_amd64.s:2086 +0x1 goroutine 793 [select]: github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendToReplicas(0xc420bdb810, 0x2b53e1e56ab8, 0xc420872c00, 0x1dcd6500, 0x0, 0x1, 0xc420c5a560, 0x1, 0x1, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1076 +0xddd github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendRPC(0xc420bdb810, 0x2b53e1e56a78, 0xc420218840, 0x1, 0xc420c5a560, 0x1, 0x1, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:340 +0x30b github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendSingleRange(0xc420bdb810, 0x2b53e1e56a78, 0xc420218840, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:409 +0x16e github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendPartialBatch(0xc420bdb810, 0x2b53e1e56a78, 0xc420218840, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:856 +0x29f github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges(0xc420bdb810, 0x2b53e1e56a78, 0xc420218840, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:738 +0x41f github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send(0xc420bdb810, 0x2b53e1e56a78, 0xc420218840, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:553 +0x2b3 github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send(0xc4203263c0, 0x2b53e1e56a78, 0xc4202187e0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:416 +0x7e9 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).send(0xc420baa6e0, 0x2b53e1e511d0, 0xc420014600, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:513 +0xbf github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Run.func1(0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:447 +0x6e github.com/cockroachdb/cockroach/pkg/internal/client.sendAndFill(0xc420c5a3e0, 0xc420c58c00, 0x0, 0xc420c5a3c0) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:418 +0x168 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Run(0xc420baa6e0, 0x2b53e1e511d0, 0xc420014600, 0xc420c58c00, 0xc420c5a3c0, 0xc4200ce320) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:449 +0xac github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).scan(0xc420baa6e0, 0x2b53e1e511d0, 0xc420014600, 0x17b2720, 0xc420c5a3a0, 0x17b2720, 0xc420c5a3c0, 0x0, 0x20c5a300, 0xc42060b4c8, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:307 +0x105 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Scan(0xc420baa6e0, 0x2b53e1e511d0, 0xc420014600, 0x17b2720, 0xc420c5a3a0, 0x17b2720, 0xc420c5a3c0, 0x0, 0x0, 0x1, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:318 +0x9a github.com/cockroachdb/cockroach/pkg/server.WaitForInitialSplits.func1(0xecfd5bf5f, 0xc407935070) /go/src/github.com/cockroachdb/cockroach/pkg/server/testserver.go:314 +0x14c github.com/cockroachdb/cockroach/pkg/util.RetryForDuration(0x2540be400, 0xc42060b620, 0x0, 0x239e180) /go/src/github.com/cockroachdb/cockroach/pkg/util/testing.go:131 +0xcd github.com/cockroachdb/cockroach/pkg/server.WaitForInitialSplits(0xc420baa6e0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/server/testserver.go:322 +0x67 github.com/cockroachdb/cockroach/pkg/server.(*TestServer).WaitForInitialSplits(0xc420bde110, 0x2b53e1e56a78, 0xc420a16600) /go/src/github.com/cockroachdb/cockroach/pkg/server/testserver.go:304 +0x3b github.com/cockroachdb/cockroach/pkg/server.(*TestServer).Start(0xc420bde110, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/server/testserver.go:283 +0x1e1 github.com/cockroachdb/cockroach/pkg/testutils/serverutils.StartServerRaw(0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/testutils/serverutils/test_server_shim.go:148 +0x134 github.com/cockroachdb/cockroach/pkg/testutils/serverutils.StartServer(0x23c0020, 0xc4204160c0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/testutils/serverutils/test_server_shim.go:118 +0x63 github.com/cockroachdb/cockroach/pkg/internal/client_test.TestClientGetAndPutProto(0xc4204160c0) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/client_test.go:375 +0xd8 testing.tRunner(0xc4204160c0, 0x1a3ea58) /usr/local/go/src/testing/testing.go:610 +0x81 created by testing.(*T).Run /usr/local/go/src/testing/testing.go:646 +0x2ec goroutine 20 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).flushDaemon(0x2874000) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1015 +0x77 created by github.com/cockroachdb/cockroach/pkg/util/log.init.1 /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:581 +0x95 goroutine 857 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Store).startGossip.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:1278 +0x39b github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420a57b30) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 868 [select]: github.com/cockroachdb/cockroach/pkg/server.(*Node).startGossip.func1() /go/src/github.com/cockroachdb/cockroach/pkg/server/node.go:609 +0x4f0 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42048e200) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 877 [IO wait]: net.runtime_pollWait(0x2b53e1e5b990, 0x72, 0x0) /usr/local/go/src/runtime/netpoll.go:160 +0x59 net.(*pollDesc).wait(0xc420b75640, 0x72, 0xc42053b608, 0xc420014230) /usr/local/go/src/net/fd_poll_runtime.go:73 +0x38 net.(*pollDesc).waitRead(0xc420b75640, 0x23a68c0, 0xc420014230) /usr/local/go/src/net/fd_poll_runtime.go:78 +0x34 net.(*netFD).accept(0xc420b755e0, 0x0, 0x23a3780, 0xc4200ce3a0) /usr/local/go/src/net/fd_unix.go:419 +0x238 net.(*TCPListener).accept(0xc420126160, 0x5fff1e, 0xc42053b6b8, 0xf7fbbb) /usr/local/go/src/net/tcpsock_posix.go:132 +0x2e net.(*TCPListener).Accept(0xc420126160, 0x1a40728, 0xc420be9200, 0x23bd1e0, 0xc420536038) /usr/local/go/src/net/tcpsock.go:222 +0x49 github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux.(*cMux).Serve(0xc420be9200, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux/cmux.go:124 +0xa4 github.com/cockroachdb/cockroach/pkg/server.(*Server).Start.func11() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:641 +0x2f github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420834d20) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 800 [chan receive]: github.com/cockroachdb/cockroach/pkg/server.(*Server).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:463 +0x71 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420a16870) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 799 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/netutil.MakeServer.func2() /go/src/github.com/cockroachdb/cockroach/pkg/util/netutil/net.go:98 +0x70 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42076a960) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 861 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:474 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493fe0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 863 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:474 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42038c020) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 804 [chan receive]: github.com/cockroachdb/cockroach/pkg/server.(*Server).Start.func6() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:490 +0x6d github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42076aba0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 797 [select]: github.com/cockroachdb/cockroach/pkg/sql.(*LeaseManager).RefreshLeases.func1() /go/src/github.com/cockroachdb/cockroach/pkg/sql/lease.go:1160 +0xcbf github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420bcfb00) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 834 [chan receive]: github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:171 +0x55 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493660) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 980 [select]: github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*http2Client).controller(0xc420812200) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/http2_client.go:999 +0x59d created by github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.newHTTP2Client /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/http2_client.go:233 +0xe78 goroutine 809 [select]: github.com/cockroachdb/cockroach/pkg/gossip.(*Gossip).manage.func1() /go/src/github.com/cockroachdb/cockroach/pkg/gossip/gossip.go:1031 +0x394 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420a0f800) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 836 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc400000033) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc4204936a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 865 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:474 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42038c060) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 849 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc40000002e) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493840) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 802 [chan receive]: github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux.(*muxListener).Accept(0xc42076aae0, 0xc420680090, 0x16e3840, 0x2383b80, 0x17851e0) <autogenerated>:9 +0x7f net/http.(*Server).Serve(0xc420926780, 0x23b3bc0, 0xc42076aae0, 0x0, 0x0) /usr/local/go/src/net/http/server.go:2273 +0x1ce github.com/cockroachdb/cockroach/pkg/server.(*Server).Start.func4() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:479 +0x3c github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42076ab40) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 846 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc40000002b) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc4204937e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 848 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc40000002d) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493820) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 862 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:474 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42038c000) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 845 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc400000030) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc4204937c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 867 [select]: github.com/cockroachdb/cockroach/pkg/server.(*Node).startComputePeriodicMetrics.func1() /go/src/github.com/cockroachdb/cockroach/pkg/server/node.go:650 +0x2bd github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42048e1e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 803 [chan receive]: github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux.(*muxListener).Accept(0xc42076ab00, 0xc420126180, 0xc420680060, 0x16e3840, 0x2383b80) <autogenerated>:9 +0x7f crypto/tls.(*listener).Accept(0xc42076ab60, 0xc420680030, 0x16e3840, 0x2383b80, 0x17851e0) /usr/local/go/src/crypto/tls/tls.go:52 +0x37 net/http.(*Server).Serve(0xc420926700, 0x23ab980, 0xc42076ab60, 0x0, 0x0) /usr/local/go/src/net/http/server.go:2273 +0x1ce github.com/cockroachdb/cockroach/pkg/server.(*Server).Start.func5() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:486 +0x42 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42076ab80) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 840 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc400000036) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493720) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 839 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc400000035) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493700) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 838 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc400000034) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc4204936e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 801 [IO wait]: net.runtime_pollWait(0x2b53e1e5be10, 0x72, 0x0) /usr/local/go/src/runtime/netpoll.go:160 +0x59 net.(*pollDesc).wait(0xc420b756b0, 0x72, 0xc420031e08, 0xc420014230) /usr/local/go/src/net/fd_poll_runtime.go:73 +0x38 net.(*pollDesc).waitRead(0xc420b756b0, 0x23a68c0, 0xc420014230) /usr/local/go/src/net/fd_poll_runtime.go:78 +0x34 net.(*netFD).accept(0xc420b75650, 0x0, 0x23a3780, 0xc420512040) /usr/local/go/src/net/fd_unix.go:419 +0x238 net.(*TCPListener).accept(0xc420126180, 0xc420031ef0, 0xc420031f00, 0x620690) /usr/local/go/src/net/tcpsock_posix.go:132 +0x2e net.(*TCPListener).Accept(0xc420126180, 0x1a40720, 0xc420be9280, 0xc420a76000, 0xc420031f88) /usr/local/go/src/net/tcpsock.go:222 +0x49 github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux.(*cMux).Serve(0xc420be9280, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux/cmux.go:124 +0xa4 github.com/cockroachdb/cockroach/pkg/server.(*Server).Start.func3() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:475 +0x2f github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42076ab20) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 795 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*StorePool).start.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store_pool.go:356 +0x3de github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42076a220) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 794 [chan receive]: github.com/cockroachdb/cockroach/pkg/rpc.NewContext.func1() /go/src/github.com/cockroachdb/cockroach/pkg/rpc/context.go:137 +0x76 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42082bee0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 907 [select]: github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*http2Server).controller(0xc420716bd0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/http2_server.go:673 +0x69c created by github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.newHTTP2Server /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/http2_server.go:139 +0x767 goroutine 852 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Store).raftTickLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3371 +0x334 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420b61c80) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 841 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc400000037) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493740) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 873 [select]: github.com/cockroachdb/cockroach/pkg/server.(*Node).startWriteSummaries.func1() /go/src/github.com/cockroachdb/cockroach/pkg/server/node.go:687 +0x26d github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc4208d03c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 875 [select]: github.com/cockroachdb/cockroach/pkg/sql.(*Executor).Start.func1() /go/src/github.com/cockroachdb/cockroach/pkg/sql/executor.go:301 +0x15f github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42048e340) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 844 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc40000002a) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc4204937a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 808 [select]: github.com/cockroachdb/cockroach/pkg/gossip.(*Gossip).bootstrap.func1() /go/src/github.com/cockroachdb/cockroach/pkg/gossip/gossip.go:993 +0x4c9 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420a0f7f0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 853 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Store).startCoalescedHeartbeatsLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3402 +0x167 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420b61c90) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 837 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc400000031) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc4204936c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 570 [IO wait]: net.runtime_pollWait(0x2b53e1e5bc90, 0x72, 0x8) /usr/local/go/src/runtime/netpoll.go:160 +0x59 net.(*pollDesc).wait(0xc420b74220, 0x72, 0xc420a9f588, 0xc420014230) /usr/local/go/src/net/fd_poll_runtime.go:73 +0x38 net.(*pollDesc).waitRead(0xc420b74220, 0x23a68c0, 0xc420014230) /usr/local/go/src/net/fd_poll_runtime.go:78 +0x34 net.(*netFD).Read(0xc420b741c0, 0xc42020dc00, 0x400, 0x400, 0x0, 0x23a68c0, 0xc420014230) /usr/local/go/src/net/fd_unix.go:243 +0x1a1 net.(*conn).Read(0xc420536038, 0xc42020dc00, 0x400, 0x400, 0x0, 0x0, 0x0) /usr/local/go/src/net/net.go:173 +0x70 github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux.(*MuxConn).Read(0xc4201798c0, 0xc42020dc00, 0x400, 0x400, 0x17d4820, 0x0, 0x2b53e1ed6280) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux/cmux.go:218 +0x113 crypto/tls.(*block).readFromUntil(0xc4204b0180, 0x2b53e1ed6280, 0xc4201798c0, 0x5, 0xc4201798c0, 0x0) /usr/local/go/src/crypto/tls/conn.go:476 +0x91 crypto/tls.(*Conn).readRecord(0xc420880380, 0x1a42617, 0xc420880488, 0x0) /usr/local/go/src/crypto/tls/conn.go:578 +0xc4 crypto/tls.(*Conn).Read(0xc420880380, 0xc420366000, 0x8000, 0x8000, 0x0, 0x0, 0x0) /usr/local/go/src/crypto/tls/conn.go:1113 +0x116 bufio.(*Reader).fill(0xc4202ef500) /usr/local/go/src/bufio/bufio.go:97 +0x10c bufio.(*Reader).Read(0xc4202ef500, 0xc420a46278, 0x9, 0x9, 0x9, 0x18, 0x28) /usr/local/go/src/bufio/bufio.go:209 +0x1bc io.ReadAtLeast(0x239d800, 0xc4202ef500, 0xc420a46278, 0x9, 0x9, 0x9, 0x5dc238, 0x30, 0xc420a9fad8) /usr/local/go/src/io/io.go:307 +0xa4 io.ReadFull(0x239d800, 0xc4202ef500, 0xc420a46278, 0x9, 0x9, 0x6af7f98fa89bc601, 0x6a00000000000000, 0x0) /usr/local/go/src/io/io.go:325 +0x58 github.com/cockroachdb/cockroach/vendor/golang.org/x/net/http2.readFrameHeader(0xc420a46278, 0x9, 0x9, 0x239d800, 0xc4202ef500, 0x0, 0xc400000000, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/net/http2/frame.go:237 +0x7b github.com/cockroachdb/cockroach/vendor/golang.org/x/net/http2.(*Framer).ReadFrame(0xc420a46240, 0x23a9200, 0xc4204b0ae0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/net/http2/frame.go:464 +0xa4 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*framer).readFrame(0xc4204b0750, 0xc4204b0ae0, 0xc4204b0ae0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/http_util.go:505 +0x2f github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*http2Server).HandleStreams(0xc420716bd0, 0xc4204b07e0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/http2_server.go:260 +0x200 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).serveStreams(0xc420aa99a0, 0x23bbf20, 0xc420716bd0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:424 +0x160 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).serveNewHTTP2Transport(0xc420aa99a0, 0x23bc340, 0xc420880380, 0x23a5b00, 0xc4204fe4d0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:411 +0x3da github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).handleRawConn(0xc420aa99a0, 0x23bd0c0, 0xc4201798c0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:388 +0x470 created by github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).Serve /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:357 +0x1a9 goroutine 889 [select]: github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*addrConn).transportMonitor(0xc420329040) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/clientconn.go:720 +0x63e github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*ClientConn).resetAddrConn.func1(0xc420329040) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/clientconn.go:479 +0x1de created by github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*ClientConn).resetAddrConn /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/clientconn.go:480 +0x3c7 goroutine 874 [select]: github.com/cockroachdb/cockroach/pkg/sql.(*SchemaChangeManager).Start.func1() /go/src/github.com/cockroachdb/cockroach/pkg/sql/schema_changer.go:751 +0x12b3 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42048e2e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 884 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*replicaScanner).waitAndProcess(0xc420326aa0, 0x2b53e1e56a78, 0xc420b9edb0, 0xecfd5bf5f, 0x153c4fd6, 0x2872aa0, 0xc420be8980, 0xc420716000, 0xc42018ad80, 0xecfd5bf5f) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scanner.go:207 +0x254 github.com/cockroachdb/cockroach/pkg/storage.(*replicaScanner).scanLoop.func1.1(0xc42018ad80, 0xc420a56f90) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scanner.go:269 +0x8e github.com/cockroachdb/cockroach/pkg/storage.(*storeReplicaVisitor).Visit(0xc420a56f90, 0xc42034ddb0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:324 +0x303 github.com/cockroachdb/cockroach/pkg/storage.(*replicaScanner).scanLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scanner.go:271 +0x2ca github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42038c0c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 901 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*idAllocator).start.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/id_alloc.go:133 +0x7c8 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc4204741a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 842 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc400000028) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493760) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 774 [select]: github.com/cockroachdb/cockroach/pkg/util/retry.(*Retry).Next(0xc4204c66e0, 0x770065) /go/src/github.com/cockroachdb/cockroach/pkg/util/retry/retry.go:128 +0x1a5 github.com/cockroachdb/cockroach/pkg/storage.(*Store).Send.func2(0xc4204c66e0, 0x148cfd9a000344ea) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:2424 +0x8d github.com/cockroachdb/cockroach/pkg/storage.(*Store).Send(0xc4204d9500, 0x2b53e1e56a78, 0xc4202193e0, 0x148cfd9a000344ea, 0x39, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:2431 +0x67e github.com/cockroachdb/cockroach/pkg/storage.(*Stores).Send(0xc420a1d1a0, 0x2b53e1e56a78, 0xc420219350, 0x0, 0x0, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/stores.go:187 +0x202 github.com/cockroachdb/cockroach/pkg/server.(*Node).batchInternal.func1(0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/server/node.go:818 +0x2a1 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunTaskWithErr(0xc420716000, 0xc420bcdcc8, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:245 +0x100 github.com/cockroachdb/cockroach/pkg/server.(*Node).batchInternal(0xc420539340, 0x2b53e1e56ab8, 0xc420872c00, 0xc42016cc08, 0xc420872c00, 0x100, 0x5dba0e) /go/src/github.com/cockroachdb/cockroach/pkg/server/node.go:829 +0x1ed github.com/cockroachdb/cockroach/pkg/server.(*Node).Batch(0xc420539340, 0x2b53e1e56ab8, 0xc420872c00, 0xc42016cc08, 0x0, 0x0, 0xc420447838) /go/src/github.com/cockroachdb/cockroach/pkg/server/node.go:851 +0xa0 github.com/cockroachdb/cockroach/pkg/kv.(*grpcTransport).SendNext.func1(0x239e8c0, 0xc420539340, 0xc4204c65f0, 0xc42016cbe0, 0xc4202a0420) /go/src/github.com/cockroachdb/cockroach/pkg/kv/transport.go:185 +0x64 created by github.com/cockroachdb/cockroach/pkg/kv.(*grpcTransport).SendNext /go/src/github.com/cockroachdb/cockroach/pkg/kv/transport.go:188 +0x1d6 goroutine 979 [IO wait]: net.runtime_pollWait(0x2b53e1e5c050, 0x72, 0x7) /usr/local/go/src/runtime/netpoll.go:160 +0x59 net.(*pollDesc).wait(0xc420b74450, 0x72, 0xc420341828, 0xc420014230) /usr/local/go/src/net/fd_poll_runtime.go:73 +0x38 net.(*pollDesc).waitRead(0xc420b74450, 0x23a68c0, 0xc420014230) /usr/local/go/src/net/fd_poll_runtime.go:78 +0x34 net.(*netFD).Read(0xc420b743f0, 0xc420d94400, 0x400, 0x400, 0x0, 0x23a68c0, 0xc420014230) /usr/local/go/src/net/fd_unix.go:243 +0x1a1 net.(*conn).Read(0xc420126010, 0xc420d94400, 0x400, 0x400, 0x0, 0x0, 0x0) /usr/local/go/src/net/net.go:173 +0x70 crypto/tls.(*block).readFromUntil(0xc4202d3920, 0x2b53e3ba3078, 0xc420126010, 0x5, 0xc420126010, 0x0) /usr/local/go/src/crypto/tls/conn.go:476 +0x91 crypto/tls.(*Conn).readRecord(0xc4204d8700, 0x1a42617, 0xc4204d8808, 0xc420816970) /usr/local/go/src/crypto/tls/conn.go:578 +0xc4 crypto/tls.(*Conn).Read(0xc4204d8700, 0xc420d78000, 0x8000, 0x8000, 0x0, 0x0, 0x0) /usr/local/go/src/crypto/tls/conn.go:1113 +0x116 bufio.(*Reader).fill(0xc4204d18c0) /usr/local/go/src/bufio/bufio.go:97 +0x10c bufio.(*Reader).Read(0xc4204d18c0, 0xc420883838, 0x9, 0x9, 0xc420341ca8, 0x620790, 0xc420476100) /usr/local/go/src/bufio/bufio.go:209 +0x1bc io.ReadAtLeast(0x239d800, 0xc4204d18c0, 0xc420883838, 0x9, 0x9, 0x9, 0xc420484760, 0xc420484758, 0x23a29c0) /usr/local/go/src/io/io.go:307 +0xa4 io.ReadFull(0x239d800, 0xc4204d18c0, 0xc420883838, 0x9, 0x9, 0xc420341eb0, 0xb01090, 0xc420233950) /usr/local/go/src/io/io.go:325 +0x58 github.com/cockroachdb/cockroach/vendor/golang.org/x/net/http2.readFrameHeader(0xc420883838, 0x9, 0x9, 0x239d800, 0xc4204d18c0, 0x0, 0xc400000000, 0xc420816960, 0xc) /go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/net/http2/frame.go:237 +0x7b github.com/cockroachdb/cockroach/vendor/golang.org/x/net/http2.(*Framer).ReadFrame(0xc420883800, 0x0, 0x0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/net/http2/frame.go:464 +0xa4 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*framer).readFrame(0xc420484690, 0xc4204b0de0, 0xc4204b0de0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/http_util.go:505 +0x2f github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*http2Client).reader(0xc420812200) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/http2_client.go:921 +0xb6 created by github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.newHTTP2Client /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/http2_client.go:203 +0x8b1 goroutine 883 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:474 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42038c0a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 870 [select]: github.com/cockroachdb/cockroach/pkg/util/retry.(*Retry).Next(0xc4208aa378, 0xc420b9f0b0) /go/src/github.com/cockroachdb/cockroach/pkg/util/retry/retry.go:128 +0x1a5 github.com/cockroachdb/cockroach/pkg/kv.(*RangeIterator).Seek(0xc4208aa6d0, 0x2b53e1e56a78, 0xc420b9f0b0, 0xc42038c180, 0xc, 0x14, 0xc4208aa400) /go/src/github.com/cockroachdb/cockroach/pkg/kv/range_iter.go:154 +0x212 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges(0xc420bdb810, 0x2b53e1e56a78, 0xc420b9f0b0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc420324b40, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:653 +0x199 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send(0xc420bdb810, 0x2b53e1e56a78, 0xc420b9f0b0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc420324b40, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:553 +0x2b3 github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send(0xc4203263c0, 0x2b53e1e56a78, 0xc420b9f0b0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc420324a80, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:416 +0x7e9 github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).sendInternal(0xc4204af340, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc4204af388, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:587 +0xfd github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).send(0xc4204af340, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:711 +0x498 github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).(github.com/cockroachdb/cockroach/pkg/internal/client.send)-fm(0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:313 +0x58 github.com/cockroachdb/cockroach/pkg/internal/client.sendAndFill(0xc4208ab610, 0xc420bf3200, 0x0, 0xc42084f720) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:418 +0x168 github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Run(0xc4204af340, 0xc420bf3200, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:313 +0xc6 github.com/cockroachdb/cockroach/pkg/storage.(*NodeLiveness).updateLiveness.func1(0xc4204af340, 0x0, 0x1) /go/src/github.com/cockroachdb/cockroach/pkg/storage/node_liveness.go:329 +0x3b2 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn.func1(0xc4204af340, 0xc42088e530, 0x4000000000000000, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:468 +0x2f github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Exec(0xc4204af340, 0xc420380101, 0x0, 0xc42088e520, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:520 +0x215 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn(0xc420baa6e0, 0x2b53e1e56a78, 0xc420b9f0b0, 0xc42038c160, 0xc420a03ad8, 0x620790) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:469 +0x1fa github.com/cockroachdb/cockroach/pkg/storage.(*NodeLiveness).updateLiveness(0xc420bdb8c0, 0x2b53e1e56a78, 0xc420b9f0b0, 0x100000001, 0xc42038c120, 0x0, 0xc4208abbd8, 0x0, 0x2faf080) /go/src/github.com/cockroachdb/cockroach/pkg/storage/node_liveness.go:330 +0xbc github.com/cockroachdb/cockroach/pkg/storage.(*NodeLiveness).heartbeat(0xc420bdb8c0, 0x2b53e1e56a78, 0xc420b9f0b0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/node_liveness.go:194 +0x2f0 github.com/cockroachdb/cockroach/pkg/storage.(*NodeLiveness).StartHeartbeat.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/node_liveness.go:146 +0x2e7 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42048e2a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 796 [select]: github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).printStatsLoop(0xc4203263c0, 0x2b53e1e56a78, 0xc420680000) /go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:207 +0xbd3 github.com/cockroachdb/cockroach/pkg/kv.NewTxnCoordSender.func1() /go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:191 +0x98 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420a0ef40) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 872 [select]: github.com/cockroachdb/cockroach/pkg/ts.(*poller).start.func1() /go/src/github.com/cockroachdb/cockroach/pkg/ts/db.go:90 +0x16f github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420a7a2e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 864 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:474 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42038c040) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 871 [select]: github.com/cockroachdb/cockroach/pkg/server.(*Server).startSampleEnvironment.func1() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:771 +0x16b github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42048e2c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 798 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/netutil.MakeServer.func2() /go/src/github.com/cockroachdb/cockroach/pkg/util/netutil/net.go:98 +0x70 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42076a8a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 876 [select]: github.com/cockroachdb/cockroach/pkg/sql/distsql.(*flowScheduler).Start.func1() /go/src/github.com/cockroachdb/cockroach/pkg/sql/distsql/flow_scheduler.go:101 +0x314 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420bde520) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 807 [chan receive]: github.com/cockroachdb/cockroach/pkg/gossip.(*server).start.func3() /go/src/github.com/cockroachdb/cockroach/pkg/gossip/server.go:381 +0x66 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42076ac40) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 810 [chan receive]: github.com/cockroachdb/cockroach/pkg/storage/engine.(*RocksDB).open.func1(0xc420baba40) /go/src/github.com/cockroachdb/cockroach/pkg/storage/engine/rocksdb.go:373 +0x44 created by github.com/cockroachdb/cockroach/pkg/storage/engine.(*RocksDB).open /go/src/github.com/cockroachdb/cockroach/pkg/storage/engine/rocksdb.go:374 +0x702 goroutine 835 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc400000032) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493680) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 805 [chan receive]: github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux.(*muxListener).Accept(0xc42076aa40, 0x1a41ab8, 0xc420aa99a0, 0x23bd0c0, 0xc4201798c0) <autogenerated>:9 +0x7f github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).Serve(0xc420aa99a0, 0x23b3bc0, 0xc42076aa40, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:348 +0x15a github.com/cockroachdb/cockroach/pkg/server.(*Server).Start.func7() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:497 +0x43 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42076abc0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 806 [chan receive]: github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux.(*muxListener).Accept(0xc42076aa20, 0xc420512060, 0x0, 0xc420111000, 0xc420448ef0) <autogenerated>:9 +0x7f github.com/cockroachdb/cockroach/pkg/util/netutil.(*Server).ServeWith(0xc420126158, 0xc420716000, 0x23b3bc0, 0xc42076aa20, 0xc420512060, 0xa97d8f, 0xc420448f88) /go/src/github.com/cockroachdb/cockroach/pkg/util/netutil/net.go:115 +0x5d github.com/cockroachdb/cockroach/pkg/server.(*Server).Start.func8() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:510 +0x13c github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420a168d0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 847 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xe0000002c) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493800) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 882 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:474 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42038c080) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 850 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc40000002f) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493860) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 855 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Store).Start.func3() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:1192 +0x165 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc4204938e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 856 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Store).startGossip.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:1278 +0x39b github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420a57b00) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 843 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc400000029) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493780) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 851 [semacquire]: sync.runtime_Semacquire(0xc420782e64) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*WaitGroup).Wait(0xc420782e58) /usr/local/go/src/sync/waitgroup.go:131 +0x97 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Wait(0xc420782dc0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:187 +0x34 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processRaft.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3345 +0x3a github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420b61c70) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 ERROR: exit status 255 1370 runs completed, 1 failures, over 11m2s Makefile:138: recipe for target 'stress' failed make: *** [stress] Error 1 ```
1.0
github.com/cockroachdb/cockroach/pkg/internal/client: TestClientGetAndPutProto failed under stress - SHA: https://github.com/cockroachdb/cockroach/commits/53ae9bb78ba376b9ec314caa5081d0ade866f3cd Parameters: ``` COCKROACH_PROPOSER_EVALUATED_KV=true TAGS=stress GOFLAGS= ``` Stress build found a failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=79293&tab=buildLog ``` I161204 07:40:15.352822 793 storage/store.go:1223 [n1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available I161204 07:40:15.352842 793 server/node.go:432 [n1] initialized store [n1,s1]: {Capacity:536870912 Available:536870912 RangeCount:1 LeaseCount:1} I161204 07:40:15.352857 793 server/node.go:317 [n1] node ID 1 initialized I161204 07:40:15.356225 793 gossip/gossip.go:290 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:36350" > attrs:<> locality:<> I161204 07:40:15.356289 793 storage/stores.go:296 [n1] read 0 node addresses from persistent storage I161204 07:40:15.356335 793 server/node.go:562 [n1] connecting to gossip network to verify cluster ID... I161204 07:40:15.356355 793 server/node.go:582 [n1] node connected via gossip and verified as part of cluster "bf3738a3-d325-4568-ab46-754e93c99ba6" I161204 07:40:15.356380 793 server/node.go:367 [n1] node=1: started with [[]=] engine(s) and attributes [] I161204 07:40:15.356409 793 sql/executor.go:291 [n1] creating distSQLPlanner with address {tcp 127.0.0.1:36350} I161204 07:40:15.361206 869 sql/event_log.go:95 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:36350} Attrs: Locality:} ClusterID:bf3738a3-d325-4568-ab46-754e93c99ba6 StartedAt:1480837215356362559} I161204 07:40:15.361721 862 storage/split_queue.go:103 [n1,split,s1,r1/1:/M{in-ax},@c42018ad80] splitting at keys [/Table/11/0 /Table/12/0 /Table/13/0 /Table/14/0] I161204 07:40:15.364411 862 storage/replica_command.go:2369 [n1,split,s1,r1/1:/M{in-ax},@c42018ad80] initiating a split of this range at key /Table/11 [r2] W161204 07:40:15.116065 914 util/hlc/hlc.go:145 backward time jump detected (-0.249007 seconds) I161204 07:40:15.122162 793 server/server.go:633 [n1] starting https server at 127.0.0.1:46912 I161204 07:40:15.122178 793 server/server.go:634 [n1] starting grpc/postgres server at 127.0.0.1:36350 I161204 07:40:15.122185 793 server/server.go:635 [n1] advertising CockroachDB node at 127.0.0.1:36350 E161204 07:40:15.131826 862 storage/queue.go:598 [n1,split,s1,r1/1:/{Min-Table/11},@c42018ad80] unable to split [n1,s1,r1/1:/{Min-Table/11}] at key "/Table/12/0": key range /Table/12/0-/Table/12/0 outside of bounds of range /Min-/Max I161204 07:40:15.131944 862 storage/split_queue.go:103 [n1,split,s1,r2/1:/{Table/11-Max},@c4202c6900] splitting at keys [/Table/12/0 /Table/13/0 /Table/14/0] I161204 07:40:15.131997 862 storage/replica_command.go:2369 [n1,split,s1,r2/1:/{Table/11-Max},@c4202c6900] initiating a split of this range at key /Table/12 [r3] E161204 07:40:15.136980 862 storage/queue.go:598 [n1,split,s1,r2/1:/Table/1{1-2},@c4202c6900] unable to split [n1,s1,r2/1:/Table/1{1-2}] at key "/Table/13/0": key range /Table/13/0-/Table/13/0 outside of bounds of range /Table/11-/Max I161204 07:40:15.137729 862 storage/split_queue.go:103 [n1,split,s1,r3/1:/{Table/12-Max},@c420099200] splitting at keys [/Table/13/0 /Table/14/0] I161204 07:40:15.137766 862 storage/replica_command.go:2369 [n1,split,s1,r3/1:/{Table/12-Max},@c420099200] initiating a split of this range at key /Table/13 [r4] E161204 07:40:15.142334 862 storage/queue.go:598 [n1,split,s1,r3/1:/Table/1{2-3},@c420099200] unable to split [n1,s1,r3/1:/Table/1{2-3}] at key "/Table/14/0": key range /Table/14/0-/Table/14/0 outside of bounds of range /Table/12-/Max I161204 07:40:15.142446 862 storage/split_queue.go:103 [n1,split,s1,r4/1:/{Table/13-Max},@c420510000] splitting at keys [/Table/14/0] I161204 07:40:15.142476 862 storage/replica_command.go:2369 [n1,split,s1,r4/1:/{Table/13-Max},@c420510000] initiating a split of this range at key /Table/14 [r5] F161204 07:40:15.148452 890 server/server.go:160 [n1] fewer than half the known nodes are within the maximum offset of 250ms (0 of 1) goroutine 890 [running]: github.com/cockroachdb/cockroach/pkg/util/log.getStacks(0x301, 0x8d93531, 0x2872aa0, 0xc420ad7a88) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:849 +0xa5 github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).outputLogEntry(0x2874000, 0xc400000004, 0x20a38bd, 0x10, 0xa0, 0xc4202ef800, 0x54) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:714 +0x7b1 github.com/cockroachdb/cockroach/pkg/util/log.addStructured(0x2b53e1e56a78, 0xc420bce930, 0x4, 0x2, 0x0, 0x0, 0xc420ad7d18, 0x1, 0x1) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/structured.go:146 +0x296 github.com/cockroachdb/cockroach/pkg/util/log.logDepth(0x2b53e1e56a78, 0xc420bce930, 0x1, 0x4, 0x0, 0x0, 0xc420ad7d18, 0x1, 0x1) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/log.go:88 +0x8c github.com/cockroachdb/cockroach/pkg/util/log.Fatal(0x2b53e1e56a78, 0xc420bce930, 0xc420ad7d18, 0x1, 0x1) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/log.go:181 +0x76 github.com/cockroachdb/cockroach/pkg/server.NewServer.func1() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:160 +0xcc github.com/cockroachdb/cockroach/pkg/rpc.(*Context).runHeartbeat(0xc420926600, 0xc42000ac00, 0xc42088ed70, 0xf, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/rpc/context.go:323 +0x47f github.com/cockroachdb/cockroach/pkg/rpc.(*Context).GRPCDial.func1.1.1() /go/src/github.com/cockroachdb/cockroach/pkg/rpc/context.go:227 +0x69 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420b9f920) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 1 [chan receive]: testing.(*T).Run(0xc4203240c0, 0x188dc17, 0x18, 0x1a3ea58, 0xc42027bc01) /usr/local/go/src/testing/testing.go:647 +0x316 testing.RunTests.func1(0xc4203240c0) /usr/local/go/src/testing/testing.go:793 +0x6d testing.tRunner(0xc4203240c0, 0xc420045e00) /usr/local/go/src/testing/testing.go:610 +0x81 testing.RunTests(0x1a41d40, 0x2395ea0, 0x31, 0x31, 0x28af6d0) /usr/local/go/src/testing/testing.go:799 +0x2f5 testing.(*M).Run(0xc420045ef8, 0x28af200) /usr/local/go/src/testing/testing.go:743 +0x85 github.com/cockroachdb/cockroach/pkg/internal/client_test.TestMain(0xc420045ef8) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/main_test.go:35 +0x76 main.main() github.com/cockroachdb/cockroach/pkg/internal/client/_test/_testmain.go:150 +0xc6 goroutine 17 [syscall, locked to thread]: runtime.goexit() /usr/local/go/src/runtime/asm_amd64.s:2086 +0x1 goroutine 793 [select]: github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendToReplicas(0xc420bdb810, 0x2b53e1e56ab8, 0xc420872c00, 0x1dcd6500, 0x0, 0x1, 0xc420c5a560, 0x1, 0x1, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1076 +0xddd github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendRPC(0xc420bdb810, 0x2b53e1e56a78, 0xc420218840, 0x1, 0xc420c5a560, 0x1, 0x1, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:340 +0x30b github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendSingleRange(0xc420bdb810, 0x2b53e1e56a78, 0xc420218840, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:409 +0x16e github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendPartialBatch(0xc420bdb810, 0x2b53e1e56a78, 0xc420218840, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:856 +0x29f github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges(0xc420bdb810, 0x2b53e1e56a78, 0xc420218840, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:738 +0x41f github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send(0xc420bdb810, 0x2b53e1e56a78, 0xc420218840, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:553 +0x2b3 github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send(0xc4203263c0, 0x2b53e1e56a78, 0xc4202187e0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:416 +0x7e9 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).send(0xc420baa6e0, 0x2b53e1e511d0, 0xc420014600, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:513 +0xbf github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Run.func1(0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:447 +0x6e github.com/cockroachdb/cockroach/pkg/internal/client.sendAndFill(0xc420c5a3e0, 0xc420c58c00, 0x0, 0xc420c5a3c0) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:418 +0x168 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Run(0xc420baa6e0, 0x2b53e1e511d0, 0xc420014600, 0xc420c58c00, 0xc420c5a3c0, 0xc4200ce320) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:449 +0xac github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).scan(0xc420baa6e0, 0x2b53e1e511d0, 0xc420014600, 0x17b2720, 0xc420c5a3a0, 0x17b2720, 0xc420c5a3c0, 0x0, 0x20c5a300, 0xc42060b4c8, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:307 +0x105 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Scan(0xc420baa6e0, 0x2b53e1e511d0, 0xc420014600, 0x17b2720, 0xc420c5a3a0, 0x17b2720, 0xc420c5a3c0, 0x0, 0x0, 0x1, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:318 +0x9a github.com/cockroachdb/cockroach/pkg/server.WaitForInitialSplits.func1(0xecfd5bf5f, 0xc407935070) /go/src/github.com/cockroachdb/cockroach/pkg/server/testserver.go:314 +0x14c github.com/cockroachdb/cockroach/pkg/util.RetryForDuration(0x2540be400, 0xc42060b620, 0x0, 0x239e180) /go/src/github.com/cockroachdb/cockroach/pkg/util/testing.go:131 +0xcd github.com/cockroachdb/cockroach/pkg/server.WaitForInitialSplits(0xc420baa6e0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/server/testserver.go:322 +0x67 github.com/cockroachdb/cockroach/pkg/server.(*TestServer).WaitForInitialSplits(0xc420bde110, 0x2b53e1e56a78, 0xc420a16600) /go/src/github.com/cockroachdb/cockroach/pkg/server/testserver.go:304 +0x3b github.com/cockroachdb/cockroach/pkg/server.(*TestServer).Start(0xc420bde110, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/server/testserver.go:283 +0x1e1 github.com/cockroachdb/cockroach/pkg/testutils/serverutils.StartServerRaw(0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/testutils/serverutils/test_server_shim.go:148 +0x134 github.com/cockroachdb/cockroach/pkg/testutils/serverutils.StartServer(0x23c0020, 0xc4204160c0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/testutils/serverutils/test_server_shim.go:118 +0x63 github.com/cockroachdb/cockroach/pkg/internal/client_test.TestClientGetAndPutProto(0xc4204160c0) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/client_test.go:375 +0xd8 testing.tRunner(0xc4204160c0, 0x1a3ea58) /usr/local/go/src/testing/testing.go:610 +0x81 created by testing.(*T).Run /usr/local/go/src/testing/testing.go:646 +0x2ec goroutine 20 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).flushDaemon(0x2874000) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1015 +0x77 created by github.com/cockroachdb/cockroach/pkg/util/log.init.1 /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:581 +0x95 goroutine 857 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Store).startGossip.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:1278 +0x39b github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420a57b30) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 868 [select]: github.com/cockroachdb/cockroach/pkg/server.(*Node).startGossip.func1() /go/src/github.com/cockroachdb/cockroach/pkg/server/node.go:609 +0x4f0 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42048e200) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 877 [IO wait]: net.runtime_pollWait(0x2b53e1e5b990, 0x72, 0x0) /usr/local/go/src/runtime/netpoll.go:160 +0x59 net.(*pollDesc).wait(0xc420b75640, 0x72, 0xc42053b608, 0xc420014230) /usr/local/go/src/net/fd_poll_runtime.go:73 +0x38 net.(*pollDesc).waitRead(0xc420b75640, 0x23a68c0, 0xc420014230) /usr/local/go/src/net/fd_poll_runtime.go:78 +0x34 net.(*netFD).accept(0xc420b755e0, 0x0, 0x23a3780, 0xc4200ce3a0) /usr/local/go/src/net/fd_unix.go:419 +0x238 net.(*TCPListener).accept(0xc420126160, 0x5fff1e, 0xc42053b6b8, 0xf7fbbb) /usr/local/go/src/net/tcpsock_posix.go:132 +0x2e net.(*TCPListener).Accept(0xc420126160, 0x1a40728, 0xc420be9200, 0x23bd1e0, 0xc420536038) /usr/local/go/src/net/tcpsock.go:222 +0x49 github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux.(*cMux).Serve(0xc420be9200, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux/cmux.go:124 +0xa4 github.com/cockroachdb/cockroach/pkg/server.(*Server).Start.func11() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:641 +0x2f github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420834d20) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 800 [chan receive]: github.com/cockroachdb/cockroach/pkg/server.(*Server).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:463 +0x71 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420a16870) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 799 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/netutil.MakeServer.func2() /go/src/github.com/cockroachdb/cockroach/pkg/util/netutil/net.go:98 +0x70 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42076a960) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 861 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:474 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493fe0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 863 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:474 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42038c020) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 804 [chan receive]: github.com/cockroachdb/cockroach/pkg/server.(*Server).Start.func6() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:490 +0x6d github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42076aba0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 797 [select]: github.com/cockroachdb/cockroach/pkg/sql.(*LeaseManager).RefreshLeases.func1() /go/src/github.com/cockroachdb/cockroach/pkg/sql/lease.go:1160 +0xcbf github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420bcfb00) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 834 [chan receive]: github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:171 +0x55 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493660) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 980 [select]: github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*http2Client).controller(0xc420812200) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/http2_client.go:999 +0x59d created by github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.newHTTP2Client /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/http2_client.go:233 +0xe78 goroutine 809 [select]: github.com/cockroachdb/cockroach/pkg/gossip.(*Gossip).manage.func1() /go/src/github.com/cockroachdb/cockroach/pkg/gossip/gossip.go:1031 +0x394 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420a0f800) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 836 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc400000033) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc4204936a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 865 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:474 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42038c060) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 849 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc40000002e) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493840) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 802 [chan receive]: github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux.(*muxListener).Accept(0xc42076aae0, 0xc420680090, 0x16e3840, 0x2383b80, 0x17851e0) <autogenerated>:9 +0x7f net/http.(*Server).Serve(0xc420926780, 0x23b3bc0, 0xc42076aae0, 0x0, 0x0) /usr/local/go/src/net/http/server.go:2273 +0x1ce github.com/cockroachdb/cockroach/pkg/server.(*Server).Start.func4() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:479 +0x3c github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42076ab40) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 846 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc40000002b) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc4204937e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 848 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc40000002d) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493820) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 862 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:474 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42038c000) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 845 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc400000030) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc4204937c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 867 [select]: github.com/cockroachdb/cockroach/pkg/server.(*Node).startComputePeriodicMetrics.func1() /go/src/github.com/cockroachdb/cockroach/pkg/server/node.go:650 +0x2bd github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42048e1e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 803 [chan receive]: github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux.(*muxListener).Accept(0xc42076ab00, 0xc420126180, 0xc420680060, 0x16e3840, 0x2383b80) <autogenerated>:9 +0x7f crypto/tls.(*listener).Accept(0xc42076ab60, 0xc420680030, 0x16e3840, 0x2383b80, 0x17851e0) /usr/local/go/src/crypto/tls/tls.go:52 +0x37 net/http.(*Server).Serve(0xc420926700, 0x23ab980, 0xc42076ab60, 0x0, 0x0) /usr/local/go/src/net/http/server.go:2273 +0x1ce github.com/cockroachdb/cockroach/pkg/server.(*Server).Start.func5() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:486 +0x42 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42076ab80) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 840 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc400000036) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493720) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 839 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc400000035) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493700) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 838 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc400000034) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc4204936e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 801 [IO wait]: net.runtime_pollWait(0x2b53e1e5be10, 0x72, 0x0) /usr/local/go/src/runtime/netpoll.go:160 +0x59 net.(*pollDesc).wait(0xc420b756b0, 0x72, 0xc420031e08, 0xc420014230) /usr/local/go/src/net/fd_poll_runtime.go:73 +0x38 net.(*pollDesc).waitRead(0xc420b756b0, 0x23a68c0, 0xc420014230) /usr/local/go/src/net/fd_poll_runtime.go:78 +0x34 net.(*netFD).accept(0xc420b75650, 0x0, 0x23a3780, 0xc420512040) /usr/local/go/src/net/fd_unix.go:419 +0x238 net.(*TCPListener).accept(0xc420126180, 0xc420031ef0, 0xc420031f00, 0x620690) /usr/local/go/src/net/tcpsock_posix.go:132 +0x2e net.(*TCPListener).Accept(0xc420126180, 0x1a40720, 0xc420be9280, 0xc420a76000, 0xc420031f88) /usr/local/go/src/net/tcpsock.go:222 +0x49 github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux.(*cMux).Serve(0xc420be9280, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux/cmux.go:124 +0xa4 github.com/cockroachdb/cockroach/pkg/server.(*Server).Start.func3() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:475 +0x2f github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42076ab20) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 795 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*StorePool).start.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store_pool.go:356 +0x3de github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42076a220) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 794 [chan receive]: github.com/cockroachdb/cockroach/pkg/rpc.NewContext.func1() /go/src/github.com/cockroachdb/cockroach/pkg/rpc/context.go:137 +0x76 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42082bee0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 907 [select]: github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*http2Server).controller(0xc420716bd0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/http2_server.go:673 +0x69c created by github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.newHTTP2Server /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/http2_server.go:139 +0x767 goroutine 852 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Store).raftTickLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3371 +0x334 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420b61c80) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 841 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc400000037) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493740) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 873 [select]: github.com/cockroachdb/cockroach/pkg/server.(*Node).startWriteSummaries.func1() /go/src/github.com/cockroachdb/cockroach/pkg/server/node.go:687 +0x26d github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc4208d03c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 875 [select]: github.com/cockroachdb/cockroach/pkg/sql.(*Executor).Start.func1() /go/src/github.com/cockroachdb/cockroach/pkg/sql/executor.go:301 +0x15f github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42048e340) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 844 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc40000002a) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc4204937a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 808 [select]: github.com/cockroachdb/cockroach/pkg/gossip.(*Gossip).bootstrap.func1() /go/src/github.com/cockroachdb/cockroach/pkg/gossip/gossip.go:993 +0x4c9 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420a0f7f0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 853 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Store).startCoalescedHeartbeatsLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3402 +0x167 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420b61c90) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 837 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc400000031) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc4204936c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 570 [IO wait]: net.runtime_pollWait(0x2b53e1e5bc90, 0x72, 0x8) /usr/local/go/src/runtime/netpoll.go:160 +0x59 net.(*pollDesc).wait(0xc420b74220, 0x72, 0xc420a9f588, 0xc420014230) /usr/local/go/src/net/fd_poll_runtime.go:73 +0x38 net.(*pollDesc).waitRead(0xc420b74220, 0x23a68c0, 0xc420014230) /usr/local/go/src/net/fd_poll_runtime.go:78 +0x34 net.(*netFD).Read(0xc420b741c0, 0xc42020dc00, 0x400, 0x400, 0x0, 0x23a68c0, 0xc420014230) /usr/local/go/src/net/fd_unix.go:243 +0x1a1 net.(*conn).Read(0xc420536038, 0xc42020dc00, 0x400, 0x400, 0x0, 0x0, 0x0) /usr/local/go/src/net/net.go:173 +0x70 github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux.(*MuxConn).Read(0xc4201798c0, 0xc42020dc00, 0x400, 0x400, 0x17d4820, 0x0, 0x2b53e1ed6280) /go/src/github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux/cmux.go:218 +0x113 crypto/tls.(*block).readFromUntil(0xc4204b0180, 0x2b53e1ed6280, 0xc4201798c0, 0x5, 0xc4201798c0, 0x0) /usr/local/go/src/crypto/tls/conn.go:476 +0x91 crypto/tls.(*Conn).readRecord(0xc420880380, 0x1a42617, 0xc420880488, 0x0) /usr/local/go/src/crypto/tls/conn.go:578 +0xc4 crypto/tls.(*Conn).Read(0xc420880380, 0xc420366000, 0x8000, 0x8000, 0x0, 0x0, 0x0) /usr/local/go/src/crypto/tls/conn.go:1113 +0x116 bufio.(*Reader).fill(0xc4202ef500) /usr/local/go/src/bufio/bufio.go:97 +0x10c bufio.(*Reader).Read(0xc4202ef500, 0xc420a46278, 0x9, 0x9, 0x9, 0x18, 0x28) /usr/local/go/src/bufio/bufio.go:209 +0x1bc io.ReadAtLeast(0x239d800, 0xc4202ef500, 0xc420a46278, 0x9, 0x9, 0x9, 0x5dc238, 0x30, 0xc420a9fad8) /usr/local/go/src/io/io.go:307 +0xa4 io.ReadFull(0x239d800, 0xc4202ef500, 0xc420a46278, 0x9, 0x9, 0x6af7f98fa89bc601, 0x6a00000000000000, 0x0) /usr/local/go/src/io/io.go:325 +0x58 github.com/cockroachdb/cockroach/vendor/golang.org/x/net/http2.readFrameHeader(0xc420a46278, 0x9, 0x9, 0x239d800, 0xc4202ef500, 0x0, 0xc400000000, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/net/http2/frame.go:237 +0x7b github.com/cockroachdb/cockroach/vendor/golang.org/x/net/http2.(*Framer).ReadFrame(0xc420a46240, 0x23a9200, 0xc4204b0ae0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/net/http2/frame.go:464 +0xa4 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*framer).readFrame(0xc4204b0750, 0xc4204b0ae0, 0xc4204b0ae0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/http_util.go:505 +0x2f github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*http2Server).HandleStreams(0xc420716bd0, 0xc4204b07e0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/http2_server.go:260 +0x200 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).serveStreams(0xc420aa99a0, 0x23bbf20, 0xc420716bd0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:424 +0x160 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).serveNewHTTP2Transport(0xc420aa99a0, 0x23bc340, 0xc420880380, 0x23a5b00, 0xc4204fe4d0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:411 +0x3da github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).handleRawConn(0xc420aa99a0, 0x23bd0c0, 0xc4201798c0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:388 +0x470 created by github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).Serve /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:357 +0x1a9 goroutine 889 [select]: github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*addrConn).transportMonitor(0xc420329040) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/clientconn.go:720 +0x63e github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*ClientConn).resetAddrConn.func1(0xc420329040) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/clientconn.go:479 +0x1de created by github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*ClientConn).resetAddrConn /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/clientconn.go:480 +0x3c7 goroutine 874 [select]: github.com/cockroachdb/cockroach/pkg/sql.(*SchemaChangeManager).Start.func1() /go/src/github.com/cockroachdb/cockroach/pkg/sql/schema_changer.go:751 +0x12b3 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42048e2e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 884 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*replicaScanner).waitAndProcess(0xc420326aa0, 0x2b53e1e56a78, 0xc420b9edb0, 0xecfd5bf5f, 0x153c4fd6, 0x2872aa0, 0xc420be8980, 0xc420716000, 0xc42018ad80, 0xecfd5bf5f) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scanner.go:207 +0x254 github.com/cockroachdb/cockroach/pkg/storage.(*replicaScanner).scanLoop.func1.1(0xc42018ad80, 0xc420a56f90) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scanner.go:269 +0x8e github.com/cockroachdb/cockroach/pkg/storage.(*storeReplicaVisitor).Visit(0xc420a56f90, 0xc42034ddb0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:324 +0x303 github.com/cockroachdb/cockroach/pkg/storage.(*replicaScanner).scanLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scanner.go:271 +0x2ca github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42038c0c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 901 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*idAllocator).start.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/id_alloc.go:133 +0x7c8 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc4204741a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 842 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc400000028) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493760) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 774 [select]: github.com/cockroachdb/cockroach/pkg/util/retry.(*Retry).Next(0xc4204c66e0, 0x770065) /go/src/github.com/cockroachdb/cockroach/pkg/util/retry/retry.go:128 +0x1a5 github.com/cockroachdb/cockroach/pkg/storage.(*Store).Send.func2(0xc4204c66e0, 0x148cfd9a000344ea) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:2424 +0x8d github.com/cockroachdb/cockroach/pkg/storage.(*Store).Send(0xc4204d9500, 0x2b53e1e56a78, 0xc4202193e0, 0x148cfd9a000344ea, 0x39, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:2431 +0x67e github.com/cockroachdb/cockroach/pkg/storage.(*Stores).Send(0xc420a1d1a0, 0x2b53e1e56a78, 0xc420219350, 0x0, 0x0, 0x100000001, 0x1, 0x1, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/storage/stores.go:187 +0x202 github.com/cockroachdb/cockroach/pkg/server.(*Node).batchInternal.func1(0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/server/node.go:818 +0x2a1 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunTaskWithErr(0xc420716000, 0xc420bcdcc8, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:245 +0x100 github.com/cockroachdb/cockroach/pkg/server.(*Node).batchInternal(0xc420539340, 0x2b53e1e56ab8, 0xc420872c00, 0xc42016cc08, 0xc420872c00, 0x100, 0x5dba0e) /go/src/github.com/cockroachdb/cockroach/pkg/server/node.go:829 +0x1ed github.com/cockroachdb/cockroach/pkg/server.(*Node).Batch(0xc420539340, 0x2b53e1e56ab8, 0xc420872c00, 0xc42016cc08, 0x0, 0x0, 0xc420447838) /go/src/github.com/cockroachdb/cockroach/pkg/server/node.go:851 +0xa0 github.com/cockroachdb/cockroach/pkg/kv.(*grpcTransport).SendNext.func1(0x239e8c0, 0xc420539340, 0xc4204c65f0, 0xc42016cbe0, 0xc4202a0420) /go/src/github.com/cockroachdb/cockroach/pkg/kv/transport.go:185 +0x64 created by github.com/cockroachdb/cockroach/pkg/kv.(*grpcTransport).SendNext /go/src/github.com/cockroachdb/cockroach/pkg/kv/transport.go:188 +0x1d6 goroutine 979 [IO wait]: net.runtime_pollWait(0x2b53e1e5c050, 0x72, 0x7) /usr/local/go/src/runtime/netpoll.go:160 +0x59 net.(*pollDesc).wait(0xc420b74450, 0x72, 0xc420341828, 0xc420014230) /usr/local/go/src/net/fd_poll_runtime.go:73 +0x38 net.(*pollDesc).waitRead(0xc420b74450, 0x23a68c0, 0xc420014230) /usr/local/go/src/net/fd_poll_runtime.go:78 +0x34 net.(*netFD).Read(0xc420b743f0, 0xc420d94400, 0x400, 0x400, 0x0, 0x23a68c0, 0xc420014230) /usr/local/go/src/net/fd_unix.go:243 +0x1a1 net.(*conn).Read(0xc420126010, 0xc420d94400, 0x400, 0x400, 0x0, 0x0, 0x0) /usr/local/go/src/net/net.go:173 +0x70 crypto/tls.(*block).readFromUntil(0xc4202d3920, 0x2b53e3ba3078, 0xc420126010, 0x5, 0xc420126010, 0x0) /usr/local/go/src/crypto/tls/conn.go:476 +0x91 crypto/tls.(*Conn).readRecord(0xc4204d8700, 0x1a42617, 0xc4204d8808, 0xc420816970) /usr/local/go/src/crypto/tls/conn.go:578 +0xc4 crypto/tls.(*Conn).Read(0xc4204d8700, 0xc420d78000, 0x8000, 0x8000, 0x0, 0x0, 0x0) /usr/local/go/src/crypto/tls/conn.go:1113 +0x116 bufio.(*Reader).fill(0xc4204d18c0) /usr/local/go/src/bufio/bufio.go:97 +0x10c bufio.(*Reader).Read(0xc4204d18c0, 0xc420883838, 0x9, 0x9, 0xc420341ca8, 0x620790, 0xc420476100) /usr/local/go/src/bufio/bufio.go:209 +0x1bc io.ReadAtLeast(0x239d800, 0xc4204d18c0, 0xc420883838, 0x9, 0x9, 0x9, 0xc420484760, 0xc420484758, 0x23a29c0) /usr/local/go/src/io/io.go:307 +0xa4 io.ReadFull(0x239d800, 0xc4204d18c0, 0xc420883838, 0x9, 0x9, 0xc420341eb0, 0xb01090, 0xc420233950) /usr/local/go/src/io/io.go:325 +0x58 github.com/cockroachdb/cockroach/vendor/golang.org/x/net/http2.readFrameHeader(0xc420883838, 0x9, 0x9, 0x239d800, 0xc4204d18c0, 0x0, 0xc400000000, 0xc420816960, 0xc) /go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/net/http2/frame.go:237 +0x7b github.com/cockroachdb/cockroach/vendor/golang.org/x/net/http2.(*Framer).ReadFrame(0xc420883800, 0x0, 0x0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/net/http2/frame.go:464 +0xa4 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*framer).readFrame(0xc420484690, 0xc4204b0de0, 0xc4204b0de0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/http_util.go:505 +0x2f github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*http2Client).reader(0xc420812200) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/http2_client.go:921 +0xb6 created by github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.newHTTP2Client /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/http2_client.go:203 +0x8b1 goroutine 883 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:474 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42038c0a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 870 [select]: github.com/cockroachdb/cockroach/pkg/util/retry.(*Retry).Next(0xc4208aa378, 0xc420b9f0b0) /go/src/github.com/cockroachdb/cockroach/pkg/util/retry/retry.go:128 +0x1a5 github.com/cockroachdb/cockroach/pkg/kv.(*RangeIterator).Seek(0xc4208aa6d0, 0x2b53e1e56a78, 0xc420b9f0b0, 0xc42038c180, 0xc, 0x14, 0xc4208aa400) /go/src/github.com/cockroachdb/cockroach/pkg/kv/range_iter.go:154 +0x212 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges(0xc420bdb810, 0x2b53e1e56a78, 0xc420b9f0b0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc420324b40, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:653 +0x199 github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send(0xc420bdb810, 0x2b53e1e56a78, 0xc420b9f0b0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc420324b40, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:553 +0x2b3 github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send(0xc4203263c0, 0x2b53e1e56a78, 0xc420b9f0b0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc420324a80, ...) /go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:416 +0x7e9 github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).sendInternal(0xc4204af340, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0xc4204af388, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:587 +0xfd github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).send(0xc4204af340, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:711 +0x498 github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).(github.com/cockroachdb/cockroach/pkg/internal/client.send)-fm(0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:313 +0x58 github.com/cockroachdb/cockroach/pkg/internal/client.sendAndFill(0xc4208ab610, 0xc420bf3200, 0x0, 0xc42084f720) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:418 +0x168 github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Run(0xc4204af340, 0xc420bf3200, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:313 +0xc6 github.com/cockroachdb/cockroach/pkg/storage.(*NodeLiveness).updateLiveness.func1(0xc4204af340, 0x0, 0x1) /go/src/github.com/cockroachdb/cockroach/pkg/storage/node_liveness.go:329 +0x3b2 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn.func1(0xc4204af340, 0xc42088e530, 0x4000000000000000, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:468 +0x2f github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Exec(0xc4204af340, 0xc420380101, 0x0, 0xc42088e520, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:520 +0x215 github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn(0xc420baa6e0, 0x2b53e1e56a78, 0xc420b9f0b0, 0xc42038c160, 0xc420a03ad8, 0x620790) /go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:469 +0x1fa github.com/cockroachdb/cockroach/pkg/storage.(*NodeLiveness).updateLiveness(0xc420bdb8c0, 0x2b53e1e56a78, 0xc420b9f0b0, 0x100000001, 0xc42038c120, 0x0, 0xc4208abbd8, 0x0, 0x2faf080) /go/src/github.com/cockroachdb/cockroach/pkg/storage/node_liveness.go:330 +0xbc github.com/cockroachdb/cockroach/pkg/storage.(*NodeLiveness).heartbeat(0xc420bdb8c0, 0x2b53e1e56a78, 0xc420b9f0b0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/node_liveness.go:194 +0x2f0 github.com/cockroachdb/cockroach/pkg/storage.(*NodeLiveness).StartHeartbeat.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/node_liveness.go:146 +0x2e7 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42048e2a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 796 [select]: github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).printStatsLoop(0xc4203263c0, 0x2b53e1e56a78, 0xc420680000) /go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:207 +0xbd3 github.com/cockroachdb/cockroach/pkg/kv.NewTxnCoordSender.func1() /go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:191 +0x98 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420a0ef40) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 872 [select]: github.com/cockroachdb/cockroach/pkg/ts.(*poller).start.func1() /go/src/github.com/cockroachdb/cockroach/pkg/ts/db.go:90 +0x16f github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420a7a2e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 864 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:474 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42038c040) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 871 [select]: github.com/cockroachdb/cockroach/pkg/server.(*Server).startSampleEnvironment.func1() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:771 +0x16b github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42048e2c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 798 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/netutil.MakeServer.func2() /go/src/github.com/cockroachdb/cockroach/pkg/util/netutil/net.go:98 +0x70 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42076a8a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 876 [select]: github.com/cockroachdb/cockroach/pkg/sql/distsql.(*flowScheduler).Start.func1() /go/src/github.com/cockroachdb/cockroach/pkg/sql/distsql/flow_scheduler.go:101 +0x314 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420bde520) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 807 [chan receive]: github.com/cockroachdb/cockroach/pkg/gossip.(*server).start.func3() /go/src/github.com/cockroachdb/cockroach/pkg/gossip/server.go:381 +0x66 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42076ac40) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 810 [chan receive]: github.com/cockroachdb/cockroach/pkg/storage/engine.(*RocksDB).open.func1(0xc420baba40) /go/src/github.com/cockroachdb/cockroach/pkg/storage/engine/rocksdb.go:373 +0x44 created by github.com/cockroachdb/cockroach/pkg/storage/engine.(*RocksDB).open /go/src/github.com/cockroachdb/cockroach/pkg/storage/engine/rocksdb.go:374 +0x702 goroutine 835 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc400000032) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493680) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 805 [chan receive]: github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux.(*muxListener).Accept(0xc42076aa40, 0x1a41ab8, 0xc420aa99a0, 0x23bd0c0, 0xc4201798c0) <autogenerated>:9 +0x7f github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).Serve(0xc420aa99a0, 0x23b3bc0, 0xc42076aa40, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:348 +0x15a github.com/cockroachdb/cockroach/pkg/server.(*Server).Start.func7() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:497 +0x43 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42076abc0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 806 [chan receive]: github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/cmux.(*muxListener).Accept(0xc42076aa20, 0xc420512060, 0x0, 0xc420111000, 0xc420448ef0) <autogenerated>:9 +0x7f github.com/cockroachdb/cockroach/pkg/util/netutil.(*Server).ServeWith(0xc420126158, 0xc420716000, 0x23b3bc0, 0xc42076aa20, 0xc420512060, 0xa97d8f, 0xc420448f88) /go/src/github.com/cockroachdb/cockroach/pkg/util/netutil/net.go:115 +0x5d github.com/cockroachdb/cockroach/pkg/server.(*Server).Start.func8() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:510 +0x13c github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420a168d0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 847 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xe0000002c) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493800) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 882 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:474 +0x383 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc42038c080) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 850 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc40000002f) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493860) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 855 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Store).Start.func3() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:1192 +0x165 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc4204938e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 856 [select]: github.com/cockroachdb/cockroach/pkg/storage.(*Store).startGossip.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:1278 +0x39b github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420a57b00) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 843 [semacquire]: sync.runtime_notifyListWait(0xc4204544d0, 0xc400000029) /usr/local/go/src/runtime/sema.go:267 +0x122 sync.(*Cond).Wait(0xc4204544c0) /usr/local/go/src/sync/cond.go:57 +0x80 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).worker(0xc420782dc0, 0xc420716000) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:212 +0xb2 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Start.func2() /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:181 +0x33 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420493780) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 goroutine 851 [semacquire]: sync.runtime_Semacquire(0xc420782e64) /usr/local/go/src/runtime/sema.go:47 +0x30 sync.(*WaitGroup).Wait(0xc420782e58) /usr/local/go/src/sync/waitgroup.go:131 +0x97 github.com/cockroachdb/cockroach/pkg/storage.(*raftScheduler).Wait(0xc420782dc0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/scheduler.go:187 +0x34 github.com/cockroachdb/cockroach/pkg/storage.(*Store).processRaft.func1() /go/src/github.com/cockroachdb/cockroach/pkg/storage/store.go:3345 +0x3a github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc420716000, 0xc420b61c70) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:196 +0x7d created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:197 +0x66 ERROR: exit status 255 1370 runs completed, 1 failures, over 11m2s Makefile:138: recipe for target 'stress' failed make: *** [stress] Error 1 ```
non_process
github com cockroachdb cockroach pkg internal client testclientgetandputproto failed under stress sha parameters cockroach proposer evaluated kv true tags stress goflags stress build found a failed test storage store go failed initial metrics computation system config not yet available server node go initialized store capacity available rangecount leasecount server node go node id initialized gossip gossip go nodedescriptor set to node id address attrs locality storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster server node go node started with engine s and attributes sql executor go creating distsqlplanner with address tcp sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality clusterid startedat storage split queue go splitting at keys storage replica command go initiating a split of this range at key table util hlc hlc go backward time jump detected seconds server server go starting https server at server server go starting grpc postgres server at server server go advertising cockroachdb node at storage queue go unable to split at key table key range table table outside of bounds of range min max storage split queue go splitting at keys storage replica command go initiating a split of this range at key table storage queue go unable to split at key table key range table table outside of bounds of range table max storage split queue go splitting at keys storage replica command go initiating a split of this range at key table storage queue go unable to split at key table key range table table outside of bounds of range table max storage split queue go splitting at keys storage replica command go initiating a split of this range at key table server server go fewer than half the known nodes are within the maximum offset of of goroutine github com cockroachdb cockroach pkg util log getstacks go src github com cockroachdb cockroach pkg util log clog go github com cockroachdb cockroach pkg util log loggingt outputlogentry go src github com cockroachdb cockroach pkg util log clog go github com cockroachdb cockroach pkg util log addstructured go src github com cockroachdb cockroach pkg util log structured go github com cockroachdb cockroach pkg util log logdepth go src github com cockroachdb cockroach pkg util log log go github com cockroachdb cockroach pkg util log fatal go src github com cockroachdb cockroach pkg util log log go github com cockroachdb cockroach pkg server newserver go src github com cockroachdb cockroach pkg server server go github com cockroachdb cockroach pkg rpc context runheartbeat go src github com cockroachdb cockroach pkg rpc context go github com cockroachdb cockroach pkg rpc context grpcdial go src github com cockroachdb cockroach pkg rpc context go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine testing t run usr local go src testing testing go testing runtests usr local go src testing testing go testing trunner usr local go src testing testing go testing runtests usr local go src testing testing go testing m run usr local go src testing testing go github com cockroachdb cockroach pkg internal client test testmain go src github com cockroachdb cockroach pkg internal client main test go main main github com cockroachdb cockroach pkg internal client test testmain go goroutine runtime goexit usr local go src runtime asm s goroutine github com cockroachdb cockroach pkg kv distsender sendtoreplicas go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendrpc go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendsinglerange go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendpartialbatch go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender divideandsendbatchtoranges go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender send go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv txncoordsender send go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg internal client db send go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client db run go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client sendandfill go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client db run go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client db scan go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client db scan go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg server waitforinitialsplits go src github com cockroachdb cockroach pkg server testserver go github com cockroachdb cockroach pkg util retryforduration go src github com cockroachdb cockroach pkg util testing go github com cockroachdb cockroach pkg server waitforinitialsplits go src github com cockroachdb cockroach pkg server testserver go github com cockroachdb cockroach pkg server testserver waitforinitialsplits go src github com cockroachdb cockroach pkg server testserver go github com cockroachdb cockroach pkg server testserver start go src github com cockroachdb cockroach pkg server testserver go github com cockroachdb cockroach pkg testutils serverutils startserverraw go src github com cockroachdb cockroach pkg testutils serverutils test server shim go github com cockroachdb cockroach pkg testutils serverutils startserver go src github com cockroachdb cockroach pkg testutils serverutils test server shim go github com cockroachdb cockroach pkg internal client test testclientgetandputproto go src github com cockroachdb cockroach pkg internal client client test go testing trunner usr local go src testing testing go created by testing t run usr local go src testing testing go goroutine github com cockroachdb cockroach pkg util log loggingt flushdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log init go src github com cockroachdb cockroach pkg util log clog go goroutine github com cockroachdb cockroach pkg storage store startgossip go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg server node startgossip go src github com cockroachdb cockroach pkg server node go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine net runtime pollwait usr local go src runtime netpoll go net polldesc wait usr local go src net fd poll runtime go net polldesc waitread usr local go src net fd poll runtime go net netfd accept usr local go src net fd unix go net tcplistener accept usr local go src net tcpsock posix go net tcplistener accept usr local go src net tcpsock go github com cockroachdb cockroach vendor github com cockroachdb cmux cmux serve go src github com cockroachdb cockroach vendor github com cockroachdb cmux cmux go github com cockroachdb cockroach pkg server server start go src github com cockroachdb cockroach pkg server server go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg server server start go src github com cockroachdb cockroach pkg server server go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg util netutil makeserver go src github com cockroachdb cockroach pkg util netutil net go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage basequeue processloop go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage basequeue processloop go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg server server start go src github com cockroachdb cockroach pkg server server go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg sql leasemanager refreshleases go src github com cockroachdb cockroach pkg sql lease go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach vendor google golang org grpc transport controller go src github com cockroachdb cockroach vendor google golang org grpc transport client go created by github com cockroachdb cockroach vendor google golang org grpc transport go src github com cockroachdb cockroach vendor google golang org grpc transport client go goroutine github com cockroachdb cockroach pkg gossip gossip manage go src github com cockroachdb cockroach pkg gossip gossip go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage basequeue processloop go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach vendor github com cockroachdb cmux muxlistener accept net http server serve usr local go src net http server go github com cockroachdb cockroach pkg server server start go src github com cockroachdb cockroach pkg server server go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage basequeue processloop go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg server node startcomputeperiodicmetrics go src github com cockroachdb cockroach pkg server node go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach vendor github com cockroachdb cmux muxlistener accept crypto tls listener accept usr local go src crypto tls tls go net http server serve usr local go src net http server go github com cockroachdb cockroach pkg server server start go src github com cockroachdb cockroach pkg server server go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine net runtime pollwait usr local go src runtime netpoll go net polldesc wait usr local go src net fd poll runtime go net polldesc waitread usr local go src net fd poll runtime go net netfd accept usr local go src net fd unix go net tcplistener accept usr local go src net tcpsock posix go net tcplistener accept usr local go src net tcpsock go github com cockroachdb cockroach vendor github com cockroachdb cmux cmux serve go src github com cockroachdb cockroach vendor github com cockroachdb cmux cmux go github com cockroachdb cockroach pkg server server start go src github com cockroachdb cockroach pkg server server go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage storepool start go src github com cockroachdb cockroach pkg storage store pool go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg rpc newcontext go src github com cockroachdb cockroach pkg rpc context go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach vendor google golang org grpc transport controller go src github com cockroachdb cockroach vendor google golang org grpc transport server go created by github com cockroachdb cockroach vendor google golang org grpc transport go src github com cockroachdb cockroach vendor google golang org grpc transport server go goroutine github com cockroachdb cockroach pkg storage store rafttickloop go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg server node startwritesummaries go src github com cockroachdb cockroach pkg server node go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg sql executor start go src github com cockroachdb cockroach pkg sql executor go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg gossip gossip bootstrap go src github com cockroachdb cockroach pkg gossip gossip go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage store startcoalescedheartbeatsloop go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine net runtime pollwait usr local go src runtime netpoll go net polldesc wait usr local go src net fd poll runtime go net polldesc waitread usr local go src net fd poll runtime go net netfd read usr local go src net fd unix go net conn read usr local go src net net go github com cockroachdb cockroach vendor github com cockroachdb cmux muxconn read go src github com cockroachdb cockroach vendor github com cockroachdb cmux cmux go crypto tls block readfromuntil usr local go src crypto tls conn go crypto tls conn readrecord usr local go src crypto tls conn go crypto tls conn read usr local go src crypto tls conn go bufio reader fill usr local go src bufio bufio go bufio reader read usr local go src bufio bufio go io readatleast usr local go src io io go io readfull usr local go src io io go github com cockroachdb cockroach vendor golang org x net readframeheader go src github com cockroachdb cockroach vendor golang org x net frame go github com cockroachdb cockroach vendor golang org x net framer readframe go src github com cockroachdb cockroach vendor golang org x net frame go github com cockroachdb cockroach vendor google golang org grpc transport framer readframe go src github com cockroachdb cockroach vendor google golang org grpc transport http util go github com cockroachdb cockroach vendor google golang org grpc transport handlestreams go src github com cockroachdb cockroach vendor google golang org grpc transport server go github com cockroachdb cockroach vendor google golang org grpc server servestreams go src github com cockroachdb cockroach vendor google golang org grpc server go github com cockroachdb cockroach vendor google golang org grpc server go src github com cockroachdb cockroach vendor google golang org grpc server go github com cockroachdb cockroach vendor google golang org grpc server handlerawconn go src github com cockroachdb cockroach vendor google golang org grpc server go created by github com cockroachdb cockroach vendor google golang org grpc server serve go src github com cockroachdb cockroach vendor google golang org grpc server go goroutine github com cockroachdb cockroach vendor google golang org grpc addrconn transportmonitor go src github com cockroachdb cockroach vendor google golang org grpc clientconn go github com cockroachdb cockroach vendor google golang org grpc clientconn resetaddrconn go src github com cockroachdb cockroach vendor google golang org grpc clientconn go created by github com cockroachdb cockroach vendor google golang org grpc clientconn resetaddrconn go src github com cockroachdb cockroach vendor google golang org grpc clientconn go goroutine github com cockroachdb cockroach pkg sql schemachangemanager start go src github com cockroachdb cockroach pkg sql schema changer go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage replicascanner waitandprocess go src github com cockroachdb cockroach pkg storage scanner go github com cockroachdb cockroach pkg storage replicascanner scanloop go src github com cockroachdb cockroach pkg storage scanner go github com cockroachdb cockroach pkg storage storereplicavisitor visit go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg storage replicascanner scanloop go src github com cockroachdb cockroach pkg storage scanner go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage idallocator start go src github com cockroachdb cockroach pkg storage id alloc go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg util retry retry next go src github com cockroachdb cockroach pkg util retry retry go github com cockroachdb cockroach pkg storage store send go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg storage store send go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg storage stores send go src github com cockroachdb cockroach pkg storage stores go github com cockroachdb cockroach pkg server node batchinternal go src github com cockroachdb cockroach pkg server node go github com cockroachdb cockroach pkg util stop stopper runtaskwitherr go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg server node batchinternal go src github com cockroachdb cockroach pkg server node go github com cockroachdb cockroach pkg server node batch go src github com cockroachdb cockroach pkg server node go github com cockroachdb cockroach pkg kv grpctransport sendnext go src github com cockroachdb cockroach pkg kv transport go created by github com cockroachdb cockroach pkg kv grpctransport sendnext go src github com cockroachdb cockroach pkg kv transport go goroutine net runtime pollwait usr local go src runtime netpoll go net polldesc wait usr local go src net fd poll runtime go net polldesc waitread usr local go src net fd poll runtime go net netfd read usr local go src net fd unix go net conn read usr local go src net net go crypto tls block readfromuntil usr local go src crypto tls conn go crypto tls conn readrecord usr local go src crypto tls conn go crypto tls conn read usr local go src crypto tls conn go bufio reader fill usr local go src bufio bufio go bufio reader read usr local go src bufio bufio go io readatleast usr local go src io io go io readfull usr local go src io io go github com cockroachdb cockroach vendor golang org x net readframeheader go src github com cockroachdb cockroach vendor golang org x net frame go github com cockroachdb cockroach vendor golang org x net framer readframe go src github com cockroachdb cockroach vendor golang org x net frame go github com cockroachdb cockroach vendor google golang org grpc transport framer readframe go src github com cockroachdb cockroach vendor google golang org grpc transport http util go github com cockroachdb cockroach vendor google golang org grpc transport reader go src github com cockroachdb cockroach vendor google golang org grpc transport client go created by github com cockroachdb cockroach vendor google golang org grpc transport go src github com cockroachdb cockroach vendor google golang org grpc transport client go goroutine github com cockroachdb cockroach pkg storage basequeue processloop go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg util retry retry next go src github com cockroachdb cockroach pkg util retry retry go github com cockroachdb cockroach pkg kv rangeiterator seek go src github com cockroachdb cockroach pkg kv range iter go github com cockroachdb cockroach pkg kv distsender divideandsendbatchtoranges go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender send go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv txncoordsender send go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg internal client txn sendinternal go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client txn send go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client txn github com cockroachdb cockroach pkg internal client send fm go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client sendandfill go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client txn run go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg storage nodeliveness updateliveness go src github com cockroachdb cockroach pkg storage node liveness go github com cockroachdb cockroach pkg internal client db txn go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client txn exec go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client db txn go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg storage nodeliveness updateliveness go src github com cockroachdb cockroach pkg storage node liveness go github com cockroachdb cockroach pkg storage nodeliveness heartbeat go src github com cockroachdb cockroach pkg storage node liveness go github com cockroachdb cockroach pkg storage nodeliveness startheartbeat go src github com cockroachdb cockroach pkg storage node liveness go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg kv txncoordsender printstatsloop go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg kv newtxncoordsender go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg ts poller start go src github com cockroachdb cockroach pkg ts db go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage basequeue processloop go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg server server startsampleenvironment go src github com cockroachdb cockroach pkg server server go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg util netutil makeserver go src github com cockroachdb cockroach pkg util netutil net go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg sql distsql flowscheduler start go src github com cockroachdb cockroach pkg sql distsql flow scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg gossip server start go src github com cockroachdb cockroach pkg gossip server go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage engine rocksdb open go src github com cockroachdb cockroach pkg storage engine rocksdb go created by github com cockroachdb cockroach pkg storage engine rocksdb open go src github com cockroachdb cockroach pkg storage engine rocksdb go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach vendor github com cockroachdb cmux muxlistener accept github com cockroachdb cockroach vendor google golang org grpc server serve go src github com cockroachdb cockroach vendor google golang org grpc server go github com cockroachdb cockroach pkg server server start go src github com cockroachdb cockroach pkg server server go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach vendor github com cockroachdb cmux muxlistener accept github com cockroachdb cockroach pkg util netutil server servewith go src github com cockroachdb cockroach pkg util netutil net go github com cockroachdb cockroach pkg server server start go src github com cockroachdb cockroach pkg server server go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage basequeue processloop go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage store start go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine github com cockroachdb cockroach pkg storage store startgossip go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime notifylistwait usr local go src runtime sema go sync cond wait usr local go src sync cond go github com cockroachdb cockroach pkg storage raftscheduler worker go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage raftscheduler start go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go goroutine sync runtime semacquire usr local go src runtime sema go sync waitgroup wait usr local go src sync waitgroup go github com cockroachdb cockroach pkg storage raftscheduler wait go src github com cockroachdb cockroach pkg storage scheduler go github com cockroachdb cockroach pkg storage store processraft go src github com cockroachdb cockroach pkg storage store go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go error exit status runs completed failures over makefile recipe for target stress failed make error
0
12,730
15,100,301,444
IssuesEvent
2021-02-08 05:10:37
pystatgen/sgkit
https://api.github.com/repos/pystatgen/sgkit
opened
Disable scheduled build/windows and validation CI on forks
process + tools
Right now some of our CI workflows are unnecessarily being run on the forks, this include: * validation workflow * scheduled build and windows This is unnecessary, and can be confusing.
1.0
Disable scheduled build/windows and validation CI on forks - Right now some of our CI workflows are unnecessarily being run on the forks, this include: * validation workflow * scheduled build and windows This is unnecessary, and can be confusing.
process
disable scheduled build windows and validation ci on forks right now some of our ci workflows are unnecessarily being run on the forks this include validation workflow scheduled build and windows this is unnecessary and can be confusing
1
196,744
22,514,405,507
IssuesEvent
2022-06-24 01:03:37
jgeraigery/mongo-csfl-encryption-java-demo
https://api.github.com/repos/jgeraigery/mongo-csfl-encryption-java-demo
opened
CVE-2022-22970 (Medium) detected in spring-beans-5.2.2.RELEASE.jar, spring-core-5.2.2.RELEASE.jar
security vulnerability
## CVE-2022-22970 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>spring-beans-5.2.2.RELEASE.jar</b>, <b>spring-core-5.2.2.RELEASE.jar</b></p></summary> <p> <details><summary><b>spring-beans-5.2.2.RELEASE.jar</b></p></summary> <p>Spring Beans</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: /mongo-csfle-enterprise/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.springframework/spring-beans/5.2.2.RELEASE/81e4d9cc2e8fac88ab4eb7325c4521bd07c6389c/spring-beans-5.2.2.RELEASE.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.springframework/spring-beans/5.2.2.RELEASE/81e4d9cc2e8fac88ab4eb7325c4521bd07c6389c/spring-beans-5.2.2.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-data-mongodb-2.2.2.RELEASE.jar (Root Library) - spring-data-mongodb-2.2.3.RELEASE.jar - :x: **spring-beans-5.2.2.RELEASE.jar** (Vulnerable Library) </details> <details><summary><b>spring-core-5.2.2.RELEASE.jar</b></p></summary> <p>Spring Core</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: /mongo-csfle-enterprise/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.springframework/spring-core/5.2.2.RELEASE/bfcf2f6d0494d89db63ae170b8491223c93a88dc/spring-core-5.2.2.RELEASE.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.springframework/spring-core/5.2.2.RELEASE/bfcf2f6d0494d89db63ae170b8491223c93a88dc/spring-core-5.2.2.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-data-mongodb-2.2.2.RELEASE.jar (Root Library) - spring-data-mongodb-2.2.3.RELEASE.jar - :x: **spring-core-5.2.2.RELEASE.jar** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In spring framework versions prior to 5.3.20+ , 5.2.22+ and old unsupported versions, applications that handle file uploads are vulnerable to DoS attack if they rely on data binding to set a MultipartFile or javax.servlet.Part to a field in a model object. <p>Publish Date: 2022-05-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-22970>CVE-2022-22970</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2022-22970">https://tanzu.vmware.com/security/cve-2022-22970</a></p> <p>Release Date: 2022-05-12</p> <p>Fix Resolution: org.springframework:spring-beans:5.2.22,5.3.20;org.springframework:spring-core:5.2.22,5.3.20</p> </p> </details> <p></p>
True
CVE-2022-22970 (Medium) detected in spring-beans-5.2.2.RELEASE.jar, spring-core-5.2.2.RELEASE.jar - ## CVE-2022-22970 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>spring-beans-5.2.2.RELEASE.jar</b>, <b>spring-core-5.2.2.RELEASE.jar</b></p></summary> <p> <details><summary><b>spring-beans-5.2.2.RELEASE.jar</b></p></summary> <p>Spring Beans</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: /mongo-csfle-enterprise/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.springframework/spring-beans/5.2.2.RELEASE/81e4d9cc2e8fac88ab4eb7325c4521bd07c6389c/spring-beans-5.2.2.RELEASE.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.springframework/spring-beans/5.2.2.RELEASE/81e4d9cc2e8fac88ab4eb7325c4521bd07c6389c/spring-beans-5.2.2.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-data-mongodb-2.2.2.RELEASE.jar (Root Library) - spring-data-mongodb-2.2.3.RELEASE.jar - :x: **spring-beans-5.2.2.RELEASE.jar** (Vulnerable Library) </details> <details><summary><b>spring-core-5.2.2.RELEASE.jar</b></p></summary> <p>Spring Core</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: /mongo-csfle-enterprise/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.springframework/spring-core/5.2.2.RELEASE/bfcf2f6d0494d89db63ae170b8491223c93a88dc/spring-core-5.2.2.RELEASE.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.springframework/spring-core/5.2.2.RELEASE/bfcf2f6d0494d89db63ae170b8491223c93a88dc/spring-core-5.2.2.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-data-mongodb-2.2.2.RELEASE.jar (Root Library) - spring-data-mongodb-2.2.3.RELEASE.jar - :x: **spring-core-5.2.2.RELEASE.jar** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In spring framework versions prior to 5.3.20+ , 5.2.22+ and old unsupported versions, applications that handle file uploads are vulnerable to DoS attack if they rely on data binding to set a MultipartFile or javax.servlet.Part to a field in a model object. <p>Publish Date: 2022-05-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-22970>CVE-2022-22970</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2022-22970">https://tanzu.vmware.com/security/cve-2022-22970</a></p> <p>Release Date: 2022-05-12</p> <p>Fix Resolution: org.springframework:spring-beans:5.2.22,5.3.20;org.springframework:spring-core:5.2.22,5.3.20</p> </p> </details> <p></p>
non_process
cve medium detected in spring beans release jar spring core release jar cve medium severity vulnerability vulnerable libraries spring beans release jar spring core release jar spring beans release jar spring beans library home page a href path to dependency file mongo csfle enterprise build gradle path to vulnerable library home wss scanner gradle caches modules files org springframework spring beans release spring beans release jar home wss scanner gradle caches modules files org springframework spring beans release spring beans release jar dependency hierarchy spring boot starter data mongodb release jar root library spring data mongodb release jar x spring beans release jar vulnerable library spring core release jar spring core library home page a href path to dependency file mongo csfle enterprise build gradle path to vulnerable library home wss scanner gradle caches modules files org springframework spring core release spring core release jar home wss scanner gradle caches modules files org springframework spring core release spring core release jar dependency hierarchy spring boot starter data mongodb release jar root library spring data mongodb release jar x spring core release jar vulnerable library found in base branch master vulnerability details in spring framework versions prior to and old unsupported versions applications that handle file uploads are vulnerable to dos attack if they rely on data binding to set a multipartfile or javax servlet part to a field in a model object publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring beans org springframework spring core
0
2,021
4,846,695,074
IssuesEvent
2016-11-10 12:42:43
Alfresco/alfresco-ng2-components
https://api.github.com/repos/Alfresco/alfresco-ng2-components
opened
If specific app is selected all processes are listed when starting a process
browser: all bug comp: activiti-processList
1. Click a specific app 2. Start a process 3. Notice all processes from all apps are displayed ![screen shot 2016-11-10 at 12 36 03](https://cloud.githubusercontent.com/assets/13200338/20177117/0690b6da-a743-11e6-88e5-9ea279a9fb17.png)
1.0
If specific app is selected all processes are listed when starting a process - 1. Click a specific app 2. Start a process 3. Notice all processes from all apps are displayed ![screen shot 2016-11-10 at 12 36 03](https://cloud.githubusercontent.com/assets/13200338/20177117/0690b6da-a743-11e6-88e5-9ea279a9fb17.png)
process
if specific app is selected all processes are listed when starting a process click a specific app start a process notice all processes from all apps are displayed
1
19,985
26,462,582,704
IssuesEvent
2023-01-16 19:14:12
kubernetes-sigs/windows-operational-readiness
https://api.github.com/repos/kubernetes-sigs/windows-operational-readiness
closed
Ability to launch hostProcess containers and shared IP
kind/feature lifecycle/rotten category/ext.hostprocess
Ability to launch hostProcess containers which share the IP address of a Windows node.
1.0
Ability to launch hostProcess containers and shared IP - Ability to launch hostProcess containers which share the IP address of a Windows node.
process
ability to launch hostprocess containers and shared ip ability to launch hostprocess containers which share the ip address of a windows node
1
1,631
4,241,342,719
IssuesEvent
2016-07-06 16:02:11
nodejs/node
https://api.github.com/repos/nodejs/node
closed
node docs issue in child_process
child_process doc
Using child_process.spawn, I believe the docs are incorrect (simple typo with drastic consequences) when using the stdio option: ```js const cp = require('child_process'); const n = cp.spawn('node',['foo.js'],{ stdio: ['ignore','ignore','ignore'], detached: true }); ``` the above should be correct, if you wish to create a child_process that is the leader of a new process group, that won't die if the parent is killed. However, the child_process docs have this: https://nodejs.org/api/child_process.html#child_process_options_detached ```js const n = cp.spawn('node',['foo.js'],{ stdio: ['ignore'], detached: true }); ``` I don't believe the above is correct, although this may work: ```js const n = cp.spawn('node',['foo.js'],{ stdio: 'ignore', detached: true }); ``` this issue had me stuck for a few days :)
1.0
node docs issue in child_process - Using child_process.spawn, I believe the docs are incorrect (simple typo with drastic consequences) when using the stdio option: ```js const cp = require('child_process'); const n = cp.spawn('node',['foo.js'],{ stdio: ['ignore','ignore','ignore'], detached: true }); ``` the above should be correct, if you wish to create a child_process that is the leader of a new process group, that won't die if the parent is killed. However, the child_process docs have this: https://nodejs.org/api/child_process.html#child_process_options_detached ```js const n = cp.spawn('node',['foo.js'],{ stdio: ['ignore'], detached: true }); ``` I don't believe the above is correct, although this may work: ```js const n = cp.spawn('node',['foo.js'],{ stdio: 'ignore', detached: true }); ``` this issue had me stuck for a few days :)
process
node docs issue in child process using child process spawn i believe the docs are incorrect simple typo with drastic consequences when using the stdio option js const cp require child process const n cp spawn node stdio detached true the above should be correct if you wish to create a child process that is the leader of a new process group that won t die if the parent is killed however the child process docs have this js const n cp spawn node stdio detached true i don t believe the above is correct although this may work js const n cp spawn node stdio ignore detached true this issue had me stuck for a few days
1
12,952
15,326,791,268
IssuesEvent
2021-02-26 04:31:45
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
Optimize relative datetime filter clauses
.Backend .Enhancement Difficulty:Medium Priority:P2 Querying/Processor Type:Performance
Related: #4043 In #10980 I added QP middleware to rewrite filters against temporal fields to generate `BETWEEN` clauses or the like and avoid temporal truncation/extraction operations that break index usages. However, this improvement only applied to filters against absolute temporal moments, e.g. "show me toucan sightings where the month is January 2020". Filters against *relative* temporal moments, e.g. "show me toucan sightings in the last month" are not currently being optimized. For example this MBQL query ```clj ;; attempted-murders is our crow sighting test dataset (mt/dataset attempted-murders (mt/mbql-query attempts {:aggregation [[:count]] :filter [:time-interval $datetime :last :month]})) ``` generates this SQL (BigQuery): ```sql SELECT count(*) AS `count` FROM `attempted_murders.attempts` WHERE datetime_trunc(`attempted_murders.attempts`.`datetime`, month) = datetime_trunc(datetime_add(current_datetime(), INTERVAL -1 month), month) ``` It seems possible to instead generate something along the lines of ```sql SELECT count(*) AS `count` FROM `attempted_murders.attempts` WHERE `attempted_murders.attempts`.`datetime` BETWEEN datetime_trunc(datetime_add(current_datetime(), INTERVAL -2 month), month) AND datetime_trunc(datetime_add(current_datetime(), INTERVAL -1 month), month) ``` I could probably get this working in a day. It would be a big performance win. Since #4043 was originally marked as P1 and #10980 didn't fully solve the entire issue I'm going to mark this as P2 and address it as soon as I get a chance. ⬇️ **Please click the 👍 reaction instead of leaving a `+1` or 👍 comment**
1.0
Optimize relative datetime filter clauses - Related: #4043 In #10980 I added QP middleware to rewrite filters against temporal fields to generate `BETWEEN` clauses or the like and avoid temporal truncation/extraction operations that break index usages. However, this improvement only applied to filters against absolute temporal moments, e.g. "show me toucan sightings where the month is January 2020". Filters against *relative* temporal moments, e.g. "show me toucan sightings in the last month" are not currently being optimized. For example this MBQL query ```clj ;; attempted-murders is our crow sighting test dataset (mt/dataset attempted-murders (mt/mbql-query attempts {:aggregation [[:count]] :filter [:time-interval $datetime :last :month]})) ``` generates this SQL (BigQuery): ```sql SELECT count(*) AS `count` FROM `attempted_murders.attempts` WHERE datetime_trunc(`attempted_murders.attempts`.`datetime`, month) = datetime_trunc(datetime_add(current_datetime(), INTERVAL -1 month), month) ``` It seems possible to instead generate something along the lines of ```sql SELECT count(*) AS `count` FROM `attempted_murders.attempts` WHERE `attempted_murders.attempts`.`datetime` BETWEEN datetime_trunc(datetime_add(current_datetime(), INTERVAL -2 month), month) AND datetime_trunc(datetime_add(current_datetime(), INTERVAL -1 month), month) ``` I could probably get this working in a day. It would be a big performance win. Since #4043 was originally marked as P1 and #10980 didn't fully solve the entire issue I'm going to mark this as P2 and address it as soon as I get a chance. ⬇️ **Please click the 👍 reaction instead of leaving a `+1` or 👍 comment**
process
optimize relative datetime filter clauses related in i added qp middleware to rewrite filters against temporal fields to generate between clauses or the like and avoid temporal truncation extraction operations that break index usages however this improvement only applied to filters against absolute temporal moments e g show me toucan sightings where the month is january filters against relative temporal moments e g show me toucan sightings in the last month are not currently being optimized for example this mbql query clj attempted murders is our crow sighting test dataset mt dataset attempted murders mt mbql query attempts aggregation filter generates this sql bigquery sql select count as count from attempted murders attempts where datetime trunc attempted murders attempts datetime month datetime trunc datetime add current datetime interval month month it seems possible to instead generate something along the lines of sql select count as count from attempted murders attempts where attempted murders attempts datetime between datetime trunc datetime add current datetime interval month month and datetime trunc datetime add current datetime interval month month i could probably get this working in a day it would be a big performance win since was originally marked as and didn t fully solve the entire issue i m going to mark this as and address it as soon as i get a chance ⬇️ please click the 👍 reaction instead of leaving a or 👍 comment
1
2,068
4,876,439,093
IssuesEvent
2016-11-16 12:53:04
neuropoly/spinalcordtoolbox
https://api.github.com/repos/neuropoly/spinalcordtoolbox
closed
[sct_process_segmentation] CSA is corrected with respect to the angle with the B0 field --> should be corrected by the angle with the slice direction
bug priority: high sct_process_segmentation
Computation of CSA is wrong if FOV is tilted --> angle should be computed with respect to the slice direction. Also in the doc it is not clear which angle is computed. Also in the doc there is a typo: ````` angle_image.nii.gz: the cord segmentation (nifti file) where each slice's value is equal to the CSA (mm^2), ````` I bypassed this issue by changing the qmatrix to a diagonal one (as if the FOV wasn't tilted):
1.0
[sct_process_segmentation] CSA is corrected with respect to the angle with the B0 field --> should be corrected by the angle with the slice direction - Computation of CSA is wrong if FOV is tilted --> angle should be computed with respect to the slice direction. Also in the doc it is not clear which angle is computed. Also in the doc there is a typo: ````` angle_image.nii.gz: the cord segmentation (nifti file) where each slice's value is equal to the CSA (mm^2), ````` I bypassed this issue by changing the qmatrix to a diagonal one (as if the FOV wasn't tilted):
process
csa is corrected with respect to the angle with the field should be corrected by the angle with the slice direction computation of csa is wrong if fov is tilted angle should be computed with respect to the slice direction also in the doc it is not clear which angle is computed also in the doc there is a typo angle image nii gz the cord segmentation nifti file where each slice s value is equal to the csa mm i bypassed this issue by changing the qmatrix to a diagonal one as if the fov wasn t tilted
1
108,134
23,538,177,291
IssuesEvent
2022-08-20 01:26:46
microsoft/vscode-cpptools
https://api.github.com/repos/microsoft/vscode-cpptools
opened
With 1.12.1, infinite memory gets used after doing a readability-else-after-return clang-tidy fix with vcFormat on Windows with a file with LF (\n) line endings
bug Language Service regression Feature: Code Formatting Feature: Code Analysis
Use a file with LF line endings on Windows with... ```cpp void CheckModified(int ii) { if (ii == 0) { return; } else { //aaa } } ``` ```json "C_Cpp.codeAnalysis.clangTidy.checks.enabled": [ "readability-else-after-return" ], "C_Cpp.formatting": "vcFormat", ``` Run Code Analysis and do the Fix option. Bug: Infinite memory gets used during some vcFormat operation -- you should close Vs Code when this happens or Windows may become unstable due to the lack of memory.
2.0
With 1.12.1, infinite memory gets used after doing a readability-else-after-return clang-tidy fix with vcFormat on Windows with a file with LF (\n) line endings - Use a file with LF line endings on Windows with... ```cpp void CheckModified(int ii) { if (ii == 0) { return; } else { //aaa } } ``` ```json "C_Cpp.codeAnalysis.clangTidy.checks.enabled": [ "readability-else-after-return" ], "C_Cpp.formatting": "vcFormat", ``` Run Code Analysis and do the Fix option. Bug: Infinite memory gets used during some vcFormat operation -- you should close Vs Code when this happens or Windows may become unstable due to the lack of memory.
non_process
with infinite memory gets used after doing a readability else after return clang tidy fix with vcformat on windows with a file with lf n line endings use a file with lf line endings on windows with cpp void checkmodified int ii if ii return else aaa json c cpp codeanalysis clangtidy checks enabled readability else after return c cpp formatting vcformat run code analysis and do the fix option bug infinite memory gets used during some vcformat operation you should close vs code when this happens or windows may become unstable due to the lack of memory
0
198,502
15,711,105,087
IssuesEvent
2021-03-27 05:23:48
Nobaiha/EECS2311_Project_G4
https://api.github.com/repos/Nobaiha/EECS2311_Project_G4
closed
Musicxml file opens automatically after pressing download
documentation
**Reported by:** Hargovind Singh **Date reported:** March 24th, 2021 **Program (or component) name:** Download screen **Configuration(s):** Operating System: Windows 10 Pro 64 Manufacturer and Model: HP Notebook Processor: Intel® Core™ i5-10210U Memory: 8 GB (RAM) JAVA version: Java-SE 15 Application Version: Downloaded from Github master March 24th. Commit key: 934e1dc **Report type:** Feature error **Reproducibility:** Yes **Severity:** Medium **Problem Summary:** After pressing the download button on the final screen, the generated ‘musicXML’ file opens automatically in a text editor. **Expected Behaviour:** The “Download” feature should simply save the file on my system in a directory of my choosing. **Steps to Reproduce:** Import the master branch in Eclipse from git import Right-click on the project -> Gradle -> Refresh Gradle Click on Run In the pop-up screen, click on Start. In the file upload screen, click on Browser -> In the file selection screen, select a text file containing a tablature. Click on convert and then click on download. Enter the name of the new ‘musicXML’ file and then click ‘Save’. You will see the musicXML open in a text editor. **New or old bug:** New
1.0
Musicxml file opens automatically after pressing download - **Reported by:** Hargovind Singh **Date reported:** March 24th, 2021 **Program (or component) name:** Download screen **Configuration(s):** Operating System: Windows 10 Pro 64 Manufacturer and Model: HP Notebook Processor: Intel® Core™ i5-10210U Memory: 8 GB (RAM) JAVA version: Java-SE 15 Application Version: Downloaded from Github master March 24th. Commit key: 934e1dc **Report type:** Feature error **Reproducibility:** Yes **Severity:** Medium **Problem Summary:** After pressing the download button on the final screen, the generated ‘musicXML’ file opens automatically in a text editor. **Expected Behaviour:** The “Download” feature should simply save the file on my system in a directory of my choosing. **Steps to Reproduce:** Import the master branch in Eclipse from git import Right-click on the project -> Gradle -> Refresh Gradle Click on Run In the pop-up screen, click on Start. In the file upload screen, click on Browser -> In the file selection screen, select a text file containing a tablature. Click on convert and then click on download. Enter the name of the new ‘musicXML’ file and then click ‘Save’. You will see the musicXML open in a text editor. **New or old bug:** New
non_process
musicxml file opens automatically after pressing download reported by hargovind singh date reported march program or component name download screen configuration s operating system windows pro manufacturer and model hp notebook processor intel® core™ memory gb ram java version java se application version downloaded from github master march commit key report type feature error reproducibility yes severity medium problem summary after pressing the download button on the final screen the generated ‘musicxml’ file opens automatically in a text editor expected behaviour the “download” feature should simply save the file on my system in a directory of my choosing steps to reproduce import the master branch in eclipse from git import right click on the project gradle refresh gradle click on run in the pop up screen click on start in the file upload screen click on browser in the file selection screen select a text file containing a tablature click on convert and then click on download enter the name of the new ‘musicxml’ file and then click ‘save’ you will see the musicxml open in a text editor new or old bug new
0
38,439
6,671,256,541
IssuesEvent
2017-10-04 06:09:58
openebs/openebs
https://api.github.com/repos/openebs/openebs
opened
Fix front page of documentation
documentation Hacktoberfest
FEATURE REQUEST **What happened**: - Index is displayed twice on the sidebar & on the right side. **What you expected to happen**: - It'll be better to see the [main content](http://openebs.readthedocs.io/en/latest/getting_started/overview.html) instead of the content page. **How to reproduce it (as minimally and precisely as possible)**: - Visit the docs [online](http://openebs.readthedocs.io/en/latest) or - build them locally
1.0
Fix front page of documentation - FEATURE REQUEST **What happened**: - Index is displayed twice on the sidebar & on the right side. **What you expected to happen**: - It'll be better to see the [main content](http://openebs.readthedocs.io/en/latest/getting_started/overview.html) instead of the content page. **How to reproduce it (as minimally and precisely as possible)**: - Visit the docs [online](http://openebs.readthedocs.io/en/latest) or - build them locally
non_process
fix front page of documentation feature request what happened index is displayed twice on the sidebar on the right side what you expected to happen it ll be better to see the instead of the content page how to reproduce it as minimally and precisely as possible visit the docs or build them locally
0
825,200
31,278,400,537
IssuesEvent
2023-08-22 08:01:31
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
LVGL Module File System Memory Leaks
bug priority: medium area: LVGL
**Describe the bug** Every time LVGL reads the file system through lvgl_fs_open(), a structure fs_file_t is allocated by malloc(). But this file structure is not released when LVGL closes the file through lvgl_fs_close(). Similar operation on directory with lvgl_fs_dir_open() & lvgl_fs_dir_close(), we can see fs_dir_t is properly malloced and released. reference: https://github.com/zephyrproject-rtos/lvgl/blob/zephyr/zephyr/lvgl_fs.c A clear and concise description of what the bug is. Please also mention any information which could help others to understand the problem you're facing: - What target platform are you using? nRF5340 - What have you tried to diagnose or workaround this issue? just added free(file) to lvgl_fs_close() - Is this a regression? If yes, have you been able to "git bisect" it to a specific commit? n/a **To Reproduce** Steps to reproduce the behavior: just create multiple LVGL image objects, the bug occurs when the platform running out of heap memory as a result of file structures not being released. **Expected behavior** LVGL library will no longer able to interact with file system as it couldn't open any files. Images would becomes blank or not being displayed at all. **Impact** not all images on file system could be shown on screen. **Logs and console output** n/a **Environment (please complete the following information):** - Toolchain (e.g Zephyr SDK, ...) Nordic Connect SDK V2.2.0 - Zephyr VERSION_MAJOR = 3 VERSION_MINOR = 2 PATCHLEVEL = 99 VERSION_TWEAK = 0 EXTRAVERSION = **Additional context** n/a
1.0
LVGL Module File System Memory Leaks - **Describe the bug** Every time LVGL reads the file system through lvgl_fs_open(), a structure fs_file_t is allocated by malloc(). But this file structure is not released when LVGL closes the file through lvgl_fs_close(). Similar operation on directory with lvgl_fs_dir_open() & lvgl_fs_dir_close(), we can see fs_dir_t is properly malloced and released. reference: https://github.com/zephyrproject-rtos/lvgl/blob/zephyr/zephyr/lvgl_fs.c A clear and concise description of what the bug is. Please also mention any information which could help others to understand the problem you're facing: - What target platform are you using? nRF5340 - What have you tried to diagnose or workaround this issue? just added free(file) to lvgl_fs_close() - Is this a regression? If yes, have you been able to "git bisect" it to a specific commit? n/a **To Reproduce** Steps to reproduce the behavior: just create multiple LVGL image objects, the bug occurs when the platform running out of heap memory as a result of file structures not being released. **Expected behavior** LVGL library will no longer able to interact with file system as it couldn't open any files. Images would becomes blank or not being displayed at all. **Impact** not all images on file system could be shown on screen. **Logs and console output** n/a **Environment (please complete the following information):** - Toolchain (e.g Zephyr SDK, ...) Nordic Connect SDK V2.2.0 - Zephyr VERSION_MAJOR = 3 VERSION_MINOR = 2 PATCHLEVEL = 99 VERSION_TWEAK = 0 EXTRAVERSION = **Additional context** n/a
non_process
lvgl module file system memory leaks describe the bug every time lvgl reads the file system through lvgl fs open a structure fs file t is allocated by malloc but this file structure is not released when lvgl closes the file through lvgl fs close similar operation on directory with lvgl fs dir open lvgl fs dir close we can see fs dir t is properly malloced and released reference a clear and concise description of what the bug is please also mention any information which could help others to understand the problem you re facing what target platform are you using what have you tried to diagnose or workaround this issue just added free file to lvgl fs close is this a regression if yes have you been able to git bisect it to a specific commit n a to reproduce steps to reproduce the behavior just create multiple lvgl image objects the bug occurs when the platform running out of heap memory as a result of file structures not being released expected behavior lvgl library will no longer able to interact with file system as it couldn t open any files images would becomes blank or not being displayed at all impact not all images on file system could be shown on screen logs and console output n a environment please complete the following information toolchain e g zephyr sdk nordic connect sdk zephyr version major version minor patchlevel version tweak extraversion additional context n a
0
11,599
14,460,248,033
IssuesEvent
2020-12-09 00:07:42
Edher-Nava/4a
https://api.github.com/repos/Edher-Nava/4a
opened
complete_size_estimating_template
process-dashboard
- with the obtained real values complete the LOC estimation form
1.0
complete_size_estimating_template - - with the obtained real values complete the LOC estimation form
process
complete size estimating template with the obtained real values complete the loc estimation form
1
324,213
27,789,322,763
IssuesEvent
2023-03-17 07:30:34
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
ccl/sqlproxyccl/acl: TestParsingErrorHandling failed
C-test-failure O-robot branch-master
ccl/sqlproxyccl/acl.TestParsingErrorHandling [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/9113225?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/9113225?buildTab=artifacts#/) on master @ [11fbbca584e3bd774ebee3ee94c49ca30a096b26](https://github.com/cockroachdb/cockroach/commits/11fbbca584e3bd774ebee3ee94c49ca30a096b26): ``` === RUN TestParsingErrorHandling --- FAIL: TestParsingErrorHandling (0.82s) === RUN TestParsingErrorHandling/error_after_update E230317 06:40:49.045938 115 ccl/sqlproxyccl/acl/file.go:71 [-] 13 error updating access control list from file /artifacts/tmp/_tmp/72c09b96526627cf742479f65b1af982/TestParsingErrorHandling3321990234/001/error_file.idk modified at 2023-03-17 06:40:49.042765008 +0000 UTC: yaml: unmarshal errors: E230317 06:40:49.045938 115 ccl/sqlproxyccl/acl/file.go:71 [-] 13 + line 1: cannot unmarshal !!str ``not yaml`` into acl.AllowlistFile file_test.go:499: Error Trace: github.com/cockroachdb/cockroach/pkg/ccl/sqlproxyccl/acl/file_test.go:499 Error: Not equal: expected: 1 actual : 0 Test: TestParsingErrorHandling/error_after_update E230317 06:40:49.863088 117 ccl/sqlproxyccl/acl/file.go:71 [-] 14 error updating access control list from file /artifacts/tmp/_tmp/72c09b96526627cf742479f65b1af982/TestParsingErrorHandling3321990234/001/allowlist.yaml modified at 2023-03-17 06:40:49.042765008 +0000 UTC: yaml: unmarshal errors: E230317 06:40:49.863088 117 ccl/sqlproxyccl/acl/file.go:71 [-] 14 + line 1: cannot unmarshal !!str ``no long...`` into acl.AllowlistFile E230317 06:40:49.863232 117 ccl/sqlproxyccl/acl/file.go:110 [-] 15 Could not read file /artifacts/tmp/_tmp/72c09b96526627cf742479f65b1af982/TestParsingErrorHandling3321990234/001/allowlist.yaml for *acl.Allowlist: yaml: unmarshal errors: E230317 06:40:49.863232 117 ccl/sqlproxyccl/acl/file.go:110 [-] 15 + line 1: cannot unmarshal !!str ``no long...`` into acl.AllowlistFile E230317 06:40:49.863289 117 ccl/sqlproxyccl/acl/file.go:100 [-] 16 WatchList daemon stopped: context canceled --- FAIL: TestParsingErrorHandling/error_after_update (0.20s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/sql-sessions @cockroachdb/server <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestParsingErrorHandling.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
1.0
ccl/sqlproxyccl/acl: TestParsingErrorHandling failed - ccl/sqlproxyccl/acl.TestParsingErrorHandling [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/9113225?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/9113225?buildTab=artifacts#/) on master @ [11fbbca584e3bd774ebee3ee94c49ca30a096b26](https://github.com/cockroachdb/cockroach/commits/11fbbca584e3bd774ebee3ee94c49ca30a096b26): ``` === RUN TestParsingErrorHandling --- FAIL: TestParsingErrorHandling (0.82s) === RUN TestParsingErrorHandling/error_after_update E230317 06:40:49.045938 115 ccl/sqlproxyccl/acl/file.go:71 [-] 13 error updating access control list from file /artifacts/tmp/_tmp/72c09b96526627cf742479f65b1af982/TestParsingErrorHandling3321990234/001/error_file.idk modified at 2023-03-17 06:40:49.042765008 +0000 UTC: yaml: unmarshal errors: E230317 06:40:49.045938 115 ccl/sqlproxyccl/acl/file.go:71 [-] 13 + line 1: cannot unmarshal !!str ``not yaml`` into acl.AllowlistFile file_test.go:499: Error Trace: github.com/cockroachdb/cockroach/pkg/ccl/sqlproxyccl/acl/file_test.go:499 Error: Not equal: expected: 1 actual : 0 Test: TestParsingErrorHandling/error_after_update E230317 06:40:49.863088 117 ccl/sqlproxyccl/acl/file.go:71 [-] 14 error updating access control list from file /artifacts/tmp/_tmp/72c09b96526627cf742479f65b1af982/TestParsingErrorHandling3321990234/001/allowlist.yaml modified at 2023-03-17 06:40:49.042765008 +0000 UTC: yaml: unmarshal errors: E230317 06:40:49.863088 117 ccl/sqlproxyccl/acl/file.go:71 [-] 14 + line 1: cannot unmarshal !!str ``no long...`` into acl.AllowlistFile E230317 06:40:49.863232 117 ccl/sqlproxyccl/acl/file.go:110 [-] 15 Could not read file /artifacts/tmp/_tmp/72c09b96526627cf742479f65b1af982/TestParsingErrorHandling3321990234/001/allowlist.yaml for *acl.Allowlist: yaml: unmarshal errors: E230317 06:40:49.863232 117 ccl/sqlproxyccl/acl/file.go:110 [-] 15 + line 1: cannot unmarshal !!str ``no long...`` into acl.AllowlistFile E230317 06:40:49.863289 117 ccl/sqlproxyccl/acl/file.go:100 [-] 16 WatchList daemon stopped: context canceled --- FAIL: TestParsingErrorHandling/error_after_update (0.20s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/sql-sessions @cockroachdb/server <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestParsingErrorHandling.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
non_process
ccl sqlproxyccl acl testparsingerrorhandling failed ccl sqlproxyccl acl testparsingerrorhandling with on master run testparsingerrorhandling fail testparsingerrorhandling run testparsingerrorhandling error after update ccl sqlproxyccl acl file go error updating access control list from file artifacts tmp tmp error file idk modified at utc yaml unmarshal errors ccl sqlproxyccl acl file go line cannot unmarshal str not yaml into acl allowlistfile file test go error trace github com cockroachdb cockroach pkg ccl sqlproxyccl acl file test go error not equal expected actual test testparsingerrorhandling error after update ccl sqlproxyccl acl file go error updating access control list from file artifacts tmp tmp allowlist yaml modified at utc yaml unmarshal errors ccl sqlproxyccl acl file go line cannot unmarshal str no long into acl allowlistfile ccl sqlproxyccl acl file go could not read file artifacts tmp tmp allowlist yaml for acl allowlist yaml unmarshal errors ccl sqlproxyccl acl file go line cannot unmarshal str no long into acl allowlistfile ccl sqlproxyccl acl file go watchlist daemon stopped context canceled fail testparsingerrorhandling error after update help see also cc cockroachdb sql sessions cockroachdb server
0
186,573
14,399,530,304
IssuesEvent
2020-12-03 11:03:23
pravega/flink-connectors
https://api.github.com/repos/pravega/flink-connectors
closed
Java 11 upgrade
area/test kind/gap priority/p0 version/0.9.0
**Problem description** Pravega has upgraded into Java 11 in the coming 0.9.0 version while connector is still using Java 8. This causes that the integration tests using Java 8 will fail to compile. We need to investigate and find out another way to deal with these tests. **Problem location** Tests or `build.gradle` **Suggestions for an improvement**
1.0
Java 11 upgrade - **Problem description** Pravega has upgraded into Java 11 in the coming 0.9.0 version while connector is still using Java 8. This causes that the integration tests using Java 8 will fail to compile. We need to investigate and find out another way to deal with these tests. **Problem location** Tests or `build.gradle` **Suggestions for an improvement**
non_process
java upgrade problem description pravega has upgraded into java in the coming version while connector is still using java this causes that the integration tests using java will fail to compile we need to investigate and find out another way to deal with these tests problem location tests or build gradle suggestions for an improvement
0
36,122
6,516,438,815
IssuesEvent
2017-08-27 09:03:51
Zimmi48/bugzilla-test
https://api.github.com/repos/Zimmi48/bugzilla-test
closed
dev/doc/changes does not detail the changes due to econstr
kind: documentation
Note: the issue was created automatically with bugzilla2github tool # Bugzilla Bug ID: 5463 Date: 2017-04-11 16:17:14 +0200 From: @_JasonGross To: CC: coq-bugs-redist@lists.gforge.inria.fr Last updated: 2017-04-27 16:17:01 +0200 ## Bugzilla Comment ID: 15923 Date: 2017-04-11 16:17:14 +0200 From: @_JasonGross ## Bugzilla Comment ID: 15924 Date: 2017-04-11 16:17:35 +0200 From: @_ppedrot ACK. ## Bugzilla Comment ID: 16054 Date: 2017-04-27 16:17:01 +0200 From: @_ppedrot Fixed by 8366b181.
1.0
dev/doc/changes does not detail the changes due to econstr - Note: the issue was created automatically with bugzilla2github tool # Bugzilla Bug ID: 5463 Date: 2017-04-11 16:17:14 +0200 From: @_JasonGross To: CC: coq-bugs-redist@lists.gforge.inria.fr Last updated: 2017-04-27 16:17:01 +0200 ## Bugzilla Comment ID: 15923 Date: 2017-04-11 16:17:14 +0200 From: @_JasonGross ## Bugzilla Comment ID: 15924 Date: 2017-04-11 16:17:35 +0200 From: @_ppedrot ACK. ## Bugzilla Comment ID: 16054 Date: 2017-04-27 16:17:01 +0200 From: @_ppedrot Fixed by 8366b181.
non_process
dev doc changes does not detail the changes due to econstr note the issue was created automatically with tool bugzilla bug id date from jasongross to cc coq bugs redist lists gforge inria fr last updated bugzilla comment id date from jasongross bugzilla comment id date from ppedrot ack bugzilla comment id date from ppedrot fixed by
0
8,242
5,534,293,906
IssuesEvent
2017-03-21 15:08:03
TillF/WASA-SED
https://api.github.com/repos/TillF/WASA-SED
opened
revise system of input files
enhancement usability
eliminate redundancies, strict relational database scheme akin to LUMP structure
True
revise system of input files - eliminate redundancies, strict relational database scheme akin to LUMP structure
non_process
revise system of input files eliminate redundancies strict relational database scheme akin to lump structure
0
21,309
28,503,232,821
IssuesEvent
2023-04-18 19:05:24
HopkinsIDD/flepiMoP
https://api.github.com/repos/HopkinsIDD/flepiMoP
opened
Simple & robust post-processing
post-processing
We need one version of the post-processing that does not depends on anything but the config. For each spatial node, it'll plot each outcome that is used for inference with the corresponding ground truth.
1.0
Simple & robust post-processing - We need one version of the post-processing that does not depends on anything but the config. For each spatial node, it'll plot each outcome that is used for inference with the corresponding ground truth.
process
simple robust post processing we need one version of the post processing that does not depends on anything but the config for each spatial node it ll plot each outcome that is used for inference with the corresponding ground truth
1
560,503
16,598,333,251
IssuesEvent
2021-06-01 15:54:15
nlpsandbox/nlpsandbox
https://api.github.com/repos/nlpsandbox/nlpsandbox
closed
Remove old Docker images from Synapse production site
Priority: Medium
https://www.synapse.org/#!Synapse:syn22277123/docker/ Keep only: - example tool images - i2b2-phi-dataset docker image I will update all the tools so that they push to Synapse instead of DockerHub. This will make life easier to developer who fork the example repo and want to push images to Synapse to easily submit the image to a submission queue. Tagging @nlpsandbox/team
1.0
Remove old Docker images from Synapse production site - https://www.synapse.org/#!Synapse:syn22277123/docker/ Keep only: - example tool images - i2b2-phi-dataset docker image I will update all the tools so that they push to Synapse instead of DockerHub. This will make life easier to developer who fork the example repo and want to push images to Synapse to easily submit the image to a submission queue. Tagging @nlpsandbox/team
non_process
remove old docker images from synapse production site keep only example tool images phi dataset docker image i will update all the tools so that they push to synapse instead of dockerhub this will make life easier to developer who fork the example repo and want to push images to synapse to easily submit the image to a submission queue tagging nlpsandbox team
0
129,705
12,417,288,299
IssuesEvent
2020-05-22 20:16:16
informalsystems/ibc-rs
https://api.github.com/repos/informalsystems/ibc-rs
closed
Clarify relayer algorithm with respect to consensus state changes via clientUpdates
documentation relayer spec
## Summary In the presence of multiple realyers or relayer threads, it can happen that a consensus update will fail if its height is smaller than the client state height. This issue is here just to facilitate discussion and its resolution will be incorporated in the Relayer ADR. ## Problem Definition The relayer scans IBC store for a source chain and creates datagrams for a destination chain. Many of these datagrams include proofs that must verify against a specified client consensus state. Due to the fact that the `appHash` for block `n` is stored in `n+1,` a typical relayer action for X will include: ``` queryX(X, h-1) -> {x, proof, h-1} getHeader(h) -> header sendMsgs(MsgClientUpdate{cl, h}, MsgX{X, proof, h}) ``` One or both of the above messages may fail. This issue is concerned with failure of `MsgClientUpdate`, when it can happen and how it can be solved. ## Proposal It is possible that there are two relayers scanning for different items, X vs Y, and they require consensus state (CS) updates on the same destination chains but at different heights. This could result on one of the relayer transactions to fail, as a `clientUpdate` for height `h` fails if the latest client height is `H >=h`. Here is a proposal of overall handling for message X relaying in general that covers this scenario. We assume relaying for A->B path, the client on B has been created and maybe updated a number of times. The relayer in this example periodically scans the source chain A, but it would work the same for event driven relayer. Here are the steps: - relayer gets the latest height `hx` on A and queries the client height on B, `ha`. - takes the maximum of the two, `h = max(hx, ha)` - queries for X at `h-1` and gets the proof `p` - with `ha` being the latest trusted header, it gets header `h` and any intermediate headers - at this point the relayer builds a number of `MsgUpdateClient(hi)` datagrams and the `MsgX(X, p, h)` datagram, and sends them in a transaction to B. - if `hx < ha` there will be no `MsgUpdateClient` datagrams - if the transaction result indicates client update failure then(*): - it queries the (new) latest state of client on A, `hA` and queries X at `ha-1` - sends `MsgX(X, p, hA)` to A If a failure is seen for the `MsgX..` transaction at either step, it is probably safe for the relayer to ignore it, as it could be caused by another "competing" relayer having submitting the same transaction for X. This needs more investigation but not part of this issue which is only for CS update failures. (*) - Not clear if the result includes failure code/ codespaces for all failed transactions (? I only see one). It is possible that `clientUpdate` fails and `MsgX` succeeds if `ha` CS is present on B. Here is a diagram for the above. ![CS_handling](https://user-images.githubusercontent.com/6361427/81168554-81477900-8f97-11ea-8a0b-ae58834cd108.jpeg) #### For Admin Use - [X] Not duplicate issue - [X] Appropriate labels applied - [X] Appropriate contributors tagged - [X] Contributor assigned/self-assigned
1.0
Clarify relayer algorithm with respect to consensus state changes via clientUpdates - ## Summary In the presence of multiple realyers or relayer threads, it can happen that a consensus update will fail if its height is smaller than the client state height. This issue is here just to facilitate discussion and its resolution will be incorporated in the Relayer ADR. ## Problem Definition The relayer scans IBC store for a source chain and creates datagrams for a destination chain. Many of these datagrams include proofs that must verify against a specified client consensus state. Due to the fact that the `appHash` for block `n` is stored in `n+1,` a typical relayer action for X will include: ``` queryX(X, h-1) -> {x, proof, h-1} getHeader(h) -> header sendMsgs(MsgClientUpdate{cl, h}, MsgX{X, proof, h}) ``` One or both of the above messages may fail. This issue is concerned with failure of `MsgClientUpdate`, when it can happen and how it can be solved. ## Proposal It is possible that there are two relayers scanning for different items, X vs Y, and they require consensus state (CS) updates on the same destination chains but at different heights. This could result on one of the relayer transactions to fail, as a `clientUpdate` for height `h` fails if the latest client height is `H >=h`. Here is a proposal of overall handling for message X relaying in general that covers this scenario. We assume relaying for A->B path, the client on B has been created and maybe updated a number of times. The relayer in this example periodically scans the source chain A, but it would work the same for event driven relayer. Here are the steps: - relayer gets the latest height `hx` on A and queries the client height on B, `ha`. - takes the maximum of the two, `h = max(hx, ha)` - queries for X at `h-1` and gets the proof `p` - with `ha` being the latest trusted header, it gets header `h` and any intermediate headers - at this point the relayer builds a number of `MsgUpdateClient(hi)` datagrams and the `MsgX(X, p, h)` datagram, and sends them in a transaction to B. - if `hx < ha` there will be no `MsgUpdateClient` datagrams - if the transaction result indicates client update failure then(*): - it queries the (new) latest state of client on A, `hA` and queries X at `ha-1` - sends `MsgX(X, p, hA)` to A If a failure is seen for the `MsgX..` transaction at either step, it is probably safe for the relayer to ignore it, as it could be caused by another "competing" relayer having submitting the same transaction for X. This needs more investigation but not part of this issue which is only for CS update failures. (*) - Not clear if the result includes failure code/ codespaces for all failed transactions (? I only see one). It is possible that `clientUpdate` fails and `MsgX` succeeds if `ha` CS is present on B. Here is a diagram for the above. ![CS_handling](https://user-images.githubusercontent.com/6361427/81168554-81477900-8f97-11ea-8a0b-ae58834cd108.jpeg) #### For Admin Use - [X] Not duplicate issue - [X] Appropriate labels applied - [X] Appropriate contributors tagged - [X] Contributor assigned/self-assigned
non_process
clarify relayer algorithm with respect to consensus state changes via clientupdates summary in the presence of multiple realyers or relayer threads it can happen that a consensus update will fail if its height is smaller than the client state height this issue is here just to facilitate discussion and its resolution will be incorporated in the relayer adr problem definition the relayer scans ibc store for a source chain and creates datagrams for a destination chain many of these datagrams include proofs that must verify against a specified client consensus state due to the fact that the apphash for block n is stored in n a typical relayer action for x will include queryx x h x proof h getheader h header sendmsgs msgclientupdate cl h msgx x proof h one or both of the above messages may fail this issue is concerned with failure of msgclientupdate when it can happen and how it can be solved proposal it is possible that there are two relayers scanning for different items x vs y and they require consensus state cs updates on the same destination chains but at different heights this could result on one of the relayer transactions to fail as a clientupdate for height h fails if the latest client height is h h here is a proposal of overall handling for message x relaying in general that covers this scenario we assume relaying for a b path the client on b has been created and maybe updated a number of times the relayer in this example periodically scans the source chain a but it would work the same for event driven relayer here are the steps relayer gets the latest height hx on a and queries the client height on b ha takes the maximum of the two h max hx ha queries for x at h and gets the proof p with ha being the latest trusted header it gets header h and any intermediate headers at this point the relayer builds a number of msgupdateclient hi datagrams and the msgx x p h datagram and sends them in a transaction to b if hx ha there will be no msgupdateclient datagrams if the transaction result indicates client update failure then it queries the new latest state of client on a ha and queries x at ha sends msgx x p ha to a if a failure is seen for the msgx transaction at either step it is probably safe for the relayer to ignore it as it could be caused by another competing relayer having submitting the same transaction for x this needs more investigation but not part of this issue which is only for cs update failures not clear if the result includes failure code codespaces for all failed transactions i only see one it is possible that clientupdate fails and msgx succeeds if ha cs is present on b here is a diagram for the above for admin use not duplicate issue appropriate labels applied appropriate contributors tagged contributor assigned self assigned
0
549,900
16,101,619,729
IssuesEvent
2021-04-27 10:00:39
googleapis/google-cloud-php
https://api.github.com/repos/googleapis/google-cloud-php
opened
Synthesis failed for bigtable
api: bigtable autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate bigtable. :broken_heart: Please investigate and fix this issue within 5 business days. While it remains broken, this library cannot be updated with changes to the bigtable API, and the library grows stale. See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md for trouble shooting tips. Here's the output from running `synth.py`: ``` urce_names_plugin' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:234:1 DEBUG: Rule 'protoc_docs_plugin' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "33b387245455775e0de45869c7355cc5a9e98b396a6fc43b02812a63b75fee20" DEBUG: Call stack for the definition of repository 'protoc_docs_plugin' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:258:1 DEBUG: Rule 'rules_python' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "48f7e716f4098b85296ad93f5a133baf712968c13fbc2fdf3a6136158fe86eac" DEBUG: Call stack for the definition of repository 'rules_python' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:42:1 DEBUG: Rule 'gapic_generator_python' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "fe995def6873fcbdc2a8764ef4bce96eb971a9d1950fe9db9be442f3c64fb3b6" DEBUG: Call stack for the definition of repository 'gapic_generator_python' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:278:1 DEBUG: Rule 'com_googleapis_gapic_generator_go' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "c0d0efba86429cee5e52baf838165b0ed7cafae1748d025abec109d25e006628" DEBUG: Call stack for the definition of repository 'com_googleapis_gapic_generator_go' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:300:1 DEBUG: Rule 'gapic_generator_php' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "3dffc5c34a5f35666843df04b42d6ce1c545b992f9c093a777ec40833b548d86" DEBUG: Call stack for the definition of repository 'gapic_generator_php' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:364:1 DEBUG: Rule 'gapic_generator_csharp' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "4db430cfb9293e4521ec8e8138f8095faf035d8e752cf332d227710d749939eb" DEBUG: Call stack for the definition of repository 'gapic_generator_csharp' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:386:1 DEBUG: Rule 'gapic_generator_ruby' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "a14ec475388542f2ea70d16d75579065758acc4b99fdd6d59463d54e1a9e4499" DEBUG: Call stack for the definition of repository 'gapic_generator_ruby' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:400:1 DEBUG: /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/rules_python/python/pip.bzl:61:5: DEPRECATED: the pip_repositories rule has been replaced with pip_install, please see rules_python 0.1 release notes DEBUG: Rule 'bazel_skylib' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "1dde365491125a3db70731e25658dfdd3bc5dbdfd11b840b3e987ecf043c7ca0" DEBUG: Call stack for the definition of repository 'bazel_skylib' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:35:1 Analyzing: target //google/bigtable/v2:google-cloud-bigtable-v2-php (1 packages loaded, 0 targets configured) ERROR: /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/upb/bazel/upb_proto_library.bzl:257:29: aspect() got unexpected keyword argument 'incompatible_use_toolchain_transition' ERROR: /home/kbuilder/.cache/synthtool/googleapis/google/bigtable/v2/BUILD.bazel:180:1: error loading package '@com_github_grpc_grpc//src/compiler': in /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/com_github_grpc_grpc/bazel/grpc_build_system.bzl: Extension file 'bazel/upb_proto_library.bzl' has errors and referenced by '//google/bigtable/v2:bigtable_php_grpc' ERROR: Analysis of target '//google/bigtable/v2:google-cloud-bigtable-v2-php' failed; build aborted: error loading package '@com_github_grpc_grpc//src/compiler': in /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/com_github_grpc_grpc/bazel/grpc_build_system.bzl: Extension file 'bazel/upb_proto_library.bzl' has errors INFO: Elapsed time: 0.419s INFO: 0 processes. FAILED: Build did NOT complete successfully (2 packages loaded, 0 targets configured) FAILED: Build did NOT complete successfully (2 packages loaded, 0 targets configured) Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/home/kbuilder/.cache/synthtool/google-cloud-php/Bigtable/synth.py", line 29, in <module> bazel_target=f'//google/bigtable/v2:google-cloud-bigtable-v2-php', File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 66, in php_library return self._generate_code(service, version, "php", clean_build, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 204, in _generate_code shell.run(bazel_run_args) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['bazel', '--max_idle_secs=240', 'build', '//google/bigtable/v2:google-cloud-bigtable-v2-php']' returned non-zero exit status 1. 2021-04-27 03:00:37,749 autosynth [ERROR] > Synthesis failed 2021-04-27 03:00:37,750 autosynth [DEBUG] > Running: git reset --hard HEAD HEAD is now at 7b571cd4 feat: Allow RequestWrapperTrait to use its auth cache when fetching ADC (#3959) 2021-04-27 03:00:37,792 autosynth [DEBUG] > Running: git checkout autosynth-bigtable Switched to branch 'autosynth-bigtable' Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 356, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 191, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 336, in _inner_main commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 68, in synthesize_loop has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest) File "/tmpfs/src/github/synthtool/autosynth/synth_toolbox.py", line 259, in synthesize_version_in_new_branch synthesizer.synthesize(synth_log_path, self.environ) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/7050f90e-eb3f-45d4-b797-dd7c5119d172/targets/github%2Fsynthtool;config=default/tests;query=google-cloud-php;failed=false).
1.0
Synthesis failed for bigtable - Hello! Autosynth couldn't regenerate bigtable. :broken_heart: Please investigate and fix this issue within 5 business days. While it remains broken, this library cannot be updated with changes to the bigtable API, and the library grows stale. See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md for trouble shooting tips. Here's the output from running `synth.py`: ``` urce_names_plugin' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:234:1 DEBUG: Rule 'protoc_docs_plugin' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "33b387245455775e0de45869c7355cc5a9e98b396a6fc43b02812a63b75fee20" DEBUG: Call stack for the definition of repository 'protoc_docs_plugin' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:258:1 DEBUG: Rule 'rules_python' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "48f7e716f4098b85296ad93f5a133baf712968c13fbc2fdf3a6136158fe86eac" DEBUG: Call stack for the definition of repository 'rules_python' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:42:1 DEBUG: Rule 'gapic_generator_python' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "fe995def6873fcbdc2a8764ef4bce96eb971a9d1950fe9db9be442f3c64fb3b6" DEBUG: Call stack for the definition of repository 'gapic_generator_python' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:278:1 DEBUG: Rule 'com_googleapis_gapic_generator_go' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "c0d0efba86429cee5e52baf838165b0ed7cafae1748d025abec109d25e006628" DEBUG: Call stack for the definition of repository 'com_googleapis_gapic_generator_go' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:300:1 DEBUG: Rule 'gapic_generator_php' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "3dffc5c34a5f35666843df04b42d6ce1c545b992f9c093a777ec40833b548d86" DEBUG: Call stack for the definition of repository 'gapic_generator_php' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:364:1 DEBUG: Rule 'gapic_generator_csharp' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "4db430cfb9293e4521ec8e8138f8095faf035d8e752cf332d227710d749939eb" DEBUG: Call stack for the definition of repository 'gapic_generator_csharp' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:386:1 DEBUG: Rule 'gapic_generator_ruby' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "a14ec475388542f2ea70d16d75579065758acc4b99fdd6d59463d54e1a9e4499" DEBUG: Call stack for the definition of repository 'gapic_generator_ruby' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:400:1 DEBUG: /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/rules_python/python/pip.bzl:61:5: DEPRECATED: the pip_repositories rule has been replaced with pip_install, please see rules_python 0.1 release notes DEBUG: Rule 'bazel_skylib' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "1dde365491125a3db70731e25658dfdd3bc5dbdfd11b840b3e987ecf043c7ca0" DEBUG: Call stack for the definition of repository 'bazel_skylib' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:35:1 Analyzing: target //google/bigtable/v2:google-cloud-bigtable-v2-php (1 packages loaded, 0 targets configured) ERROR: /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/upb/bazel/upb_proto_library.bzl:257:29: aspect() got unexpected keyword argument 'incompatible_use_toolchain_transition' ERROR: /home/kbuilder/.cache/synthtool/googleapis/google/bigtable/v2/BUILD.bazel:180:1: error loading package '@com_github_grpc_grpc//src/compiler': in /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/com_github_grpc_grpc/bazel/grpc_build_system.bzl: Extension file 'bazel/upb_proto_library.bzl' has errors and referenced by '//google/bigtable/v2:bigtable_php_grpc' ERROR: Analysis of target '//google/bigtable/v2:google-cloud-bigtable-v2-php' failed; build aborted: error loading package '@com_github_grpc_grpc//src/compiler': in /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/com_github_grpc_grpc/bazel/grpc_build_system.bzl: Extension file 'bazel/upb_proto_library.bzl' has errors INFO: Elapsed time: 0.419s INFO: 0 processes. FAILED: Build did NOT complete successfully (2 packages loaded, 0 targets configured) FAILED: Build did NOT complete successfully (2 packages loaded, 0 targets configured) Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/home/kbuilder/.cache/synthtool/google-cloud-php/Bigtable/synth.py", line 29, in <module> bazel_target=f'//google/bigtable/v2:google-cloud-bigtable-v2-php', File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 66, in php_library return self._generate_code(service, version, "php", clean_build, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 204, in _generate_code shell.run(bazel_run_args) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['bazel', '--max_idle_secs=240', 'build', '//google/bigtable/v2:google-cloud-bigtable-v2-php']' returned non-zero exit status 1. 2021-04-27 03:00:37,749 autosynth [ERROR] > Synthesis failed 2021-04-27 03:00:37,750 autosynth [DEBUG] > Running: git reset --hard HEAD HEAD is now at 7b571cd4 feat: Allow RequestWrapperTrait to use its auth cache when fetching ADC (#3959) 2021-04-27 03:00:37,792 autosynth [DEBUG] > Running: git checkout autosynth-bigtable Switched to branch 'autosynth-bigtable' Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 356, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 191, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 336, in _inner_main commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 68, in synthesize_loop has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest) File "/tmpfs/src/github/synthtool/autosynth/synth_toolbox.py", line 259, in synthesize_version_in_new_branch synthesizer.synthesize(synth_log_path, self.environ) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/7050f90e-eb3f-45d4-b797-dd7c5119d172/targets/github%2Fsynthtool;config=default/tests;query=google-cloud-php;failed=false).
non_process
synthesis failed for bigtable hello autosynth couldn t regenerate bigtable broken heart please investigate and fix this issue within business days while it remains broken this library cannot be updated with changes to the bigtable api and the library grows stale see for trouble shooting tips here s the output from running synth py urce names plugin which is a http archive rule definition at home kbuilder cache bazel bazel kbuilder external bazel tools tools build defs repo http bzl home kbuilder cache synthtool googleapis workspace debug rule protoc docs plugin indicated that a canonical reproducible form can be obtained by modifying arguments debug call stack for the definition of repository protoc docs plugin which is a http archive rule definition at home kbuilder cache bazel bazel kbuilder external bazel tools tools build defs repo http bzl home kbuilder cache synthtool googleapis workspace debug rule rules python indicated that a canonical reproducible form can be obtained by modifying arguments debug call stack for the definition of repository rules python which is a http archive rule definition at home kbuilder cache bazel bazel kbuilder external bazel tools tools build defs repo http bzl home kbuilder cache synthtool googleapis workspace debug rule gapic generator python indicated that a canonical reproducible form can be obtained by modifying arguments debug call stack for the definition of repository gapic generator python which is a http archive rule definition at home kbuilder cache bazel bazel kbuilder external bazel tools tools build defs repo http bzl home kbuilder cache synthtool googleapis workspace debug rule com googleapis gapic generator go indicated that a canonical reproducible form can be obtained by modifying arguments debug call stack for the definition of repository com googleapis gapic generator go which is a http archive rule definition at home kbuilder cache bazel bazel kbuilder external bazel tools tools build defs repo http bzl home kbuilder cache synthtool googleapis workspace debug rule gapic generator php indicated that a canonical reproducible form can be obtained by modifying arguments debug call stack for the definition of repository gapic generator php which is a http archive rule definition at home kbuilder cache bazel bazel kbuilder external bazel tools tools build defs repo http bzl home kbuilder cache synthtool googleapis workspace debug rule gapic generator csharp indicated that a canonical reproducible form can be obtained by modifying arguments debug call stack for the definition of repository gapic generator csharp which is a http archive rule definition at home kbuilder cache bazel bazel kbuilder external bazel tools tools build defs repo http bzl home kbuilder cache synthtool googleapis workspace debug rule gapic generator ruby indicated that a canonical reproducible form can be obtained by modifying arguments debug call stack for the definition of repository gapic generator ruby which is a http archive rule definition at home kbuilder cache bazel bazel kbuilder external bazel tools tools build defs repo http bzl home kbuilder cache synthtool googleapis workspace debug home kbuilder cache bazel bazel kbuilder external rules python python pip bzl deprecated the pip repositories rule has been replaced with pip install please see rules python release notes debug rule bazel skylib indicated that a canonical reproducible form can be obtained by modifying arguments debug call stack for the definition of repository bazel skylib which is a http archive rule definition at home kbuilder cache bazel bazel kbuilder external bazel tools tools build defs repo http bzl home kbuilder cache synthtool googleapis workspace analyzing target google bigtable google cloud bigtable php packages loaded targets configured error home kbuilder cache bazel bazel kbuilder external upb bazel upb proto library bzl aspect got unexpected keyword argument incompatible use toolchain transition error home kbuilder cache synthtool googleapis google bigtable build bazel error loading package com github grpc grpc src compiler in home kbuilder cache bazel bazel kbuilder external com github grpc grpc bazel grpc build system bzl extension file bazel upb proto library bzl has errors and referenced by google bigtable bigtable php grpc error analysis of target google bigtable google cloud bigtable php failed build aborted error loading package com github grpc grpc src compiler in home kbuilder cache bazel bazel kbuilder external com github grpc grpc bazel grpc build system bzl extension file bazel upb proto library bzl has errors info elapsed time info processes failed build did not complete successfully packages loaded targets configured failed build did not complete successfully packages loaded targets configured traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file home kbuilder cache synthtool google cloud php bigtable synth py line in bazel target f google bigtable google cloud bigtable php file tmpfs src github synthtool synthtool gcp gapic bazel py line in php library return self generate code service version php clean build kwargs file tmpfs src github synthtool synthtool gcp gapic bazel py line in generate code shell run bazel run args file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status autosynth synthesis failed autosynth running git reset hard head head is now at feat allow requestwrappertrait to use its auth cache when fetching adc autosynth running git checkout autosynth bigtable switched to branch autosynth bigtable traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main commit count synthesize loop x multiple prs change pusher synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize loop has changes toolbox synthesize version in new branch synthesizer youngest file tmpfs src github synthtool autosynth synth toolbox py line in synthesize version in new branch synthesizer synthesize synth log path self environ file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
0
97,740
16,242,452,168
IssuesEvent
2021-05-07 11:11:45
NixOS/nixpkgs
https://api.github.com/repos/NixOS/nixpkgs
closed
Vulnerability roundup 100: keycloak-12.0.4: 4 advisories [6.5]
1.severity: security
[search](https://search.nix.gsc.io/?q=keycloak&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=keycloak+in%3Apath&type=Code) * [ ] [CVE-2020-27838](https://nvd.nist.gov/vuln/detail/CVE-2020-27838) CVSSv3=6.5 (nixos-20.09, nixos-unstable) * [ ] [CVE-2020-1725](https://nvd.nist.gov/vuln/detail/CVE-2020-1725) CVSSv3=5.4 (nixos-20.09, nixos-unstable) * [ ] [CVE-2020-10770](https://nvd.nist.gov/vuln/detail/CVE-2020-10770) CVSSv3=5.3 (nixos-20.09, nixos-unstable) * [ ] [CVE-2020-14302](https://nvd.nist.gov/vuln/detail/CVE-2020-14302) CVSSv3=4.9 (nixos-20.09, nixos-unstable) Scanned versions: nixos-20.09: 12d9950bf47; nixos-unstable: 1f77a4c8c74. Cc @ngerstle Cc @talyz
True
Vulnerability roundup 100: keycloak-12.0.4: 4 advisories [6.5] - [search](https://search.nix.gsc.io/?q=keycloak&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=keycloak+in%3Apath&type=Code) * [ ] [CVE-2020-27838](https://nvd.nist.gov/vuln/detail/CVE-2020-27838) CVSSv3=6.5 (nixos-20.09, nixos-unstable) * [ ] [CVE-2020-1725](https://nvd.nist.gov/vuln/detail/CVE-2020-1725) CVSSv3=5.4 (nixos-20.09, nixos-unstable) * [ ] [CVE-2020-10770](https://nvd.nist.gov/vuln/detail/CVE-2020-10770) CVSSv3=5.3 (nixos-20.09, nixos-unstable) * [ ] [CVE-2020-14302](https://nvd.nist.gov/vuln/detail/CVE-2020-14302) CVSSv3=4.9 (nixos-20.09, nixos-unstable) Scanned versions: nixos-20.09: 12d9950bf47; nixos-unstable: 1f77a4c8c74. Cc @ngerstle Cc @talyz
non_process
vulnerability roundup keycloak advisories nixos nixos unstable nixos nixos unstable nixos nixos unstable nixos nixos unstable scanned versions nixos nixos unstable cc ngerstle cc talyz
0
2,462
5,242,369,847
IssuesEvent
2017-01-31 17:53:16
jlm2017/jlm-video-subtitles
https://api.github.com/repos/jlm2017/jlm-video-subtitles
closed
[subtitles] [eng] «Nous devrons réduire la part de consommation de protéines carnées» - Mélenchon
Language: English Process: [6] Approved
# Video title «Nous devrons réduire la part de consommation de protéines carnées» - Mélenchon # URL https://www.youtube.com/watch?v=0_kz-S1E8S8 # Youtube subtitle language Anglais # Duration 2:11 # URL subtitles https://www.youtube.com/timedtext_editor?ref=player&tab=captions&v=0_kz-S1E8S8&bl=vmp&lang=en&ui=hd&action_mde_edit_form=1
1.0
[subtitles] [eng] «Nous devrons réduire la part de consommation de protéines carnées» - Mélenchon - # Video title «Nous devrons réduire la part de consommation de protéines carnées» - Mélenchon # URL https://www.youtube.com/watch?v=0_kz-S1E8S8 # Youtube subtitle language Anglais # Duration 2:11 # URL subtitles https://www.youtube.com/timedtext_editor?ref=player&tab=captions&v=0_kz-S1E8S8&bl=vmp&lang=en&ui=hd&action_mde_edit_form=1
process
«nous devrons réduire la part de consommation de protéines carnées» mélenchon video title «nous devrons réduire la part de consommation de protéines carnées» mélenchon url youtube subtitle language anglais duration url subtitles
1
1,466
4,045,811,819
IssuesEvent
2016-05-22 08:38:48
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
@keyscope on the root <map> element causes builds to fail
bug P1 preprocess/keyref
It seems that using `@keyscope` on the root `<map>` element causes builds to fail. (Put aside for the moment the fact that a keyspace on `<map>` is a bad practice. The spec allows it, so it should work: "Map authors define the boundaries for key scopes by specifying the `@keyscope` attribute on `<map>` elements, `<topicref>` elements, or elements that are specializations of `<map>` or `<topicref>`. ") Sample map: ```xml <map keyscope="widget"> <title>Keyscope test 4</title> <keydef keys="module"> <topicmeta> <keywords> <keyword>Widget</keyword> </keywords> </topicmeta> </keydef> <topicgroup keyscope="gadget"> <keydef keys="module"> <topicmeta> <keywords> <keyword>Gadget</keyword> </keywords> </topicmeta> </keydef> <topicref href="topic1.xml" keyscope="doodad"> <keydef keys="module"> <topicmeta> <keywords> <keyword>Doodad</keyword> </keywords> </topicmeta> </keydef> </topicref> </topicgroup> </map> ``` With a simple keyref such as `<ph keyref="module"/>` in topic1.xml, the build fails. With a pathed keyref such as `<ph keyref="widget.gadget.doodad.module/>` in topic7.xml, the build still fails. Removing the keyspace definition from the `<map>` and using `<ph keyref="module"/>` in topic1.xml, the key resolves to "Widget," as expected. Using `<ph keyref="gadget.doodad.module"/>`, the key resolves to "Doodad," as expected.
1.0
@keyscope on the root <map> element causes builds to fail - It seems that using `@keyscope` on the root `<map>` element causes builds to fail. (Put aside for the moment the fact that a keyspace on `<map>` is a bad practice. The spec allows it, so it should work: "Map authors define the boundaries for key scopes by specifying the `@keyscope` attribute on `<map>` elements, `<topicref>` elements, or elements that are specializations of `<map>` or `<topicref>`. ") Sample map: ```xml <map keyscope="widget"> <title>Keyscope test 4</title> <keydef keys="module"> <topicmeta> <keywords> <keyword>Widget</keyword> </keywords> </topicmeta> </keydef> <topicgroup keyscope="gadget"> <keydef keys="module"> <topicmeta> <keywords> <keyword>Gadget</keyword> </keywords> </topicmeta> </keydef> <topicref href="topic1.xml" keyscope="doodad"> <keydef keys="module"> <topicmeta> <keywords> <keyword>Doodad</keyword> </keywords> </topicmeta> </keydef> </topicref> </topicgroup> </map> ``` With a simple keyref such as `<ph keyref="module"/>` in topic1.xml, the build fails. With a pathed keyref such as `<ph keyref="widget.gadget.doodad.module/>` in topic7.xml, the build still fails. Removing the keyspace definition from the `<map>` and using `<ph keyref="module"/>` in topic1.xml, the key resolves to "Widget," as expected. Using `<ph keyref="gadget.doodad.module"/>`, the key resolves to "Doodad," as expected.
process
keyscope on the root element causes builds to fail it seems that using keyscope on the root element causes builds to fail put aside for the moment the fact that a keyspace on is a bad practice the spec allows it so it should work map authors define the boundaries for key scopes by specifying the keyscope attribute on elements elements or elements that are specializations of or sample map xml keyscope test widget gadget doodad with a simple keyref such as in xml the build fails with a pathed keyref such as in xml the build still fails removing the keyspace definition from the and using in xml the key resolves to widget as expected using the key resolves to doodad as expected
1
11,155
13,957,693,603
IssuesEvent
2020-10-24 08:11:01
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
PT: Harvesting
Geoportal Harvesting process PT - Portugal
Geoportal Team, Can you please start a harvesting to the Portuguese catalogue? Thank you!
1.0
PT: Harvesting - Geoportal Team, Can you please start a harvesting to the Portuguese catalogue? Thank you!
process
pt harvesting geoportal team can you please start a harvesting to the portuguese catalogue thank you
1
5,798
8,641,430,966
IssuesEvent
2018-11-24 17:40:15
allan-walpy/vulmy
https://api.github.com/repos/allan-walpy/vulmy
closed
add auth [server side]
issue://feature issue://refactor project://Vulmy.Server stage://in_process stage://info
- [x] users info; - [x] database; - [x] service for database; - [x] login; - [x] login with password => get token; - [x] login with id from database using service; - [x] claims; - [x] custom claims; - [x] field for rights; - [x] tree like rights; - [x] ultimate right admin with all rights; - [ ] third party; - [ ] login with google and github *[questionable]*; - [ ] using auth0 *[double questionable]*;
1.0
add auth [server side] - - [x] users info; - [x] database; - [x] service for database; - [x] login; - [x] login with password => get token; - [x] login with id from database using service; - [x] claims; - [x] custom claims; - [x] field for rights; - [x] tree like rights; - [x] ultimate right admin with all rights; - [ ] third party; - [ ] login with google and github *[questionable]*; - [ ] using auth0 *[double questionable]*;
process
add auth users info database service for database login login with password get token login with id from database using service claims custom claims field for rights tree like rights ultimate right admin with all rights third party login with google and github using
1
11,123
13,957,686,249
IssuesEvent
2020-10-24 08:08:53
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
NO: Dataset feeds with multiple links in each entry
Geoportal Harvesting process NO - Norway
Hi! We have some issues regarding dataset entries for atom feeds with multiple files. Looks like the rel-attribute causes some trouble when we are mixing datasets with national coverage and regional coverages in the same dataset entry. We use rel=alternate for the dataset with spatial scope like national and rel=section for those datasets with spatial scope like regional. We now try to set all &quot;rel&quot; attributes to &quot;section&quot; where we have mulitple links in the entries in the dataset feed. The linkage to the datasets from the feed should be fine, but I guess this issue prevents correct linkage result, both in the linkage checker and in the harvested results in the Geoportal? Best regards Lars-Inge Arnevik
1.0
NO: Dataset feeds with multiple links in each entry - Hi! We have some issues regarding dataset entries for atom feeds with multiple files. Looks like the rel-attribute causes some trouble when we are mixing datasets with national coverage and regional coverages in the same dataset entry. We use rel=alternate for the dataset with spatial scope like national and rel=section for those datasets with spatial scope like regional. We now try to set all &quot;rel&quot; attributes to &quot;section&quot; where we have mulitple links in the entries in the dataset feed. The linkage to the datasets from the feed should be fine, but I guess this issue prevents correct linkage result, both in the linkage checker and in the harvested results in the Geoportal? Best regards Lars-Inge Arnevik
process
no dataset feeds with multiple links in each entry hi we have some issues regarding dataset entries for atom feeds with multiple files looks like the rel attribute causes some trouble when we are mixing datasets with national coverage and regional coverages in the same dataset entry we use rel alternate for the dataset with spatial scope like national and rel section for those datasets with spatial scope like regional we now try to set all quot rel quot attributes to quot section quot where we have mulitple links in the entries in the dataset feed the linkage to the datasets from the feed should be fine but i guess this issue prevents correct linkage result both in the linkage checker and in the harvested results in the geoportal best regards lars inge arnevik
1
9,520
12,499,511,769
IssuesEvent
2020-06-01 20:18:21
googleapis/google-cloud-cpp
https://api.github.com/repos/googleapis/google-cloud-cpp
closed
Update README.md in old repo to say the project moved
api: spanner type: process
Update README.md in the OLD repo to LOUDLY document the new location for the code. (is there any type of redirect we can/should put up here?)
1.0
Update README.md in old repo to say the project moved - Update README.md in the OLD repo to LOUDLY document the new location for the code. (is there any type of redirect we can/should put up here?)
process
update readme md in old repo to say the project moved update readme md in the old repo to loudly document the new location for the code is there any type of redirect we can should put up here
1
276,129
23,968,179,508
IssuesEvent
2022-09-13 04:33:39
statechannels/go-nitro
https://api.github.com/repos/statechannels/go-nitro
closed
The message service should buffer more messages
Testground
Right now a message service is using a relatively[small buffered channel](https://github.com/statechannels/go-nitro/blob/220ef0ef8418380c74aa9fc12530dc85f2332621/client/engine/messageservice/simple-tcp/simpletcp-messageservice.go#L38) for in and out. With such small buffers it's easy for them to get filled up which means the message service or engine can get blocked on writing to these channels. I think we can even run into a deadlock situation. Using a larger buffered channel for `in` and `out` helps address this but it seems like a code smell to just increase it. Should the message service be doing it's own buffering somehow?
1.0
The message service should buffer more messages - Right now a message service is using a relatively[small buffered channel](https://github.com/statechannels/go-nitro/blob/220ef0ef8418380c74aa9fc12530dc85f2332621/client/engine/messageservice/simple-tcp/simpletcp-messageservice.go#L38) for in and out. With such small buffers it's easy for them to get filled up which means the message service or engine can get blocked on writing to these channels. I think we can even run into a deadlock situation. Using a larger buffered channel for `in` and `out` helps address this but it seems like a code smell to just increase it. Should the message service be doing it's own buffering somehow?
non_process
the message service should buffer more messages right now a message service is using a relatively for in and out with such small buffers it s easy for them to get filled up which means the message service or engine can get blocked on writing to these channels i think we can even run into a deadlock situation using a larger buffered channel for in and out helps address this but it seems like a code smell to just increase it should the message service be doing it s own buffering somehow
0
9,386
12,392,245,198
IssuesEvent
2020-05-20 13:44:32
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
`findMany` fails to `select` an enum list
process/candidate
I was tempted to post this as a comment on #2442. I have a feeling this, #2442 & #2501 are all the same issue but can't prove it, so I'm creating separate ones. Please move if you think otherwise! Again, I first discovered this while trying to reproduce a report on Studio: https://github.com/prisma/studio/issues/406 ## Bug description For a model with an enum list, certain combinations of `select` crash ## How to reproduce <details> <summary>Use this schema:</summary> ``` generator client { provider = "prisma-client-js" } datasource db { provider = "postgresql" url = "postgresql://siddhant@localhost:5432/experiments" } model Startup { id Int @id name String Contact Contact[] } enum ScoutingFile { One Two } model Contact { id Int @default(autoincrement()) @id firstName String? lastName String? email String @unique startup Startup @relation(fields: [startupId], references: [id]) startupId Int validEmail Boolean @default(true) scoutingFiles ScoutingFile[] } ``` </details> <details> <summary>And this script:</summary> ``` const { PrismaClient } = require("@prisma/client"); const prisma = new PrismaClient(); const main = async () => { const contacts = await prisma.contact.create({ data: { email: "something@example.com", startup: { create: { id: 1, name: "Startup 1", }, }, }, select: { id: true, firstName: true, lastName: true, email: true, startup: true, startupId: true, validEmail: true, scoutingFiles: true, }, }); console.log(contacts); }; main() .catch((e) => console.log(e)) .finally(() => prisma.disconnect()); ``` </details> </details> <details> <summary>You should see this error:</summary> ``` PrismaClientUnknownRequestError: Invalid `prisma.contact.create()` invocation in /Users/siddhant/Code/Tests/experiments/index.ts:6:41 2 3 const prisma = new PrismaClient(); 4 5 const main = async () => { → 6 const contacts = await prisma.contact.create(Error occurred during query execution: ConnectorError(ConnectorError { user_facing_error: None, kind: ConversionError(Custom { kind: InvalidData, error: "List field did not return an Array from database. Type identifier was Enum(\"ScoutingFile\"). Value was Boolean(true)." } 0: backtrace::backtrace::trace 1: backtrace::capture::Backtrace::new_unresolved 2: failure::backtrace::internal::InternalBacktrace::new 3: failure::backtrace::Backtrace::new 4: <quaint::connector::result_set::result_row::ResultRow as sql_query_connector::row::ToSqlRow>::to_sql_row 5: <std::future::GenFuture<T> as core::future::future::Future>::poll 6: <std::future::GenFuture<T> as core::future::future::Future>::poll 7: std::future::poll_with_tls_context 8: <std::future::GenFuture<T> as core::future::future::Future>::poll 9: <std::future::GenFuture<T> as core::future::future::Future>::poll 10: <std::future::GenFuture<T> as core::future::future::Future>::poll 11: <std::future::GenFuture<T> as core::future::future::Future>::poll 12: <std::future::GenFuture<T> as core::future::future::Future>::poll 13: <std::future::GenFuture<T> as core::future::future::Future>::poll 14: <std::future::GenFuture<T> as core::future::future::Future>::poll 15: <std::future::GenFuture<T> as core::future::future::Future>::poll 16: <std::future::GenFuture<T> as core::future::future::Future>::poll 17: <std::future::GenFuture<T> as core::future::future::Future>::poll 18: std::future::poll_with_tls_context 19: <std::future::GenFuture<T> as core::future::future::Future>::poll 20: <std::future::GenFuture<T> as core::future::future::Future>::poll 21: std::panicking::try::do_call 22: __rust_maybe_catch_panic 23: <std::future::GenFuture<T> as core::future::future::Future>::poll 24: <std::future::GenFuture<T> as core::future::future::Future>::poll 25: <std::future::GenFuture<T> as core::future::future::Future>::poll 26: hyper::proto::h1::dispatch::Dispatcher<D,Bs,I,T>::poll_catch 27: <hyper::server::conn::upgrades::UpgradeableConnection<I,S,E> as core::future::future::Future>::poll 28: <hyper::server::conn::spawn_all::NewSvcTask<I,N,S,E,W> as core::future::future::Future>::poll 29: tokio::task::core::Core<T>::poll 30: std::panicking::try::do_call 31: __rust_maybe_catch_panic 32: tokio::task::harness::Harness<T,S>::poll 33: tokio::runtime::thread_pool::worker::GenerationGuard::run_task 34: tokio::runtime::thread_pool::worker::GenerationGuard::run 35: std::thread::local::LocalKey<T>::with 36: tokio::runtime::thread_pool::worker::Worker::run 37: tokio::task::core::Core<T>::poll 38: std::panicking::try::do_call 39: __rust_maybe_catch_panic 40: tokio::task::harness::Harness<T,S>::poll 41: tokio::runtime::blocking::pool::Inner::run 42: tokio::runtime::context::enter 43: std::sys_common::backtrace::__rust_begin_short_backtrace 44: std::panicking::try::do_call 45: __rust_maybe_catch_panic 46: core::ops::function::FnOnce::call_once{{vtable.shim}} 47: <alloc::boxed::Box<F> as core::ops::function::FnOnce<A>>::call_once 48: std::sys::unix::thread::Thread::new::thread_start 49: _ZL12preoptimized ) }) at PrismaClientFetcher.message (/Users/siddhant/Code/Tests/experiments/node_modules/@prisma/client/src/runtime/getPrismaClient.ts:649:46) at processTicksAndRejections (internal/process/task_queues.js:97:5) ``` </details> So the script above seems to only crash when both `startupId` & `scoutingFIles` exist in `select`. I think this is also a `select` bug rather than a `create` bug, because even if you create the records from Postico or something, Client cannot fetch it. ## Expected behavior Script shouldn't crash ## Environment & setup <!-- In which environment does the problem occur --> - OS: macOS - Database: Postgres - Prisma version: ``` @prisma/cli : 2.0.0-beta.5 Current platform : darwin Query Engine : query-engine 0c2ec197653b278b2978845ef958db88824cd82e Migration Engine : migration-engine-cli 0c2ec197653b278b2978845ef958db88824cd82e Introspection Engine : introspection-core 0c2ec197653b278b2978845ef958db88824cd82e Format Binary : prisma-fmt 0c2ec197653b278b2978845ef958db88824cd82e ``` - Node.js version: v12.16.3
1.0
`findMany` fails to `select` an enum list - I was tempted to post this as a comment on #2442. I have a feeling this, #2442 & #2501 are all the same issue but can't prove it, so I'm creating separate ones. Please move if you think otherwise! Again, I first discovered this while trying to reproduce a report on Studio: https://github.com/prisma/studio/issues/406 ## Bug description For a model with an enum list, certain combinations of `select` crash ## How to reproduce <details> <summary>Use this schema:</summary> ``` generator client { provider = "prisma-client-js" } datasource db { provider = "postgresql" url = "postgresql://siddhant@localhost:5432/experiments" } model Startup { id Int @id name String Contact Contact[] } enum ScoutingFile { One Two } model Contact { id Int @default(autoincrement()) @id firstName String? lastName String? email String @unique startup Startup @relation(fields: [startupId], references: [id]) startupId Int validEmail Boolean @default(true) scoutingFiles ScoutingFile[] } ``` </details> <details> <summary>And this script:</summary> ``` const { PrismaClient } = require("@prisma/client"); const prisma = new PrismaClient(); const main = async () => { const contacts = await prisma.contact.create({ data: { email: "something@example.com", startup: { create: { id: 1, name: "Startup 1", }, }, }, select: { id: true, firstName: true, lastName: true, email: true, startup: true, startupId: true, validEmail: true, scoutingFiles: true, }, }); console.log(contacts); }; main() .catch((e) => console.log(e)) .finally(() => prisma.disconnect()); ``` </details> </details> <details> <summary>You should see this error:</summary> ``` PrismaClientUnknownRequestError: Invalid `prisma.contact.create()` invocation in /Users/siddhant/Code/Tests/experiments/index.ts:6:41 2 3 const prisma = new PrismaClient(); 4 5 const main = async () => { → 6 const contacts = await prisma.contact.create(Error occurred during query execution: ConnectorError(ConnectorError { user_facing_error: None, kind: ConversionError(Custom { kind: InvalidData, error: "List field did not return an Array from database. Type identifier was Enum(\"ScoutingFile\"). Value was Boolean(true)." } 0: backtrace::backtrace::trace 1: backtrace::capture::Backtrace::new_unresolved 2: failure::backtrace::internal::InternalBacktrace::new 3: failure::backtrace::Backtrace::new 4: <quaint::connector::result_set::result_row::ResultRow as sql_query_connector::row::ToSqlRow>::to_sql_row 5: <std::future::GenFuture<T> as core::future::future::Future>::poll 6: <std::future::GenFuture<T> as core::future::future::Future>::poll 7: std::future::poll_with_tls_context 8: <std::future::GenFuture<T> as core::future::future::Future>::poll 9: <std::future::GenFuture<T> as core::future::future::Future>::poll 10: <std::future::GenFuture<T> as core::future::future::Future>::poll 11: <std::future::GenFuture<T> as core::future::future::Future>::poll 12: <std::future::GenFuture<T> as core::future::future::Future>::poll 13: <std::future::GenFuture<T> as core::future::future::Future>::poll 14: <std::future::GenFuture<T> as core::future::future::Future>::poll 15: <std::future::GenFuture<T> as core::future::future::Future>::poll 16: <std::future::GenFuture<T> as core::future::future::Future>::poll 17: <std::future::GenFuture<T> as core::future::future::Future>::poll 18: std::future::poll_with_tls_context 19: <std::future::GenFuture<T> as core::future::future::Future>::poll 20: <std::future::GenFuture<T> as core::future::future::Future>::poll 21: std::panicking::try::do_call 22: __rust_maybe_catch_panic 23: <std::future::GenFuture<T> as core::future::future::Future>::poll 24: <std::future::GenFuture<T> as core::future::future::Future>::poll 25: <std::future::GenFuture<T> as core::future::future::Future>::poll 26: hyper::proto::h1::dispatch::Dispatcher<D,Bs,I,T>::poll_catch 27: <hyper::server::conn::upgrades::UpgradeableConnection<I,S,E> as core::future::future::Future>::poll 28: <hyper::server::conn::spawn_all::NewSvcTask<I,N,S,E,W> as core::future::future::Future>::poll 29: tokio::task::core::Core<T>::poll 30: std::panicking::try::do_call 31: __rust_maybe_catch_panic 32: tokio::task::harness::Harness<T,S>::poll 33: tokio::runtime::thread_pool::worker::GenerationGuard::run_task 34: tokio::runtime::thread_pool::worker::GenerationGuard::run 35: std::thread::local::LocalKey<T>::with 36: tokio::runtime::thread_pool::worker::Worker::run 37: tokio::task::core::Core<T>::poll 38: std::panicking::try::do_call 39: __rust_maybe_catch_panic 40: tokio::task::harness::Harness<T,S>::poll 41: tokio::runtime::blocking::pool::Inner::run 42: tokio::runtime::context::enter 43: std::sys_common::backtrace::__rust_begin_short_backtrace 44: std::panicking::try::do_call 45: __rust_maybe_catch_panic 46: core::ops::function::FnOnce::call_once{{vtable.shim}} 47: <alloc::boxed::Box<F> as core::ops::function::FnOnce<A>>::call_once 48: std::sys::unix::thread::Thread::new::thread_start 49: _ZL12preoptimized ) }) at PrismaClientFetcher.message (/Users/siddhant/Code/Tests/experiments/node_modules/@prisma/client/src/runtime/getPrismaClient.ts:649:46) at processTicksAndRejections (internal/process/task_queues.js:97:5) ``` </details> So the script above seems to only crash when both `startupId` & `scoutingFIles` exist in `select`. I think this is also a `select` bug rather than a `create` bug, because even if you create the records from Postico or something, Client cannot fetch it. ## Expected behavior Script shouldn't crash ## Environment & setup <!-- In which environment does the problem occur --> - OS: macOS - Database: Postgres - Prisma version: ``` @prisma/cli : 2.0.0-beta.5 Current platform : darwin Query Engine : query-engine 0c2ec197653b278b2978845ef958db88824cd82e Migration Engine : migration-engine-cli 0c2ec197653b278b2978845ef958db88824cd82e Introspection Engine : introspection-core 0c2ec197653b278b2978845ef958db88824cd82e Format Binary : prisma-fmt 0c2ec197653b278b2978845ef958db88824cd82e ``` - Node.js version: v12.16.3
process
findmany fails to select an enum list i was tempted to post this as a comment on i have a feeling this are all the same issue but can t prove it so i m creating separate ones please move if you think otherwise again i first discovered this while trying to reproduce a report on studio bug description for a model with an enum list certain combinations of select crash how to reproduce use this schema generator client provider prisma client js datasource db provider postgresql url postgresql siddhant localhost experiments model startup id int id name string contact contact enum scoutingfile one two model contact id int default autoincrement id firstname string lastname string email string unique startup startup relation fields references startupid int validemail boolean default true scoutingfiles scoutingfile and this script const prismaclient require prisma client const prisma new prismaclient const main async const contacts await prisma contact create data email something example com startup create id name startup select id true firstname true lastname true email true startup true startupid true validemail true scoutingfiles true console log contacts main catch e console log e finally prisma disconnect you should see this error prismaclientunknownrequesterror invalid prisma contact create invocation in users siddhant code tests experiments index ts const prisma new prismaclient const main async → const contacts await prisma contact create error occurred during query execution connectorerror connectorerror user facing error none kind conversionerror custom kind invaliddata error list field did not return an array from database type identifier was enum scoutingfile value was boolean true backtrace backtrace trace backtrace capture backtrace new unresolved failure backtrace internal internalbacktrace new failure backtrace backtrace new to sql row as core future future future poll as core future future future poll std future poll with tls context as core future future future poll as core future future future poll as core future future future poll as core future future future poll as core future future future poll as core future future future poll as core future future future poll as core future future future poll as core future future future poll as core future future future poll std future poll with tls context as core future future future poll as core future future future poll std panicking try do call rust maybe catch panic as core future future future poll as core future future future poll as core future future future poll hyper proto dispatch dispatcher poll catch as core future future future poll as core future future future poll tokio task core core poll std panicking try do call rust maybe catch panic tokio task harness harness poll tokio runtime thread pool worker generationguard run task tokio runtime thread pool worker generationguard run std thread local localkey with tokio runtime thread pool worker worker run tokio task core core poll std panicking try do call rust maybe catch panic tokio task harness harness poll tokio runtime blocking pool inner run tokio runtime context enter std sys common backtrace rust begin short backtrace std panicking try do call rust maybe catch panic core ops function fnonce call once vtable shim as core ops function fnonce call once std sys unix thread thread new thread start at prismaclientfetcher message users siddhant code tests experiments node modules prisma client src runtime getprismaclient ts at processticksandrejections internal process task queues js so the script above seems to only crash when both startupid scoutingfiles exist in select i think this is also a select bug rather than a create bug because even if you create the records from postico or something client cannot fetch it expected behavior script shouldn t crash environment setup os macos database postgres prisma version prisma cli beta current platform darwin query engine query engine migration engine migration engine cli introspection engine introspection core format binary prisma fmt node js version
1
18,478
24,550,717,615
IssuesEvent
2022-10-12 12:24:24
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] Locations > Search bar placeholder text should be changed
Bug P2 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
Actual: "Search location" Expected: "Search by location ID or name" Note: Functionality wise, search results are getting displayed if admin searches by location ID or location name and hence we have to specify both in placeholder ![Screenshot_4](https://user-images.githubusercontent.com/60386291/180806981-1aff1033-3c93-49eb-b32f-d9f2795fbee5.png)
3.0
[PM] Locations > Search bar placeholder text should be changed - Actual: "Search location" Expected: "Search by location ID or name" Note: Functionality wise, search results are getting displayed if admin searches by location ID or location name and hence we have to specify both in placeholder ![Screenshot_4](https://user-images.githubusercontent.com/60386291/180806981-1aff1033-3c93-49eb-b32f-d9f2795fbee5.png)
process
locations search bar placeholder text should be changed actual search location expected search by location id or name note functionality wise search results are getting displayed if admin searches by location id or location name and hence we have to specify both in placeholder
1
10,148
13,044,162,547
IssuesEvent
2020-07-29 03:47:33
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `RowCount` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `RowCount` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @lonng ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `RowCount` from TiDB - ## Description Port the scalar function `RowCount` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @lonng ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function rowcount from tidb description port the scalar function rowcount from tidb to coprocessor score mentor s lonng recommended skills rust programming learning materials already implemented expressions ported from tidb
1
6,091
8,951,501,866
IssuesEvent
2019-01-25 14:08:53
jasonblais/mattermost-community
https://api.github.com/repos/jasonblais/mattermost-community
opened
Add "PR Submitted" label for issues that already have a PR submitted
Contributor Journey Process
This could be automated for those PRs that reference a given help wanted issue.
1.0
Add "PR Submitted" label for issues that already have a PR submitted - This could be automated for those PRs that reference a given help wanted issue.
process
add pr submitted label for issues that already have a pr submitted this could be automated for those prs that reference a given help wanted issue
1
227,797
18,100,721,721
IssuesEvent
2021-09-22 13:58:32
SPW-DIG/metawal-core-geonetwork
https://api.github.com/repos/SPW-DIG/metawal-core-geonetwork
opened
Impossible d'instaurer un point de moissonnage
criticité.majeur Env test - NOK
Erreur quand j'essaie d'instaurer un nouveau point de moissonnage. ![image](https://user-images.githubusercontent.com/6093597/134357576-ca33c1dd-1382-4411-8fa7-52e19bebc8c4.png)
1.0
Impossible d'instaurer un point de moissonnage - Erreur quand j'essaie d'instaurer un nouveau point de moissonnage. ![image](https://user-images.githubusercontent.com/6093597/134357576-ca33c1dd-1382-4411-8fa7-52e19bebc8c4.png)
non_process
impossible d instaurer un point de moissonnage erreur quand j essaie d instaurer un nouveau point de moissonnage
0
22,612
31,835,739,034
IssuesEvent
2023-09-14 13:25:39
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Status of Bazel 7.0.0-pre.20230906.2
P1 type: process release team-OSS
- Expected release date: 2023-09-13 Task list: - [x] Pick release baseline: [08070932](https://github.com/bazelbuild/bazel/commit/08070932379cd3dafaefe5b546c84ad26cd72951) with cherrypick [4fb701ad](https://github.com/bazelbuild/bazel/commit/4fb701adb5cdf4a87d7457bfe75b76338a8d351a) - [x] Create release candidate: https://releases.bazel.build/7.0.0/rolling/7.0.0-pre.20230906.2rc1/index.html - [x] Post-submit: https://buildkite.com/bazel/bazel-bazel/builds?branch=release-7.0.0-pre.20230906.2rc1 - [x] Push the release: https://releases.bazel.build/7.0.0/rolling/7.0.0-pre.20230906.2/index.html - [x] Update the [release page](https://github.com/bazelbuild/bazel/releases/)
1.0
Status of Bazel 7.0.0-pre.20230906.2 - - Expected release date: 2023-09-13 Task list: - [x] Pick release baseline: [08070932](https://github.com/bazelbuild/bazel/commit/08070932379cd3dafaefe5b546c84ad26cd72951) with cherrypick [4fb701ad](https://github.com/bazelbuild/bazel/commit/4fb701adb5cdf4a87d7457bfe75b76338a8d351a) - [x] Create release candidate: https://releases.bazel.build/7.0.0/rolling/7.0.0-pre.20230906.2rc1/index.html - [x] Post-submit: https://buildkite.com/bazel/bazel-bazel/builds?branch=release-7.0.0-pre.20230906.2rc1 - [x] Push the release: https://releases.bazel.build/7.0.0/rolling/7.0.0-pre.20230906.2/index.html - [x] Update the [release page](https://github.com/bazelbuild/bazel/releases/)
process
status of bazel pre expected release date task list pick release baseline with cherrypick create release candidate post submit push the release update the
1
91,772
8,318,340,440
IssuesEvent
2018-09-25 14:27:31
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
teamcity: failed test: TestSystemZoneConfigs
C-test-failure O-robot
The following tests appear to have failed on master (testrace): TestSystemZoneConfigs You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+TestSystemZoneConfigs). [#924878](https://teamcity.cockroachdb.com/viewLog.html?buildId=924878): ``` TestSystemZoneConfigs ...another peer W180925 14:06:06.633448 93828 rpc/nodedialer/nodedialer.go:89 [ct-client] unable to connect to n6: context canceled W180925 14:06:06.640413 93782 rpc/nodedialer/nodedialer.go:89 [ct-client] unable to connect to n6: context canceled W180925 14:06:06.663728 83534 storage/store.go:1657 [n4,s4,r1/2:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer W180925 14:06:06.701065 82356 storage/store.go:1657 [n1,s1,r1/1:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer W180925 14:06:06.750555 93359 storage/store.go:1838 [n3,s3,r9/10:/Table/1{2-3}] unable to gossip on capacity change: node unavailable; try another peer W180925 14:06:06.775287 83982 ts/db.go:195 [n5,ts-poll] error writing time series data: node unavailable; try another peer W180925 14:06:06.776282 83982 ts/db.go:198 [n5,ts-poll] node unavailable; try another peer I180925 14:06:06.795303 93359 storage/replica_raftstorage.go:809 [n3,s3,r9/10:/Table/1{2-3}] applied Raft snapshot in 2753ms [clear=0ms batch=0ms entries=2706ms commit=1ms] I180925 14:06:06.823772 82365 kv/transport_race.go:91 transport race promotion: ran 103 iterations on up to 7166 requests I180925 14:06:06.824468 93589 util/stop/stopper.go:537 quiescing; tasks left: 1 ts.poller: poll I180925 14:06:06.838366 93591 util/stop/stopper.go:537 quiescing; tasks left: 1 ts.poller: poll W180925 14:06:06.839360 82195 gossip/gossip.go:1513 [n1] no incoming or outgoing connections I180925 14:06:06.848932 83770 gossip/gossip.go:938 [n5] gossip connectivity n1 [sentinel]; n2 -> n1; n3 -> n1; n4 -> n1; n6 -> n4; n7 -> n3; W180925 14:06:06.857675 84263 ts/db.go:195 [n6,ts-poll] error writing time series data: node unavailable; try another peer W180925 14:06:06.864928 84263 ts/db.go:198 [n6,ts-poll] node unavailable; try another peer W180925 14:06:06.875917 83534 storage/store.go:1657 [n4,s4,r1/2:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer W180925 14:06:06.884896 83549 ts/db.go:195 [n4,ts-poll] error writing time series data: node unavailable; try another peer W180925 14:06:06.885481 83549 ts/db.go:198 [n4,ts-poll] node unavailable; try another peer W180925 14:06:06.885855 84850 storage/store.go:1657 [n7,s7,r1/3:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer W180925 14:06:06.897020 82925 ts/db.go:195 [n2,ts-poll] error writing time series data: node unavailable; try another peer W180925 14:06:06.897551 82925 ts/db.go:198 [n2,ts-poll] node unavailable; try another peer W180925 14:06:06.919917 83191 storage/store.go:1657 [n3,s3,r1/5:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer W180925 14:06:06.944004 84850 storage/store.go:1657 [n7,s7,r1/3:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer W180925 14:06:06.951333 83206 ts/db.go:195 [n3,ts-poll] error writing time series data: node unavailable; try another peer W180925 14:06:06.951767 83206 ts/db.go:198 [n3,ts-poll] node unavailable; try another peer W180925 14:06:07.000634 84879 ts/db.go:195 [n7,ts-poll] error writing time series data: node unavailable; try another peer W180925 14:06:07.001219 84879 ts/db.go:198 [n7,ts-poll] node unavailable; try another peer W180925 14:06:07.017024 82386 ts/db.go:195 [n1,ts-poll] error writing time series data: node unavailable; try another peer W180925 14:06:07.039198 84850 storage/store.go:1657 [n7,s7,r1/3:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer I180925 14:06:07.048364 82805 gossip/gossip.go:938 [n2] gossip connectivity n1 [sentinel]; n3 -> n1; n4 -> n1; n5 -> n3; n6 -> n4; n7 -> n3; W180925 14:06:07.237411 84850 storage/store.go:1657 [n7,s7,r1/3:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer TestSystemZoneConfigs ... 5 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.335960 86934 util/stop/stopper.go:537 quiescing; tasks left: 4 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.336013 86938 util/stop/stopper.go:537 quiescing; tasks left: 5 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.336301 86934 util/stop/stopper.go:537 quiescing; tasks left: 3 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.336319 86934 util/stop/stopper.go:537 quiescing; tasks left: 3 [async] closedts-subscription I180925 13:50:12.336332 86934 util/stop/stopper.go:537 quiescing; tasks left: 2 [async] closedts-subscription I180925 13:50:12.336362 86935 util/stop/stopper.go:537 quiescing; tasks left: 5 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.336396 86935 util/stop/stopper.go:537 quiescing; tasks left: 4 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.336412 86934 util/stop/stopper.go:537 quiescing; tasks left: 1 [async] closedts-subscription I180925 13:50:12.336427 86935 util/stop/stopper.go:537 quiescing; tasks left: 3 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.336442 86935 util/stop/stopper.go:537 quiescing; tasks left: 2 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.336454 86935 util/stop/stopper.go:537 quiescing; tasks left: 2 [async] closedts-subscription I180925 13:50:12.336468 86935 util/stop/stopper.go:537 quiescing; tasks left: 1 [async] closedts-subscription I180925 13:50:12.346829 84311 storage/split_queue.go:116 [n6,split,s6,r1/3:/{Min-System/}] shouldQueue: shouldQ=false priority=0.0 I180925 13:50:12.347137 83892 storage/split_queue.go:116 [n5,split,s5,r4/2:/System/{NodeLive…-tsd}] shouldQueue: shouldQ=false priority=0.0 I180925 13:50:12.348460 86937 util/stop/stopper.go:537 quiescing; tasks left: 6 [async] closedts-subscription I180925 13:50:12.348493 86937 util/stop/stopper.go:537 quiescing; tasks left: 5 [async] closedts-subscription I180925 13:50:12.348511 86937 util/stop/stopper.go:537 quiescing; tasks left: 1 [async] closedts-subscription I180925 13:50:12.348607 86939 util/stop/stopper.go:537 quiescing; tasks left: 5 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.348658 86939 util/stop/stopper.go:537 quiescing; tasks left: 4 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.348705 86939 util/stop/stopper.go:537 quiescing; tasks left: 3 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.348783 86939 util/stop/stopper.go:537 quiescing; tasks left: 2 [async] closedts-subscription I180925 13:50:12.348891 86940 util/stop/stopper.go:537 quiescing; tasks left: 5 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.349308 86936 util/stop/stopper.go:537 quiescing; tasks left: 5 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.349380 86936 util/stop/stopper.go:537 quiescing; tasks left: 4 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.349407 86936 util/stop/stopper.go:537 quiescing; tasks left: 3 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.349447 86936 util/stop/stopper.go:537 quiescing; tasks left: 2 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.349746 86940 util/stop/stopper.go:537 quiescing; tasks left: 1 [async] closedts-subscription ``` Please assign, take a look and update the issue accordingly.
1.0
teamcity: failed test: TestSystemZoneConfigs - The following tests appear to have failed on master (testrace): TestSystemZoneConfigs You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+TestSystemZoneConfigs). [#924878](https://teamcity.cockroachdb.com/viewLog.html?buildId=924878): ``` TestSystemZoneConfigs ...another peer W180925 14:06:06.633448 93828 rpc/nodedialer/nodedialer.go:89 [ct-client] unable to connect to n6: context canceled W180925 14:06:06.640413 93782 rpc/nodedialer/nodedialer.go:89 [ct-client] unable to connect to n6: context canceled W180925 14:06:06.663728 83534 storage/store.go:1657 [n4,s4,r1/2:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer W180925 14:06:06.701065 82356 storage/store.go:1657 [n1,s1,r1/1:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer W180925 14:06:06.750555 93359 storage/store.go:1838 [n3,s3,r9/10:/Table/1{2-3}] unable to gossip on capacity change: node unavailable; try another peer W180925 14:06:06.775287 83982 ts/db.go:195 [n5,ts-poll] error writing time series data: node unavailable; try another peer W180925 14:06:06.776282 83982 ts/db.go:198 [n5,ts-poll] node unavailable; try another peer I180925 14:06:06.795303 93359 storage/replica_raftstorage.go:809 [n3,s3,r9/10:/Table/1{2-3}] applied Raft snapshot in 2753ms [clear=0ms batch=0ms entries=2706ms commit=1ms] I180925 14:06:06.823772 82365 kv/transport_race.go:91 transport race promotion: ran 103 iterations on up to 7166 requests I180925 14:06:06.824468 93589 util/stop/stopper.go:537 quiescing; tasks left: 1 ts.poller: poll I180925 14:06:06.838366 93591 util/stop/stopper.go:537 quiescing; tasks left: 1 ts.poller: poll W180925 14:06:06.839360 82195 gossip/gossip.go:1513 [n1] no incoming or outgoing connections I180925 14:06:06.848932 83770 gossip/gossip.go:938 [n5] gossip connectivity n1 [sentinel]; n2 -> n1; n3 -> n1; n4 -> n1; n6 -> n4; n7 -> n3; W180925 14:06:06.857675 84263 ts/db.go:195 [n6,ts-poll] error writing time series data: node unavailable; try another peer W180925 14:06:06.864928 84263 ts/db.go:198 [n6,ts-poll] node unavailable; try another peer W180925 14:06:06.875917 83534 storage/store.go:1657 [n4,s4,r1/2:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer W180925 14:06:06.884896 83549 ts/db.go:195 [n4,ts-poll] error writing time series data: node unavailable; try another peer W180925 14:06:06.885481 83549 ts/db.go:198 [n4,ts-poll] node unavailable; try another peer W180925 14:06:06.885855 84850 storage/store.go:1657 [n7,s7,r1/3:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer W180925 14:06:06.897020 82925 ts/db.go:195 [n2,ts-poll] error writing time series data: node unavailable; try another peer W180925 14:06:06.897551 82925 ts/db.go:198 [n2,ts-poll] node unavailable; try another peer W180925 14:06:06.919917 83191 storage/store.go:1657 [n3,s3,r1/5:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer W180925 14:06:06.944004 84850 storage/store.go:1657 [n7,s7,r1/3:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer W180925 14:06:06.951333 83206 ts/db.go:195 [n3,ts-poll] error writing time series data: node unavailable; try another peer W180925 14:06:06.951767 83206 ts/db.go:198 [n3,ts-poll] node unavailable; try another peer W180925 14:06:07.000634 84879 ts/db.go:195 [n7,ts-poll] error writing time series data: node unavailable; try another peer W180925 14:06:07.001219 84879 ts/db.go:198 [n7,ts-poll] node unavailable; try another peer W180925 14:06:07.017024 82386 ts/db.go:195 [n1,ts-poll] error writing time series data: node unavailable; try another peer W180925 14:06:07.039198 84850 storage/store.go:1657 [n7,s7,r1/3:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer I180925 14:06:07.048364 82805 gossip/gossip.go:938 [n2] gossip connectivity n1 [sentinel]; n3 -> n1; n4 -> n1; n5 -> n3; n6 -> n4; n7 -> n3; W180925 14:06:07.237411 84850 storage/store.go:1657 [n7,s7,r1/3:/{Min-System/}] could not gossip first range descriptor: node unavailable; try another peer TestSystemZoneConfigs ... 5 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.335960 86934 util/stop/stopper.go:537 quiescing; tasks left: 4 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.336013 86938 util/stop/stopper.go:537 quiescing; tasks left: 5 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.336301 86934 util/stop/stopper.go:537 quiescing; tasks left: 3 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.336319 86934 util/stop/stopper.go:537 quiescing; tasks left: 3 [async] closedts-subscription I180925 13:50:12.336332 86934 util/stop/stopper.go:537 quiescing; tasks left: 2 [async] closedts-subscription I180925 13:50:12.336362 86935 util/stop/stopper.go:537 quiescing; tasks left: 5 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.336396 86935 util/stop/stopper.go:537 quiescing; tasks left: 4 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.336412 86934 util/stop/stopper.go:537 quiescing; tasks left: 1 [async] closedts-subscription I180925 13:50:12.336427 86935 util/stop/stopper.go:537 quiescing; tasks left: 3 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.336442 86935 util/stop/stopper.go:537 quiescing; tasks left: 2 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.336454 86935 util/stop/stopper.go:537 quiescing; tasks left: 2 [async] closedts-subscription I180925 13:50:12.336468 86935 util/stop/stopper.go:537 quiescing; tasks left: 1 [async] closedts-subscription I180925 13:50:12.346829 84311 storage/split_queue.go:116 [n6,split,s6,r1/3:/{Min-System/}] shouldQueue: shouldQ=false priority=0.0 I180925 13:50:12.347137 83892 storage/split_queue.go:116 [n5,split,s5,r4/2:/System/{NodeLive…-tsd}] shouldQueue: shouldQ=false priority=0.0 I180925 13:50:12.348460 86937 util/stop/stopper.go:537 quiescing; tasks left: 6 [async] closedts-subscription I180925 13:50:12.348493 86937 util/stop/stopper.go:537 quiescing; tasks left: 5 [async] closedts-subscription I180925 13:50:12.348511 86937 util/stop/stopper.go:537 quiescing; tasks left: 1 [async] closedts-subscription I180925 13:50:12.348607 86939 util/stop/stopper.go:537 quiescing; tasks left: 5 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.348658 86939 util/stop/stopper.go:537 quiescing; tasks left: 4 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.348705 86939 util/stop/stopper.go:537 quiescing; tasks left: 3 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.348783 86939 util/stop/stopper.go:537 quiescing; tasks left: 2 [async] closedts-subscription I180925 13:50:12.348891 86940 util/stop/stopper.go:537 quiescing; tasks left: 5 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.349308 86936 util/stop/stopper.go:537 quiescing; tasks left: 5 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.349380 86936 util/stop/stopper.go:537 quiescing; tasks left: 4 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.349407 86936 util/stop/stopper.go:537 quiescing; tasks left: 3 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.349447 86936 util/stop/stopper.go:537 quiescing; tasks left: 2 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I180925 13:50:12.349746 86940 util/stop/stopper.go:537 quiescing; tasks left: 1 [async] closedts-subscription ``` Please assign, take a look and update the issue accordingly.
non_process
teamcity failed test testsystemzoneconfigs the following tests appear to have failed on master testrace testsystemzoneconfigs you may want to check testsystemzoneconfigs another peer rpc nodedialer nodedialer go unable to connect to context canceled rpc nodedialer nodedialer go unable to connect to context canceled storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer storage store go unable to gossip on capacity change node unavailable try another peer ts db go error writing time series data node unavailable try another peer ts db go node unavailable try another peer storage replica raftstorage go applied raft snapshot in kv transport race go transport race promotion ran iterations on up to requests util stop stopper go quiescing tasks left ts poller poll util stop stopper go quiescing tasks left ts poller poll gossip gossip go no incoming or outgoing connections gossip gossip go gossip connectivity ts db go error writing time series data node unavailable try another peer ts db go node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer ts db go error writing time series data node unavailable try another peer ts db go node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer ts db go error writing time series data node unavailable try another peer ts db go node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer ts db go error writing time series data node unavailable try another peer ts db go node unavailable try another peer ts db go error writing time series data node unavailable try another peer ts db go node unavailable try another peer ts db go error writing time series data node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer gossip gossip go gossip connectivity storage store go could not gossip first range descriptor node unavailable try another peer testsystemzoneconfigs closedts subscription closedts rangefeed subscriber util stop stopper go quiescing tasks left closedts subscription closedts rangefeed subscriber util stop stopper go quiescing tasks left closedts subscription closedts rangefeed subscriber util stop stopper go quiescing tasks left closedts subscription closedts rangefeed subscriber util stop stopper go quiescing tasks left closedts subscription util stop stopper go quiescing tasks left closedts subscription util stop stopper go quiescing tasks left closedts subscription closedts rangefeed subscriber util stop stopper go quiescing tasks left closedts subscription closedts rangefeed subscriber util stop stopper go quiescing tasks left closedts subscription util stop stopper go quiescing tasks left closedts subscription closedts rangefeed subscriber util stop stopper go quiescing tasks left closedts subscription closedts rangefeed subscriber util stop stopper go quiescing tasks left closedts subscription util stop stopper go quiescing tasks left closedts subscription storage split queue go shouldqueue shouldq false priority storage split queue go shouldqueue shouldq false priority util stop stopper go quiescing tasks left closedts subscription util stop stopper go quiescing tasks left closedts subscription util stop stopper go quiescing tasks left closedts subscription util stop stopper go quiescing tasks left closedts subscription closedts rangefeed subscriber util stop stopper go quiescing tasks left closedts subscription closedts rangefeed subscriber util stop stopper go quiescing tasks left closedts subscription closedts rangefeed subscriber util stop stopper go quiescing tasks left closedts subscription util stop stopper go quiescing tasks left closedts subscription closedts rangefeed subscriber util stop stopper go quiescing tasks left closedts subscription closedts rangefeed subscriber util stop stopper go quiescing tasks left closedts subscription closedts rangefeed subscriber util stop stopper go quiescing tasks left closedts subscription closedts rangefeed subscriber util stop stopper go quiescing tasks left closedts subscription closedts rangefeed subscriber util stop stopper go quiescing tasks left closedts subscription please assign take a look and update the issue accordingly
0
181,960
14,894,793,310
IssuesEvent
2021-01-21 08:08:56
neo4j/graph-data-science
https://api.github.com/repos/neo4j/graph-data-science
closed
Graph Data Science "Node Similarity" algorithm documentation is partially unclear
BUG documentation
In https://neo4j.com/docs/graph-data-science/current/algorithms/node-similarity/ similarity algorithm documentation there are a lot of config properties in the "General configuration for algorithm execution on an anonymous graph" section, that are not well explained. For example: "nodeProjection", "relationshipProjection", "nodeQuery", "relationshipQuery", "nodeProperties","relationshipProperties" etc. How to use them? What should those query properties (nodeQuery, relationshipQuery) return?? How they work?? Are they for filtering? Could not find examples explaining their usage (besides "nodeProjection", "relationshipProjection" properties, which are used in the dedicated graph creation configs example in the same doc page, but would be great to see more capabilities of those properties too). "relationshipWeightProperty" configuration property is used in the example in the end, but if we search "relationshipWeightProperty" word on the same page, there is nowhere written about it. It is being clear just logically from the example. Maybe there are other properties that are also acceptable but not written in the documentation??? Also, there is no chapter about the "Similarity algorithms" in the free book of "Graph Algorithms". I mean no one can find info even from that resource. Let's take the example nodes and relationships are these (as in the example in doc page) Also, let's also consider the "strength" field of relationship. Could you give an example using all those config properties in the documentation?? ``` CREATE (alice:Person {name: 'Alice'}), (bob:Person {name: 'Bob'}), (carol:Person {name: 'Carol'}), (dave:Person {name: 'Dave'}), (eve:Person {name: 'Eve'}), (guitar:Instrument {name: 'Guitar'}), (synth:Instrument {name: 'Synthesizer'}), (bongos:Instrument {name: 'Bongos'}), (trumpet:Instrument {name: 'Trumpet'}), (alice)-[:LIKES]->(guitar), (alice)-[:LIKES]->(synth), (alice)-[:LIKES {strength: 0.5}]->(bongos), (bob)-[:LIKES]->(guitar), (bob)-[:LIKES]->(synth), (carol)-[:LIKES]->(bongos), (dave)-[:LIKES]->(guitar), (dave)-[:LIKES]->(synth), (dave)-[:LIKES]->(bongos); ```
1.0
Graph Data Science "Node Similarity" algorithm documentation is partially unclear - In https://neo4j.com/docs/graph-data-science/current/algorithms/node-similarity/ similarity algorithm documentation there are a lot of config properties in the "General configuration for algorithm execution on an anonymous graph" section, that are not well explained. For example: "nodeProjection", "relationshipProjection", "nodeQuery", "relationshipQuery", "nodeProperties","relationshipProperties" etc. How to use them? What should those query properties (nodeQuery, relationshipQuery) return?? How they work?? Are they for filtering? Could not find examples explaining their usage (besides "nodeProjection", "relationshipProjection" properties, which are used in the dedicated graph creation configs example in the same doc page, but would be great to see more capabilities of those properties too). "relationshipWeightProperty" configuration property is used in the example in the end, but if we search "relationshipWeightProperty" word on the same page, there is nowhere written about it. It is being clear just logically from the example. Maybe there are other properties that are also acceptable but not written in the documentation??? Also, there is no chapter about the "Similarity algorithms" in the free book of "Graph Algorithms". I mean no one can find info even from that resource. Let's take the example nodes and relationships are these (as in the example in doc page) Also, let's also consider the "strength" field of relationship. Could you give an example using all those config properties in the documentation?? ``` CREATE (alice:Person {name: 'Alice'}), (bob:Person {name: 'Bob'}), (carol:Person {name: 'Carol'}), (dave:Person {name: 'Dave'}), (eve:Person {name: 'Eve'}), (guitar:Instrument {name: 'Guitar'}), (synth:Instrument {name: 'Synthesizer'}), (bongos:Instrument {name: 'Bongos'}), (trumpet:Instrument {name: 'Trumpet'}), (alice)-[:LIKES]->(guitar), (alice)-[:LIKES]->(synth), (alice)-[:LIKES {strength: 0.5}]->(bongos), (bob)-[:LIKES]->(guitar), (bob)-[:LIKES]->(synth), (carol)-[:LIKES]->(bongos), (dave)-[:LIKES]->(guitar), (dave)-[:LIKES]->(synth), (dave)-[:LIKES]->(bongos); ```
non_process
graph data science node similarity algorithm documentation is partially unclear in similarity algorithm documentation there are a lot of config properties in the general configuration for algorithm execution on an anonymous graph section that are not well explained for example nodeprojection relationshipprojection nodequery relationshipquery nodeproperties relationshipproperties etc how to use them what should those query properties nodequery relationshipquery return how they work are they for filtering could not find examples explaining their usage besides nodeprojection relationshipprojection properties which are used in the dedicated graph creation configs example in the same doc page but would be great to see more capabilities of those properties too relationshipweightproperty configuration property is used in the example in the end but if we search relationshipweightproperty word on the same page there is nowhere written about it it is being clear just logically from the example maybe there are other properties that are also acceptable but not written in the documentation also there is no chapter about the similarity algorithms in the free book of graph algorithms i mean no one can find info even from that resource let s take the example nodes and relationships are these as in the example in doc page also let s also consider the strength field of relationship could you give an example using all those config properties in the documentation create alice person name alice bob person name bob carol person name carol dave person name dave eve person name eve guitar instrument name guitar synth instrument name synthesizer bongos instrument name bongos trumpet instrument name trumpet alice guitar alice synth alice bongos bob guitar bob synth carol bongos dave guitar dave synth dave bongos
0
15,431
19,622,341,197
IssuesEvent
2022-01-07 08:42:06
arcus-azure/arcus.messaging
https://api.github.com/repos/arcus-azure/arcus.messaging
closed
Provide message handler support for Azure Functions
enhancement area:message-processing
**Is your feature request related to a problem? Please describe.** Provide message handler support for Azure Functions which is similar to our message pump scenario, but then by using Azure Function Service Bus bindings. This would allow us to use Azure Functions as the message pump, but still use message handlers with it. **Describe the solution you'd like** Let's work on an API design together on this based on a POC which can become a template as well then.
1.0
Provide message handler support for Azure Functions - **Is your feature request related to a problem? Please describe.** Provide message handler support for Azure Functions which is similar to our message pump scenario, but then by using Azure Function Service Bus bindings. This would allow us to use Azure Functions as the message pump, but still use message handlers with it. **Describe the solution you'd like** Let's work on an API design together on this based on a POC which can become a template as well then.
process
provide message handler support for azure functions is your feature request related to a problem please describe provide message handler support for azure functions which is similar to our message pump scenario but then by using azure function service bus bindings this would allow us to use azure functions as the message pump but still use message handlers with it describe the solution you d like let s work on an api design together on this based on a poc which can become a template as well then
1
2,439
5,219,603,486
IssuesEvent
2017-01-26 19:32:23
opentrials/opentrials
https://api.github.com/repos/opentrials/opentrials
opened
Deduplicate trials via title
2. Ready for Development data cleaning Processors
We currently only consider two trials to be the same if they share a common identifier. That gets many cases, but there're many others that we miss. We want to deduplicate as much trials as possible, but the cost of a false positive is high (as it adds noise to our discrepancies feature, for example), so we want to be conservative. The current code lives on https://github.com/opentrials/processors/blob/4c359df7f258c6eb2dd64a6d4a1ccab3ba2410bc/processors/base/writers/trial.py, and uses `helpers.find_trial_by_identifiers()` to do the deduplication. In general, what we want to do in that method is: 1. Find if the trial already exists (currently only via identifiers, but potentially using other attributes as the title) 2. If it does, check if the current trial is higher priority than the one we have 2.1. If so, overwrite it 2.2. If not, move on. 3. If it doesn't, write the new trial This task is to change step `1`, looking not only for the same identifiers but also the title. Look for exact matches at first. After that's done and merged, we can start developing fuzzy matching strategies. # Requirements * The number of false positives must be very low (ideally zero). If, in your tests, you see that there're many different trials with the same title, look for a different solution (e.g. adding other columns) and write your findings here. * Log when a match was successful, and via which method (identifiers or title)
1.0
Deduplicate trials via title - We currently only consider two trials to be the same if they share a common identifier. That gets many cases, but there're many others that we miss. We want to deduplicate as much trials as possible, but the cost of a false positive is high (as it adds noise to our discrepancies feature, for example), so we want to be conservative. The current code lives on https://github.com/opentrials/processors/blob/4c359df7f258c6eb2dd64a6d4a1ccab3ba2410bc/processors/base/writers/trial.py, and uses `helpers.find_trial_by_identifiers()` to do the deduplication. In general, what we want to do in that method is: 1. Find if the trial already exists (currently only via identifiers, but potentially using other attributes as the title) 2. If it does, check if the current trial is higher priority than the one we have 2.1. If so, overwrite it 2.2. If not, move on. 3. If it doesn't, write the new trial This task is to change step `1`, looking not only for the same identifiers but also the title. Look for exact matches at first. After that's done and merged, we can start developing fuzzy matching strategies. # Requirements * The number of false positives must be very low (ideally zero). If, in your tests, you see that there're many different trials with the same title, look for a different solution (e.g. adding other columns) and write your findings here. * Log when a match was successful, and via which method (identifiers or title)
process
deduplicate trials via title we currently only consider two trials to be the same if they share a common identifier that gets many cases but there re many others that we miss we want to deduplicate as much trials as possible but the cost of a false positive is high as it adds noise to our discrepancies feature for example so we want to be conservative the current code lives on and uses helpers find trial by identifiers to do the deduplication in general what we want to do in that method is find if the trial already exists currently only via identifiers but potentially using other attributes as the title if it does check if the current trial is higher priority than the one we have if so overwrite it if not move on if it doesn t write the new trial this task is to change step looking not only for the same identifiers but also the title look for exact matches at first after that s done and merged we can start developing fuzzy matching strategies requirements the number of false positives must be very low ideally zero if in your tests you see that there re many different trials with the same title look for a different solution e g adding other columns and write your findings here log when a match was successful and via which method identifiers or title
1
4,539
7,374,158,889
IssuesEvent
2018-03-13 19:25:53
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
reopened
Activity Web1 failed: Response Content is not a valid JObject
cxp data-factory doc-bug in-process triaged
Web activity fails if response is valid JSON array. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 34b3d465-6a2b-88ad-6e80-4fb784cd489e * Version Independent ID: 51b70888-d089-7352-4abd-7d587726700a * Content: [Web Activity in Azure Data Factory | Microsoft Docs](https://docs.microsoft.com/en-us/azure/data-factory/control-flow-web-activity) * Content Source: [articles/data-factory/control-flow-web-activity.md](https://github.com/Microsoft/azure-docs/blob/master/articles/data-factory/control-flow-web-activity.md) * Service: **data-factory** * GitHub Login: @sharonlo101 * Microsoft Alias: **shlo**
1.0
Activity Web1 failed: Response Content is not a valid JObject - Web activity fails if response is valid JSON array. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 34b3d465-6a2b-88ad-6e80-4fb784cd489e * Version Independent ID: 51b70888-d089-7352-4abd-7d587726700a * Content: [Web Activity in Azure Data Factory | Microsoft Docs](https://docs.microsoft.com/en-us/azure/data-factory/control-flow-web-activity) * Content Source: [articles/data-factory/control-flow-web-activity.md](https://github.com/Microsoft/azure-docs/blob/master/articles/data-factory/control-flow-web-activity.md) * Service: **data-factory** * GitHub Login: @sharonlo101 * Microsoft Alias: **shlo**
process
activity failed response content is not a valid jobject web activity fails if response is valid json array document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service data factory github login microsoft alias shlo
1
559,328
16,555,994,767
IssuesEvent
2021-05-28 14:00:49
eclipse-sirius/sirius-components
https://api.github.com/repos/eclipse-sirius/sirius-components
closed
Improve the Domain modeler
area: backend difficulty: medium 👩‍💻 priority: high type: enhancement ✨
The current Domain modeler is a first great step but it misses some features: * _Direct Edit_: to be able to change label directly from the diagram * _Advanced Direct Edit_: if I type "myAttribute:String", it will set the type of the attribute to string, we could be inspired by EcoreTools here. It provides many shorctuts to make the life easier, especially regarding references and attributes. Typing "1" will make the reference or the attribute mandatory. Typing "*" will make it a "many". Just typing a name will only change the name, but typing ": someTypeName" will set the type. * _Set Attributes Types_: Today all the attributes are String because of an `error with the expression 'aql:domain::Type::STRING': Invalid enum literal` * _Improve the Design_: we could do netter than this yellow :), icons are low quality, some are missing the default ecore one is used * _Make visible the different attributes on the diagram_: optional, many, abstract class... * _Show Super Types_ * Need a tool in the palette to create Super types
1.0
Improve the Domain modeler - The current Domain modeler is a first great step but it misses some features: * _Direct Edit_: to be able to change label directly from the diagram * _Advanced Direct Edit_: if I type "myAttribute:String", it will set the type of the attribute to string, we could be inspired by EcoreTools here. It provides many shorctuts to make the life easier, especially regarding references and attributes. Typing "1" will make the reference or the attribute mandatory. Typing "*" will make it a "many". Just typing a name will only change the name, but typing ": someTypeName" will set the type. * _Set Attributes Types_: Today all the attributes are String because of an `error with the expression 'aql:domain::Type::STRING': Invalid enum literal` * _Improve the Design_: we could do netter than this yellow :), icons are low quality, some are missing the default ecore one is used * _Make visible the different attributes on the diagram_: optional, many, abstract class... * _Show Super Types_ * Need a tool in the palette to create Super types
non_process
improve the domain modeler the current domain modeler is a first great step but it misses some features direct edit to be able to change label directly from the diagram advanced direct edit if i type myattribute string it will set the type of the attribute to string we could be inspired by ecoretools here it provides many shorctuts to make the life easier especially regarding references and attributes typing will make the reference or the attribute mandatory typing will make it a many just typing a name will only change the name but typing sometypename will set the type set attributes types today all the attributes are string because of an error with the expression aql domain type string invalid enum literal improve the design we could do netter than this yellow icons are low quality some are missing the default ecore one is used make visible the different attributes on the diagram optional many abstract class show super types need a tool in the palette to create super types
0
112,097
17,068,671,811
IssuesEvent
2021-07-07 10:30:49
PrimalSystemStudio/cannibal-cafes
https://api.github.com/repos/PrimalSystemStudio/cannibal-cafes
opened
CVE-2020-35911 (Medium) detected in lock_api-0.3.4.crate
security vulnerability
## CVE-2020-35911 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lock_api-0.3.4.crate</b></p></summary> <p>Wrappers to create fully-featured Mutex and RwLock types. Compatible with no_std.</p> <p>Library home page: <a href="https://crates.io/api/v1/crates/lock_api/0.3.4/download">https://crates.io/api/v1/crates/lock_api/0.3.4/download</a></p> <p> Dependency Hierarchy: - coffee-0.4.1.crate (Root Library) - glutin-0.24.1.crate - winit-0.22.2.crate - parking_lot-0.10.2.crate - :x: **lock_api-0.3.4.crate** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/PrimalSystemStudio/cannibal-cafes/commit/812d9d669ba2b19e9f66446504242712d1c1f8db">812d9d669ba2b19e9f66446504242712d1c1f8db</a></p> <p>Found in base branch: <b>CannibalCafes</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the lock_api crate before 0.4.2 for Rust. A data race can occur because of MappedRwLockReadGuard unsoundness. <p>Publish Date: 2020-12-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35911>CVE-2020-35911</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/Amanieu/parking_lot/releases/tag/lock_api-0.4.2">https://github.com/Amanieu/parking_lot/releases/tag/lock_api-0.4.2</a></p> <p>Release Date: 2020-12-31</p> <p>Fix Resolution: lock_api-0.4.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-35911 (Medium) detected in lock_api-0.3.4.crate - ## CVE-2020-35911 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lock_api-0.3.4.crate</b></p></summary> <p>Wrappers to create fully-featured Mutex and RwLock types. Compatible with no_std.</p> <p>Library home page: <a href="https://crates.io/api/v1/crates/lock_api/0.3.4/download">https://crates.io/api/v1/crates/lock_api/0.3.4/download</a></p> <p> Dependency Hierarchy: - coffee-0.4.1.crate (Root Library) - glutin-0.24.1.crate - winit-0.22.2.crate - parking_lot-0.10.2.crate - :x: **lock_api-0.3.4.crate** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/PrimalSystemStudio/cannibal-cafes/commit/812d9d669ba2b19e9f66446504242712d1c1f8db">812d9d669ba2b19e9f66446504242712d1c1f8db</a></p> <p>Found in base branch: <b>CannibalCafes</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the lock_api crate before 0.4.2 for Rust. A data race can occur because of MappedRwLockReadGuard unsoundness. <p>Publish Date: 2020-12-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35911>CVE-2020-35911</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/Amanieu/parking_lot/releases/tag/lock_api-0.4.2">https://github.com/Amanieu/parking_lot/releases/tag/lock_api-0.4.2</a></p> <p>Release Date: 2020-12-31</p> <p>Fix Resolution: lock_api-0.4.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in lock api crate cve medium severity vulnerability vulnerable library lock api crate wrappers to create fully featured mutex and rwlock types compatible with no std library home page a href dependency hierarchy coffee crate root library glutin crate winit crate parking lot crate x lock api crate vulnerable library found in head commit a href found in base branch cannibalcafes vulnerability details an issue was discovered in the lock api crate before for rust a data race can occur because of mappedrwlockreadguard unsoundness publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lock api step up your open source security game with whitesource
0
243,973
20,600,362,510
IssuesEvent
2022-03-06 06:24:07
metaplex-foundation/metaplex
https://api.github.com/repos/metaplex-foundation/metaplex
closed
hey guys can anyone help me with this error code 0x177d[Bug]:
needs tests bug
### Which package is this bug report for? candy machine cli ### Issue description i cant find any thing on google about this error code 0x177d i cant update the candy machine and i cant upload any new assets. ### Command ```shell C:\Users\user\Desktop\solana-nft\metaplex>ts-node C:\Users\user\Desktop\solana-nft\metaplex\js\packages\cli\src\candy-machine-v2-cli.ts update_candy_machine -e devnet -k C:\Users\user\Desktop\solana-nft\devnet-wallet.json -cp C:\Users\user\Desktop\solana-nft\config.json -c example C:\Users\user\Desktop\solana-nft\assets ``` ### Relevant log output ```shell wallet public key: EHP1Gtd6RRfAxj6EJcASxSk5hhCqQWRZQZjdkU7sexCe Transaction simulation failed: Error processing Instruction 0: custom program error: 0x177d Program cndy3Z4yapfJBmL3ShUp5exZKqR3z33thTzeNMm2gRZ invoke [1] Program log: Instruction: UpdateCandyMachine Program log: Custom program error: 0x177d Program cndy3Z4yapfJBmL3ShUp5exZKqR3z33thTzeNMm2gRZ consumed 7584 of 200000 compute units Program cndy3Z4yapfJBmL3ShUp5exZKqR3z33thTzeNMm2gRZ failed: custom program error: 0x177d Translating error SendTransactionError: failed to send transaction: Transaction simulation failed: Error processing Instruction 0: custom program error: 0x177d at Connection.sendEncodedTransaction (C:\Users\user\Desktop\solana-nft\metaplex\js\node_modules\@solana\web3.js\src\connection.ts:3961:13) at processTicksAndRejections (node:internal/process/task_queues:96:5) at async Connection.sendRawTransaction (C:\Users\user\Desktop\solana-nft\metaplex\js\node_modules\@solana\web3.js\src\connection.ts:3918:20) at async sendAndConfirmRawTransaction (C:\Users\user\Desktop\solana-nft\metaplex\js\node_modules\@solana\web3.js\src\util\send-and-confirm-raw-transaction.ts:27:21) at async Provider.send (C:\Users\user\Desktop\solana-nft\metaplex\js\packages\cli\node_modules\@project-serum\anchor\src\provider.ts:114:18) at async Object.rpc [as updateCandyMachine] (C:\Users\user\Desktop\solana-nft\metaplex\js\packages\cli\node_modules\@project-serum\anchor\src\program\namespace\rpc.ts:19:23) at async Command.<anonymous> (C:\Users\user\Desktop\solana-nft\metaplex\js\packages\cli\src\candy-machine-v2-cli.ts:799:16) { logs: [ 'Program cndy3Z4yapfJBmL3ShUp5exZKqR3z33thTzeNMm2gRZ invoke [1]', 'Program log: Instruction: UpdateCandyMachine', 'Program log: Custom program error: 0x177d', 'Program cndy3Z4yapfJBmL3ShUp5exZKqR3z33thTzeNMm2gRZ consumed 7584 of 200000 compute units', 'Program cndy3Z4yapfJBmL3ShUp5exZKqR3z33thTzeNMm2gRZ failed: custom program error: 0x177d' ] } C:\Users\user\Desktop\solana-nft\metaplex\js\packages\cli\node_modules\@project-serum\anchor\src\error.ts:35 return new ProgramError(errorCode, errorMsg, errorCode + ": " + errorMsg); ^ ProgramError: 6013: Cannot change number of lines unless is a hidden config at Function.parse (C:\Users\user\Desktop\solana-nft\metaplex\js\packages\cli\node_modules\@project-serum\anchor\src\error.ts:35:14) at Object.rpc [as updateCandyMachine] (C:\Users\user\Desktop\solana-nft\metaplex\js\packages\cli\node_modules\@project-serum\anchor\src\program\namespace\rpc.ts:23:42) at processTicksAndRejections (node:internal/process/task_queues:96:5) at async Command.<anonymous> (C:\Users\user\Desktop\solana-nft\metaplex\js\packages\cli\src\candy-machine-v2-cli.ts:799:16) { code: 6013, msg: 'Cannot change number of lines unless is a hidden config' } ``` ### Operating system windows ### Priority this issue should have High (immediate attention needed) ### Check the Docs First - [X] I have checked the docs and it didn't solve my issue
1.0
hey guys can anyone help me with this error code 0x177d[Bug]: - ### Which package is this bug report for? candy machine cli ### Issue description i cant find any thing on google about this error code 0x177d i cant update the candy machine and i cant upload any new assets. ### Command ```shell C:\Users\user\Desktop\solana-nft\metaplex>ts-node C:\Users\user\Desktop\solana-nft\metaplex\js\packages\cli\src\candy-machine-v2-cli.ts update_candy_machine -e devnet -k C:\Users\user\Desktop\solana-nft\devnet-wallet.json -cp C:\Users\user\Desktop\solana-nft\config.json -c example C:\Users\user\Desktop\solana-nft\assets ``` ### Relevant log output ```shell wallet public key: EHP1Gtd6RRfAxj6EJcASxSk5hhCqQWRZQZjdkU7sexCe Transaction simulation failed: Error processing Instruction 0: custom program error: 0x177d Program cndy3Z4yapfJBmL3ShUp5exZKqR3z33thTzeNMm2gRZ invoke [1] Program log: Instruction: UpdateCandyMachine Program log: Custom program error: 0x177d Program cndy3Z4yapfJBmL3ShUp5exZKqR3z33thTzeNMm2gRZ consumed 7584 of 200000 compute units Program cndy3Z4yapfJBmL3ShUp5exZKqR3z33thTzeNMm2gRZ failed: custom program error: 0x177d Translating error SendTransactionError: failed to send transaction: Transaction simulation failed: Error processing Instruction 0: custom program error: 0x177d at Connection.sendEncodedTransaction (C:\Users\user\Desktop\solana-nft\metaplex\js\node_modules\@solana\web3.js\src\connection.ts:3961:13) at processTicksAndRejections (node:internal/process/task_queues:96:5) at async Connection.sendRawTransaction (C:\Users\user\Desktop\solana-nft\metaplex\js\node_modules\@solana\web3.js\src\connection.ts:3918:20) at async sendAndConfirmRawTransaction (C:\Users\user\Desktop\solana-nft\metaplex\js\node_modules\@solana\web3.js\src\util\send-and-confirm-raw-transaction.ts:27:21) at async Provider.send (C:\Users\user\Desktop\solana-nft\metaplex\js\packages\cli\node_modules\@project-serum\anchor\src\provider.ts:114:18) at async Object.rpc [as updateCandyMachine] (C:\Users\user\Desktop\solana-nft\metaplex\js\packages\cli\node_modules\@project-serum\anchor\src\program\namespace\rpc.ts:19:23) at async Command.<anonymous> (C:\Users\user\Desktop\solana-nft\metaplex\js\packages\cli\src\candy-machine-v2-cli.ts:799:16) { logs: [ 'Program cndy3Z4yapfJBmL3ShUp5exZKqR3z33thTzeNMm2gRZ invoke [1]', 'Program log: Instruction: UpdateCandyMachine', 'Program log: Custom program error: 0x177d', 'Program cndy3Z4yapfJBmL3ShUp5exZKqR3z33thTzeNMm2gRZ consumed 7584 of 200000 compute units', 'Program cndy3Z4yapfJBmL3ShUp5exZKqR3z33thTzeNMm2gRZ failed: custom program error: 0x177d' ] } C:\Users\user\Desktop\solana-nft\metaplex\js\packages\cli\node_modules\@project-serum\anchor\src\error.ts:35 return new ProgramError(errorCode, errorMsg, errorCode + ": " + errorMsg); ^ ProgramError: 6013: Cannot change number of lines unless is a hidden config at Function.parse (C:\Users\user\Desktop\solana-nft\metaplex\js\packages\cli\node_modules\@project-serum\anchor\src\error.ts:35:14) at Object.rpc [as updateCandyMachine] (C:\Users\user\Desktop\solana-nft\metaplex\js\packages\cli\node_modules\@project-serum\anchor\src\program\namespace\rpc.ts:23:42) at processTicksAndRejections (node:internal/process/task_queues:96:5) at async Command.<anonymous> (C:\Users\user\Desktop\solana-nft\metaplex\js\packages\cli\src\candy-machine-v2-cli.ts:799:16) { code: 6013, msg: 'Cannot change number of lines unless is a hidden config' } ``` ### Operating system windows ### Priority this issue should have High (immediate attention needed) ### Check the Docs First - [X] I have checked the docs and it didn't solve my issue
non_process
hey guys can anyone help me with this error code which package is this bug report for candy machine cli issue description i cant find any thing on google about this error code i cant update the candy machine and i cant upload any new assets command shell c users user desktop solana nft metaplex ts node c users user desktop solana nft metaplex js packages cli src candy machine cli ts update candy machine e devnet k c users user desktop solana nft devnet wallet json cp c users user desktop solana nft config json c example c users user desktop solana nft assets relevant log output shell wallet public key transaction simulation failed error processing instruction custom program error program invoke program log instruction updatecandymachine program log custom program error program consumed of compute units program failed custom program error translating error sendtransactionerror failed to send transaction transaction simulation failed error processing instruction custom program error at connection sendencodedtransaction c users user desktop solana nft metaplex js node modules solana js src connection ts at processticksandrejections node internal process task queues at async connection sendrawtransaction c users user desktop solana nft metaplex js node modules solana js src connection ts at async sendandconfirmrawtransaction c users user desktop solana nft metaplex js node modules solana js src util send and confirm raw transaction ts at async provider send c users user desktop solana nft metaplex js packages cli node modules project serum anchor src provider ts at async object rpc c users user desktop solana nft metaplex js packages cli node modules project serum anchor src program namespace rpc ts at async command c users user desktop solana nft metaplex js packages cli src candy machine cli ts logs program invoke program log instruction updatecandymachine program log custom program error program consumed of compute units program failed custom program error c users user desktop solana nft metaplex js packages cli node modules project serum anchor src error ts return new programerror errorcode errormsg errorcode errormsg programerror cannot change number of lines unless is a hidden config at function parse c users user desktop solana nft metaplex js packages cli node modules project serum anchor src error ts at object rpc c users user desktop solana nft metaplex js packages cli node modules project serum anchor src program namespace rpc ts at processticksandrejections node internal process task queues at async command c users user desktop solana nft metaplex js packages cli src candy machine cli ts code msg cannot change number of lines unless is a hidden config operating system windows priority this issue should have high immediate attention needed check the docs first i have checked the docs and it didn t solve my issue
0
7,432
10,548,718,617
IssuesEvent
2019-10-03 06:49:20
Altinn/altinn-studio
https://api.github.com/repos/Altinn/altinn-studio
closed
Set step related configuration for an app
altinn-studio process
**Description** There is some app configuration that naturally could be put as workflow configration instead of general metadata. This task is to identify those configuration and define how to set it. **Configuration examples** - Form should be stored in reportee archive?
1.0
Set step related configuration for an app - **Description** There is some app configuration that naturally could be put as workflow configration instead of general metadata. This task is to identify those configuration and define how to set it. **Configuration examples** - Form should be stored in reportee archive?
process
set step related configuration for an app description there is some app configuration that naturally could be put as workflow configration instead of general metadata this task is to identify those configuration and define how to set it configuration examples form should be stored in reportee archive
1
665,316
22,308,568,052
IssuesEvent
2022-06-13 14:59:22
keyko-io/filecoin-verifier-frontend
https://api.github.com/repos/keyko-io/filecoin-verifier-frontend
closed
LOGS: add the search query in the searchbox of the browser
frontend priority:low
add the search query in the searchbox of the browser, so if you input this query in the searchbox you'll have the result of the research
1.0
LOGS: add the search query in the searchbox of the browser - add the search query in the searchbox of the browser, so if you input this query in the searchbox you'll have the result of the research
non_process
logs add the search query in the searchbox of the browser add the search query in the searchbox of the browser so if you input this query in the searchbox you ll have the result of the research
0
479,237
13,793,275,670
IssuesEvent
2020-10-09 14:44:04
Azure/autorest.typescript
https://api.github.com/repos/Azure/autorest.typescript
closed
Wrong code generation for 'formData'
priority-1
Full swagger.yaml ```yaml swagger: "2.0" info: version: 1.0.0 title: Swagger Petstore host: petstore.swagger.io basePath: /v2 consumes: - multipart/form-data paths: /pet: post: summary: Add a new pet to the store description: "" operationId: addPet parameters: - in: formData name: title description: Pet object that needs to be added to the store type: number required: true - in: formData name: subtitle description: Pet object that needs to be added to the store type: string required: true responses: 405: description: Invalid input put: summary: Update an existing pet description: "" operationId: updatePet parameters: - in: formData name: title description: Pet object that needs to be added to the store type: number required: true - in: formData name: subtitle description: Pet object that needs to be added to the store type: string required: true responses: 400: description: Bad request ``` result ```typescript /** * @summary Add a new pet to the store * @param title Pet object that needs to be added to the store * @param [options] The optional parameters * @returns Promise<msRest.RestResponse> */ addPet(title: string, options?: msRest.RequestOptionsBase): Promise<msRest.RestResponse>; ``` expected ```typescript /** * @summary Add a new pet to the store * @param title Pet object that needs to be added to the store * @param subtitle Pet object that needs to be added to the store * @param [options] The optional parameters * @returns Promise<msRest.RestResponse> */ addPet(title: string, subtitle: string, options?: msRest.RequestOptionsBase): Promise<msRest.RestResponse>; ```
1.0
Wrong code generation for 'formData' - Full swagger.yaml ```yaml swagger: "2.0" info: version: 1.0.0 title: Swagger Petstore host: petstore.swagger.io basePath: /v2 consumes: - multipart/form-data paths: /pet: post: summary: Add a new pet to the store description: "" operationId: addPet parameters: - in: formData name: title description: Pet object that needs to be added to the store type: number required: true - in: formData name: subtitle description: Pet object that needs to be added to the store type: string required: true responses: 405: description: Invalid input put: summary: Update an existing pet description: "" operationId: updatePet parameters: - in: formData name: title description: Pet object that needs to be added to the store type: number required: true - in: formData name: subtitle description: Pet object that needs to be added to the store type: string required: true responses: 400: description: Bad request ``` result ```typescript /** * @summary Add a new pet to the store * @param title Pet object that needs to be added to the store * @param [options] The optional parameters * @returns Promise<msRest.RestResponse> */ addPet(title: string, options?: msRest.RequestOptionsBase): Promise<msRest.RestResponse>; ``` expected ```typescript /** * @summary Add a new pet to the store * @param title Pet object that needs to be added to the store * @param subtitle Pet object that needs to be added to the store * @param [options] The optional parameters * @returns Promise<msRest.RestResponse> */ addPet(title: string, subtitle: string, options?: msRest.RequestOptionsBase): Promise<msRest.RestResponse>; ```
non_process
wrong code generation for formdata full swagger yaml yaml swagger info version title swagger petstore host petstore swagger io basepath consumes multipart form data paths pet post summary add a new pet to the store description operationid addpet parameters in formdata name title description pet object that needs to be added to the store type number required true in formdata name subtitle description pet object that needs to be added to the store type string required true responses description invalid input put summary update an existing pet description operationid updatepet parameters in formdata name title description pet object that needs to be added to the store type number required true in formdata name subtitle description pet object that needs to be added to the store type string required true responses description bad request result typescript summary add a new pet to the store param title pet object that needs to be added to the store param the optional parameters returns promise addpet title string options msrest requestoptionsbase promise expected typescript summary add a new pet to the store param title pet object that needs to be added to the store param subtitle pet object that needs to be added to the store param the optional parameters returns promise addpet title string subtitle string options msrest requestoptionsbase promise
0
5,999
8,808,779,764
IssuesEvent
2018-12-27 16:15:30
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
Processing providers and algorithm - standard hint for programmers
Processing Processing help User Manual enhancement
## Description In the "Processing providers and algorithms" section of the user guide (https://docs.qgis.org/testing/en/docs/user_manual/processing_algs/index.html), algorithms and their parameters are described. The parameters are listed with their user interface names. For a developer / script creator that wants to use an algorithm, it is more useful to know the "real" names of the parameters (that have to be used when running the algorithm from Python). Including all these "real" names automatically may be complicated, and doing it manually would probably be too much of an effort. A standard explanation of how to find these names using the Python console could be an alternative. For instance: > To get the names of the algorithm parameters for use in Python scripts and plugins, see "19.7 Using processing algorithms from the console". ## Checklist <!-- This is the issue/bug trackers for [QGIS Documentation](https://docs.qgis.org). Cleaning the queue is a process done by project maintainers, mostly on a volunteer basis. We try to keep the overhead as small as possible and appreciate if you help us to do so by completing the following items. Replace the space between square brackets by an x if it's appropriate. --> - [x] I'm aware this repository is about QGIS Documentation and the issue I'm reporting is related to its usage. If it's related to QGIS application or website, please refer to http://qgis.org/en/site/getinvolved/development/bugreporting.html for the right issue tracker or to the [support channels](http://qgis.org/en/site/forusers/support.html) - [x] I have added a link to the page concerned by the issue
2.0
Processing providers and algorithm - standard hint for programmers - ## Description In the "Processing providers and algorithms" section of the user guide (https://docs.qgis.org/testing/en/docs/user_manual/processing_algs/index.html), algorithms and their parameters are described. The parameters are listed with their user interface names. For a developer / script creator that wants to use an algorithm, it is more useful to know the "real" names of the parameters (that have to be used when running the algorithm from Python). Including all these "real" names automatically may be complicated, and doing it manually would probably be too much of an effort. A standard explanation of how to find these names using the Python console could be an alternative. For instance: > To get the names of the algorithm parameters for use in Python scripts and plugins, see "19.7 Using processing algorithms from the console". ## Checklist <!-- This is the issue/bug trackers for [QGIS Documentation](https://docs.qgis.org). Cleaning the queue is a process done by project maintainers, mostly on a volunteer basis. We try to keep the overhead as small as possible and appreciate if you help us to do so by completing the following items. Replace the space between square brackets by an x if it's appropriate. --> - [x] I'm aware this repository is about QGIS Documentation and the issue I'm reporting is related to its usage. If it's related to QGIS application or website, please refer to http://qgis.org/en/site/getinvolved/development/bugreporting.html for the right issue tracker or to the [support channels](http://qgis.org/en/site/forusers/support.html) - [x] I have added a link to the page concerned by the issue
process
processing providers and algorithm standard hint for programmers description in the processing providers and algorithms section of the user guide algorithms and their parameters are described the parameters are listed with their user interface names for a developer script creator that wants to use an algorithm it is more useful to know the real names of the parameters that have to be used when running the algorithm from python including all these real names automatically may be complicated and doing it manually would probably be too much of an effort a standard explanation of how to find these names using the python console could be an alternative for instance to get the names of the algorithm parameters for use in python scripts and plugins see using processing algorithms from the console checklist this is the issue bug trackers for cleaning the queue is a process done by project maintainers mostly on a volunteer basis we try to keep the overhead as small as possible and appreciate if you help us to do so by completing the following items replace the space between square brackets by an x if it s appropriate i m aware this repository is about qgis documentation and the issue i m reporting is related to its usage if it s related to qgis application or website please refer to for the right issue tracker or to the i have added a link to the page concerned by the issue
1
21,492
29,658,780,902
IssuesEvent
2023-06-10 00:05:55
googleapis/java-logging-logback
https://api.github.com/repos/googleapis/java-logging-logback
opened
Warning: a recent release failed
type: process
The following release PRs may have failed: * #1105 - The release job is 'autorelease: pending', but expected 'autorelease: published'.
1.0
Warning: a recent release failed - The following release PRs may have failed: * #1105 - The release job is 'autorelease: pending', but expected 'autorelease: published'.
process
warning a recent release failed the following release prs may have failed the release job is autorelease pending but expected autorelease published
1
124
3,398,042,355
IssuesEvent
2015-12-02 00:49:13
gillesdegottex/dfasma
https://api.github.com/repos/gillesdegottex/dfasma
closed
qmake5 fails on Centos6
Portability/Distribution
Hello Gilles, on our centos 6 work stations I get the following error when running qmake-qt5 503> qmake-qt5 -makefile -cache cache.tmp "INCLUDEPATH+=/u/formes/share/include" "LIBS+=-L/u/formes/share/lib/x86_64-Linux-rh65" dfasma.pro Project MESSAGE: CONFIG=lex yacc debug exceptions depend_includepath testcase_targets import_plugins import_qpa_plugin qt warn_on release link_prl incremental shared qpa no_mocdepend release qt_no_framework linux unix posix gcc fft_fftw3 file_audio_libsndfile file_sdif precision_double Project MESSAGE: Git: Version: v1.3.4 Project MESSAGE: Git: Branch: HEAD Project MESSAGE: PREFIX=/usr/local Project MESSAGE: PREFIXSHORTCUT=/usr/local Project MESSAGE: For Linux Project MESSAGE: Using GCC compiler Project MESSAGE: For 64bits Project MESSAGE: With double precision Project MESSAGE: Audio file reader: libsndfile Project MESSAGE: FFT Implementation: FFTW3 Project MESSAGE: Files: SDIF support: YES Project ERROR: Unknown module(s) in QT: multimedia we have libQTMultimedia available for Qt5 - in fact I had compîled dfasma 1.2 on the very same machines a few weeks ago. Any idea what could go wrong here? I have checked out tag 1.3.4 Thanks Axel
True
qmake5 fails on Centos6 - Hello Gilles, on our centos 6 work stations I get the following error when running qmake-qt5 503> qmake-qt5 -makefile -cache cache.tmp "INCLUDEPATH+=/u/formes/share/include" "LIBS+=-L/u/formes/share/lib/x86_64-Linux-rh65" dfasma.pro Project MESSAGE: CONFIG=lex yacc debug exceptions depend_includepath testcase_targets import_plugins import_qpa_plugin qt warn_on release link_prl incremental shared qpa no_mocdepend release qt_no_framework linux unix posix gcc fft_fftw3 file_audio_libsndfile file_sdif precision_double Project MESSAGE: Git: Version: v1.3.4 Project MESSAGE: Git: Branch: HEAD Project MESSAGE: PREFIX=/usr/local Project MESSAGE: PREFIXSHORTCUT=/usr/local Project MESSAGE: For Linux Project MESSAGE: Using GCC compiler Project MESSAGE: For 64bits Project MESSAGE: With double precision Project MESSAGE: Audio file reader: libsndfile Project MESSAGE: FFT Implementation: FFTW3 Project MESSAGE: Files: SDIF support: YES Project ERROR: Unknown module(s) in QT: multimedia we have libQTMultimedia available for Qt5 - in fact I had compîled dfasma 1.2 on the very same machines a few weeks ago. Any idea what could go wrong here? I have checked out tag 1.3.4 Thanks Axel
non_process
fails on hello gilles on our centos work stations i get the following error when running qmake qmake makefile cache cache tmp includepath u formes share include libs l u formes share lib linux dfasma pro project message config lex yacc debug exceptions depend includepath testcase targets import plugins import qpa plugin qt warn on release link prl incremental shared qpa no mocdepend release qt no framework linux unix posix gcc fft file audio libsndfile file sdif precision double project message git version project message git branch head project message prefix usr local project message prefixshortcut usr local project message for linux project message using gcc compiler project message for project message with double precision project message audio file reader libsndfile project message fft implementation project message files sdif support yes project error unknown module s in qt multimedia we have libqtmultimedia available for in fact i had compîled dfasma on the very same machines a few weeks ago any idea what could go wrong here i have checked out tag thanks axel
0
4,374
7,260,516,089
IssuesEvent
2018-02-18 10:54:46
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
[FEATURE][processing] Snap geometries algorithm allows snapping to other layer types, supports point/line layers
Automatic new feature Processing
Original commit: https://github.com/qgis/QGIS/commit/dae0a017617d3082b6e89f8a3e75c2c9bcd0ebf5 by nyalldawson Fix #14791, #15313
1.0
[FEATURE][processing] Snap geometries algorithm allows snapping to other layer types, supports point/line layers - Original commit: https://github.com/qgis/QGIS/commit/dae0a017617d3082b6e89f8a3e75c2c9bcd0ebf5 by nyalldawson Fix #14791, #15313
process
snap geometries algorithm allows snapping to other layer types supports point line layers original commit by nyalldawson fix
1
18,986
24,977,626,393
IssuesEvent
2022-11-02 09:11:38
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
meiotic G2 phase obsolete
PomBase cell cycle and DNA processes obsoletion
Please provide as much information as you can: * **GO term ID and Label** meiotic G2 phase The cell cycle 'gap' phase which is the interval between the completion of DNA synthesis and the beginning of DNA segregation by meiosis. * **Reason for deprecation** I had never heard of this but Googling I found: PMID: 16546077 by A Hochwagen Angelika Amon· 2006 · Cited by 169 — ... after premeiotic DNA replication (Figure 1B), during a stage that has variably been called meiotic prophase and meiotic G2 phase. * **"Consider" term(s) (ID and label)** GO:0007128 meiotic prophase I * **Are there annotations to this term?** - How many EXP: Not that the check here will need to be for EXTENSIONS that use this term because there are no direct annotations to 'phase' terms. * **Are there mappings and cross references to this term? (InterPro, Keywords; check QuickGO cross-references section)** * **Is this term in a subset? (check the AmiGO page for that term)** * **Any other information**
1.0
meiotic G2 phase obsolete - Please provide as much information as you can: * **GO term ID and Label** meiotic G2 phase The cell cycle 'gap' phase which is the interval between the completion of DNA synthesis and the beginning of DNA segregation by meiosis. * **Reason for deprecation** I had never heard of this but Googling I found: PMID: 16546077 by A Hochwagen Angelika Amon· 2006 · Cited by 169 — ... after premeiotic DNA replication (Figure 1B), during a stage that has variably been called meiotic prophase and meiotic G2 phase. * **"Consider" term(s) (ID and label)** GO:0007128 meiotic prophase I * **Are there annotations to this term?** - How many EXP: Not that the check here will need to be for EXTENSIONS that use this term because there are no direct annotations to 'phase' terms. * **Are there mappings and cross references to this term? (InterPro, Keywords; check QuickGO cross-references section)** * **Is this term in a subset? (check the AmiGO page for that term)** * **Any other information**
process
meiotic phase obsolete please provide as much information as you can go term id and label meiotic phase the cell cycle gap phase which is the interval between the completion of dna synthesis and the beginning of dna segregation by meiosis reason for deprecation i had never heard of this but googling i found pmid by a hochwagen angelika amon· · cited by — after premeiotic dna replication figure during a stage that has variably been called meiotic prophase and meiotic phase consider term s id and label go meiotic prophase i are there annotations to this term how many exp not that the check here will need to be for extensions that use this term because there are no direct annotations to phase terms are there mappings and cross references to this term interpro keywords check quickgo cross references section is this term in a subset check the amigo page for that term any other information
1
8,144
11,352,593,060
IssuesEvent
2020-01-24 13:58:28
googleapis/google-cloud-python
https://api.github.com/repos/googleapis/google-cloud-python
opened
Storage: bump google-auth dependency to 1.11.0+
api: storage type: process
The `google-auth` [release](https://github.com/googleapis/google-auth-library-python/releases/tag/v1.11.0) version 1.11.0 fixes the issue with too eager Timeout errors in cases when the underlying request takes a lot of time, but still succeeds and does not timeout itself. In order to benefit from it, the version pin needs to be updated.
1.0
Storage: bump google-auth dependency to 1.11.0+ - The `google-auth` [release](https://github.com/googleapis/google-auth-library-python/releases/tag/v1.11.0) version 1.11.0 fixes the issue with too eager Timeout errors in cases when the underlying request takes a lot of time, but still succeeds and does not timeout itself. In order to benefit from it, the version pin needs to be updated.
process
storage bump google auth dependency to the google auth version fixes the issue with too eager timeout errors in cases when the underlying request takes a lot of time but still succeeds and does not timeout itself in order to benefit from it the version pin needs to be updated
1
59,255
14,369,091,538
IssuesEvent
2020-12-01 09:19:08
ignatandrei/stankins
https://api.github.com/repos/ignatandrei/stankins
closed
WS-2020-0091 (High) detected in http-proxy-1.17.0.tgz
security vulnerability
## WS-2020-0091 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>http-proxy-1.17.0.tgz</b></p></summary> <p>HTTP proxying for the masses</p> <p>Library home page: <a href="https://registry.npmjs.org/http-proxy/-/http-proxy-1.17.0.tgz">https://registry.npmjs.org/http-proxy/-/http-proxy-1.17.0.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsAliveAngular/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsDataWebAngular/node_modules/http-proxy/package.json,/tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsDataWebAngular/node_modules/http-proxy/package.json</p> <p> Dependency Hierarchy: - karma-3.0.0.tgz (Root Library) - :x: **http-proxy-1.17.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ignatandrei/stankins/commit/525550ef1e023c62d5d53d2f2bce03d5d168d46e">525550ef1e023c62d5d53d2f2bce03d5d168d46e</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions of http-proxy prior to 1.18.1 are vulnerable to Denial of Service. An HTTP request with a long body triggers an ERR_HTTP_HEADERS_SENT unhandled exception that crashes the proxy server. This is only possible when the proxy server sets headers in the proxy request using the proxyReq.setHeader function. <p>Publish Date: 2020-05-14 <p>URL: <a href=https://github.com/http-party/node-http-proxy/pull/1447>WS-2020-0091</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1486">https://www.npmjs.com/advisories/1486</a></p> <p>Release Date: 2020-05-26</p> <p>Fix Resolution: http-proxy - 1.18.1 </p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2020-0091 (High) detected in http-proxy-1.17.0.tgz - ## WS-2020-0091 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>http-proxy-1.17.0.tgz</b></p></summary> <p>HTTP proxying for the masses</p> <p>Library home page: <a href="https://registry.npmjs.org/http-proxy/-/http-proxy-1.17.0.tgz">https://registry.npmjs.org/http-proxy/-/http-proxy-1.17.0.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsAliveAngular/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsDataWebAngular/node_modules/http-proxy/package.json,/tmp/ws-scm/stankins/stankinsv2/solution/StankinsV2/StankinsDataWebAngular/node_modules/http-proxy/package.json</p> <p> Dependency Hierarchy: - karma-3.0.0.tgz (Root Library) - :x: **http-proxy-1.17.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ignatandrei/stankins/commit/525550ef1e023c62d5d53d2f2bce03d5d168d46e">525550ef1e023c62d5d53d2f2bce03d5d168d46e</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions of http-proxy prior to 1.18.1 are vulnerable to Denial of Service. An HTTP request with a long body triggers an ERR_HTTP_HEADERS_SENT unhandled exception that crashes the proxy server. This is only possible when the proxy server sets headers in the proxy request using the proxyReq.setHeader function. <p>Publish Date: 2020-05-14 <p>URL: <a href=https://github.com/http-party/node-http-proxy/pull/1447>WS-2020-0091</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1486">https://www.npmjs.com/advisories/1486</a></p> <p>Release Date: 2020-05-26</p> <p>Fix Resolution: http-proxy - 1.18.1 </p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws high detected in http proxy tgz ws high severity vulnerability vulnerable library http proxy tgz http proxying for the masses library home page a href path to dependency file tmp ws scm stankins solution stankinsaliveangular package json path to vulnerable library tmp ws scm stankins solution stankinsdatawebangular node modules http proxy package json tmp ws scm stankins solution stankinsdatawebangular node modules http proxy package json dependency hierarchy karma tgz root library x http proxy tgz vulnerable library found in head commit a href vulnerability details versions of http proxy prior to are vulnerable to denial of service an http request with a long body triggers an err http headers sent unhandled exception that crashes the proxy server this is only possible when the proxy server sets headers in the proxy request using the proxyreq setheader function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution http proxy step up your open source security game with whitesource
0
143,328
5,514,334,459
IssuesEvent
2017-03-17 14:54:32
vtyulb/BSA-Analytics
https://api.github.com/repos/vtyulb/BSA-Analytics
closed
баг картинки
bug Normal priority
Уже говорил об этом баге: 1) Внизу картинки виден обрезанный адрес. 2) Почему-то не воспроизвелось склонение источника, которое должно отражаться при выведении исходных бинари файлов. ![default](https://cloud.githubusercontent.com/assets/11756008/24045813/67bb04ea-0b31-11e7-97fc-c908edb887ba.jpg)
1.0
баг картинки - Уже говорил об этом баге: 1) Внизу картинки виден обрезанный адрес. 2) Почему-то не воспроизвелось склонение источника, которое должно отражаться при выведении исходных бинари файлов. ![default](https://cloud.githubusercontent.com/assets/11756008/24045813/67bb04ea-0b31-11e7-97fc-c908edb887ba.jpg)
non_process
баг картинки уже говорил об этом баге внизу картинки виден обрезанный адрес почему то не воспроизвелось склонение источника которое должно отражаться при выведении исходных бинари файлов
0
172,444
14,360,956,639
IssuesEvent
2020-11-30 17:33:27
namsnath/money_manager
https://api.github.com/repos/namsnath/money_manager
opened
[Doc] Add database design documentation
documentation
# [Doc] Add database design documentation ## What is required? - [ ] Add some sort of database design diagram ## Why is it required? - Makes for a good reference to people new to the project, and for ease of use - Easy way to document the database structure
1.0
[Doc] Add database design documentation - # [Doc] Add database design documentation ## What is required? - [ ] Add some sort of database design diagram ## Why is it required? - Makes for a good reference to people new to the project, and for ease of use - Easy way to document the database structure
non_process
add database design documentation add database design documentation what is required add some sort of database design diagram why is it required makes for a good reference to people new to the project and for ease of use easy way to document the database structure
0
1,759
2,571,619,816
IssuesEvent
2015-02-10 17:31:05
elasticsearch/elasticsearch-river-twitter
https://api.github.com/repos/elasticsearch/elasticsearch-river-twitter
closed
[Test] ThreadLeak detection while running tests
2.4.2 test
When we run tests, the test framework detects Thread leaks: ``` Throwable #2: com.carrotsearch.randomizedtesting.ThreadLeakError: There are still zombie threads that couldn't be terminated: > 1) Thread[id=317, name=Twitter4J Async Dispatcher[0], state=WAITING, group=TGRP-TwitterIntegrationTest] > at sun.misc.Unsafe.park(Native Method) > at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186) > at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043) > at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) > at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1068) > at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130) > at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:745) > at __randomizedtesting.SeedInfo.seed([7EBEF43EC89FBD41]:0) ``` It sounds like it's caused by Twitter4J which run one Thread `Twitter4J Async Dispatcher` and can not stop it quickly enough.
1.0
[Test] ThreadLeak detection while running tests - When we run tests, the test framework detects Thread leaks: ``` Throwable #2: com.carrotsearch.randomizedtesting.ThreadLeakError: There are still zombie threads that couldn't be terminated: > 1) Thread[id=317, name=Twitter4J Async Dispatcher[0], state=WAITING, group=TGRP-TwitterIntegrationTest] > at sun.misc.Unsafe.park(Native Method) > at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186) > at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043) > at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) > at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1068) > at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130) > at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:745) > at __randomizedtesting.SeedInfo.seed([7EBEF43EC89FBD41]:0) ``` It sounds like it's caused by Twitter4J which run one Thread `Twitter4J Async Dispatcher` and can not stop it quickly enough.
non_process
threadleak detection while running tests when we run tests the test framework detects thread leaks throwable com carrotsearch randomizedtesting threadleakerror there are still zombie threads that couldn t be terminated thread state waiting group tgrp twitterintegrationtest at sun misc unsafe park native method at java util concurrent locks locksupport park locksupport java at java util concurrent locks abstractqueuedsynchronizer conditionobject await abstractqueuedsynchronizer java at java util concurrent linkedblockingqueue take linkedblockingqueue java at java util concurrent threadpoolexecutor gettask threadpoolexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java at randomizedtesting seedinfo seed it sounds like it s caused by which run one thread async dispatcher and can not stop it quickly enough
0
35,766
14,877,398,274
IssuesEvent
2021-01-20 03:03:23
Azure/azure-cli
https://api.github.com/repos/Azure/azure-cli
closed
az webapp up failed with 502 in eastus2euap
Service Attention Web Apps
> ### `az feedback` auto-generates most of the information requested below, as of CLI version 2.0.62 **Describe the bug** I am following https://docs.microsoft.com/en-us/azure/app-service/tutorial-python-postgresql-app?tabs=bash%2Cclone#41-create-the-app-service-app to create a web app in Eastus2euap(this region is specific for my business requirement) and I already registed feature Microsoft.Resouces/EUAPParticipation and Microsoft.Web/EastUS2EUAPRegionEnabled request failed: Error occurred in request., RetryError: HTTPSConnectionPool(host='management.azure.com', port=443): Max retries exceeded with url: /subscriptions/937bc588-a144-4083-8612-5f9ffbbddb14/resourceGroups/DjangoPostgres-tutorial-rg/providers/Microsoft.Web/serverfarms/DjangoPostgres-tutorial-plan?api-version=2019-08-01 (Caused by ResponseError('too many 502 error responses',)) **To Reproduce** <!--- Steps to reproduce the behavior. ---> az webapp up --resource-group DjangoPostgres-tutorial-rg --location eastus2euap --plan DjangoPostgres-tutorial-plan --sku B1 --name najitest using my subscription 937bc588-a144-4083-8612-5f9ffbbddb14 **Expected behavior** <!--- A clear and concise description of what you expected to happen. ---> A web app in eastus2 euap created **Environment summary** <!--- Install Method (e.g. pip, interactive script, apt-get, Docker, MSI, edge build) / CLI version (`az --version`) / OS version / Shell Type (e.g. bash, cmd.exe, Bash on Windows) ---> azure-cli 2.11.1 core 2.11.1 telemetry 1.0.5 Extensions: connect 0.0.1 db-up 0.2.0 dev-spaces 1.0.5 **Additional context** <!--- Add any other context about the problem here. --->
1.0
az webapp up failed with 502 in eastus2euap - > ### `az feedback` auto-generates most of the information requested below, as of CLI version 2.0.62 **Describe the bug** I am following https://docs.microsoft.com/en-us/azure/app-service/tutorial-python-postgresql-app?tabs=bash%2Cclone#41-create-the-app-service-app to create a web app in Eastus2euap(this region is specific for my business requirement) and I already registed feature Microsoft.Resouces/EUAPParticipation and Microsoft.Web/EastUS2EUAPRegionEnabled request failed: Error occurred in request., RetryError: HTTPSConnectionPool(host='management.azure.com', port=443): Max retries exceeded with url: /subscriptions/937bc588-a144-4083-8612-5f9ffbbddb14/resourceGroups/DjangoPostgres-tutorial-rg/providers/Microsoft.Web/serverfarms/DjangoPostgres-tutorial-plan?api-version=2019-08-01 (Caused by ResponseError('too many 502 error responses',)) **To Reproduce** <!--- Steps to reproduce the behavior. ---> az webapp up --resource-group DjangoPostgres-tutorial-rg --location eastus2euap --plan DjangoPostgres-tutorial-plan --sku B1 --name najitest using my subscription 937bc588-a144-4083-8612-5f9ffbbddb14 **Expected behavior** <!--- A clear and concise description of what you expected to happen. ---> A web app in eastus2 euap created **Environment summary** <!--- Install Method (e.g. pip, interactive script, apt-get, Docker, MSI, edge build) / CLI version (`az --version`) / OS version / Shell Type (e.g. bash, cmd.exe, Bash on Windows) ---> azure-cli 2.11.1 core 2.11.1 telemetry 1.0.5 Extensions: connect 0.0.1 db-up 0.2.0 dev-spaces 1.0.5 **Additional context** <!--- Add any other context about the problem here. --->
non_process
az webapp up failed with in az feedback auto generates most of the information requested below as of cli version describe the bug i am following to create a web app in this region is specific for my business requirement and i already registed feature microsoft resouces euapparticipation and microsoft web request failed error occurred in request retryerror httpsconnectionpool host management azure com port max retries exceeded with url subscriptions resourcegroups djangopostgres tutorial rg providers microsoft web serverfarms djangopostgres tutorial plan api version caused by responseerror too many error responses to reproduce az webapp up resource group djangopostgres tutorial rg location plan djangopostgres tutorial plan sku name najitest using my subscription expected behavior a web app in euap created environment summary azure cli core telemetry extensions connect db up dev spaces additional context
0
12,819
15,192,794,344
IssuesEvent
2021-02-15 22:54:37
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Duplicate parameter coordinates registered for alg v.net.visibility
Bug Processing
<!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue --> **Describe the bug** The Log Messages panel shows the following warning: `WARNING Duplicate parameter coordinates registered for alg v.net.visibility` **QGIS and OS versions** QGIS 3.10.13 - QGIS 3.16.2 **Additional context** This issue was fixed in 3.2 with 5436ce9a3583fa77e335931dd69828ca950c0894 (https://github.com/qgis/QGIS/pull/6649) and then reintroduced since 3.10.11 and 3.16.0 with 3119eb7de2f2d1bf27001d6b5160af9295b3b355 (https://github.com/qgis/QGIS/pull/38966)
1.0
Duplicate parameter coordinates registered for alg v.net.visibility - <!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue --> **Describe the bug** The Log Messages panel shows the following warning: `WARNING Duplicate parameter coordinates registered for alg v.net.visibility` **QGIS and OS versions** QGIS 3.10.13 - QGIS 3.16.2 **Additional context** This issue was fixed in 3.2 with 5436ce9a3583fa77e335931dd69828ca950c0894 (https://github.com/qgis/QGIS/pull/6649) and then reintroduced since 3.10.11 and 3.16.0 with 3119eb7de2f2d1bf27001d6b5160af9295b3b355 (https://github.com/qgis/QGIS/pull/38966)
process
duplicate parameter coordinates registered for alg v net visibility bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue describe the bug the log messages panel shows the following warning warning duplicate parameter coordinates registered for alg v net visibility qgis and os versions qgis qgis additional context this issue was fixed in with and then reintroduced since and with
1
11,562
14,440,035,382
IssuesEvent
2020-12-07 15:05:42
jhu-idc/iDC-general
https://api.github.com/repos/jhu-idc/iDC-general
closed
Implement the ability to revert Drupal nodes by the handler.
Graph Processor ingest
This might be able to be accomplished asynchronously. Estimate: 2 days
1.0
Implement the ability to revert Drupal nodes by the handler. - This might be able to be accomplished asynchronously. Estimate: 2 days
process
implement the ability to revert drupal nodes by the handler this might be able to be accomplished asynchronously estimate days
1
81,861
10,195,317,964
IssuesEvent
2019-08-12 17:50:08
aspnet/EntityFrameworkCore
https://api.github.com/repos/aspnet/EntityFrameworkCore
closed
Query: Cleanup pattern matching to FromSql method in Core
needs-design type-enhancement
Issues: Our query roots are entityQueryables in expression tree that is how we identify starting entityType during navigation expansion. - FromSql is also query root but relational method so core does not know about it. So query filters are supposed to be applied after FromSql method but does not happen. Few possible solutions: - Make FromSql a query root in the tree by introducing FromSqlEntityQueryable - FromSql method has parameters which needs to be funcletized. If FromSqlEntityQueryable is injected when the method is called then funcletizer needs to know about it and requires relational specific version. - If we inject FromSqlEntityQueryable after funcletizer part then FromSql in defining query wouldn't be changed. (The injector would be relational specific and navigation expansion cannot run a relational visitor). - Relational specific nav expansion which is only for to inject query root for FromSql method. This requires us to expose a lot of private code in nav expansion publicly. For now, it is pattern matching on method in core level (since we cannot use actual MethodInfo). Given it is our own method, chances of it being broken is very low.
1.0
Query: Cleanup pattern matching to FromSql method in Core - Issues: Our query roots are entityQueryables in expression tree that is how we identify starting entityType during navigation expansion. - FromSql is also query root but relational method so core does not know about it. So query filters are supposed to be applied after FromSql method but does not happen. Few possible solutions: - Make FromSql a query root in the tree by introducing FromSqlEntityQueryable - FromSql method has parameters which needs to be funcletized. If FromSqlEntityQueryable is injected when the method is called then funcletizer needs to know about it and requires relational specific version. - If we inject FromSqlEntityQueryable after funcletizer part then FromSql in defining query wouldn't be changed. (The injector would be relational specific and navigation expansion cannot run a relational visitor). - Relational specific nav expansion which is only for to inject query root for FromSql method. This requires us to expose a lot of private code in nav expansion publicly. For now, it is pattern matching on method in core level (since we cannot use actual MethodInfo). Given it is our own method, chances of it being broken is very low.
non_process
query cleanup pattern matching to fromsql method in core issues our query roots are entityqueryables in expression tree that is how we identify starting entitytype during navigation expansion fromsql is also query root but relational method so core does not know about it so query filters are supposed to be applied after fromsql method but does not happen few possible solutions make fromsql a query root in the tree by introducing fromsqlentityqueryable fromsql method has parameters which needs to be funcletized if fromsqlentityqueryable is injected when the method is called then funcletizer needs to know about it and requires relational specific version if we inject fromsqlentityqueryable after funcletizer part then fromsql in defining query wouldn t be changed the injector would be relational specific and navigation expansion cannot run a relational visitor relational specific nav expansion which is only for to inject query root for fromsql method this requires us to expose a lot of private code in nav expansion publicly for now it is pattern matching on method in core level since we cannot use actual methodinfo given it is our own method chances of it being broken is very low
0
109,625
13,794,782,905
IssuesEvent
2020-10-09 16:52:40
trezor/trezor-suite
https://api.github.com/repos/trezor/trezor-suite
opened
Update closed beta landing page
Copy Design
This is to be worried about after the Public release. Prerequisite for Closed beta release. - [ ] Update text (Closed beta, Trezor Suite etc.) - [ ] Update logo (Trezor Suite) - [ ] Update link from the 1st secondary button (destination: suite.trezor.io)
1.0
Update closed beta landing page - This is to be worried about after the Public release. Prerequisite for Closed beta release. - [ ] Update text (Closed beta, Trezor Suite etc.) - [ ] Update logo (Trezor Suite) - [ ] Update link from the 1st secondary button (destination: suite.trezor.io)
non_process
update closed beta landing page this is to be worried about after the public release prerequisite for closed beta release update text closed beta trezor suite etc update logo trezor suite update link from the secondary button destination suite trezor io
0
350,820
10,509,088,825
IssuesEvent
2019-09-27 10:07:08
telstra/open-kilda
https://api.github.com/repos/telstra/open-kilda
closed
System should save switch port information
priority/3-normal
As a user, I want to have ability to know the switch port (not only switch address), so that I can more precisely identify a switch. AC: 1. Switch port information is saved to DB 2. Switch port information is returned from Northbound when dumping switch(es)
1.0
System should save switch port information - As a user, I want to have ability to know the switch port (not only switch address), so that I can more precisely identify a switch. AC: 1. Switch port information is saved to DB 2. Switch port information is returned from Northbound when dumping switch(es)
non_process
system should save switch port information as a user i want to have ability to know the switch port not only switch address so that i can more precisely identify a switch ac switch port information is saved to db switch port information is returned from northbound when dumping switch es
0
13,907
16,664,955,272
IssuesEvent
2021-06-07 00:54:34
krystinli/Time_Management
https://api.github.com/repos/krystinli/Time_Management
closed
update experience
process
making daily update is not easy enough - need to have a more robust way of making updates
1.0
update experience - making daily update is not easy enough - need to have a more robust way of making updates
process
update experience making daily update is not easy enough need to have a more robust way of making updates
1
22,456
31,234,011,731
IssuesEvent
2023-08-20 03:19:32
pyanodon/pybugreports
https://api.github.com/repos/pyanodon/pybugreports
closed
Creosote to aromatics is more expensive than tar to aromatics
balance mod:pycoalprocessing
### Mod source PyAE Beta ### Which mod are you having an issue with? - [ ] pyalienlife - [ ] pyalternativeenergy - [ ] pycoalprocessing - [ ] pyfusionenergy - [ ] pyhightech - [ ] pyindustry - [X] pypetroleumhandling - [ ] pypostprocessing - [ ] pyrawores ### Operating system >=Windows 10 ### What kind of issue is this? - [ ] Compatibility - [ ] Locale (names, descriptions, unknown keys) - [ ] Graphical - [ ] Crash - [ ] Progression - [X] Balance - [ ] Pypostprocessing failure - [ ] Other ### What is the problem? Early game you can create aromatics in 2 ways: 375 tar -> 300 creosote -> 100 aromatics (Requires 16 buildings) OR 350 tar -> 100 aromatics (plus 100 CO2, 10 rich clay, 500 flue gas) (Requires 3 buildings) The second option is far superior to the first. And the funny part is the creosote to aromatics recipe's description says "Cheap way to get some aromatics". ### Steps to reproduce _No response_ ### Additional context Creosote recipe as shown in YAFC ![image](https://user-images.githubusercontent.com/246963/177655819-1e00140d-ed9f-4d25-89f4-010348ffb815.png) Tar recipe as shown in YAFC ![image](https://user-images.githubusercontent.com/246963/177655847-bd7c10b2-2d19-470e-a6a8-c50c4935f580.png) ### Log file _No response_
1.0
Creosote to aromatics is more expensive than tar to aromatics - ### Mod source PyAE Beta ### Which mod are you having an issue with? - [ ] pyalienlife - [ ] pyalternativeenergy - [ ] pycoalprocessing - [ ] pyfusionenergy - [ ] pyhightech - [ ] pyindustry - [X] pypetroleumhandling - [ ] pypostprocessing - [ ] pyrawores ### Operating system >=Windows 10 ### What kind of issue is this? - [ ] Compatibility - [ ] Locale (names, descriptions, unknown keys) - [ ] Graphical - [ ] Crash - [ ] Progression - [X] Balance - [ ] Pypostprocessing failure - [ ] Other ### What is the problem? Early game you can create aromatics in 2 ways: 375 tar -> 300 creosote -> 100 aromatics (Requires 16 buildings) OR 350 tar -> 100 aromatics (plus 100 CO2, 10 rich clay, 500 flue gas) (Requires 3 buildings) The second option is far superior to the first. And the funny part is the creosote to aromatics recipe's description says "Cheap way to get some aromatics". ### Steps to reproduce _No response_ ### Additional context Creosote recipe as shown in YAFC ![image](https://user-images.githubusercontent.com/246963/177655819-1e00140d-ed9f-4d25-89f4-010348ffb815.png) Tar recipe as shown in YAFC ![image](https://user-images.githubusercontent.com/246963/177655847-bd7c10b2-2d19-470e-a6a8-c50c4935f580.png) ### Log file _No response_
process
creosote to aromatics is more expensive than tar to aromatics mod source pyae beta which mod are you having an issue with pyalienlife pyalternativeenergy pycoalprocessing pyfusionenergy pyhightech pyindustry pypetroleumhandling pypostprocessing pyrawores operating system windows what kind of issue is this compatibility locale names descriptions unknown keys graphical crash progression balance pypostprocessing failure other what is the problem early game you can create aromatics in ways tar creosote aromatics requires buildings or tar aromatics plus rich clay flue gas requires buildings the second option is far superior to the first and the funny part is the creosote to aromatics recipe s description says cheap way to get some aromatics steps to reproduce no response additional context creosote recipe as shown in yafc tar recipe as shown in yafc log file no response
1
4,873
7,753,290,770
IssuesEvent
2018-05-30 23:47:09
StackSavingsTeam/stacksavings.com_templates
https://api.github.com/repos/StackSavingsTeam/stacksavings.com_templates
opened
Modificar Layout
On Process
Hay que modificar el siguiente layout segun los requerimientos de @Baleizer : http://stacksavings.com/detail-post/iniciando-como-internet-marketer-en-stacksavings-1pQoTf/es?skipcache - Quiero quitar la sección donde aparece lo de "Working Hours", "Satisfied Clients Percent", etc... - Quiero remover el botón de "Contact Us" en la penúltima sección de "Comienza a trabajar con nosotros"- Quiero cambiar la última sección donde dice "We’re nuts about improving your business so we ensure that we track results from day one. Sure, we come up with great creative ideas that engage and captivate your customers." junto al logo de StackSaving, en lugar de eso quiero colocar un formulario para que el usuario llene y envíe, el formulario debe contener los siguientes campos:• Nombre: _(Campo de texto)_ • Correo electrónico: _(Campo de e-mail)_ • ¿En qué tipo de Internet Marketing tienes experiencia? _(Campo de texto)_ • ¿Tienes conocimientos de inglés? _(Campo de selección simple: `Sí` o `No`)_ • ¿Tienes un estimado de aspiración salarial por hora en USD? _(Campo de texto)_ • Por favor menciona otro tipo de experiencia relevante que poseas: _(Campo de texto)_Eso debería contener un botón de `Enviar`, creo (editado) Oh, también, no se está mostrando la última imagen que debería aparecer debajo de "Comienza a trabajar con nosotros"
1.0
Modificar Layout - Hay que modificar el siguiente layout segun los requerimientos de @Baleizer : http://stacksavings.com/detail-post/iniciando-como-internet-marketer-en-stacksavings-1pQoTf/es?skipcache - Quiero quitar la sección donde aparece lo de "Working Hours", "Satisfied Clients Percent", etc... - Quiero remover el botón de "Contact Us" en la penúltima sección de "Comienza a trabajar con nosotros"- Quiero cambiar la última sección donde dice "We’re nuts about improving your business so we ensure that we track results from day one. Sure, we come up with great creative ideas that engage and captivate your customers." junto al logo de StackSaving, en lugar de eso quiero colocar un formulario para que el usuario llene y envíe, el formulario debe contener los siguientes campos:• Nombre: _(Campo de texto)_ • Correo electrónico: _(Campo de e-mail)_ • ¿En qué tipo de Internet Marketing tienes experiencia? _(Campo de texto)_ • ¿Tienes conocimientos de inglés? _(Campo de selección simple: `Sí` o `No`)_ • ¿Tienes un estimado de aspiración salarial por hora en USD? _(Campo de texto)_ • Por favor menciona otro tipo de experiencia relevante que poseas: _(Campo de texto)_Eso debería contener un botón de `Enviar`, creo (editado) Oh, también, no se está mostrando la última imagen que debería aparecer debajo de "Comienza a trabajar con nosotros"
process
modificar layout hay que modificar el siguiente layout segun los requerimientos de baleizer quiero quitar la sección donde aparece lo de working hours satisfied clients percent etc quiero remover el botón de contact us en la penúltima sección de comienza a trabajar con nosotros quiero cambiar la última sección donde dice we’re nuts about improving your business so we ensure that we track results from day one sure we come up with great creative ideas that engage and captivate your customers junto al logo de stacksaving en lugar de eso quiero colocar un formulario para que el usuario llene y envíe el formulario debe contener los siguientes campos • nombre campo de texto • correo electrónico campo de e mail • ¿en qué tipo de internet marketing tienes experiencia campo de texto • ¿tienes conocimientos de inglés campo de selección simple sí o no • ¿tienes un estimado de aspiración salarial por hora en usd campo de texto • por favor menciona otro tipo de experiencia relevante que poseas campo de texto eso debería contener un botón de enviar creo editado oh también no se está mostrando la última imagen que debería aparecer debajo de comienza a trabajar con nosotros
1
373,577
26,076,165,408
IssuesEvent
2022-12-24 15:00:40
freesewing/freesewing
https://api.github.com/repos/freesewing/freesewing
opened
[docs]: Undocumented Custom Tags and properties
:book: documentation :+1: good first issue
### Where can we see the problem? 🤔 https://freesewing.dev/guides/markdown/custom-components ### Are you a FreeSewing patron? 😃 No, I am not 😞 ### Suggested content In the MDX code, there exists some undocumented custom tags: - <Link> which works like the other popout tags like <Related>. - <Tab> and <Tabs> - <Examples> - <pre>, <img>, and <table> There is also an undocumented 'hideable' property that works with the popout tags. It is unclear whether these undocumented items should be documented for use or whether they were intended to remain undocumented. ### Additional context _No response_
1.0
[docs]: Undocumented Custom Tags and properties - ### Where can we see the problem? 🤔 https://freesewing.dev/guides/markdown/custom-components ### Are you a FreeSewing patron? 😃 No, I am not 😞 ### Suggested content In the MDX code, there exists some undocumented custom tags: - <Link> which works like the other popout tags like <Related>. - <Tab> and <Tabs> - <Examples> - <pre>, <img>, and <table> There is also an undocumented 'hideable' property that works with the popout tags. It is unclear whether these undocumented items should be documented for use or whether they were intended to remain undocumented. ### Additional context _No response_
non_process
undocumented custom tags and properties where can we see the problem 🤔 are you a freesewing patron 😃 no i am not 😞 suggested content in the mdx code there exists some undocumented custom tags which works like the other popout tags like and and there is also an undocumented hideable property that works with the popout tags it is unclear whether these undocumented items should be documented for use or whether they were intended to remain undocumented additional context no response
0
374,779
11,095,312,923
IssuesEvent
2019-12-16 08:49:48
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.empik.com - see bug description
browser-firefox-mobile engine-gecko priority-important type-tracking-protection-basic
<!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @extra_labels: type-tracking-protection-basic --> **URL**: https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Something else **Description**: cannot do anything - there is something like black curtain **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2019/12/e705d4bf-427a-40d8-88be-0ded585ca709-thumb.jpeg)](https://webcompat.com/uploads/2019/12/e705d4bf-427a-40d8-88be-0ded585ca709.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20191202144737</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: true (basic)</li> </ul> <p>Console Messages:</p> <pre> ['[JavaScript Warning: "The resource at https://cdn.branch.io/branch-latest.min.js was blocked because content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 0}]', '[JavaScript Warning: "Wczytanie si nie powiodo dla elementu script ze rdem https://www.googletagmanager.com/gtm.js?id=GTM-PS28TV." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 1}]', '[JavaScript Warning: "Content Security Policy: Ignoring \'unsafe-inline\' within script-src or style-src: nonce-source or hash-source specified"]', '[JavaScript Error: "Nie udao si odczyta rejestracji wtku usugowego: dostp do przechowywania danych jest ograniczony wtym kontekcie zpowodu ustawie uytkownika lub trybu przegldania prywatnego." {file: "https://www.gstatic.com/_/mss/boq-payments-consumer/_/js/k=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.en.CXqtjmj3U4g.es5.O/ck=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.C4Thx8gwKBQ.L.F4.O/am=Qg/d=1/exm=_b,_tp,wmwg8b/excm=_b,_tp,payframeview/ed=1/wt=2/ct=zgms/rs=AMitfrgcnbW3ajxKAURZOie1pG4oswouPg/m=Das5Le,IZT63,PrPYRd,Ru0Pgb,SF3gsd,Y2UGcc,ZyYHPb,hc6Ubd,p8L0ob,vfuNJf,wI7Sfc,ws9Tlc" line: 260}]', '[JavaScript Warning: "The resource at https://www.google-analytics.com/analytics.js was blocked because content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 0}]', '[JavaScript Error: "SecurityError: The operation is insecure." {file: "https://www.gstatic.com/_/mss/boq-payments-consumer/_/js/k=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.en.CXqtjmj3U4g.es5.O/ck=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.C4Thx8gwKBQ.L.F4.O/am=Qg/d=1/exm=_b,_tp,wmwg8b/excm=_b,_tp,payframeview/ed=1/wt=2/ct=zgms/rs=AMitfrgcnbW3ajxKAURZOie1pG4oswouPg/m=Das5Le,IZT63,PrPYRd,Ru0Pgb,SF3gsd,Y2UGcc,ZyYHPb,hc6Ubd,p8L0ob,vfuNJf,wI7Sfc,ws9Tlc" line: 260}]', '[JavaScript Warning: "Request to access cookie or storage on https://pay.google.com/gp/p/ui/payframe?origin=https%3A%2F%2Fwww.empik.com&mid= was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 258 column: 344 source: "https://www.gstatic.com/_/mss/boq-payments-consumer/_/js/k=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.en.CXqtjmj3U4g.es5.O/ck=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.C4Thx8gwKBQ.L.F4.O/am=Qg/d=1/exm=_b,_tp,wmwg8b/excm=_b,_tp,payframeview/ed=1/wt=2/ct=zgms/rs=AMitfrgcnbW3ajxKAURZOie1pG4oswouPg/m=Das5Le,IZT63,PrPYRd,Ru0Pgb,SF3gsd,Y2UGcc,ZyYHPb,hc6Ubd,p8L0ob,vfuNJf,wI7Sfc,ws9Tlc"}]', '[JavaScript Warning: "Request to access cookie or storage on https://pay.google.com/gp/p/ui/payframe?origin=https%3A%2F%2Fwww.empik.com&mid= was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 260 column: 240 source: "https://www.gstatic.com/_/mss/boq-payments-consumer/_/js/k=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.en.CXqtjmj3U4g.es5.O/ck=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.C4Thx8gwKBQ.L.F4.O/am=Qg/d=1/exm=_b,_tp,wmwg8b/excm=_b,_tp,payframeview/ed=1/wt=2/ct=zgms/rs=AMitfrgcnbW3ajxKAURZOie1pG4oswouPg/m=Das5Le,IZT63,PrPYRd,Ru0Pgb,SF3gsd,Y2UGcc,ZyYHPb,hc6Ubd,p8L0ob,vfuNJf,wI7Sfc,ws9Tlc"}]', '[JavaScript Warning: "Request to access cookie or storage on https://pay.google.com/gp/p/ui/payframe?origin=https%3A%2F%2Fwww.empik.com&mid= was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 60 column: 41 source: "https://www.gstatic.com/_/mss/boq-payments-consumer/_/js/k=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.en.CXqtjmj3U4g.es5.O/ck=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.C4Thx8gwKBQ.L.F4.O/am=Qg/d=1/exm=_b,_tp,wmwg8b/excm=_b,_tp,payframeview/ed=1/wt=2/ct=zgms/rs=AMitfrgcnbW3ajxKAURZOie1pG4oswouPg/m=Das5Le,IZT63,PrPYRd,Ru0Pgb,SF3gsd,Y2UGcc,ZyYHPb,hc6Ubd,p8L0ob,vfuNJf,wI7Sfc,ws9Tlc"}]', '[JavaScript Warning: "Request to access cookie or storage on https://pay.google.com/gp/p/ui/payframe?origin=https%3A%2F%2Fwww.empik.com&mid= was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 60 column: 41 source: "https://www.gstatic.com/_/mss/boq-payments-consumer/_/js/k=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.en.CXqtjmj3U4g.es5.O/ck=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.C4Thx8gwKBQ.L.F4.O/am=Qg/d=1/exm=_b,_tp,wmwg8b/excm=_b,_tp,payframeview/ed=1/wt=2/ct=zgms/rs=AMitfrgcnbW3ajxKAURZOie1pG4oswouPg/m=Das5Le,IZT63,PrPYRd,Ru0Pgb,SF3gsd,Y2UGcc,ZyYHPb,hc6Ubd,p8L0ob,vfuNJf,wI7Sfc,ws9Tlc"}]', '[JavaScript Warning: "Request to access cookie or storage on https://pay.google.com/gp/p/ui/payframe?origin=https%3A%2F%2Fwww.empik.com&mid= was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 60 column: 41 source: "https://www.gstatic.com/_/mss/boq-payments-consumer/_/js/k=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.en.CXqtjmj3U4g.es5.O/ck=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.C4Thx8gwKBQ.L.F4.O/am=Qg/d=1/exm=_b,_tp,wmwg8b/excm=_b,_tp,payframeview/ed=1/wt=2/ct=zgms/rs=AMitfrgcnbW3ajxKAURZOie1pG4oswouPg/m=Das5Le,IZT63,PrPYRd,Ru0Pgb,SF3gsd,Y2UGcc,ZyYHPb,hc6Ubd,p8L0ob,vfuNJf,wI7Sfc,ws9Tlc"}]', '[JavaScript Warning: "Request to access cookie or storage on https://pay.google.com/gp/p/ui/payframe?origin=https%3A%2F%2Fwww.empik.com&mid= was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 60 column: 41 source: "https://www.gstatic.com/_/mss/boq-payments-consumer/_/js/k=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.en.CXqtjmj3U4g.es5.O/ck=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.C4Thx8gwKBQ.L.F4.O/am=Qg/d=1/exm=_b,_tp,wmwg8b/excm=_b,_tp,payframeview/ed=1/wt=2/ct=zgms/rs=AMitfrgcnbW3ajxKAURZOie1pG4oswouPg/m=Das5Le,IZT63,PrPYRd,Ru0Pgb,SF3gsd,Y2UGcc,ZyYHPb,hc6Ubd,p8L0ob,vfuNJf,wI7Sfc,ws9Tlc"}]', '[JavaScript Warning: "Content Security Policy: Ignoring \'unsafe-inline\' within script-src: strict-dynamic specified"]', '[JavaScript Warning: "Content Security Policy: Ignoring https: within script-src: strict-dynamic specified"]', '[JavaScript Warning: "Content Security Policy: Ignoring http: within script-src: strict-dynamic specified"]', '[JavaScript Warning: "Request to access cookie or storage on https://pay.google.com/gp/p/ui/payframe?origin=https%3A%2F%2Fwww.empik.com&mid= was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 58 column: 91 source: "https://www.gstatic.com/_/mss/boq-payments-consumer/_/js/k=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.en.CXqtjmj3U4g.es5.O/ck=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.C4Thx8gwKBQ.L.F4.O/am=Qg/d=1/exm=_b,_tp,wmwg8b/excm=_b,_tp,payframeview/ed=1/wt=2/ct=zgms/rs=AMitfrgcnbW3ajxKAURZOie1pG4oswouPg/m=Das5Le,IZT63,PrPYRd,Ru0Pgb,SF3gsd,Y2UGcc,ZyYHPb,hc6Ubd,p8L0ob,vfuNJf,wI7Sfc,ws9Tlc"}]', '[JavaScript Warning: "Request to access cookie or storage on https://www.google.com/recaptcha/api2/anchor?ar=1&k=6LdPQMMUAAAAAHgxRX9XF__AfZeKF54Xjxzpranj&co=aHR0cHM6Ly93d3cuZW1waWsuY29tOjQ0Mw..&hl=pl&v=mhgGrlTs_PbFQOW4ejlxlxZn&size=normal&cb=h4llkmjyj803 was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 328 column: 213 source: "https://www.gstatic.com/recaptcha/releases/mhgGrlTs_PbFQOW4ejlxlxZn/recaptcha__pl.js"}]', '[JavaScript Warning: "Request to access cookie or storage on https://www.google.com/recaptcha/api2/anchor?ar=1&k=6LdPQMMUAAAAAHgxRX9XF__AfZeKF54Xjxzpranj&co=aHR0cHM6Ly93d3cuZW1waWsuY29tOjQ0Mw..&hl=pl&v=mhgGrlTs_PbFQOW4ejlxlxZn&size=normal&cb=h4llkmjyj803 was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 320 column: 152 source: "https://www.gstatic.com/recaptcha/releases/mhgGrlTs_PbFQOW4ejlxlxZn/recaptcha__pl.js"}]', '[JavaScript Warning: "Request to access cookie or storage on https://www.google.com/recaptcha/api2/anchor?ar=1&k=6LdPQMMUAAAAAHgxRX9XF__AfZeKF54Xjxzpranj&co=aHR0cHM6Ly93d3cuZW1waWsuY29tOjQ0Mw..&hl=pl&v=mhgGrlTs_PbFQOW4ejlxlxZn&size=normal&cb=h4llkmjyj803 was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 320 column: 152 source: "https://www.gstatic.com/recaptcha/releases/mhgGrlTs_PbFQOW4ejlxlxZn/recaptcha__pl.js"}]', '[JavaScript Warning: "Request to access cookie or storage on https://www.google.com/recaptcha/api2/anchor?ar=1&k=6LdPQMMUAAAAAHgxRX9XF__AfZeKF54Xjxzpranj&co=aHR0cHM6Ly93d3cuZW1waWsuY29tOjQ0Mw..&hl=pl&v=mhgGrlTs_PbFQOW4ejlxlxZn&size=normal&cb=h4llkmjyj803 was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 321 column: 184 source: "https://www.gstatic.com/recaptcha/releases/mhgGrlTs_PbFQOW4ejlxlxZn/recaptcha__pl.js"}]', '[JavaScript Warning: "Content Security Policy: Ignoring \'unsafe-inline\' within script-src: strict-dynamic specified"]', '[JavaScript Warning: "Content Security Policy: Ignoring https: within script-src: strict-dynamic specified"]', '[JavaScript Warning: "Content Security Policy: Ignoring http: within script-src: strict-dynamic specified"]', '[JavaScript Warning: "Wczytanie si nie powiodo dla elementu script ze rdem https://nebula-cdn.kampyle.com/we/263992/onsite/embed.js." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 1}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.empik.com - see bug description - <!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @extra_labels: type-tracking-protection-basic --> **URL**: https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Something else **Description**: cannot do anything - there is something like black curtain **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2019/12/e705d4bf-427a-40d8-88be-0ded585ca709-thumb.jpeg)](https://webcompat.com/uploads/2019/12/e705d4bf-427a-40d8-88be-0ded585ca709.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20191202144737</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: true (basic)</li> </ul> <p>Console Messages:</p> <pre> ['[JavaScript Warning: "The resource at https://cdn.branch.io/branch-latest.min.js was blocked because content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 0}]', '[JavaScript Warning: "Wczytanie si nie powiodo dla elementu script ze rdem https://www.googletagmanager.com/gtm.js?id=GTM-PS28TV." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 1}]', '[JavaScript Warning: "Content Security Policy: Ignoring \'unsafe-inline\' within script-src or style-src: nonce-source or hash-source specified"]', '[JavaScript Error: "Nie udao si odczyta rejestracji wtku usugowego: dostp do przechowywania danych jest ograniczony wtym kontekcie zpowodu ustawie uytkownika lub trybu przegldania prywatnego." {file: "https://www.gstatic.com/_/mss/boq-payments-consumer/_/js/k=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.en.CXqtjmj3U4g.es5.O/ck=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.C4Thx8gwKBQ.L.F4.O/am=Qg/d=1/exm=_b,_tp,wmwg8b/excm=_b,_tp,payframeview/ed=1/wt=2/ct=zgms/rs=AMitfrgcnbW3ajxKAURZOie1pG4oswouPg/m=Das5Le,IZT63,PrPYRd,Ru0Pgb,SF3gsd,Y2UGcc,ZyYHPb,hc6Ubd,p8L0ob,vfuNJf,wI7Sfc,ws9Tlc" line: 260}]', '[JavaScript Warning: "The resource at https://www.google-analytics.com/analytics.js was blocked because content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 0}]', '[JavaScript Error: "SecurityError: The operation is insecure." {file: "https://www.gstatic.com/_/mss/boq-payments-consumer/_/js/k=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.en.CXqtjmj3U4g.es5.O/ck=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.C4Thx8gwKBQ.L.F4.O/am=Qg/d=1/exm=_b,_tp,wmwg8b/excm=_b,_tp,payframeview/ed=1/wt=2/ct=zgms/rs=AMitfrgcnbW3ajxKAURZOie1pG4oswouPg/m=Das5Le,IZT63,PrPYRd,Ru0Pgb,SF3gsd,Y2UGcc,ZyYHPb,hc6Ubd,p8L0ob,vfuNJf,wI7Sfc,ws9Tlc" line: 260}]', '[JavaScript Warning: "Request to access cookie or storage on https://pay.google.com/gp/p/ui/payframe?origin=https%3A%2F%2Fwww.empik.com&mid= was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 258 column: 344 source: "https://www.gstatic.com/_/mss/boq-payments-consumer/_/js/k=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.en.CXqtjmj3U4g.es5.O/ck=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.C4Thx8gwKBQ.L.F4.O/am=Qg/d=1/exm=_b,_tp,wmwg8b/excm=_b,_tp,payframeview/ed=1/wt=2/ct=zgms/rs=AMitfrgcnbW3ajxKAURZOie1pG4oswouPg/m=Das5Le,IZT63,PrPYRd,Ru0Pgb,SF3gsd,Y2UGcc,ZyYHPb,hc6Ubd,p8L0ob,vfuNJf,wI7Sfc,ws9Tlc"}]', '[JavaScript Warning: "Request to access cookie or storage on https://pay.google.com/gp/p/ui/payframe?origin=https%3A%2F%2Fwww.empik.com&mid= was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 260 column: 240 source: "https://www.gstatic.com/_/mss/boq-payments-consumer/_/js/k=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.en.CXqtjmj3U4g.es5.O/ck=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.C4Thx8gwKBQ.L.F4.O/am=Qg/d=1/exm=_b,_tp,wmwg8b/excm=_b,_tp,payframeview/ed=1/wt=2/ct=zgms/rs=AMitfrgcnbW3ajxKAURZOie1pG4oswouPg/m=Das5Le,IZT63,PrPYRd,Ru0Pgb,SF3gsd,Y2UGcc,ZyYHPb,hc6Ubd,p8L0ob,vfuNJf,wI7Sfc,ws9Tlc"}]', '[JavaScript Warning: "Request to access cookie or storage on https://pay.google.com/gp/p/ui/payframe?origin=https%3A%2F%2Fwww.empik.com&mid= was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 60 column: 41 source: "https://www.gstatic.com/_/mss/boq-payments-consumer/_/js/k=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.en.CXqtjmj3U4g.es5.O/ck=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.C4Thx8gwKBQ.L.F4.O/am=Qg/d=1/exm=_b,_tp,wmwg8b/excm=_b,_tp,payframeview/ed=1/wt=2/ct=zgms/rs=AMitfrgcnbW3ajxKAURZOie1pG4oswouPg/m=Das5Le,IZT63,PrPYRd,Ru0Pgb,SF3gsd,Y2UGcc,ZyYHPb,hc6Ubd,p8L0ob,vfuNJf,wI7Sfc,ws9Tlc"}]', '[JavaScript Warning: "Request to access cookie or storage on https://pay.google.com/gp/p/ui/payframe?origin=https%3A%2F%2Fwww.empik.com&mid= was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 60 column: 41 source: "https://www.gstatic.com/_/mss/boq-payments-consumer/_/js/k=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.en.CXqtjmj3U4g.es5.O/ck=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.C4Thx8gwKBQ.L.F4.O/am=Qg/d=1/exm=_b,_tp,wmwg8b/excm=_b,_tp,payframeview/ed=1/wt=2/ct=zgms/rs=AMitfrgcnbW3ajxKAURZOie1pG4oswouPg/m=Das5Le,IZT63,PrPYRd,Ru0Pgb,SF3gsd,Y2UGcc,ZyYHPb,hc6Ubd,p8L0ob,vfuNJf,wI7Sfc,ws9Tlc"}]', '[JavaScript Warning: "Request to access cookie or storage on https://pay.google.com/gp/p/ui/payframe?origin=https%3A%2F%2Fwww.empik.com&mid= was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 60 column: 41 source: "https://www.gstatic.com/_/mss/boq-payments-consumer/_/js/k=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.en.CXqtjmj3U4g.es5.O/ck=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.C4Thx8gwKBQ.L.F4.O/am=Qg/d=1/exm=_b,_tp,wmwg8b/excm=_b,_tp,payframeview/ed=1/wt=2/ct=zgms/rs=AMitfrgcnbW3ajxKAURZOie1pG4oswouPg/m=Das5Le,IZT63,PrPYRd,Ru0Pgb,SF3gsd,Y2UGcc,ZyYHPb,hc6Ubd,p8L0ob,vfuNJf,wI7Sfc,ws9Tlc"}]', '[JavaScript Warning: "Request to access cookie or storage on https://pay.google.com/gp/p/ui/payframe?origin=https%3A%2F%2Fwww.empik.com&mid= was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 60 column: 41 source: "https://www.gstatic.com/_/mss/boq-payments-consumer/_/js/k=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.en.CXqtjmj3U4g.es5.O/ck=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.C4Thx8gwKBQ.L.F4.O/am=Qg/d=1/exm=_b,_tp,wmwg8b/excm=_b,_tp,payframeview/ed=1/wt=2/ct=zgms/rs=AMitfrgcnbW3ajxKAURZOie1pG4oswouPg/m=Das5Le,IZT63,PrPYRd,Ru0Pgb,SF3gsd,Y2UGcc,ZyYHPb,hc6Ubd,p8L0ob,vfuNJf,wI7Sfc,ws9Tlc"}]', '[JavaScript Warning: "Content Security Policy: Ignoring \'unsafe-inline\' within script-src: strict-dynamic specified"]', '[JavaScript Warning: "Content Security Policy: Ignoring https: within script-src: strict-dynamic specified"]', '[JavaScript Warning: "Content Security Policy: Ignoring http: within script-src: strict-dynamic specified"]', '[JavaScript Warning: "Request to access cookie or storage on https://pay.google.com/gp/p/ui/payframe?origin=https%3A%2F%2Fwww.empik.com&mid= was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 58 column: 91 source: "https://www.gstatic.com/_/mss/boq-payments-consumer/_/js/k=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.en.CXqtjmj3U4g.es5.O/ck=boq-payments-consumer.InstantbuyFrontendBuyflowPayframeUi.C4Thx8gwKBQ.L.F4.O/am=Qg/d=1/exm=_b,_tp,wmwg8b/excm=_b,_tp,payframeview/ed=1/wt=2/ct=zgms/rs=AMitfrgcnbW3ajxKAURZOie1pG4oswouPg/m=Das5Le,IZT63,PrPYRd,Ru0Pgb,SF3gsd,Y2UGcc,ZyYHPb,hc6Ubd,p8L0ob,vfuNJf,wI7Sfc,ws9Tlc"}]', '[JavaScript Warning: "Request to access cookie or storage on https://www.google.com/recaptcha/api2/anchor?ar=1&k=6LdPQMMUAAAAAHgxRX9XF__AfZeKF54Xjxzpranj&co=aHR0cHM6Ly93d3cuZW1waWsuY29tOjQ0Mw..&hl=pl&v=mhgGrlTs_PbFQOW4ejlxlxZn&size=normal&cb=h4llkmjyj803 was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 328 column: 213 source: "https://www.gstatic.com/recaptcha/releases/mhgGrlTs_PbFQOW4ejlxlxZn/recaptcha__pl.js"}]', '[JavaScript Warning: "Request to access cookie or storage on https://www.google.com/recaptcha/api2/anchor?ar=1&k=6LdPQMMUAAAAAHgxRX9XF__AfZeKF54Xjxzpranj&co=aHR0cHM6Ly93d3cuZW1waWsuY29tOjQ0Mw..&hl=pl&v=mhgGrlTs_PbFQOW4ejlxlxZn&size=normal&cb=h4llkmjyj803 was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 320 column: 152 source: "https://www.gstatic.com/recaptcha/releases/mhgGrlTs_PbFQOW4ejlxlxZn/recaptcha__pl.js"}]', '[JavaScript Warning: "Request to access cookie or storage on https://www.google.com/recaptcha/api2/anchor?ar=1&k=6LdPQMMUAAAAAHgxRX9XF__AfZeKF54Xjxzpranj&co=aHR0cHM6Ly93d3cuZW1waWsuY29tOjQ0Mw..&hl=pl&v=mhgGrlTs_PbFQOW4ejlxlxZn&size=normal&cb=h4llkmjyj803 was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 320 column: 152 source: "https://www.gstatic.com/recaptcha/releases/mhgGrlTs_PbFQOW4ejlxlxZn/recaptcha__pl.js"}]', '[JavaScript Warning: "Request to access cookie or storage on https://www.google.com/recaptcha/api2/anchor?ar=1&k=6LdPQMMUAAAAAHgxRX9XF__AfZeKF54Xjxzpranj&co=aHR0cHM6Ly93d3cuZW1waWsuY29tOjQ0Mw..&hl=pl&v=mhgGrlTs_PbFQOW4ejlxlxZn&size=normal&cb=h4llkmjyj803 was blocked because we are blocking all third-party storage access requests and content blocking is enabled." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 321 column: 184 source: "https://www.gstatic.com/recaptcha/releases/mhgGrlTs_PbFQOW4ejlxlxZn/recaptcha__pl.js"}]', '[JavaScript Warning: "Content Security Policy: Ignoring \'unsafe-inline\' within script-src: strict-dynamic specified"]', '[JavaScript Warning: "Content Security Policy: Ignoring https: within script-src: strict-dynamic specified"]', '[JavaScript Warning: "Content Security Policy: Ignoring http: within script-src: strict-dynamic specified"]', '[JavaScript Warning: "Wczytanie si nie powiodo dla elementu script ze rdem https://nebula-cdn.kampyle.com/we/263992/onsite/embed.js." {file: "https://www.empik.com/wielka-czworka-ukryte-dna-amazon-apple-facebook-i-google-galloway-scott,p1208727762,ksiazka-p" line: 1}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
see bug description url browser version firefox mobile operating system android tested another browser no problem type something else description cannot do anything there is something like black curtain steps to reproduce browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked true basic console messages from with ❤️
0
12,413
14,920,122,946
IssuesEvent
2021-01-23 02:58:53
e4exp/paper_manager_abstract
https://api.github.com/repos/e4exp/paper_manager_abstract
opened
Prefix-Tuning: Optimizing Continuous Prompts for Generation
2020 Fine Tuning Natural Language Processing Pretraining _read_later
* https://arxiv.org/abs/2101.00190 * 2021 微調整は、事前に学習された大規模な言語モデルを活用して下流のタスクを実行するための事実上の方法です。 しかし,言語モデルのパラメータをすべて変更してしまうため,タスクごとに完全なコピーを保存しておく必要がある. この論文では、自然言語生成タスクのための微調整に代わる軽量な方法として、言語モデルパラメータを凍結したまま、タスク固有の小さな連続ベクトル(プレフィックスと呼ばれる)を最適化するプレフィックスチューニングを提案する。 プレフィックスチューニングはプロンプトからヒントを得ており、後続のトークンが「仮想トークン」であるかのように、このプレフィックスに注目することを可能にします。 我々は、テーブルからテキストへの生成にGPT-2を、要約にBARTを適用した。 その結果、0.1%のパラメータを学習するだけで、フルデータ環境では同等の性能が得られ、低データ環境では微調整よりも優れた性能が得られ、学習中には見られなかったトピックを持つ例に対しても優れた外挿が可能であることがわかった。
1.0
Prefix-Tuning: Optimizing Continuous Prompts for Generation - * https://arxiv.org/abs/2101.00190 * 2021 微調整は、事前に学習された大規模な言語モデルを活用して下流のタスクを実行するための事実上の方法です。 しかし,言語モデルのパラメータをすべて変更してしまうため,タスクごとに完全なコピーを保存しておく必要がある. この論文では、自然言語生成タスクのための微調整に代わる軽量な方法として、言語モデルパラメータを凍結したまま、タスク固有の小さな連続ベクトル(プレフィックスと呼ばれる)を最適化するプレフィックスチューニングを提案する。 プレフィックスチューニングはプロンプトからヒントを得ており、後続のトークンが「仮想トークン」であるかのように、このプレフィックスに注目することを可能にします。 我々は、テーブルからテキストへの生成にGPT-2を、要約にBARTを適用した。 その結果、0.1%のパラメータを学習するだけで、フルデータ環境では同等の性能が得られ、低データ環境では微調整よりも優れた性能が得られ、学習中には見られなかったトピックを持つ例に対しても優れた外挿が可能であることがわかった。
process
prefix tuning optimizing continuous prompts for generation 微調整は、事前に学習された大規模な言語モデルを活用して下流のタスクを実行するための事実上の方法です。 しかし,言語モデルのパラメータをすべて変更してしまうため,タスクごとに完全なコピーを保存しておく必要がある. この論文では、自然言語生成タスクのための微調整に代わる軽量な方法として、言語モデルパラメータを凍結したまま、タスク固有の小さな連続ベクトル(プレフィックスと呼ばれる)を最適化するプレフィックスチューニングを提案する。 プレフィックスチューニングはプロンプトからヒントを得ており、後続のトークンが「仮想トークン」であるかのように、このプレフィックスに注目することを可能にします。 我々は、テーブルからテキストへの生成にgpt 、要約にbartを適用した。 その結果、 のパラメータを学習するだけで、フルデータ環境では同等の性能が得られ、低データ環境では微調整よりも優れた性能が得られ、学習中には見られなかったトピックを持つ例に対しても優れた外挿が可能であることがわかった。
1
145,873
22,823,677,917
IssuesEvent
2022-07-12 06:29:23
TeamHappic/happic-iOS
https://api.github.com/repos/TeamHappic/happic-iOS
opened
[Chore] Literal 익스텐션 변경 및 추가
⚙️ Setting 🖍 Design 수빈✨
## 💡 이슈 요약 - Literal 익스텐션 변경 및 추가 ## 📌 체크 리스트 <!-- 해야 할 일을 적어주세요. --> - [ ] 명명 변정 - [ ] ImageLiteral 추가
1.0
[Chore] Literal 익스텐션 변경 및 추가 - ## 💡 이슈 요약 - Literal 익스텐션 변경 및 추가 ## 📌 체크 리스트 <!-- 해야 할 일을 적어주세요. --> - [ ] 명명 변정 - [ ] ImageLiteral 추가
non_process
literal 익스텐션 변경 및 추가 💡 이슈 요약 literal 익스텐션 변경 및 추가 📌 체크 리스트 명명 변정 imageliteral 추가
0