Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
689,340
| 23,617,233,954
|
IssuesEvent
|
2022-08-24 16:56:51
|
bigbinary/neeto-editor-tiptap
|
https://api.github.com/repos/bigbinary/neeto-editor-tiptap
|
closed
|
Insert nodes created from FixedMenu at the end
|
working PR low-priority 0.5D
|
Elements from the fixed menu are inserted at the beginning of the content at the moment. This can be cumbersome for the user when there is already some content in the editor. Ideally, such elements should be inserted at the end.
https://vimeo.com/737784830/c7a0074df8
@AbhayVAshokan _a Please take a look.
|
1.0
|
Insert nodes created from FixedMenu at the end - Elements from the fixed menu are inserted at the beginning of the content at the moment. This can be cumbersome for the user when there is already some content in the editor. Ideally, such elements should be inserted at the end.
https://vimeo.com/737784830/c7a0074df8
@AbhayVAshokan _a Please take a look.
|
non_process
|
insert nodes created from fixedmenu at the end elements from the fixed menu are inserted at the beginning of the content at the moment this can be cumbersome for the user when there is already some content in the editor ideally such elements should be inserted at the end abhayvashokan a please take a look
| 0
|
537,853
| 15,755,540,373
|
IssuesEvent
|
2021-03-31 01:57:49
|
OnTopicCMS/OnTopic-Editor-AspNetCore
|
https://api.github.com/repos/OnTopicCMS/OnTopic-Editor-AspNetCore
|
closed
|
Default Export Options
|
Area: Interface Priority: 3 Severity 0: Nice to Have Status 5: Complete Type: Improvement
|
When exporting a topic (tree) from the OnTopic Editor, set the following defaults on the options:
- [x] Include Child Topics?
- [x] Include Nested Topics?
- [ ] Include External Associations?
- [ ] Translate Legacy Topic Pointers?
|
1.0
|
Default Export Options - When exporting a topic (tree) from the OnTopic Editor, set the following defaults on the options:
- [x] Include Child Topics?
- [x] Include Nested Topics?
- [ ] Include External Associations?
- [ ] Translate Legacy Topic Pointers?
|
non_process
|
default export options when exporting a topic tree from the ontopic editor set the following defaults on the options include child topics include nested topics include external associations translate legacy topic pointers
| 0
|
300,917
| 26,002,566,947
|
IssuesEvent
|
2022-12-20 16:25:45
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
sql/importer: TestImportDefaultWithResume failed [ingestKVs hangs with BulkAdderFlushesEveryBatch]
|
C-test-failure O-robot branch-master T-disaster-recovery
|
sql/importer.TestImportDefaultWithResume [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7804159?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7804159?buildTab=artifacts#/) on master @ [7cb778506d75bbef2eb90abccaa75b9dc7e3fb91](https://github.com/cockroachdb/cockroach/commits/7cb778506d75bbef2eb90abccaa75b9dc7e3fb91):
```
test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/73dfce30b9a5630b1b4dabed3c94b32c/logTestImportDefaultWithResume2888010391
test_log_scope.go:79: use -show-logs to present logs inline
=== CONT TestImportDefaultWithResume
import_stmt_test.go:4712: -- test log scope end --
import_stmt_test.go:4712: Leaked goroutine: goroutine 298969 [semacquire]:
sync.runtime_Semacquire(0x11bbf05?)
GOROOT/src/runtime/sema.go:62 +0x25
sync.(*WaitGroup).Wait(0xc0043bbdf0?)
GOROOT/src/sync/waitgroup.go:139 +0x52
golang.org/x/sync/errgroup.(*Group).Wait(0xc00487fbc0)
golang.org/x/sync/errgroup/external/org_golang_x_sync/errgroup/errgroup.go:53 +0x27
github.com/cockroachdb/cockroach/pkg/util/ctxgroup.Group.Wait({0xc00487fbc0?, {0x6b20dd0?, 0xc00487fb80?}})
github.com/cockroachdb/cockroach/pkg/util/ctxgroup/ctxgroup.go:144 +0x4a
github.com/cockroachdb/cockroach/pkg/sql/importer.runImport({0x6b20e78, 0xc007ce73e0}, 0xc002c47400, 0xc005abb8d0, 0xc008b5ba40, 0xc00487fa40?)
github.com/cockroachdb/cockroach/pkg/sql/importer/read_import_base.go:127 +0x7b9
github.com/cockroachdb/cockroach/pkg/sql/importer.(*readImportDataProcessor).Start.func1()
github.com/cockroachdb/cockroach/pkg/sql/importer/import_processor.go:183 +0x8b
created by github.com/cockroachdb/cockroach/pkg/sql/importer.(*readImportDataProcessor).Start
github.com/cockroachdb/cockroach/pkg/sql/importer/import_processor.go:181 +0xeb
Leaked goroutine: goroutine 298971 [semacquire]:
sync.runtime_Semacquire(0x11bbf05?)
GOROOT/src/runtime/sema.go:62 +0x25
sync.(*WaitGroup).Wait(0xc002c8b808?)
GOROOT/src/sync/waitgroup.go:139 +0x52
golang.org/x/sync/errgroup.(*Group).Wait(0xc00487fcc0)
golang.org/x/sync/errgroup/external/org_golang_x_sync/errgroup/errgroup.go:53 +0x27
github.com/cockroachdb/cockroach/pkg/util/ctxgroup.Group.Wait({0xc00487fcc0?, {0x6b20dd0?, 0xc00487fc80?}})
github.com/cockroachdb/cockroach/pkg/util/ctxgroup/ctxgroup.go:144 +0x4a
github.com/cockroachdb/cockroach/pkg/sql/importer.ingestKvs({0x6b20dd0?, 0xc00487fb80?}, 0xc002c47400, 0xc005abb8d0, 0xc008b5ba40, 0xc004855ec0)
github.com/cockroachdb/cockroach/pkg/sql/importer/import_processor.go:553 +0x1028
github.com/cockroachdb/cockroach/pkg/sql/importer.runImport.func2({0x6b20dd0, 0xc00487fb80})
github.com/cockroachdb/cockroach/pkg/sql/importer/read_import_base.go:108 +0x6a
github.com/cockroachdb/cockroach/pkg/util/ctxgroup.Group.GoCtx.func1()
github.com/cockroachdb/cockroach/pkg/util/ctxgroup/ctxgroup.go:168 +0x25
golang.org/x/sync/errgroup.(*Group).Go.func1()
golang.org/x/sync/errgroup/external/org_golang_x_sync/errgroup/errgroup.go:75 +0x64
created by golang.org/x/sync/errgroup.(*Group).Go
golang.org/x/sync/errgroup/external/org_golang_x_sync/errgroup/errgroup.go:72 +0xa5
Leaked goroutine: goroutine 298973 [chan send]:
github.com/cockroachdb/cockroach/pkg/sql/importer.ingestKvs.func3()
github.com/cockroachdb/cockroach/pkg/sql/importer/import_processor.go:468 +0x311
github.com/cockroachdb/cockroach/pkg/sql/importer.ingestKvs.func5({0x6b20dd0, 0xc00487fc80})
github.com/cockroachdb/cockroach/pkg/sql/importer/import_processor.go:547 +0x2cf
github.com/cockroachdb/cockroach/pkg/util/ctxgroup.Group.GoCtx.func1()
github.com/cockroachdb/cockroach/pkg/util/ctxgroup/ctxgroup.go:168 +0x25
golang.org/x/sync/errgroup.(*Group).Go.func1()
golang.org/x/sync/errgroup/external/org_golang_x_sync/errgroup/errgroup.go:75 +0x64
created by golang.org/x/sync/errgroup.(*Group).Go
golang.org/x/sync/errgroup/external/org_golang_x_sync/errgroup/errgroup.go:72 +0xa5
--- FAIL: TestImportDefaultWithResume (5.74s)
```
<p>Parameters: <code>TAGS=bazel,gss</code>
</p>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/sql-experience
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestImportDefaultWithResume.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-22045
|
1.0
|
sql/importer: TestImportDefaultWithResume failed [ingestKVs hangs with BulkAdderFlushesEveryBatch] - sql/importer.TestImportDefaultWithResume [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7804159?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7804159?buildTab=artifacts#/) on master @ [7cb778506d75bbef2eb90abccaa75b9dc7e3fb91](https://github.com/cockroachdb/cockroach/commits/7cb778506d75bbef2eb90abccaa75b9dc7e3fb91):
```
test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/73dfce30b9a5630b1b4dabed3c94b32c/logTestImportDefaultWithResume2888010391
test_log_scope.go:79: use -show-logs to present logs inline
=== CONT TestImportDefaultWithResume
import_stmt_test.go:4712: -- test log scope end --
import_stmt_test.go:4712: Leaked goroutine: goroutine 298969 [semacquire]:
sync.runtime_Semacquire(0x11bbf05?)
GOROOT/src/runtime/sema.go:62 +0x25
sync.(*WaitGroup).Wait(0xc0043bbdf0?)
GOROOT/src/sync/waitgroup.go:139 +0x52
golang.org/x/sync/errgroup.(*Group).Wait(0xc00487fbc0)
golang.org/x/sync/errgroup/external/org_golang_x_sync/errgroup/errgroup.go:53 +0x27
github.com/cockroachdb/cockroach/pkg/util/ctxgroup.Group.Wait({0xc00487fbc0?, {0x6b20dd0?, 0xc00487fb80?}})
github.com/cockroachdb/cockroach/pkg/util/ctxgroup/ctxgroup.go:144 +0x4a
github.com/cockroachdb/cockroach/pkg/sql/importer.runImport({0x6b20e78, 0xc007ce73e0}, 0xc002c47400, 0xc005abb8d0, 0xc008b5ba40, 0xc00487fa40?)
github.com/cockroachdb/cockroach/pkg/sql/importer/read_import_base.go:127 +0x7b9
github.com/cockroachdb/cockroach/pkg/sql/importer.(*readImportDataProcessor).Start.func1()
github.com/cockroachdb/cockroach/pkg/sql/importer/import_processor.go:183 +0x8b
created by github.com/cockroachdb/cockroach/pkg/sql/importer.(*readImportDataProcessor).Start
github.com/cockroachdb/cockroach/pkg/sql/importer/import_processor.go:181 +0xeb
Leaked goroutine: goroutine 298971 [semacquire]:
sync.runtime_Semacquire(0x11bbf05?)
GOROOT/src/runtime/sema.go:62 +0x25
sync.(*WaitGroup).Wait(0xc002c8b808?)
GOROOT/src/sync/waitgroup.go:139 +0x52
golang.org/x/sync/errgroup.(*Group).Wait(0xc00487fcc0)
golang.org/x/sync/errgroup/external/org_golang_x_sync/errgroup/errgroup.go:53 +0x27
github.com/cockroachdb/cockroach/pkg/util/ctxgroup.Group.Wait({0xc00487fcc0?, {0x6b20dd0?, 0xc00487fc80?}})
github.com/cockroachdb/cockroach/pkg/util/ctxgroup/ctxgroup.go:144 +0x4a
github.com/cockroachdb/cockroach/pkg/sql/importer.ingestKvs({0x6b20dd0?, 0xc00487fb80?}, 0xc002c47400, 0xc005abb8d0, 0xc008b5ba40, 0xc004855ec0)
github.com/cockroachdb/cockroach/pkg/sql/importer/import_processor.go:553 +0x1028
github.com/cockroachdb/cockroach/pkg/sql/importer.runImport.func2({0x6b20dd0, 0xc00487fb80})
github.com/cockroachdb/cockroach/pkg/sql/importer/read_import_base.go:108 +0x6a
github.com/cockroachdb/cockroach/pkg/util/ctxgroup.Group.GoCtx.func1()
github.com/cockroachdb/cockroach/pkg/util/ctxgroup/ctxgroup.go:168 +0x25
golang.org/x/sync/errgroup.(*Group).Go.func1()
golang.org/x/sync/errgroup/external/org_golang_x_sync/errgroup/errgroup.go:75 +0x64
created by golang.org/x/sync/errgroup.(*Group).Go
golang.org/x/sync/errgroup/external/org_golang_x_sync/errgroup/errgroup.go:72 +0xa5
Leaked goroutine: goroutine 298973 [chan send]:
github.com/cockroachdb/cockroach/pkg/sql/importer.ingestKvs.func3()
github.com/cockroachdb/cockroach/pkg/sql/importer/import_processor.go:468 +0x311
github.com/cockroachdb/cockroach/pkg/sql/importer.ingestKvs.func5({0x6b20dd0, 0xc00487fc80})
github.com/cockroachdb/cockroach/pkg/sql/importer/import_processor.go:547 +0x2cf
github.com/cockroachdb/cockroach/pkg/util/ctxgroup.Group.GoCtx.func1()
github.com/cockroachdb/cockroach/pkg/util/ctxgroup/ctxgroup.go:168 +0x25
golang.org/x/sync/errgroup.(*Group).Go.func1()
golang.org/x/sync/errgroup/external/org_golang_x_sync/errgroup/errgroup.go:75 +0x64
created by golang.org/x/sync/errgroup.(*Group).Go
golang.org/x/sync/errgroup/external/org_golang_x_sync/errgroup/errgroup.go:72 +0xa5
--- FAIL: TestImportDefaultWithResume (5.74s)
```
<p>Parameters: <code>TAGS=bazel,gss</code>
</p>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/sql-experience
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestImportDefaultWithResume.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-22045
|
non_process
|
sql importer testimportdefaultwithresume failed sql importer testimportdefaultwithresume with on master test log scope go test logs captured to artifacts tmp tmp test log scope go use show logs to present logs inline cont testimportdefaultwithresume import stmt test go test log scope end import stmt test go leaked goroutine goroutine sync runtime semacquire goroot src runtime sema go sync waitgroup wait goroot src sync waitgroup go golang org x sync errgroup group wait golang org x sync errgroup external org golang x sync errgroup errgroup go github com cockroachdb cockroach pkg util ctxgroup group wait github com cockroachdb cockroach pkg util ctxgroup ctxgroup go github com cockroachdb cockroach pkg sql importer runimport github com cockroachdb cockroach pkg sql importer read import base go github com cockroachdb cockroach pkg sql importer readimportdataprocessor start github com cockroachdb cockroach pkg sql importer import processor go created by github com cockroachdb cockroach pkg sql importer readimportdataprocessor start github com cockroachdb cockroach pkg sql importer import processor go leaked goroutine goroutine sync runtime semacquire goroot src runtime sema go sync waitgroup wait goroot src sync waitgroup go golang org x sync errgroup group wait golang org x sync errgroup external org golang x sync errgroup errgroup go github com cockroachdb cockroach pkg util ctxgroup group wait github com cockroachdb cockroach pkg util ctxgroup ctxgroup go github com cockroachdb cockroach pkg sql importer ingestkvs github com cockroachdb cockroach pkg sql importer import processor go github com cockroachdb cockroach pkg sql importer runimport github com cockroachdb cockroach pkg sql importer read import base go github com cockroachdb cockroach pkg util ctxgroup group goctx github com cockroachdb cockroach pkg util ctxgroup ctxgroup go golang org x sync errgroup group go golang org x sync errgroup external org golang x sync errgroup errgroup go created by golang org x sync errgroup group go golang org x sync errgroup external org golang x sync errgroup errgroup go leaked goroutine goroutine github com cockroachdb cockroach pkg sql importer ingestkvs github com cockroachdb cockroach pkg sql importer import processor go github com cockroachdb cockroach pkg sql importer ingestkvs github com cockroachdb cockroach pkg sql importer import processor go github com cockroachdb cockroach pkg util ctxgroup group goctx github com cockroachdb cockroach pkg util ctxgroup ctxgroup go golang org x sync errgroup group go golang org x sync errgroup external org golang x sync errgroup errgroup go created by golang org x sync errgroup group go golang org x sync errgroup external org golang x sync errgroup errgroup go fail testimportdefaultwithresume parameters tags bazel gss help see also cc cockroachdb sql experience jira issue crdb
| 0
|
199,751
| 6,994,045,001
|
IssuesEvent
|
2017-12-15 13:58:01
|
go-gitea/gitea
|
https://api.github.com/repos/go-gitea/gitea
|
opened
|
Repository home page show error file last commit message.
|
kind/bug priority/critical
|
Compare github's home page
https://github.com/go-gitea/gitea/tree/release/v1.0

vs https://try.gitea.io/gitea/gitea/src/branch/release/v1.0

|
1.0
|
Repository home page show error file last commit message. - Compare github's home page
https://github.com/go-gitea/gitea/tree/release/v1.0

vs https://try.gitea.io/gitea/gitea/src/branch/release/v1.0

|
non_process
|
repository home page show error file last commit message compare github s home page vs
| 0
|
21,788
| 30,297,176,715
|
IssuesEvent
|
2023-07-10 00:32:30
|
h4sh5/pypi-auto-scanner
|
https://api.github.com/repos/h4sh5/pypi-auto-scanner
|
opened
|
dbretina 2.2.10 has 1 GuardDog issues
|
guarddog silent-process-execution
|
https://pypi.org/project/dbretina
https://inspector.pypi.io/project/dbretina
```{
"dependency": "dbretina",
"version": "2.2.10",
"result": {
"issues": 1,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "kSpider2/ks_filter.py/kSpider2/ks_filter.py:15",
"code": " subprocess.run([\"awk\"], stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmp6f_pfe0h/dbretina"
}
}```
|
1.0
|
dbretina 2.2.10 has 1 GuardDog issues - https://pypi.org/project/dbretina
https://inspector.pypi.io/project/dbretina
```{
"dependency": "dbretina",
"version": "2.2.10",
"result": {
"issues": 1,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "kSpider2/ks_filter.py/kSpider2/ks_filter.py:15",
"code": " subprocess.run([\"awk\"], stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmp6f_pfe0h/dbretina"
}
}```
|
process
|
dbretina has guarddog issues dependency dbretina version result issues errors results silent process execution location ks filter py ks filter py code subprocess run stdin subprocess devnull n stdout subprocess devnull stderr subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp dbretina
| 1
|
17,890
| 23,864,912,348
|
IssuesEvent
|
2022-09-07 10:10:01
|
jasp-stats/jasp-issues
|
https://api.github.com/repos/jasp-stats/jasp-issues
|
closed
|
[Bug]: unexpected error in Process Capability Study of jaspProcessControl
|
waiting for requester Module: jaspProcessControl
|
### JASP Version
0.16
### Commit ID
_No response_
### JASP Module
jaspNetwork
### What analysis are you seeing the problem on?
_No response_
### What OS are you seeing the problem on?
windows 7
### Bug Description
IMPOSSIBLE TO SHOW UP THE RESULTS. I NEED TO FIX IT ASAP
### Expected Behaviour
[](url)


### Steps to Reproduce
1.
2.
3.
...

### Log (if any)
_No response_
### Final Checklist
- [X] I have included a screenshot showcasing the issue, if any.
- [X] I have included a JASP file that causes the crash/bug, if any.
- [X] I have accurately described the bug, and steps to reproduce it.
|
1.0
|
[Bug]: unexpected error in Process Capability Study of jaspProcessControl - ### JASP Version
0.16
### Commit ID
_No response_
### JASP Module
jaspNetwork
### What analysis are you seeing the problem on?
_No response_
### What OS are you seeing the problem on?
windows 7
### Bug Description
IMPOSSIBLE TO SHOW UP THE RESULTS. I NEED TO FIX IT ASAP
### Expected Behaviour
[](url)


### Steps to Reproduce
1.
2.
3.
...

### Log (if any)
_No response_
### Final Checklist
- [X] I have included a screenshot showcasing the issue, if any.
- [X] I have included a JASP file that causes the crash/bug, if any.
- [X] I have accurately described the bug, and steps to reproduce it.
|
process
|
unexpected error in process capability study of jaspprocesscontrol jasp version commit id no response jasp module jaspnetwork what analysis are you seeing the problem on no response what os are you seeing the problem on windows bug description impossible to show up the results i need to fix it asap expected behaviour url steps to reproduce log if any no response final checklist i have included a screenshot showcasing the issue if any i have included a jasp file that causes the crash bug if any i have accurately described the bug and steps to reproduce it
| 1
|
217,069
| 24,312,769,084
|
IssuesEvent
|
2022-09-30 01:17:23
|
sesong11/example
|
https://api.github.com/repos/sesong11/example
|
opened
|
CVE-2020-1935 (Medium) detected in tomcat-embed-core-9.0.21.jar
|
security vulnerability
|
## CVE-2020-1935 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.21.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p>
<p>Path to dependency file: /example/spring-statemachine-jpa/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/9.0.21/tomcat-embed-core-9.0.21.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.1.6.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-2.1.6.RELEASE.jar
- :x: **tomcat-embed-core-9.0.21.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Apache Tomcat 9.0.0.M1 to 9.0.30, 8.5.0 to 8.5.50 and 7.0.0 to 7.0.99 the HTTP header parsing code used an approach to end-of-line parsing that allowed some invalid HTTP headers to be parsed as valid. This led to a possibility of HTTP Request Smuggling if Tomcat was located behind a reverse proxy that incorrectly handled the invalid Transfer-Encoding header in a particular manner. Such a reverse proxy is considered unlikely.
<p>Publish Date: 2020-02-24
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-1935>CVE-2020-1935</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-6v7p-v754-j89v">https://github.com/advisories/GHSA-6v7p-v754-j89v</a></p>
<p>Release Date: 2020-02-24</p>
<p>Fix Resolution (org.apache.tomcat.embed:tomcat-embed-core): 9.0.31</p>
<p>Direct dependency fix Resolution (org.springframework.boot:spring-boot-starter-web): 2.1.13.RELEASE</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-1935 (Medium) detected in tomcat-embed-core-9.0.21.jar - ## CVE-2020-1935 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.21.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p>
<p>Path to dependency file: /example/spring-statemachine-jpa/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/9.0.21/tomcat-embed-core-9.0.21.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.1.6.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-2.1.6.RELEASE.jar
- :x: **tomcat-embed-core-9.0.21.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Apache Tomcat 9.0.0.M1 to 9.0.30, 8.5.0 to 8.5.50 and 7.0.0 to 7.0.99 the HTTP header parsing code used an approach to end-of-line parsing that allowed some invalid HTTP headers to be parsed as valid. This led to a possibility of HTTP Request Smuggling if Tomcat was located behind a reverse proxy that incorrectly handled the invalid Transfer-Encoding header in a particular manner. Such a reverse proxy is considered unlikely.
<p>Publish Date: 2020-02-24
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-1935>CVE-2020-1935</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-6v7p-v754-j89v">https://github.com/advisories/GHSA-6v7p-v754-j89v</a></p>
<p>Release Date: 2020-02-24</p>
<p>Fix Resolution (org.apache.tomcat.embed:tomcat-embed-core): 9.0.31</p>
<p>Direct dependency fix Resolution (org.springframework.boot:spring-boot-starter-web): 2.1.13.RELEASE</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in tomcat embed core jar cve medium severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation library home page a href path to dependency file example spring statemachine jpa pom xml path to vulnerable library root repository org apache tomcat embed tomcat embed core tomcat embed core jar dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar x tomcat embed core jar vulnerable library vulnerability details in apache tomcat to to and to the http header parsing code used an approach to end of line parsing that allowed some invalid http headers to be parsed as valid this led to a possibility of http request smuggling if tomcat was located behind a reverse proxy that incorrectly handled the invalid transfer encoding header in a particular manner such a reverse proxy is considered unlikely publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat embed tomcat embed core direct dependency fix resolution org springframework boot spring boot starter web release step up your open source security game with mend
| 0
|
90,803
| 10,697,802,933
|
IssuesEvent
|
2019-10-23 17:18:40
|
jan-molak/serenity-js
|
https://api.github.com/repos/jan-molak/serenity-js
|
closed
|
Suggestion required: Restricting Abilities based on Persona
|
documentation question
|
Hi @jan-molak ,
We have started using SerenityJs 2.0 in our Projects. We have figured out personas to be used by teams for our Workflow testing. We do not want everyone to create his or her own personas, instead use from the list defined by stakeholders.
We currently implemented this by creating a core package with Personas and their abilities and plan is to create a package and publish so that every team is using defined set of personas and abilities.
**-- Our Current Thought Process --
core/persona.ts**
```
/**
* List of Personas that can be used in framework
*/
export enum Persona {
ABC,
/**
* ABC: a Patient
*/
XYZ,
/**
* XYZ: Is an Doctor
*/
MNO,
/**
* MNO: Is an Nurse
*/
}
```
**Preparing Actor with abilities**
_**core/customestage.ts**_
```
export class customStage implements DressingRoom {
prepare(actor: Actor): Actor {
switch (actor.name) {
case 'ABC':
return Actor.named(actor.name).whoCan(
BrowseTheWeb.using(protractor.browser));
case 'MNO':
return Actor.named(actor.name).whoCan(
BrowseTheWeb.using(protractor.browser));
case 'XYZ':
return Actor.named(actor.name).whoCan(
BrowseTheWeb.using(protractor.browser),
Search.with("ourapplication", ''));//custom ability
default:
return Actor.named(actor.name).whoCan(
BrowseTheWeb.using(protractor.browser),);
}
}
}
```
**_core/setup.ts_**
```
export function customactor(persona: string): Actor {
return serenity.callToStageFor(new customStage()).theActorCalled(persona)
}
setWorldConstructor(customactor)
```
**_stepdefination.ts_**
```
let xyz = customactor(Persona[Persona.XYZ])
Given(/^that XYZ is an authorized user of our system$/, async ()=> {
return xyz.attemptsTo(Be.onOurSystem())
});
```
Need your inputs on if this is the correct approach. With above implementation, we lose flexibility to reuse step_definition as we are preparing actor and providing them with the abilities based on the names in persona enum. But can enforce that no one creating their own random personas/abilities across organization.
If we go with the feature file and step_definition shown in your example, It can work for any actor. Wondering how we can handle the restrictions we want to impose in similar way. Also we wanted to keep step definition precise ` return xyz.attemptsTo(Be.onOurSystem())`
```
Given(/(.*) decides to use the Super Calculator/, function (this: WithStage, actorName: string) {
return this.stage.theActorCalled(actorName).attemptsTo(
Navigate.to('/protractor-demo/'),
);
});
```
Thank You for all your great work and help!
|
1.0
|
Suggestion required: Restricting Abilities based on Persona - Hi @jan-molak ,
We have started using SerenityJs 2.0 in our Projects. We have figured out personas to be used by teams for our Workflow testing. We do not want everyone to create his or her own personas, instead use from the list defined by stakeholders.
We currently implemented this by creating a core package with Personas and their abilities and plan is to create a package and publish so that every team is using defined set of personas and abilities.
**-- Our Current Thought Process --
core/persona.ts**
```
/**
* List of Personas that can be used in framework
*/
export enum Persona {
ABC,
/**
* ABC: a Patient
*/
XYZ,
/**
* XYZ: Is an Doctor
*/
MNO,
/**
* MNO: Is an Nurse
*/
}
```
**Preparing Actor with abilities**
_**core/customestage.ts**_
```
export class customStage implements DressingRoom {
prepare(actor: Actor): Actor {
switch (actor.name) {
case 'ABC':
return Actor.named(actor.name).whoCan(
BrowseTheWeb.using(protractor.browser));
case 'MNO':
return Actor.named(actor.name).whoCan(
BrowseTheWeb.using(protractor.browser));
case 'XYZ':
return Actor.named(actor.name).whoCan(
BrowseTheWeb.using(protractor.browser),
Search.with("ourapplication", ''));//custom ability
default:
return Actor.named(actor.name).whoCan(
BrowseTheWeb.using(protractor.browser),);
}
}
}
```
**_core/setup.ts_**
```
export function customactor(persona: string): Actor {
return serenity.callToStageFor(new customStage()).theActorCalled(persona)
}
setWorldConstructor(customactor)
```
**_stepdefination.ts_**
```
let xyz = customactor(Persona[Persona.XYZ])
Given(/^that XYZ is an authorized user of our system$/, async ()=> {
return xyz.attemptsTo(Be.onOurSystem())
});
```
Need your inputs on if this is the correct approach. With above implementation, we lose flexibility to reuse step_definition as we are preparing actor and providing them with the abilities based on the names in persona enum. But can enforce that no one creating their own random personas/abilities across organization.
If we go with the feature file and step_definition shown in your example, It can work for any actor. Wondering how we can handle the restrictions we want to impose in similar way. Also we wanted to keep step definition precise ` return xyz.attemptsTo(Be.onOurSystem())`
```
Given(/(.*) decides to use the Super Calculator/, function (this: WithStage, actorName: string) {
return this.stage.theActorCalled(actorName).attemptsTo(
Navigate.to('/protractor-demo/'),
);
});
```
Thank You for all your great work and help!
|
non_process
|
suggestion required restricting abilities based on persona hi jan molak we have started using serenityjs in our projects we have figured out personas to be used by teams for our workflow testing we do not want everyone to create his or her own personas instead use from the list defined by stakeholders we currently implemented this by creating a core package with personas and their abilities and plan is to create a package and publish so that every team is using defined set of personas and abilities our current thought process core persona ts list of personas that can be used in framework export enum persona abc abc a patient xyz xyz is an doctor mno mno is an nurse preparing actor with abilities core customestage ts export class customstage implements dressingroom prepare actor actor actor switch actor name case abc return actor named actor name whocan browsetheweb using protractor browser case mno return actor named actor name whocan browsetheweb using protractor browser case xyz return actor named actor name whocan browsetheweb using protractor browser search with ourapplication custom ability default return actor named actor name whocan browsetheweb using protractor browser core setup ts export function customactor persona string actor return serenity calltostagefor new customstage theactorcalled persona setworldconstructor customactor stepdefination ts let xyz customactor persona given that xyz is an authorized user of our system async return xyz attemptsto be onoursystem need your inputs on if this is the correct approach with above implementation we lose flexibility to reuse step definition as we are preparing actor and providing them with the abilities based on the names in persona enum but can enforce that no one creating their own random personas abilities across organization if we go with the feature file and step definition shown in your example it can work for any actor wondering how we can handle the restrictions we want to impose in similar way also we wanted to keep step definition precise return xyz attemptsto be onoursystem given decides to use the super calculator function this withstage actorname string return this stage theactorcalled actorname attemptsto navigate to protractor demo thank you for all your great work and help
| 0
|
46,481
| 19,252,476,886
|
IssuesEvent
|
2021-12-09 07:36:16
|
DimoDimchev/Trivial
|
https://api.github.com/repos/DimoDimchev/Trivial
|
closed
|
Research asynchronous operations between different services and external API's
|
microservices
|
Decide where and if async functionality is needed for communication between cervices and async API's.
* Django and Celery
* Flask needs research
|
1.0
|
Research asynchronous operations between different services and external API's - Decide where and if async functionality is needed for communication between cervices and async API's.
* Django and Celery
* Flask needs research
|
non_process
|
research asynchronous operations between different services and external api s decide where and if async functionality is needed for communication between cervices and async api s django and celery flask needs research
| 0
|
11,117
| 13,957,683,594
|
IssuesEvent
|
2020-10-24 08:08:05
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
ES: Discovery service unavailable
|
ES - Spain Geoportal Harvesting process
|
Dear Paloma,
we are unable to harvest from your registered endpoint. It returns a HTTP status 503 UNAVAILABLE.
We already tried on Dec 13th, 16th and today, and each time, after four attempts with a increasing timeout (up to 18sec), the service replies HTTP status 503 Service Unavailable to our POST GetRecords request.
Since we follow the specification inside the GetCapabilities document
http://www.idee.es/csw-codsi-idee/srv/spa/csw?request=GetCapabilities&service=CSW&version=2.0.2
we notice that every provided URLs contain the specification of the port 8080.
<ows:Operation name="GetRecords">
<ows:DCP>
<ows:HTTP>
<ows:Get xlink:href="http://www.idee.es:8080/csw-codsi-idee/srv/eng/csw" />
<ows:Post xlink:href="http://www.idee.es:8080/csw-codsi-idee/srv/eng/csw" />
</ows:HTTP>
</ows:DCP>
Just for test, we tried to execute a POST GetRecords without the mentioned port and it succeed.
Could you please investigate from your side and fix it? We may suggest to update the GetCapabilities URLs by removing the "internal" port specification.
Best regards,
Davide on behalf of JRC INSPIRE Support team
|
1.0
|
ES: Discovery service unavailable - Dear Paloma,
we are unable to harvest from your registered endpoint. It returns a HTTP status 503 UNAVAILABLE.
We already tried on Dec 13th, 16th and today, and each time, after four attempts with a increasing timeout (up to 18sec), the service replies HTTP status 503 Service Unavailable to our POST GetRecords request.
Since we follow the specification inside the GetCapabilities document
http://www.idee.es/csw-codsi-idee/srv/spa/csw?request=GetCapabilities&service=CSW&version=2.0.2
we notice that every provided URLs contain the specification of the port 8080.
<ows:Operation name="GetRecords">
<ows:DCP>
<ows:HTTP>
<ows:Get xlink:href="http://www.idee.es:8080/csw-codsi-idee/srv/eng/csw" />
<ows:Post xlink:href="http://www.idee.es:8080/csw-codsi-idee/srv/eng/csw" />
</ows:HTTP>
</ows:DCP>
Just for test, we tried to execute a POST GetRecords without the mentioned port and it succeed.
Could you please investigate from your side and fix it? We may suggest to update the GetCapabilities URLs by removing the "internal" port specification.
Best regards,
Davide on behalf of JRC INSPIRE Support team
|
process
|
es discovery service unavailable dear paloma we are unable to harvest from your registered endpoint it returns a http status unavailable we already tried on dec and today and each time after four attempts with a increasing timeout up to the service replies http status service unavailable to our post getrecords request since we follow the specification inside the getcapabilities document we notice that every provided urls contain the specification of the port lt ows operation name quot getrecords quot gt lt ows dcp gt lt ows http gt lt ows get xlink href quot gt lt ows post xlink href quot gt lt ows http gt lt ows dcp gt just for test we tried to execute a post getrecords without the mentioned port and it succeed could you please investigate from your side and fix it we may suggest to update the getcapabilities urls by removing the quot internal quot port specification best regards davide on behalf of jrc inspire support team
| 1
|
152,863
| 12,128,299,035
|
IssuesEvent
|
2020-04-22 20:15:00
|
physiopy/phys2bids
|
https://api.github.com/repos/physiopy/phys2bids
|
closed
|
Add Windows CI testing
|
Testing
|
We are currently not testing (nor developing) `phys2bids` on Windows systems, leading to issues like e.g., #171. I think we should look into adding Windows CI testing using a service like [AppVeyor](https://www.appveyor.com/) or [Azure Pipelines](https://azure.microsoft.com/en-us/services/devops/pipelines/) (since TravisCI currently only has [minimal](https://docs.travis-ci.com/user/reference/windows/) Windows support).
## Context / Motivation
We shouldn't limit our users to those researchers working on Mac or Linux systems, so adding AppVeyor or Azure support / testing so that we can ensure our primary workflows are viable on Windows is critical.
## Possible Implementation
`nipy/nibabel` recently moved to Azure Pipelines from AppVeyor (relevant files: [1](https://github.com/nipy/nibabel/blob/master/.azure-pipelines/windows.yml) [2](https://github.com/nipy/nibabel/blob/master/azure-pipelines.yml)); we could look into simply copying the relevant portions of their framework for `phys2bids`.
|
1.0
|
Add Windows CI testing - We are currently not testing (nor developing) `phys2bids` on Windows systems, leading to issues like e.g., #171. I think we should look into adding Windows CI testing using a service like [AppVeyor](https://www.appveyor.com/) or [Azure Pipelines](https://azure.microsoft.com/en-us/services/devops/pipelines/) (since TravisCI currently only has [minimal](https://docs.travis-ci.com/user/reference/windows/) Windows support).
## Context / Motivation
We shouldn't limit our users to those researchers working on Mac or Linux systems, so adding AppVeyor or Azure support / testing so that we can ensure our primary workflows are viable on Windows is critical.
## Possible Implementation
`nipy/nibabel` recently moved to Azure Pipelines from AppVeyor (relevant files: [1](https://github.com/nipy/nibabel/blob/master/.azure-pipelines/windows.yml) [2](https://github.com/nipy/nibabel/blob/master/azure-pipelines.yml)); we could look into simply copying the relevant portions of their framework for `phys2bids`.
|
non_process
|
add windows ci testing we are currently not testing nor developing on windows systems leading to issues like e g i think we should look into adding windows ci testing using a service like or since travisci currently only has windows support context motivation we shouldn t limit our users to those researchers working on mac or linux systems so adding appveyor or azure support testing so that we can ensure our primary workflows are viable on windows is critical possible implementation nipy nibabel recently moved to azure pipelines from appveyor relevant files we could look into simply copying the relevant portions of their framework for
| 0
|
3,389
| 6,515,961,352
|
IssuesEvent
|
2017-08-26 23:31:44
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
test: implement a way to regularly test `abort` behaviour on all platform
|
process test
|
<!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify affected core module name
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you are able.
-->
* **Version**: *
* **Platform**: *
* **Subsystem**: process,test
<!-- Enter your issue details below this comment. -->
Since testing `abort()` and `--abort-on-uncaught-exception` generate core-dump that take allot of storage they have been moved to `test/abort/` and are not regularly tested on the full repertoire of platforms that we support.
/cc @nodejs/testing @nodejs/build
|
1.0
|
test: implement a way to regularly test `abort` behaviour on all platform - <!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify affected core module name
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you are able.
-->
* **Version**: *
* **Platform**: *
* **Subsystem**: process,test
<!-- Enter your issue details below this comment. -->
Since testing `abort()` and `--abort-on-uncaught-exception` generate core-dump that take allot of storage they have been moved to `test/abort/` and are not regularly tested on the full repertoire of platforms that we support.
/cc @nodejs/testing @nodejs/build
|
process
|
test implement a way to regularly test abort behaviour on all platform thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able version output of node v platform output of uname a unix or version and or bit windows subsystem if known please specify affected core module name if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you are able version platform subsystem process test since testing abort and abort on uncaught exception generate core dump that take allot of storage they have been moved to test abort and are not regularly tested on the full repertoire of platforms that we support cc nodejs testing nodejs build
| 1
|
227,336
| 17,380,694,735
|
IssuesEvent
|
2021-07-31 16:47:14
|
open-contracting/standard
|
https://api.github.com/repos/open-contracting/standard
|
opened
|
Worked example: Minimal updates in change history
|
Focus - Documentation
|
#1302 rewrites the Getting Started section as a Primer, which included removing the "Repeating previous information" section from the previous version of the Releases and Records page, accessible here: https://github.com/open-contracting/standard/blob/106f4c4/docs/getting_started/releases_and_records.md#repeating-previous-information
This content should be re-introduced as a worked example.
|
1.0
|
Worked example: Minimal updates in change history - #1302 rewrites the Getting Started section as a Primer, which included removing the "Repeating previous information" section from the previous version of the Releases and Records page, accessible here: https://github.com/open-contracting/standard/blob/106f4c4/docs/getting_started/releases_and_records.md#repeating-previous-information
This content should be re-introduced as a worked example.
|
non_process
|
worked example minimal updates in change history rewrites the getting started section as a primer which included removing the repeating previous information section from the previous version of the releases and records page accessible here this content should be re introduced as a worked example
| 0
|
304,546
| 26,287,240,688
|
IssuesEvent
|
2023-01-08 00:36:47
|
devssa/onde-codar-em-salvador
|
https://api.github.com/repos/devssa/onde-codar-em-salvador
|
closed
|
[TESTES] [ESPECIALISTA] [REMOTO] Testes Automatizados (Especialista) na [SOFIST]
|
BANCO DE DADOS JAVASCRIPT BDD TESTE DE INTEGRAÇÃO TYPESCRIPT REMOTO TESTES AUTOMATIZADOS HELP WANTED ESPECIALISTA TESTES DE API TESTES DE UNIDADE Stale
|
<!--
==================================================
POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS!
Use: "Desenvolvedor Front-end" ao invés de
"Front-End Developer" \o/
Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]`
==================================================
-->
## Descrição da vaga
- Sabemos que testes de regressão precisam ser executados rapidamente. Quando realizados 100% manualmente, os testes de regressão impedem os QAs de realmente realizarem o que importa, que é encontrar novas formas de explorar novos bugs nos aplicativos. Imagine agora produtos que são utilizados por milhões de pessoas, em centenas de dispositivos distintos: há muitas situações para serem automatizadas! Seu papel será apoiar nosso time de QA automatizando cada vez mais testes e inúmeras situações de uso de aplicativos Android e iOS e sistemas WEB, de forma a acelerar o processo de liberação de novas versões dos apps a cada sprint. Faremos nosso trabalho utilizando BDD (Cucumber), Appium, Selenium Webdriver (com Page Objects) e, quando aplicável, nuvens de dispositivos, portanto saber aplicar isto é muito importante para o sucesso de todos no projeto.
- Na Sofist valorizamos pessoas que saibam trabalhar em equipe, entendam como o trabalho de cada um contribui para que os resultados sejam alcançados, estejam em constante aprimoramento profissional, sejam apaixonadas pelo que fazem e entendam que ações (não ideias) podem mudar a vida das pessoas.
RESPONSABILIDADES E ATRIBUIÇÕES
DESCRIÇÃO DO TRABALHO
- Automatizar testes em projetos de apps Android e iOS e sistemas WEB utilizando BDD (Cucumber), Selenium Webdriver, Appium e nuvens de dispositivos
- Apoiar o desenvolvimento de features e APIs que simulem ações dos usuários para uso durante os testes automatizados e testes exploratórios manuais
- Ajudar na manutenção dos ambientes de testes de integração
- Auxiliar na capacitação de outros membros do time no processo e tecnologias usadas
- Propor ao time melhorias no processo que aprimorem o resultado que entregamos
- Introduzir ferramentas e metodologias que tornem o trabalho mais eficiente
## Local
- Remoto
## Benefícios
- Vale Refeição
- Vale Alimentação
- Plano de saúde (Unimed)
- Plano odontológico (Amil)
- Gympass
- Seguro de vida
- PLR
- Horário Flexível
- Auxílio Creche
- Licença Maternidade e Paternidade Estendida
- Day Off
## Requisitos
**Obrigatórios:**
- Ter atuado com linguagens JavaScript e/ou TypeScript
- Conhecimento prévio para elaboração e escrita dos cenários utilizando BDD
- Ter atuado com Testes de Unidade
- Ter atuado com Testes de API
- Ter atuado com Testes de Integração
- Conhecimento em Banco de dados
- Conhecimento na utilização de Docker
**Diferenciais:**
- Conhecimentos a práticas de CI/CD
- Conhecimento na utilização do Gitlab,
- Conhecimento em Testes Unitários
- Conhecimento em Mocks e Stubs
- Conhecimento em Testes de Contrato)
## Contratação
- a combinar
## Nossa empresa
- Somos uma empresa que promove troca de conhecimento entre os colaboradores, entendemos que cada Sofister tem uma habilidade que nos ajuda a crescer.
- Somos feitos por pessoas, por isso nossa prioridade é cuidar delas. Não é a toa que fomos, pela terceira vez consecutiva, certificados como um Great Place to Work. E foi com muito orgulho que conquistamos um espaço entre as 20 melhores empresa para trabalhar na nossa categoria, em 2019.
- Aqui você encontrará um horário de trabalho flexível, day off de aniversário, café da manhã na empresa, não terá que se preocupar com dress code, e temos um ambiente que promove a diversidade, focado no respeito e colaboração. Aliás, é com enorme orgulho que ressaltamos que em 2020 ficamos com nota 100 (máxima) no GPTW em itens como: as pessoas aqui são bem tratadas independentemente de sua orientação sexual, cor, etnia, sexo e idade.
- Isso só reforça nosso empenho e preocupação genuína com as pessoas.
- Seja durante sua atuação remota ou no escritório físico, nosso clima organizacional é pautado em uma cultura de relações saudáveis, promovendo o desenvolvimento e a autonomia.
- Trabalhando aqui você fará parte de um time que se ajuda e apoia, com liberdade para colocar suas ideias em prática, além de um programa de desenvolvimento apoiado por mentores técnicos, que são os melhores QA’s do mercado!
- Se você ama o universo QA, você vai amar ser um(a) Sofister. Somos parceiros de grandes nomes do mercado de tecnologia, participamos de projetos incríveis, que vão te desafiar todos os dias a enriquecer seu conhecimento sobre o mundo de testes.
- Vem pra Sofist!
## Como se candidatar
- [Clique aqui para se candidatar](https://vemprasofist.gupy.io/jobs/487457?jobBoardSource=gupy_public_page)
|
4.0
|
[TESTES] [ESPECIALISTA] [REMOTO] Testes Automatizados (Especialista) na [SOFIST] - <!--
==================================================
POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS!
Use: "Desenvolvedor Front-end" ao invés de
"Front-End Developer" \o/
Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]`
==================================================
-->
## Descrição da vaga
- Sabemos que testes de regressão precisam ser executados rapidamente. Quando realizados 100% manualmente, os testes de regressão impedem os QAs de realmente realizarem o que importa, que é encontrar novas formas de explorar novos bugs nos aplicativos. Imagine agora produtos que são utilizados por milhões de pessoas, em centenas de dispositivos distintos: há muitas situações para serem automatizadas! Seu papel será apoiar nosso time de QA automatizando cada vez mais testes e inúmeras situações de uso de aplicativos Android e iOS e sistemas WEB, de forma a acelerar o processo de liberação de novas versões dos apps a cada sprint. Faremos nosso trabalho utilizando BDD (Cucumber), Appium, Selenium Webdriver (com Page Objects) e, quando aplicável, nuvens de dispositivos, portanto saber aplicar isto é muito importante para o sucesso de todos no projeto.
- Na Sofist valorizamos pessoas que saibam trabalhar em equipe, entendam como o trabalho de cada um contribui para que os resultados sejam alcançados, estejam em constante aprimoramento profissional, sejam apaixonadas pelo que fazem e entendam que ações (não ideias) podem mudar a vida das pessoas.
RESPONSABILIDADES E ATRIBUIÇÕES
DESCRIÇÃO DO TRABALHO
- Automatizar testes em projetos de apps Android e iOS e sistemas WEB utilizando BDD (Cucumber), Selenium Webdriver, Appium e nuvens de dispositivos
- Apoiar o desenvolvimento de features e APIs que simulem ações dos usuários para uso durante os testes automatizados e testes exploratórios manuais
- Ajudar na manutenção dos ambientes de testes de integração
- Auxiliar na capacitação de outros membros do time no processo e tecnologias usadas
- Propor ao time melhorias no processo que aprimorem o resultado que entregamos
- Introduzir ferramentas e metodologias que tornem o trabalho mais eficiente
## Local
- Remoto
## Benefícios
- Vale Refeição
- Vale Alimentação
- Plano de saúde (Unimed)
- Plano odontológico (Amil)
- Gympass
- Seguro de vida
- PLR
- Horário Flexível
- Auxílio Creche
- Licença Maternidade e Paternidade Estendida
- Day Off
## Requisitos
**Obrigatórios:**
- Ter atuado com linguagens JavaScript e/ou TypeScript
- Conhecimento prévio para elaboração e escrita dos cenários utilizando BDD
- Ter atuado com Testes de Unidade
- Ter atuado com Testes de API
- Ter atuado com Testes de Integração
- Conhecimento em Banco de dados
- Conhecimento na utilização de Docker
**Diferenciais:**
- Conhecimentos a práticas de CI/CD
- Conhecimento na utilização do Gitlab,
- Conhecimento em Testes Unitários
- Conhecimento em Mocks e Stubs
- Conhecimento em Testes de Contrato)
## Contratação
- a combinar
## Nossa empresa
- Somos uma empresa que promove troca de conhecimento entre os colaboradores, entendemos que cada Sofister tem uma habilidade que nos ajuda a crescer.
- Somos feitos por pessoas, por isso nossa prioridade é cuidar delas. Não é a toa que fomos, pela terceira vez consecutiva, certificados como um Great Place to Work. E foi com muito orgulho que conquistamos um espaço entre as 20 melhores empresa para trabalhar na nossa categoria, em 2019.
- Aqui você encontrará um horário de trabalho flexível, day off de aniversário, café da manhã na empresa, não terá que se preocupar com dress code, e temos um ambiente que promove a diversidade, focado no respeito e colaboração. Aliás, é com enorme orgulho que ressaltamos que em 2020 ficamos com nota 100 (máxima) no GPTW em itens como: as pessoas aqui são bem tratadas independentemente de sua orientação sexual, cor, etnia, sexo e idade.
- Isso só reforça nosso empenho e preocupação genuína com as pessoas.
- Seja durante sua atuação remota ou no escritório físico, nosso clima organizacional é pautado em uma cultura de relações saudáveis, promovendo o desenvolvimento e a autonomia.
- Trabalhando aqui você fará parte de um time que se ajuda e apoia, com liberdade para colocar suas ideias em prática, além de um programa de desenvolvimento apoiado por mentores técnicos, que são os melhores QA’s do mercado!
- Se você ama o universo QA, você vai amar ser um(a) Sofister. Somos parceiros de grandes nomes do mercado de tecnologia, participamos de projetos incríveis, que vão te desafiar todos os dias a enriquecer seu conhecimento sobre o mundo de testes.
- Vem pra Sofist!
## Como se candidatar
- [Clique aqui para se candidatar](https://vemprasofist.gupy.io/jobs/487457?jobBoardSource=gupy_public_page)
|
non_process
|
testes automatizados especialista na por favor só poste se a vaga for para salvador e cidades vizinhas use desenvolvedor front end ao invés de front end developer o exemplo desenvolvedor front end na descrição da vaga sabemos que testes de regressão precisam ser executados rapidamente quando realizados manualmente os testes de regressão impedem os qas de realmente realizarem o que importa que é encontrar novas formas de explorar novos bugs nos aplicativos imagine agora produtos que são utilizados por milhões de pessoas em centenas de dispositivos distintos há muitas situações para serem automatizadas seu papel será apoiar nosso time de qa automatizando cada vez mais testes e inúmeras situações de uso de aplicativos android e ios e sistemas web de forma a acelerar o processo de liberação de novas versões dos apps a cada sprint faremos nosso trabalho utilizando bdd cucumber appium selenium webdriver com page objects e quando aplicável nuvens de dispositivos portanto saber aplicar isto é muito importante para o sucesso de todos no projeto na sofist valorizamos pessoas que saibam trabalhar em equipe entendam como o trabalho de cada um contribui para que os resultados sejam alcançados estejam em constante aprimoramento profissional sejam apaixonadas pelo que fazem e entendam que ações não ideias podem mudar a vida das pessoas responsabilidades e atribuições descrição do trabalho automatizar testes em projetos de apps android e ios e sistemas web utilizando bdd cucumber selenium webdriver appium e nuvens de dispositivos apoiar o desenvolvimento de features e apis que simulem ações dos usuários para uso durante os testes automatizados e testes exploratórios manuais ajudar na manutenção dos ambientes de testes de integração auxiliar na capacitação de outros membros do time no processo e tecnologias usadas propor ao time melhorias no processo que aprimorem o resultado que entregamos introduzir ferramentas e metodologias que tornem o trabalho mais eficiente local remoto benefícios vale refeição vale alimentação plano de saúde unimed plano odontológico amil gympass seguro de vida plr horário flexível auxílio creche licença maternidade e paternidade estendida day off requisitos obrigatórios ter atuado com linguagens javascript e ou typescript conhecimento prévio para elaboração e escrita dos cenários utilizando bdd ter atuado com testes de unidade ter atuado com testes de api ter atuado com testes de integração conhecimento em banco de dados conhecimento na utilização de docker diferenciais conhecimentos a práticas de ci cd conhecimento na utilização do gitlab conhecimento em testes unitários conhecimento em mocks e stubs conhecimento em testes de contrato contratação a combinar nossa empresa somos uma empresa que promove troca de conhecimento entre os colaboradores entendemos que cada sofister tem uma habilidade que nos ajuda a crescer somos feitos por pessoas por isso nossa prioridade é cuidar delas não é a toa que fomos pela terceira vez consecutiva certificados como um great place to work e foi com muito orgulho que conquistamos um espaço entre as melhores empresa para trabalhar na nossa categoria em aqui você encontrará um horário de trabalho flexível day off de aniversário café da manhã na empresa não terá que se preocupar com dress code e temos um ambiente que promove a diversidade focado no respeito e colaboração aliás é com enorme orgulho que ressaltamos que em ficamos com nota máxima no gptw em itens como as pessoas aqui são bem tratadas independentemente de sua orientação sexual cor etnia sexo e idade isso só reforça nosso empenho e preocupação genuína com as pessoas seja durante sua atuação remota ou no escritório físico nosso clima organizacional é pautado em uma cultura de relações saudáveis promovendo o desenvolvimento e a autonomia trabalhando aqui você fará parte de um time que se ajuda e apoia com liberdade para colocar suas ideias em prática além de um programa de desenvolvimento apoiado por mentores técnicos que são os melhores qa’s do mercado se você ama o universo qa você vai amar ser um a sofister somos parceiros de grandes nomes do mercado de tecnologia participamos de projetos incríveis que vão te desafiar todos os dias a enriquecer seu conhecimento sobre o mundo de testes vem pra sofist como se candidatar
| 0
|
742,720
| 25,867,124,987
|
IssuesEvent
|
2022-12-13 21:58:55
|
jesus-collective/mobile
|
https://api.github.com/repos/jesus-collective/mobile
|
closed
|
How to invite user to specific membership
|
Secondary Priority login 50-60 Hrs
|
* [ ] Allow admin user to create custom emails for different platform invites
* [ ] Allow invites to specific products
* [ ] Admin user needs to be able to connect a custom email to a product
* [ ] When a user is invited to a specific product, they face a specific product payment screen
* [ ] Admin should be able to choose whether to invite an org or an individual

┆Issue is synchronized with this [Wrike Task](https://www.wrike.com/open.htm?id=713148447)
|
1.0
|
How to invite user to specific membership - * [ ] Allow admin user to create custom emails for different platform invites
* [ ] Allow invites to specific products
* [ ] Admin user needs to be able to connect a custom email to a product
* [ ] When a user is invited to a specific product, they face a specific product payment screen
* [ ] Admin should be able to choose whether to invite an org or an individual

┆Issue is synchronized with this [Wrike Task](https://www.wrike.com/open.htm?id=713148447)
|
non_process
|
how to invite user to specific membership allow admin user to create custom emails for different platform invites allow invites to specific products admin user needs to be able to connect a custom email to a product when a user is invited to a specific product they face a specific product payment screen admin should be able to choose whether to invite an org or an individual ┆issue is synchronized with this
| 0
|
359
| 2,794,984,374
|
IssuesEvent
|
2015-05-11 19:34:21
|
scieloorg/search-journals
|
https://api.github.com/repos/scieloorg/search-journals
|
opened
|
Capacidade de remover registro inexistentes
|
Processamento Tarefa
|
Atualmente o processamento do sistema de busca não tem a capacidade de remover e otimizar os registros, portanto para que não tenhamos registros fantasma no índice é importante incorporar essas opções no processamento, essa remoção deve considera o resultado da diferença entre o article meta e o índice.
Sugiro o parâmetro ``-o`` para otimizar e o parâmetro ``-r`` para remover.
|
1.0
|
Capacidade de remover registro inexistentes - Atualmente o processamento do sistema de busca não tem a capacidade de remover e otimizar os registros, portanto para que não tenhamos registros fantasma no índice é importante incorporar essas opções no processamento, essa remoção deve considera o resultado da diferença entre o article meta e o índice.
Sugiro o parâmetro ``-o`` para otimizar e o parâmetro ``-r`` para remover.
|
process
|
capacidade de remover registro inexistentes atualmente o processamento do sistema de busca não tem a capacidade de remover e otimizar os registros portanto para que não tenhamos registros fantasma no índice é importante incorporar essas opções no processamento essa remoção deve considera o resultado da diferença entre o article meta e o índice sugiro o parâmetro o para otimizar e o parâmetro r para remover
| 1
|
237,254
| 19,604,668,958
|
IssuesEvent
|
2022-01-06 07:49:38
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/transform/creation_runtime_mappings·ts - transform creation with runtime mappings batch transform with unique rt_airline_lower and sort by time and runtime mappings runs the transform and displays it correctly in Discover page
|
:ml failed-test skipped-test
|
A test failed on a tracked branch
```
Error: Timeout of 360000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (/var/lib/buildkite-agent/builds/kb-cigroup-6-04cd8299cf7a5667/elastic/kibana-hourly/kibana/x-pack/test/functional/apps/transform/creation_runtime_mappings.ts)
at listOnTimeout (internal/timers.js:557:17)
at processTimers (internal/timers.js:500:7)
```
First failure: [CI Build](https://buildkite.com/elastic/kibana-hourly/builds/916#20f30ec4-c3c1-4444-9914-d0cbb40a8b12)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/transform/creation_runtime_mappings·ts","test.name":"transform creation with runtime mappings batch transform with unique rt_airline_lower and sort by time and runtime mappings runs the transform and displays it correctly in Discover page","test.failCount":18}} -->
|
2.0
|
Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/transform/creation_runtime_mappings·ts - transform creation with runtime mappings batch transform with unique rt_airline_lower and sort by time and runtime mappings runs the transform and displays it correctly in Discover page - A test failed on a tracked branch
```
Error: Timeout of 360000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (/var/lib/buildkite-agent/builds/kb-cigroup-6-04cd8299cf7a5667/elastic/kibana-hourly/kibana/x-pack/test/functional/apps/transform/creation_runtime_mappings.ts)
at listOnTimeout (internal/timers.js:557:17)
at processTimers (internal/timers.js:500:7)
```
First failure: [CI Build](https://buildkite.com/elastic/kibana-hourly/builds/916#20f30ec4-c3c1-4444-9914-d0cbb40a8b12)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/transform/creation_runtime_mappings·ts","test.name":"transform creation with runtime mappings batch transform with unique rt_airline_lower and sort by time and runtime mappings runs the transform and displays it correctly in Discover page","test.failCount":18}} -->
|
non_process
|
failing test chrome x pack ui functional tests x pack test functional apps transform creation runtime mappings·ts transform creation with runtime mappings batch transform with unique rt airline lower and sort by time and runtime mappings runs the transform and displays it correctly in discover page a test failed on a tracked branch error timeout of exceeded for async tests and hooks ensure done is called if returning a promise ensure it resolves var lib buildkite agent builds kb cigroup elastic kibana hourly kibana x pack test functional apps transform creation runtime mappings ts at listontimeout internal timers js at processtimers internal timers js first failure
| 0
|
20,439
| 27,099,901,118
|
IssuesEvent
|
2023-02-15 07:45:23
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
"Join by nearest" regression on 3.28 (from 3.22) when multiple geometry colums are present
|
Processing Regression Bug
|
### What is the bug or the crash?
Having a nodes layer with multiple geometries I usually did on 3.22 LTR the join by nearest algorithm to find out nodes that are closer than 0.003 m.
With 3.28.2 there's an error now:
```
Feature could not be written to Joined_layer_.....: Could not store attribute "geometry_alt1": Could not convert value "" to target type "string"
Could not write feature into OUTPUT
Execution failed after 0.11 seconds
```
### Steps to reproduce the issue
1. Add the following SQL to your PostgreSQL database:
```
CREATE TABLE public.nodes (
id integer NOT NULL,
fk_district integer,
fk_pressurezone integer,
fk_printmap integer[],
fk_precision integer,
fk_precisionalti integer,
fk_object_reference integer,
altitude numeric(10,3) DEFAULT NULL::numeric,
_printmaps text,
_geometry_alt1_used boolean,
_geometry_alt2_used boolean,
_pipe_orientation double precision DEFAULT 0,
_pipe_schema_visible boolean DEFAULT false,
geometry public.geometry(PointZ,3844) NOT NULL,
geometry_alt1 public.geometry(PointZ,3844),
geometry_alt2 public.geometry(PointZ,3844)
);
INSERT INTO public.nodes VALUES (58957, NULL, NULL, '{}', 101, 101, 101, NULL, NULL, false, false, -8.627881391886133, true, '01010000A0040F00000000000018451B4100000000044018410000000000000000', NULL, NULL);
INSERT INTO public.nodes VALUES (58971, NULL, NULL, '{}', 101, 101, 101, NULL, NULL, false, false, -18.95382931168005, true, '01010000A0040F00000000000018451B4100000000084018410000000000000000', NULL, NULL);
ALTER TABLE ONLY public.nodes
ADD CONSTRAINT node_pkey PRIMARY KEY (id);
CREATE INDEX node_geoidx ON public.nodes USING gist (geometry);
CREATE INDEX node_geoidx_alt1 ON public.nodes USING gist (geometry_alt1);
CREATE INDEX node_geoidx_alt2 ON public.nodes USING gist (geometry_alt2);
```
2. Run the join by nearest algo:

3. The error is shown:
```
Feature could not be written to Joined_layer_.....: Could not store attribute "geometry_alt1": Could not convert value "" to target type "string"
Could not write feature into OUTPUT
Execution failed after 0.11 seconds
```
### Versions
<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0//EN" "http://www.w3.org/TR/REC-html40/strict.dtd">
<html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8" /><style type="text/css">
p, li { white-space: pre-wrap; }
</style></head><body>
QGIS version | 3.28.2-Firenze | QGIS code revision | d33b8750f3
-- | -- | -- | --
Qt version | 5.15.8
Python version | 3.11.1
GDAL/OGR version | 3.5.2
PROJ version | 9.0.1
EPSG Registry database version | v10.064 (2022-05-19)
GEOS version | 3.11.0-CAPI-1.17.0
SQLite version | 3.40.0
PDAL version | 2.4.3
PostgreSQL client version | unknown
SpatiaLite version | 5.0.1
QWT version | 6.1.5
QScintilla2 version | 2.13.0
OS version | Fedora Linux 37 (Thirty Seven)
| | |
Active Python plugins
slyr_community | 4.0.6
QuickWKT | 3.1
project_export_inspire | 1.0.0
water_topology | 0.1
plugin_reloader | 0.9.3
db_manager | 0.1.20
processing | 2.12.99
sagaprovider | 2.12.99
grassprovider | 2.12.99
MetaSearch | 0.3.6
</body></html>QGIS version
3.28.2-Firenze
QGIS code revision
[d33b8750f3](https://github.com/qgis/QGIS/commit/d33b8750f3)
Qt version
5.15.8
Python version
3.11.1
GDAL/OGR version
3.5.2
PROJ version
9.0.1
EPSG Registry database version
v10.064 (2022-05-19)
GEOS version
3.11.0-CAPI-1.17.0
SQLite version
3.40.0
PDAL version
2.4.3
PostgreSQL client version
unknown
SpatiaLite version
5.0.1
QWT version
6.1.5
QScintilla2 version
2.13.0
OS version
Fedora Linux 37 (Thirty Seven)
Active Python plugins
slyr_community
4.0.6
QuickWKT
3.1
project_export_inspire
1.0.0
water_topology
0.1
plugin_reloader
0.9.3
db_manager
0.1.20
processing
2.12.99
sagaprovider
2.12.99
grassprovider
2.12.99
MetaSearch
0.3.6
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
_No response_
|
1.0
|
"Join by nearest" regression on 3.28 (from 3.22) when multiple geometry colums are present - ### What is the bug or the crash?
Having a nodes layer with multiple geometries I usually did on 3.22 LTR the join by nearest algorithm to find out nodes that are closer than 0.003 m.
With 3.28.2 there's an error now:
```
Feature could not be written to Joined_layer_.....: Could not store attribute "geometry_alt1": Could not convert value "" to target type "string"
Could not write feature into OUTPUT
Execution failed after 0.11 seconds
```
### Steps to reproduce the issue
1. Add the following SQL to your PostgreSQL database:
```
CREATE TABLE public.nodes (
id integer NOT NULL,
fk_district integer,
fk_pressurezone integer,
fk_printmap integer[],
fk_precision integer,
fk_precisionalti integer,
fk_object_reference integer,
altitude numeric(10,3) DEFAULT NULL::numeric,
_printmaps text,
_geometry_alt1_used boolean,
_geometry_alt2_used boolean,
_pipe_orientation double precision DEFAULT 0,
_pipe_schema_visible boolean DEFAULT false,
geometry public.geometry(PointZ,3844) NOT NULL,
geometry_alt1 public.geometry(PointZ,3844),
geometry_alt2 public.geometry(PointZ,3844)
);
INSERT INTO public.nodes VALUES (58957, NULL, NULL, '{}', 101, 101, 101, NULL, NULL, false, false, -8.627881391886133, true, '01010000A0040F00000000000018451B4100000000044018410000000000000000', NULL, NULL);
INSERT INTO public.nodes VALUES (58971, NULL, NULL, '{}', 101, 101, 101, NULL, NULL, false, false, -18.95382931168005, true, '01010000A0040F00000000000018451B4100000000084018410000000000000000', NULL, NULL);
ALTER TABLE ONLY public.nodes
ADD CONSTRAINT node_pkey PRIMARY KEY (id);
CREATE INDEX node_geoidx ON public.nodes USING gist (geometry);
CREATE INDEX node_geoidx_alt1 ON public.nodes USING gist (geometry_alt1);
CREATE INDEX node_geoidx_alt2 ON public.nodes USING gist (geometry_alt2);
```
2. Run the join by nearest algo:

3. The error is shown:
```
Feature could not be written to Joined_layer_.....: Could not store attribute "geometry_alt1": Could not convert value "" to target type "string"
Could not write feature into OUTPUT
Execution failed after 0.11 seconds
```
### Versions
<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0//EN" "http://www.w3.org/TR/REC-html40/strict.dtd">
<html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8" /><style type="text/css">
p, li { white-space: pre-wrap; }
</style></head><body>
QGIS version | 3.28.2-Firenze | QGIS code revision | d33b8750f3
-- | -- | -- | --
Qt version | 5.15.8
Python version | 3.11.1
GDAL/OGR version | 3.5.2
PROJ version | 9.0.1
EPSG Registry database version | v10.064 (2022-05-19)
GEOS version | 3.11.0-CAPI-1.17.0
SQLite version | 3.40.0
PDAL version | 2.4.3
PostgreSQL client version | unknown
SpatiaLite version | 5.0.1
QWT version | 6.1.5
QScintilla2 version | 2.13.0
OS version | Fedora Linux 37 (Thirty Seven)
| | |
Active Python plugins
slyr_community | 4.0.6
QuickWKT | 3.1
project_export_inspire | 1.0.0
water_topology | 0.1
plugin_reloader | 0.9.3
db_manager | 0.1.20
processing | 2.12.99
sagaprovider | 2.12.99
grassprovider | 2.12.99
MetaSearch | 0.3.6
</body></html>QGIS version
3.28.2-Firenze
QGIS code revision
[d33b8750f3](https://github.com/qgis/QGIS/commit/d33b8750f3)
Qt version
5.15.8
Python version
3.11.1
GDAL/OGR version
3.5.2
PROJ version
9.0.1
EPSG Registry database version
v10.064 (2022-05-19)
GEOS version
3.11.0-CAPI-1.17.0
SQLite version
3.40.0
PDAL version
2.4.3
PostgreSQL client version
unknown
SpatiaLite version
5.0.1
QWT version
6.1.5
QScintilla2 version
2.13.0
OS version
Fedora Linux 37 (Thirty Seven)
Active Python plugins
slyr_community
4.0.6
QuickWKT
3.1
project_export_inspire
1.0.0
water_topology
0.1
plugin_reloader
0.9.3
db_manager
0.1.20
processing
2.12.99
sagaprovider
2.12.99
grassprovider
2.12.99
MetaSearch
0.3.6
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
_No response_
|
process
|
join by nearest regression on from when multiple geometry colums are present what is the bug or the crash having a nodes layer with multiple geometries i usually did on ltr the join by nearest algorithm to find out nodes that are closer than m with there s an error now feature could not be written to joined layer could not store attribute geometry could not convert value to target type string could not write feature into output execution failed after seconds steps to reproduce the issue add the following sql to your postgresql database create table public nodes id integer not null fk district integer fk pressurezone integer fk printmap integer fk precision integer fk precisionalti integer fk object reference integer altitude numeric default null numeric printmaps text geometry used boolean geometry used boolean pipe orientation double precision default pipe schema visible boolean default false geometry public geometry pointz not null geometry public geometry pointz geometry public geometry pointz insert into public nodes values null null null null false false true null null insert into public nodes values null null null null false false true null null alter table only public nodes add constraint node pkey primary key id create index node geoidx on public nodes using gist geometry create index node geoidx on public nodes using gist geometry create index node geoidx on public nodes using gist geometry run the join by nearest algo the error is shown feature could not be written to joined layer could not store attribute geometry could not convert value to target type string could not write feature into output execution failed after seconds versions doctype html public dtd html en p li white space pre wrap qgis version firenze qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version unknown spatialite version qwt version version os version fedora linux thirty seven active python plugins slyr community quickwkt project export inspire water topology plugin reloader db manager processing sagaprovider grassprovider metasearch qgis version firenze qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version unknown spatialite version qwt version version os version fedora linux thirty seven active python plugins slyr community quickwkt project export inspire water topology plugin reloader db manager processing sagaprovider grassprovider metasearch supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
| 1
|
500,057
| 14,485,386,014
|
IssuesEvent
|
2020-12-10 17:31:17
|
canonical-web-and-design/cookie-policy
|
https://api.github.com/repos/canonical-web-and-design/cookie-policy
|
closed
|
Disable cookie policy on legal page /legal/data-pricacy
|
Priority: High
|
We might want to remove any form of analytics on the data-privacy page since the cookie policy links to this page if I didn't accept yet the tracking.

|
1.0
|
Disable cookie policy on legal page /legal/data-pricacy - We might want to remove any form of analytics on the data-privacy page since the cookie policy links to this page if I didn't accept yet the tracking.

|
non_process
|
disable cookie policy on legal page legal data pricacy we might want to remove any form of analytics on the data privacy page since the cookie policy links to this page if i didn t accept yet the tracking
| 0
|
182,876
| 31,019,860,682
|
IssuesEvent
|
2023-08-10 03:49:41
|
WordPress/wporg-developer
|
https://api.github.com/repos/WordPress/wporg-developer
|
closed
|
PHP deprecation notice: get_page_by_title is deprecated since version 6.2.0
|
[Type] Bug Redesign
|
> Deprecated: Function get_page_by_title is deprecated since version 6.2.0! Use WP_Query instead.
Looks like we're using this function in a few places:
https://github.com/WordPress/wporg-developer/blob/3f5c6b8c68e88cfafac9fb5bee7b7a5d55466c16/source/wp-content/themes/wporg-developer-2023/inc/formatting.php#L415
https://github.com/WordPress/wporg-developer/blob/3f5c6b8c68e88cfafac9fb5bee7b7a5d55466c16/source/wp-content/themes/wporg-developer-2023/inc/formatting.php#L426
https://github.com/WordPress/wporg-developer/blob/3f5c6b8c68e88cfafac9fb5bee7b7a5d55466c16/source/wp-content/themes/wporg-developer-2023/inc/formatting.php#L468
https://github.com/WordPress/wporg-developer/blob/3f5c6b8c68e88cfafac9fb5bee7b7a5d55466c16/source/wp-content/themes/wporg-developer-2023/inc/template-tags.php#L879
https://github.com/WordPress/wporg-developer/blob/3f5c6b8c68e88cfafac9fb5bee7b7a5d55466c16/source/wp-content/themes/wporg-developer-2023/inc/redirects.php#L68
|
1.0
|
PHP deprecation notice: get_page_by_title is deprecated since version 6.2.0 - > Deprecated: Function get_page_by_title is deprecated since version 6.2.0! Use WP_Query instead.
Looks like we're using this function in a few places:
https://github.com/WordPress/wporg-developer/blob/3f5c6b8c68e88cfafac9fb5bee7b7a5d55466c16/source/wp-content/themes/wporg-developer-2023/inc/formatting.php#L415
https://github.com/WordPress/wporg-developer/blob/3f5c6b8c68e88cfafac9fb5bee7b7a5d55466c16/source/wp-content/themes/wporg-developer-2023/inc/formatting.php#L426
https://github.com/WordPress/wporg-developer/blob/3f5c6b8c68e88cfafac9fb5bee7b7a5d55466c16/source/wp-content/themes/wporg-developer-2023/inc/formatting.php#L468
https://github.com/WordPress/wporg-developer/blob/3f5c6b8c68e88cfafac9fb5bee7b7a5d55466c16/source/wp-content/themes/wporg-developer-2023/inc/template-tags.php#L879
https://github.com/WordPress/wporg-developer/blob/3f5c6b8c68e88cfafac9fb5bee7b7a5d55466c16/source/wp-content/themes/wporg-developer-2023/inc/redirects.php#L68
|
non_process
|
php deprecation notice get page by title is deprecated since version deprecated function get page by title is deprecated since version use wp query instead looks like we re using this function in a few places
| 0
|
34,754
| 7,859,882,451
|
IssuesEvent
|
2018-06-21 18:03:02
|
MatheusRich/SIGS-GCES
|
https://api.github.com/repos/MatheusRich/SIGS-GCES
|
closed
|
Refatorar features/
|
clean code
|
## Descrição do Problema
Há muita repetição de código nas features. O uso de `Backgrounds` ajudará nisto.
|
1.0
|
Refatorar features/ - ## Descrição do Problema
Há muita repetição de código nas features. O uso de `Backgrounds` ajudará nisto.
|
non_process
|
refatorar features descrição do problema há muita repetição de código nas features o uso de backgrounds ajudará nisto
| 0
|
62,720
| 17,173,450,595
|
IssuesEvent
|
2021-07-15 08:28:57
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
opened
|
Once I add the TO_TIMESTAMP_TZ time function to the SQL query, the LIMIT doesn't work any more.
|
Type: Defect
|
**Describe the bug**
Once I add the TO_TIMESTAMP_TZ time function to the SQL query, the LIMIT doesn't work any more.
This query returns 10 rows, as per the limit clause:
`select trades.symbol, trades.price, trades."timestamp", symbols.this from trades JOIN symbols ON symbols.__key = trades.symbol WHERE symbol = 'ELECU' AND price > 2500 ORDER BY "timestamp" DESC LIMIT 10;`
This query returns all rows, ignoring the limit:
`select trades.symbol, trades.price, TO_TIMESTAMP_TZ(trades."timestamp"), symbols.this from trades JOIN symbols ON symbols.__key = trades.symbol WHERE symbol = 'ELECU' AND price > 2500 ORDER BY "timestamp" DESC LIMIT 10;`
**Expected behavior**
LIMIT should kick in and reduce the output to 10 rows in both cases.
**To Reproduce**
1. Get the the [realtime-trade-monitor](https://github.com/hazelcast/hazelcast-demos/tree/master/realtime-trade-monitor)
2. Upgrade the Hazelcast dependency in the demo from 5.0-SNAPSHOT to 5.0-BETA-1
3. Add the hazelcast-sql dependency to the jet-server module
4. Follow the instructions to load symbols and ingest trades.
5. Start an SQL console (`hz-cli sql`).
6. Try both queries
**Additional context**
- I've used the hz-cli from 5.0-BETA-1 ZIP distribution
|
1.0
|
Once I add the TO_TIMESTAMP_TZ time function to the SQL query, the LIMIT doesn't work any more. - **Describe the bug**
Once I add the TO_TIMESTAMP_TZ time function to the SQL query, the LIMIT doesn't work any more.
This query returns 10 rows, as per the limit clause:
`select trades.symbol, trades.price, trades."timestamp", symbols.this from trades JOIN symbols ON symbols.__key = trades.symbol WHERE symbol = 'ELECU' AND price > 2500 ORDER BY "timestamp" DESC LIMIT 10;`
This query returns all rows, ignoring the limit:
`select trades.symbol, trades.price, TO_TIMESTAMP_TZ(trades."timestamp"), symbols.this from trades JOIN symbols ON symbols.__key = trades.symbol WHERE symbol = 'ELECU' AND price > 2500 ORDER BY "timestamp" DESC LIMIT 10;`
**Expected behavior**
LIMIT should kick in and reduce the output to 10 rows in both cases.
**To Reproduce**
1. Get the the [realtime-trade-monitor](https://github.com/hazelcast/hazelcast-demos/tree/master/realtime-trade-monitor)
2. Upgrade the Hazelcast dependency in the demo from 5.0-SNAPSHOT to 5.0-BETA-1
3. Add the hazelcast-sql dependency to the jet-server module
4. Follow the instructions to load symbols and ingest trades.
5. Start an SQL console (`hz-cli sql`).
6. Try both queries
**Additional context**
- I've used the hz-cli from 5.0-BETA-1 ZIP distribution
|
non_process
|
once i add the to timestamp tz time function to the sql query the limit doesn t work any more describe the bug once i add the to timestamp tz time function to the sql query the limit doesn t work any more this query returns rows as per the limit clause select trades symbol trades price trades timestamp symbols this from trades join symbols on symbols key trades symbol where symbol elecu and price order by timestamp desc limit this query returns all rows ignoring the limit select trades symbol trades price to timestamp tz trades timestamp symbols this from trades join symbols on symbols key trades symbol where symbol elecu and price order by timestamp desc limit expected behavior limit should kick in and reduce the output to rows in both cases to reproduce get the the upgrade the hazelcast dependency in the demo from snapshot to beta add the hazelcast sql dependency to the jet server module follow the instructions to load symbols and ingest trades start an sql console hz cli sql try both queries additional context i ve used the hz cli from beta zip distribution
| 0
|
21,798
| 30,311,554,883
|
IssuesEvent
|
2023-07-10 13:08:31
|
openfoodfacts/openfoodfacts-server
|
https://api.github.com/repos/openfoodfacts/openfoodfacts-server
|
closed
|
Do not require HTACCESS login/password on API requests for .net
|
API API auth needs reprocessing htaccess
|
### Task
- Do not require HTACCESS login/password on API requests for .net as it forces developers to tweak behaviors between prod and staging, bringing challenges for debugging, CI, implementation…
### Part of
- #5523
cc @raphael0202
|
1.0
|
Do not require HTACCESS login/password on API requests for .net - ### Task
- Do not require HTACCESS login/password on API requests for .net as it forces developers to tweak behaviors between prod and staging, bringing challenges for debugging, CI, implementation…
### Part of
- #5523
cc @raphael0202
|
process
|
do not require htaccess login password on api requests for net task do not require htaccess login password on api requests for net as it forces developers to tweak behaviors between prod and staging bringing challenges for debugging ci implementation… part of cc
| 1
|
242,624
| 26,277,756,973
|
IssuesEvent
|
2023-01-07 01:06:46
|
AlexRogalskiy/AlexRogalskiy
|
https://api.github.com/repos/AlexRogalskiy/AlexRogalskiy
|
opened
|
CVE-2021-32810 (High) detected in crossbeam-deque-0.7.3.crate
|
security vulnerability
|
## CVE-2021-32810 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>crossbeam-deque-0.7.3.crate</b></p></summary>
<p>Concurrent work-stealing deque</p>
<p>Library home page: <a href="https://crates.io/api/v1/crates/crossbeam-deque/0.7.3/download">https://crates.io/api/v1/crates/crossbeam-deque/0.7.3/download</a></p>
<p>
Dependency Hierarchy:
- rss-1.9.0.crate (Root Library)
- reqwest-0.9.24.crate
- tokio-threadpool-0.1.18.crate
- :x: **crossbeam-deque-0.7.3.crate** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
crossbeam-deque is a package of work-stealing deques for building task schedulers when programming in Rust. In versions prior to 0.7.4 and 0.8.0, the result of the race condition is that one or more tasks in the worker queue can be popped twice instead of other tasks that are forgotten and never popped. If tasks are allocated on the heap, this can cause double free and a memory leak. If not, this still can cause a logical bug. Crates using `Stealer::steal`, `Stealer::steal_batch`, or `Stealer::steal_batch_and_pop` are affected by this issue. This has been fixed in crossbeam-deque 0.8.1 and 0.7.4.
<p>Publish Date: 2021-08-02
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-32810>CVE-2021-32810</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/crossbeam-rs/crossbeam/security/advisories/GHSA-pqqp-xmhj-wgcw">https://github.com/crossbeam-rs/crossbeam/security/advisories/GHSA-pqqp-xmhj-wgcw</a></p>
<p>Release Date: 2021-08-02</p>
<p>Fix Resolution: crossbeam-deque - 0.7.4, 0.8.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-32810 (High) detected in crossbeam-deque-0.7.3.crate - ## CVE-2021-32810 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>crossbeam-deque-0.7.3.crate</b></p></summary>
<p>Concurrent work-stealing deque</p>
<p>Library home page: <a href="https://crates.io/api/v1/crates/crossbeam-deque/0.7.3/download">https://crates.io/api/v1/crates/crossbeam-deque/0.7.3/download</a></p>
<p>
Dependency Hierarchy:
- rss-1.9.0.crate (Root Library)
- reqwest-0.9.24.crate
- tokio-threadpool-0.1.18.crate
- :x: **crossbeam-deque-0.7.3.crate** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
crossbeam-deque is a package of work-stealing deques for building task schedulers when programming in Rust. In versions prior to 0.7.4 and 0.8.0, the result of the race condition is that one or more tasks in the worker queue can be popped twice instead of other tasks that are forgotten and never popped. If tasks are allocated on the heap, this can cause double free and a memory leak. If not, this still can cause a logical bug. Crates using `Stealer::steal`, `Stealer::steal_batch`, or `Stealer::steal_batch_and_pop` are affected by this issue. This has been fixed in crossbeam-deque 0.8.1 and 0.7.4.
<p>Publish Date: 2021-08-02
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-32810>CVE-2021-32810</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/crossbeam-rs/crossbeam/security/advisories/GHSA-pqqp-xmhj-wgcw">https://github.com/crossbeam-rs/crossbeam/security/advisories/GHSA-pqqp-xmhj-wgcw</a></p>
<p>Release Date: 2021-08-02</p>
<p>Fix Resolution: crossbeam-deque - 0.7.4, 0.8.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in crossbeam deque crate cve high severity vulnerability vulnerable library crossbeam deque crate concurrent work stealing deque library home page a href dependency hierarchy rss crate root library reqwest crate tokio threadpool crate x crossbeam deque crate vulnerable library found in base branch master vulnerability details crossbeam deque is a package of work stealing deques for building task schedulers when programming in rust in versions prior to and the result of the race condition is that one or more tasks in the worker queue can be popped twice instead of other tasks that are forgotten and never popped if tasks are allocated on the heap this can cause double free and a memory leak if not this still can cause a logical bug crates using stealer steal stealer steal batch or stealer steal batch and pop are affected by this issue this has been fixed in crossbeam deque and publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution crossbeam deque step up your open source security game with mend
| 0
|
13,869
| 16,627,145,432
|
IssuesEvent
|
2021-06-03 11:03:59
|
arcus-azure/arcus.messaging
|
https://api.github.com/repos/arcus-azure/arcus.messaging
|
closed
|
Remove the 'old' Azure Service Bus message pump workings once the Arcus.BackgroundJobs package is updated
|
area:message-processing dependencies enhancement integration:service-bus
|
**Is your feature request related to a problem? Please describe.**
Since we rely in the Arcus.Messaging.ServiceBus.KeyRotation project on the Arcus.BackgroundJobs.* package (that uses again our messaging package), we have 'method not found's exceptions when we break the contract of the Azure Service Bus message pump.
**Describe the solution you'd like**
Once the Arcus.BackgroundJobs.* package is updated with our new messaging package, we can remove the 'old' workings of the message pump.
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Additional context**
PR with backwards compatible Azure Service Bus message pump #175
|
1.0
|
Remove the 'old' Azure Service Bus message pump workings once the Arcus.BackgroundJobs package is updated - **Is your feature request related to a problem? Please describe.**
Since we rely in the Arcus.Messaging.ServiceBus.KeyRotation project on the Arcus.BackgroundJobs.* package (that uses again our messaging package), we have 'method not found's exceptions when we break the contract of the Azure Service Bus message pump.
**Describe the solution you'd like**
Once the Arcus.BackgroundJobs.* package is updated with our new messaging package, we can remove the 'old' workings of the message pump.
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Additional context**
PR with backwards compatible Azure Service Bus message pump #175
|
process
|
remove the old azure service bus message pump workings once the arcus backgroundjobs package is updated is your feature request related to a problem please describe since we rely in the arcus messaging servicebus keyrotation project on the arcus backgroundjobs package that uses again our messaging package we have method not found s exceptions when we break the contract of the azure service bus message pump describe the solution you d like once the arcus backgroundjobs package is updated with our new messaging package we can remove the old workings of the message pump describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context pr with backwards compatible azure service bus message pump
| 1
|
15,719
| 19,862,086,691
|
IssuesEvent
|
2022-01-22 01:54:57
|
ooi-data/CE07SHSP-SP001-04-OPTAAJ000-recovered_cspp-optaa_dj_cspp_instrument_recovered
|
https://api.github.com/repos/ooi-data/CE07SHSP-SP001-04-OPTAAJ000-recovered_cspp-optaa_dj_cspp_instrument_recovered
|
opened
|
🛑 Processing failed: ValueError
|
process
|
## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-22T01:54:56.535140.
## Details
Flow name: `CE07SHSP-SP001-04-OPTAAJ000-recovered_cspp-optaa_dj_cspp_instrument_recovered`
Task name: `processing_task`
Error type: `ValueError`
Error message: cannot reindex or align along dimension 'wavelength' because the index has duplicate values
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 157, in processing
process_dataset(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 147, in process_dataset
append_to_zarr(mod_ds, store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 355, in append_to_zarr
mod_ds = mod_ds.reindex(dim_indexer)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/dataset.py", line 2943, in reindex
return self._reindex(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/dataset.py", line 2972, in _reindex
variables, indexes = alignment.reindex_variables(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/alignment.py", line 584, in reindex_variables
raise ValueError(
ValueError: cannot reindex or align along dimension 'wavelength' because the index has duplicate values
```
</details>
|
1.0
|
🛑 Processing failed: ValueError - ## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-22T01:54:56.535140.
## Details
Flow name: `CE07SHSP-SP001-04-OPTAAJ000-recovered_cspp-optaa_dj_cspp_instrument_recovered`
Task name: `processing_task`
Error type: `ValueError`
Error message: cannot reindex or align along dimension 'wavelength' because the index has duplicate values
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 157, in processing
process_dataset(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 147, in process_dataset
append_to_zarr(mod_ds, store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 355, in append_to_zarr
mod_ds = mod_ds.reindex(dim_indexer)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/dataset.py", line 2943, in reindex
return self._reindex(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/dataset.py", line 2972, in _reindex
variables, indexes = alignment.reindex_variables(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/alignment.py", line 584, in reindex_variables
raise ValueError(
ValueError: cannot reindex or align along dimension 'wavelength' because the index has duplicate values
```
</details>
|
process
|
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered cspp optaa dj cspp instrument recovered task name processing task error type valueerror error message cannot reindex or align along dimension wavelength because the index has duplicate values traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing process dataset file srv conda envs notebook lib site packages ooi harvester processor init py line in process dataset append to zarr mod ds store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr mod ds mod ds reindex dim indexer file srv conda envs notebook lib site packages xarray core dataset py line in reindex return self reindex file srv conda envs notebook lib site packages xarray core dataset py line in reindex variables indexes alignment reindex variables file srv conda envs notebook lib site packages xarray core alignment py line in reindex variables raise valueerror valueerror cannot reindex or align along dimension wavelength because the index has duplicate values
| 1
|
20,864
| 27,645,588,455
|
IssuesEvent
|
2023-03-10 22:35:14
|
cse442-at-ub/project_s23-iweatherify
|
https://api.github.com/repos/cse442-at-ub/project_s23-iweatherify
|
closed
|
Convert HTML/CSS Homepage into a Vue.js Homepage for Desktop and Mobile
|
Processing Task Sprint 2
|
**Tests**
*Test 1*
1) Go to https://github.com/cse442-at-ub/project_s23-iweatherify/tree/Bugfix_Static_Homepage_To_Vue
2) Click on <> Code and download ZIP
3) Unzip the code
4) Enter the folder via terminal and run `npm install` (Or via IDE)
**Make sure you have npm installed or just run a localhost version on cheshire**
5) Run the command `npm start` (Or via IDE)
6) Open via localhost, which in my case it's (http://localhost:1234), but it should be provided by the output from `npm start`
*Test 2*
1) Do Test 1
2) Visit the path https://[local host here]/#/homepage-logged-in
3) Confirm that there's a logged-in version of the site!
*Test 3*
1) Do Test 1
2) Visit the homepage
3) Go to inspect elements
4) Click on Toggle Device Toolbar
5) Check Responsiveness on a variety of devices
*Test 4*
1) Do Test 1
2) Visit the path https://[local host here]/#/homepage-logged-in
3) Go to inspect elements
4) Click on Toggle Device Toolbar
5) Check Responsiveness on a variety of devices
|
1.0
|
Convert HTML/CSS Homepage into a Vue.js Homepage for Desktop and Mobile - **Tests**
*Test 1*
1) Go to https://github.com/cse442-at-ub/project_s23-iweatherify/tree/Bugfix_Static_Homepage_To_Vue
2) Click on <> Code and download ZIP
3) Unzip the code
4) Enter the folder via terminal and run `npm install` (Or via IDE)
**Make sure you have npm installed or just run a localhost version on cheshire**
5) Run the command `npm start` (Or via IDE)
6) Open via localhost, which in my case it's (http://localhost:1234), but it should be provided by the output from `npm start`
*Test 2*
1) Do Test 1
2) Visit the path https://[local host here]/#/homepage-logged-in
3) Confirm that there's a logged-in version of the site!
*Test 3*
1) Do Test 1
2) Visit the homepage
3) Go to inspect elements
4) Click on Toggle Device Toolbar
5) Check Responsiveness on a variety of devices
*Test 4*
1) Do Test 1
2) Visit the path https://[local host here]/#/homepage-logged-in
3) Go to inspect elements
4) Click on Toggle Device Toolbar
5) Check Responsiveness on a variety of devices
|
process
|
convert html css homepage into a vue js homepage for desktop and mobile tests test go to click on code and download zip unzip the code enter the folder via terminal and run npm install or via ide make sure you have npm installed or just run a localhost version on cheshire run the command npm start or via ide open via localhost which in my case it s but it should be provided by the output from npm start test do test visit the path https homepage logged in confirm that there s a logged in version of the site test do test visit the homepage go to inspect elements click on toggle device toolbar check responsiveness on a variety of devices test do test visit the path https homepage logged in go to inspect elements click on toggle device toolbar check responsiveness on a variety of devices
| 1
|
11,749
| 14,583,463,217
|
IssuesEvent
|
2020-12-18 14:02:19
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
Adjust "no records found" text to have padding
|
Bug P2 Participant manager Process: Tested QA Process: Tested dev UI
|
Add space between the "no records found" text and the top of the panel.

|
2.0
|
Adjust "no records found" text to have padding - Add space between the "no records found" text and the top of the panel.

|
process
|
adjust no records found text to have padding add space between the no records found text and the top of the panel
| 1
|
5,292
| 3,917,125,976
|
IssuesEvent
|
2016-04-21 06:43:56
|
kolliSuman/issues
|
https://api.github.com/repos/kolliSuman/issues
|
closed
|
QA_Decoder with 7-Segment Display _Experiment_p1
|
Category: Usability Developed By: VLEAD Release Number: Production Severity: S2 Status: Open
|
Defect Description :
In the simulation page of "Decoder with 7-Segment Display " experiment, the input values are not clearly visible instead the input values should be clearly visible to the user
Actual Result :
In the simulation page of "Decoder with 7-Segment Display " experiment, the input values are not clearly visible
Environment :
OS: Windows 7, Linux
Browsers: Firefox,Chrome
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM ,
Processor:i5
Test Step Link:
https://github.com/Virtual-Labs/digital-logic-design-iiith/blob/master/test-cases/integration_test-cases/Decoder%20with%207-Segment%20Display%20/Decoder%20with%207-Segment%20Display%20_06_Experiment_p1.org
|
True
|
QA_Decoder with 7-Segment Display _Experiment_p1 -
Defect Description :
In the simulation page of "Decoder with 7-Segment Display " experiment, the input values are not clearly visible instead the input values should be clearly visible to the user
Actual Result :
In the simulation page of "Decoder with 7-Segment Display " experiment, the input values are not clearly visible
Environment :
OS: Windows 7, Linux
Browsers: Firefox,Chrome
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM ,
Processor:i5
Test Step Link:
https://github.com/Virtual-Labs/digital-logic-design-iiith/blob/master/test-cases/integration_test-cases/Decoder%20with%207-Segment%20Display%20/Decoder%20with%207-Segment%20Display%20_06_Experiment_p1.org
|
non_process
|
qa decoder with segment display experiment defect description in the simulation page of decoder with segment display experiment the input values are not clearly visible instead the input values should be clearly visible to the user actual result in the simulation page of decoder with segment display experiment the input values are not clearly visible environment os windows linux browsers firefox chrome bandwidth hardware configuration processor test step link
| 0
|
91,618
| 10,724,061,678
|
IssuesEvent
|
2019-10-27 23:06:36
|
letscontrolit/ESPEasy
|
https://api.github.com/repos/letscontrolit/ESPEasy
|
closed
|
Split command handling + allow to generate list of commands
|
Type: Documentation Type: Enhancement
|
In the command.ino file, there is a long-long-loooooong list of if-the-else statements to filter the commands.
There should be a single function to parse all command names and convert them to an enum.
And the other way around (can be done using a map or something like that) to generate a list of commands as help function on the webpage and via serial terminal.
Then handling the commands is just a single switch statement, which is a lot easier to maintain, to read, etc. Also may benefit memory use, since the lookup could be stored almost entirely in flash.
Another benefit is that there is someone looking through these commands and may optimize some parts, check documentation, etc.
|
1.0
|
Split command handling + allow to generate list of commands - In the command.ino file, there is a long-long-loooooong list of if-the-else statements to filter the commands.
There should be a single function to parse all command names and convert them to an enum.
And the other way around (can be done using a map or something like that) to generate a list of commands as help function on the webpage and via serial terminal.
Then handling the commands is just a single switch statement, which is a lot easier to maintain, to read, etc. Also may benefit memory use, since the lookup could be stored almost entirely in flash.
Another benefit is that there is someone looking through these commands and may optimize some parts, check documentation, etc.
|
non_process
|
split command handling allow to generate list of commands in the command ino file there is a long long loooooong list of if the else statements to filter the commands there should be a single function to parse all command names and convert them to an enum and the other way around can be done using a map or something like that to generate a list of commands as help function on the webpage and via serial terminal then handling the commands is just a single switch statement which is a lot easier to maintain to read etc also may benefit memory use since the lookup could be stored almost entirely in flash another benefit is that there is someone looking through these commands and may optimize some parts check documentation etc
| 0
|
13,306
| 15,780,587,809
|
IssuesEvent
|
2021-04-01 10:08:41
|
unicode-org/icu4x
|
https://api.github.com/repos/unicode-org/icu4x
|
opened
|
Higher-fidelity triaging
|
C-process discuss
|
Now that we have over 100 open issues, I want to discuss ways to help organize these issues better.
We currently have the component and type labels. I think component is very useful; we could probably clean up the types a bit.
The main thing is that I think we should add one more axis of required labels. I think we should add a "time required" or "size" label with several options:
- `S-tiny` = doable in less than an hour (updating API docs, trivial code fix)
- `S-small` = doable in an afternoon (small bug fix or tech debt)
- `S-medium` = doable in 1-3 days (smaller new feature, larger bug fix or tech debt)
- `S-large` = doable in 1-2 weeks (larger new feature, major refactoring effort)
- `S-epic` = tracking issue for major project (should have child issues for smaller pieces of work)
With this new label, we should endeavor to keep the number of `S-tiny` and `S-small` issues to a minimum, and it will also help both us and external people gauge at a glance how much time an issue will take to complete.
We could also consider a priority label, corresponding to the commonly used P0 through P4. I don't know if we're quite yet at the point where we need this. Everything right now is either equally high priority or backlog. Once we reach 1.0, we may want to consider a priority label.
|
1.0
|
Higher-fidelity triaging - Now that we have over 100 open issues, I want to discuss ways to help organize these issues better.
We currently have the component and type labels. I think component is very useful; we could probably clean up the types a bit.
The main thing is that I think we should add one more axis of required labels. I think we should add a "time required" or "size" label with several options:
- `S-tiny` = doable in less than an hour (updating API docs, trivial code fix)
- `S-small` = doable in an afternoon (small bug fix or tech debt)
- `S-medium` = doable in 1-3 days (smaller new feature, larger bug fix or tech debt)
- `S-large` = doable in 1-2 weeks (larger new feature, major refactoring effort)
- `S-epic` = tracking issue for major project (should have child issues for smaller pieces of work)
With this new label, we should endeavor to keep the number of `S-tiny` and `S-small` issues to a minimum, and it will also help both us and external people gauge at a glance how much time an issue will take to complete.
We could also consider a priority label, corresponding to the commonly used P0 through P4. I don't know if we're quite yet at the point where we need this. Everything right now is either equally high priority or backlog. Once we reach 1.0, we may want to consider a priority label.
|
process
|
higher fidelity triaging now that we have over open issues i want to discuss ways to help organize these issues better we currently have the component and type labels i think component is very useful we could probably clean up the types a bit the main thing is that i think we should add one more axis of required labels i think we should add a time required or size label with several options s tiny doable in less than an hour updating api docs trivial code fix s small doable in an afternoon small bug fix or tech debt s medium doable in days smaller new feature larger bug fix or tech debt s large doable in weeks larger new feature major refactoring effort s epic tracking issue for major project should have child issues for smaller pieces of work with this new label we should endeavor to keep the number of s tiny and s small issues to a minimum and it will also help both us and external people gauge at a glance how much time an issue will take to complete we could also consider a priority label corresponding to the commonly used through i don t know if we re quite yet at the point where we need this everything right now is either equally high priority or backlog once we reach we may want to consider a priority label
| 1
|
13,385
| 15,862,968,890
|
IssuesEvent
|
2021-04-08 12:16:47
|
didi/mpx
|
https://api.github.com/repos/didi/mpx
|
closed
|
[Bug report]scrollview的 binddragend 事件无效
|
processing
|
**问题描述**
1. 在wx小程序工具 不能触发
2. 真机预览 可以触发
3. 转成web后 不能触发
**期望**
web中可以触发binddragend事件
**环境信息描述**
1. mac
3. "@mpxjs/core": "^2.6.51",
4. "@mpxjs/webpack-plugin": "^2.6.52",
3. 小程序开发者工具信息(小程序平台 wx、开发者工具版本1.05.2102010,基础库版本 2.16.0)
|
1.0
|
[Bug report]scrollview的 binddragend 事件无效 - **问题描述**
1. 在wx小程序工具 不能触发
2. 真机预览 可以触发
3. 转成web后 不能触发
**期望**
web中可以触发binddragend事件
**环境信息描述**
1. mac
3. "@mpxjs/core": "^2.6.51",
4. "@mpxjs/webpack-plugin": "^2.6.52",
3. 小程序开发者工具信息(小程序平台 wx、开发者工具版本1.05.2102010,基础库版本 2.16.0)
|
process
|
scrollview的 binddragend 事件无效 问题描述 在wx小程序工具 不能触发 真机预览 可以触发 转成web后 不能触发 期望 web中可以触发binddragend事件 环境信息描述 mac mpxjs core mpxjs webpack plugin 小程序开发者工具信息 小程序平台 wx、 基础库版本 )
| 1
|
151,619
| 5,824,798,432
|
IssuesEvent
|
2017-05-07 16:04:36
|
jasonwebb/tc-maker-4x4-router
|
https://api.github.com/repos/jasonwebb/tc-maker-4x4-router
|
closed
|
Initiate discussion of relocation project
|
High Priority Relocation to Scary Bathroom
|
Jon Alt (wood shop manager) has requested that we move (or develop a plan to move) the CNC machine into the adjacent "Scary Bathroom" space.
This would have clear benefits to both the wood shop and to the CNC machine, specifically:
1. Wood shop gaining much-needed floor space to better serve its members.
2. CNC machine being better isolated from dust and noise.
3. CNC machine having a dedicated and clearly-defined workspace so that members can keep relevant materials and tools on-hand as well as stay focused on their projects without disruption (or causing disruption) to wood shop.
Some of the challenges of moving the machine include:
1. Readying the bathroom to be capable of housing the CNC machine. Dust collection, power (120V and 240V), and evaluating practical space requirements for working with materials.
2. Physically moving the machine into the room, possibly requiring disassembly, reassembly, and recalibration.. This is potentially the most crucial topic, requiring strong coordination and planning for how to get done.
Given the complexity of the machine and the delicate nature of its calibration, we ought to figure this out before continuing on any further work, especially involving mechanical tuning and tweaking.
Jon and Pete need to discuss options and plans, and Jason will be involved as desired to facilitate discussion and coordinate work.
Given that this discussion involves the re-appropriation of shop space (the bathroom in particular), I feel it would be best to be as transparent as possible with the membership base and the board. Before plans are acted upon, they should perhaps be approved by the membership and board (though maybe that doesn't need to be done?).
|
1.0
|
Initiate discussion of relocation project - Jon Alt (wood shop manager) has requested that we move (or develop a plan to move) the CNC machine into the adjacent "Scary Bathroom" space.
This would have clear benefits to both the wood shop and to the CNC machine, specifically:
1. Wood shop gaining much-needed floor space to better serve its members.
2. CNC machine being better isolated from dust and noise.
3. CNC machine having a dedicated and clearly-defined workspace so that members can keep relevant materials and tools on-hand as well as stay focused on their projects without disruption (or causing disruption) to wood shop.
Some of the challenges of moving the machine include:
1. Readying the bathroom to be capable of housing the CNC machine. Dust collection, power (120V and 240V), and evaluating practical space requirements for working with materials.
2. Physically moving the machine into the room, possibly requiring disassembly, reassembly, and recalibration.. This is potentially the most crucial topic, requiring strong coordination and planning for how to get done.
Given the complexity of the machine and the delicate nature of its calibration, we ought to figure this out before continuing on any further work, especially involving mechanical tuning and tweaking.
Jon and Pete need to discuss options and plans, and Jason will be involved as desired to facilitate discussion and coordinate work.
Given that this discussion involves the re-appropriation of shop space (the bathroom in particular), I feel it would be best to be as transparent as possible with the membership base and the board. Before plans are acted upon, they should perhaps be approved by the membership and board (though maybe that doesn't need to be done?).
|
non_process
|
initiate discussion of relocation project jon alt wood shop manager has requested that we move or develop a plan to move the cnc machine into the adjacent scary bathroom space this would have clear benefits to both the wood shop and to the cnc machine specifically wood shop gaining much needed floor space to better serve its members cnc machine being better isolated from dust and noise cnc machine having a dedicated and clearly defined workspace so that members can keep relevant materials and tools on hand as well as stay focused on their projects without disruption or causing disruption to wood shop some of the challenges of moving the machine include readying the bathroom to be capable of housing the cnc machine dust collection power and and evaluating practical space requirements for working with materials physically moving the machine into the room possibly requiring disassembly reassembly and recalibration this is potentially the most crucial topic requiring strong coordination and planning for how to get done given the complexity of the machine and the delicate nature of its calibration we ought to figure this out before continuing on any further work especially involving mechanical tuning and tweaking jon and pete need to discuss options and plans and jason will be involved as desired to facilitate discussion and coordinate work given that this discussion involves the re appropriation of shop space the bathroom in particular i feel it would be best to be as transparent as possible with the membership base and the board before plans are acted upon they should perhaps be approved by the membership and board though maybe that doesn t need to be done
| 0
|
55,463
| 7,988,361,740
|
IssuesEvent
|
2018-07-19 10:46:48
|
debops/debops
|
https://api.github.com/repos/debops/debops
|
closed
|
Include `../../../includes/global.rst` in all `defaults/main.yml` role files
|
Documentation enhancement
|
Status: Waiting
Waiting for: No PR open against debops/debops which touch `defaults/main.yml`
Currently anonymous links are being used for new docs. This might be quicker in the short run but is more difficult to maintain and can have inconsistencies, especially with different authors. Include `../../../includes/global.rst` in all `defaults/main.yml` so that all links defined there can be used without trouble. This change can be done with yaml4rst as called by `debops-optimize`. `debops-optimize` will need a minor change to pass `../../../includes/global.rst` as config option to yaml4rst or I will make it the default in yaml4rst. I would propose to wait shortly before the release or at least for a time when there are no PRs open which change a `defaults/main.yml` file. `debops-optimize` can do the change automatically so all manual PRs are to be preferred when merging because yaml4rst will probably create a big changeset.
Related to: #316
|
1.0
|
Include `../../../includes/global.rst` in all `defaults/main.yml` role files - Status: Waiting
Waiting for: No PR open against debops/debops which touch `defaults/main.yml`
Currently anonymous links are being used for new docs. This might be quicker in the short run but is more difficult to maintain and can have inconsistencies, especially with different authors. Include `../../../includes/global.rst` in all `defaults/main.yml` so that all links defined there can be used without trouble. This change can be done with yaml4rst as called by `debops-optimize`. `debops-optimize` will need a minor change to pass `../../../includes/global.rst` as config option to yaml4rst or I will make it the default in yaml4rst. I would propose to wait shortly before the release or at least for a time when there are no PRs open which change a `defaults/main.yml` file. `debops-optimize` can do the change automatically so all manual PRs are to be preferred when merging because yaml4rst will probably create a big changeset.
Related to: #316
|
non_process
|
include includes global rst in all defaults main yml role files status waiting waiting for no pr open against debops debops which touch defaults main yml currently anonymous links are being used for new docs this might be quicker in the short run but is more difficult to maintain and can have inconsistencies especially with different authors include includes global rst in all defaults main yml so that all links defined there can be used without trouble this change can be done with as called by debops optimize debops optimize will need a minor change to pass includes global rst as config option to or i will make it the default in i would propose to wait shortly before the release or at least for a time when there are no prs open which change a defaults main yml file debops optimize can do the change automatically so all manual prs are to be preferred when merging because will probably create a big changeset related to
| 0
|
15,483
| 19,693,135,301
|
IssuesEvent
|
2022-01-12 09:23:03
|
bisq-network/bisq
|
https://api.github.com/repos/bisq-network/bisq
|
closed
|
Improve user experience once mediation has been accepted by both parties
|
in:gui a:feature in:trade-process
|
<!--
SUPPORT REQUESTS: This is for reporting bugs in the Bisq app.
If you have a support request, please join #support on Bisq's
Keybase team over at https://keybase.io/team/Bisq
-->
### Description
When mediation for a given trade is accepted by both parties the trade moves from 'open trades' to 'history' and the payout transaction ID is broadcast.
This is how it should be but the user experience could be achieved by keeping the trade in 'open trades' for longer and give the user the option to withdraw funds to Bisq wallet or external wallet. This would add an additional step that is the same as what happens currently when trades in 'open trades' are successfully completed ie, the user is notified and then they make the choice to keep funds in Bisq wallet or withdraw them. This would improve the user experience consistency.
#### Version
v 1.6.5
### Steps to reproduce
**From the perspective of 'Trader 1 (initial trader to accept mediation)'**
A trade goes to mediation.
1. Trader 1 accepts mediation proposal.
2. Trader 2 accepts mediation proposal.
Trader 1 is not informed mediation has been accepted, there is no notification. Next time they log in the trade will have moved from open trades to history and the broadcast payout tx id will hopefully have been paid out. Trader 1 will need to check funds > transactions to see the status of the trade.
**From the perspective of 'Trader 2 (second trader to accept mediation)'**
A trade goes to mediation.
1. Trader 1 accepts mediation proposal.
2. Trader 2 accepts mediation proposal.
Trader 2 is not informed mediation has been accepted, there is no notification. The trade will have moved from open trades to history and the payout tx id will be broadcast.
### Expected behaviour
**From the perspective of 'Trader 1 (initial trader to accept mediation)'**
A trade goes to mediation.
1. Trader 1 accepts mediation proposal.
2. Trader 2 accepts mediation proposal.
Trader 1 is not informed mediation has been accepted, there is a notification. The mediated trade will still be in open trades and the Trader 1 will have the option to keep the funds in their Bisq wallet or withdraw to an external wallet. The trade will then move to history.
**From the perspective of 'Trader 2 (second trader to accept mediation)'**
A trade goes to mediation.
1. Trader 1 accepts mediation proposal.
2. Trader 2 accepts mediation proposal.
Trader 2 is informed mediation has been accepted, by a notification. The trade will still be in open trades and Trader 2 will have the option to keep the funds in their Bisq wallet or withdraw to an external wallet. The trade will then move to history.
|
1.0
|
Improve user experience once mediation has been accepted by both parties - <!--
SUPPORT REQUESTS: This is for reporting bugs in the Bisq app.
If you have a support request, please join #support on Bisq's
Keybase team over at https://keybase.io/team/Bisq
-->
### Description
When mediation for a given trade is accepted by both parties the trade moves from 'open trades' to 'history' and the payout transaction ID is broadcast.
This is how it should be but the user experience could be achieved by keeping the trade in 'open trades' for longer and give the user the option to withdraw funds to Bisq wallet or external wallet. This would add an additional step that is the same as what happens currently when trades in 'open trades' are successfully completed ie, the user is notified and then they make the choice to keep funds in Bisq wallet or withdraw them. This would improve the user experience consistency.
#### Version
v 1.6.5
### Steps to reproduce
**From the perspective of 'Trader 1 (initial trader to accept mediation)'**
A trade goes to mediation.
1. Trader 1 accepts mediation proposal.
2. Trader 2 accepts mediation proposal.
Trader 1 is not informed mediation has been accepted, there is no notification. Next time they log in the trade will have moved from open trades to history and the broadcast payout tx id will hopefully have been paid out. Trader 1 will need to check funds > transactions to see the status of the trade.
**From the perspective of 'Trader 2 (second trader to accept mediation)'**
A trade goes to mediation.
1. Trader 1 accepts mediation proposal.
2. Trader 2 accepts mediation proposal.
Trader 2 is not informed mediation has been accepted, there is no notification. The trade will have moved from open trades to history and the payout tx id will be broadcast.
### Expected behaviour
**From the perspective of 'Trader 1 (initial trader to accept mediation)'**
A trade goes to mediation.
1. Trader 1 accepts mediation proposal.
2. Trader 2 accepts mediation proposal.
Trader 1 is not informed mediation has been accepted, there is a notification. The mediated trade will still be in open trades and the Trader 1 will have the option to keep the funds in their Bisq wallet or withdraw to an external wallet. The trade will then move to history.
**From the perspective of 'Trader 2 (second trader to accept mediation)'**
A trade goes to mediation.
1. Trader 1 accepts mediation proposal.
2. Trader 2 accepts mediation proposal.
Trader 2 is informed mediation has been accepted, by a notification. The trade will still be in open trades and Trader 2 will have the option to keep the funds in their Bisq wallet or withdraw to an external wallet. The trade will then move to history.
|
process
|
improve user experience once mediation has been accepted by both parties support requests this is for reporting bugs in the bisq app if you have a support request please join support on bisq s keybase team over at description when mediation for a given trade is accepted by both parties the trade moves from open trades to history and the payout transaction id is broadcast this is how it should be but the user experience could be achieved by keeping the trade in open trades for longer and give the user the option to withdraw funds to bisq wallet or external wallet this would add an additional step that is the same as what happens currently when trades in open trades are successfully completed ie the user is notified and then they make the choice to keep funds in bisq wallet or withdraw them this would improve the user experience consistency version v steps to reproduce from the perspective of trader initial trader to accept mediation a trade goes to mediation trader accepts mediation proposal trader accepts mediation proposal trader is not informed mediation has been accepted there is no notification next time they log in the trade will have moved from open trades to history and the broadcast payout tx id will hopefully have been paid out trader will need to check funds transactions to see the status of the trade from the perspective of trader second trader to accept mediation a trade goes to mediation trader accepts mediation proposal trader accepts mediation proposal trader is not informed mediation has been accepted there is no notification the trade will have moved from open trades to history and the payout tx id will be broadcast expected behaviour from the perspective of trader initial trader to accept mediation a trade goes to mediation trader accepts mediation proposal trader accepts mediation proposal trader is not informed mediation has been accepted there is a notification the mediated trade will still be in open trades and the trader will have the option to keep the funds in their bisq wallet or withdraw to an external wallet the trade will then move to history from the perspective of trader second trader to accept mediation a trade goes to mediation trader accepts mediation proposal trader accepts mediation proposal trader is informed mediation has been accepted by a notification the trade will still be in open trades and trader will have the option to keep the funds in their bisq wallet or withdraw to an external wallet the trade will then move to history
| 1
|
10,449
| 8,951,939,313
|
IssuesEvent
|
2019-01-25 15:17:28
|
terraform-providers/terraform-provider-aws
|
https://api.github.com/repos/terraform-providers/terraform-provider-aws
|
closed
|
aws_rds_cluster_parameter_group does not generate a unique name when specifying name_prefix
|
bug service/rds
|
### Terraform Version
```
$ terraform --version
Terraform v0.9.11
Your version of Terraform is out of date! The latest version
is 0.10.6. You can update by downloading from www.terraform.io
```
### Affected Resource(s)
Please list the resources as a list, for example:
- aws_rds_cluster_parameter_group
### Terraform Configuration Files
```hcl
####
# RDS Cluster Paramater Group
resource "aws_rds_cluster_parameter_group" "fa-arora" {
name_prefix = "${var.env}-aurora-params"
family = "aurora5.6"
description = "RDS ${var.env} cluster parameter group"
parameter {
name = "character_set_server"
value = "utf8"
}
parameter {
name = "character_set_client"
value = "utf8"
}
}
# End of RDS Cluster Paramater Group
####
```
### Debug Output
https://gist.github.com/majormoses/f373836f0219ae37a0079f3938868ce9
### Expected Behavior
Terraform should have computed a unique name (using the prefix) and submitted it with the request.
### Actual Behavior
It appears that the plan believes it should compute the value but either does and does not submit it or does not compute it. This results in a 400 because the body did not match the requirements:
```
* aws_rds_cluster_parameter_group.fa-arora: Error modifying DB Cluster Parameter Group: InvalidParameterValue: The parameter DBClusterParameterGroupName must be provided and must not be blank.
status code: 400, request id: 40a2e32e-a094-11e7-a4f7-21baf61bfe2b
```
### Steps to Reproduce
Please list the steps required to reproduce the issue, for example:
1. `terraform plan`
1. `terraform apply`
### Important Factoids
None
### References
Are there any other GitHub issues (open or closed) or Pull Requests that should be linked here?
Not that I saw
|
1.0
|
aws_rds_cluster_parameter_group does not generate a unique name when specifying name_prefix - ### Terraform Version
```
$ terraform --version
Terraform v0.9.11
Your version of Terraform is out of date! The latest version
is 0.10.6. You can update by downloading from www.terraform.io
```
### Affected Resource(s)
Please list the resources as a list, for example:
- aws_rds_cluster_parameter_group
### Terraform Configuration Files
```hcl
####
# RDS Cluster Paramater Group
resource "aws_rds_cluster_parameter_group" "fa-arora" {
name_prefix = "${var.env}-aurora-params"
family = "aurora5.6"
description = "RDS ${var.env} cluster parameter group"
parameter {
name = "character_set_server"
value = "utf8"
}
parameter {
name = "character_set_client"
value = "utf8"
}
}
# End of RDS Cluster Paramater Group
####
```
### Debug Output
https://gist.github.com/majormoses/f373836f0219ae37a0079f3938868ce9
### Expected Behavior
Terraform should have computed a unique name (using the prefix) and submitted it with the request.
### Actual Behavior
It appears that the plan believes it should compute the value but either does and does not submit it or does not compute it. This results in a 400 because the body did not match the requirements:
```
* aws_rds_cluster_parameter_group.fa-arora: Error modifying DB Cluster Parameter Group: InvalidParameterValue: The parameter DBClusterParameterGroupName must be provided and must not be blank.
status code: 400, request id: 40a2e32e-a094-11e7-a4f7-21baf61bfe2b
```
### Steps to Reproduce
Please list the steps required to reproduce the issue, for example:
1. `terraform plan`
1. `terraform apply`
### Important Factoids
None
### References
Are there any other GitHub issues (open or closed) or Pull Requests that should be linked here?
Not that I saw
|
non_process
|
aws rds cluster parameter group does not generate a unique name when specifying name prefix terraform version terraform version terraform your version of terraform is out of date the latest version is you can update by downloading from affected resource s please list the resources as a list for example aws rds cluster parameter group terraform configuration files hcl rds cluster paramater group resource aws rds cluster parameter group fa arora name prefix var env aurora params family description rds var env cluster parameter group parameter name character set server value parameter name character set client value end of rds cluster paramater group debug output expected behavior terraform should have computed a unique name using the prefix and submitted it with the request actual behavior it appears that the plan believes it should compute the value but either does and does not submit it or does not compute it this results in a because the body did not match the requirements aws rds cluster parameter group fa arora error modifying db cluster parameter group invalidparametervalue the parameter dbclusterparametergroupname must be provided and must not be blank status code request id steps to reproduce please list the steps required to reproduce the issue for example terraform plan terraform apply important factoids none references are there any other github issues open or closed or pull requests that should be linked here not that i saw
| 0
|
37,115
| 8,214,080,720
|
IssuesEvent
|
2018-09-04 21:43:32
|
shanalikhan/code-settings-sync
|
https://api.github.com/repos/shanalikhan/code-settings-sync
|
closed
|
Limitation on Windows behind a PAC corporate proxy
|
code-limitation feature-request
|
Hello
Current version of setting sync suffer some limitations when using on windows behind a PAC proxy.
When launching vscode without any argument, it requests automatically the proxy credential in an "Authentication required" window.
This can be skipped if vscode has been launched with the `--proxy-pac-url=<pac-file-url>` argument, using the corporate PAC url.
However, code-setting-sync does not seem to use it and raise javascript error (401 host not found). The only way I found to make it work under windows is to set an User Environment variable `http_proxy` to something like `http://username:password@proxyip:port`.
Expected behavior: propagate the proxy setting from vscode defined in https://code.visualstudio.com/docs/setup/network.
|
1.0
|
Limitation on Windows behind a PAC corporate proxy - Hello
Current version of setting sync suffer some limitations when using on windows behind a PAC proxy.
When launching vscode without any argument, it requests automatically the proxy credential in an "Authentication required" window.
This can be skipped if vscode has been launched with the `--proxy-pac-url=<pac-file-url>` argument, using the corporate PAC url.
However, code-setting-sync does not seem to use it and raise javascript error (401 host not found). The only way I found to make it work under windows is to set an User Environment variable `http_proxy` to something like `http://username:password@proxyip:port`.
Expected behavior: propagate the proxy setting from vscode defined in https://code.visualstudio.com/docs/setup/network.
|
non_process
|
limitation on windows behind a pac corporate proxy hello current version of setting sync suffer some limitations when using on windows behind a pac proxy when launching vscode without any argument it requests automatically the proxy credential in an authentication required window this can be skipped if vscode has been launched with the proxy pac url argument using the corporate pac url however code setting sync does not seem to use it and raise javascript error host not found the only way i found to make it work under windows is to set an user environment variable http proxy to something like expected behavior propagate the proxy setting from vscode defined in
| 0
|
11,527
| 14,403,079,750
|
IssuesEvent
|
2020-12-03 15:38:26
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
Order in DMMF is unstable
|
process/candidate team/client
|
Reported by @MichalLytek, apparently our DMMF not always has the same order for:
- `outputObjectTypes`: Sometimes `model` comes first, sometimes `prisma`
- `enumTypes`: The order of `model` and `prisma` is also sometimes different
|
1.0
|
Order in DMMF is unstable - Reported by @MichalLytek, apparently our DMMF not always has the same order for:
- `outputObjectTypes`: Sometimes `model` comes first, sometimes `prisma`
- `enumTypes`: The order of `model` and `prisma` is also sometimes different
|
process
|
order in dmmf is unstable reported by michallytek apparently our dmmf not always has the same order for outputobjecttypes sometimes model comes first sometimes prisma enumtypes the order of model and prisma is also sometimes different
| 1
|
20,491
| 10,521,253,890
|
IssuesEvent
|
2019-09-30 05:15:05
|
emilwareus/NodeGoat
|
https://api.github.com/repos/emilwareus/NodeGoat
|
opened
|
CVE-2011-4969 (Medium) detected in jquery-1.4.4.min.js
|
security vulnerability
|
## CVE-2011-4969 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.4.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.4/jquery.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/NodeGoat/node_modules/selenium-webdriver/lib/test/data/droppableItems.html</p>
<p>Path to vulnerable library: /NodeGoat/node_modules/selenium-webdriver/lib/test/data/js/jquery-1.4.4.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.4.4.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/emilwareus/NodeGoat/commit/6330909928926144add7fc2fa4dd11e0a24a7469">6330909928926144add7fc2fa4dd11e0a24a7469</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Cross-site scripting (XSS) vulnerability in jQuery before 1.6.3, when using location.hash to select elements, allows remote attackers to inject arbitrary web script or HTML via a crafted tag.
<p>Publish Date: 2013-03-08
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2011-4969>CVE-2011-4969</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>4.3</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2011-4969">https://nvd.nist.gov/vuln/detail/CVE-2011-4969</a></p>
<p>Release Date: 2013-03-08</p>
<p>Fix Resolution: 1.6.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2011-4969 (Medium) detected in jquery-1.4.4.min.js - ## CVE-2011-4969 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.4.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.4/jquery.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/NodeGoat/node_modules/selenium-webdriver/lib/test/data/droppableItems.html</p>
<p>Path to vulnerable library: /NodeGoat/node_modules/selenium-webdriver/lib/test/data/js/jquery-1.4.4.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.4.4.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/emilwareus/NodeGoat/commit/6330909928926144add7fc2fa4dd11e0a24a7469">6330909928926144add7fc2fa4dd11e0a24a7469</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Cross-site scripting (XSS) vulnerability in jQuery before 1.6.3, when using location.hash to select elements, allows remote attackers to inject arbitrary web script or HTML via a crafted tag.
<p>Publish Date: 2013-03-08
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2011-4969>CVE-2011-4969</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>4.3</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2011-4969">https://nvd.nist.gov/vuln/detail/CVE-2011-4969</a></p>
<p>Release Date: 2013-03-08</p>
<p>Fix Resolution: 1.6.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file tmp ws scm nodegoat node modules selenium webdriver lib test data droppableitems html path to vulnerable library nodegoat node modules selenium webdriver lib test data js jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details cross site scripting xss vulnerability in jquery before when using location hash to select elements allows remote attackers to inject arbitrary web script or html via a crafted tag publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
822,273
| 30,862,267,365
|
IssuesEvent
|
2023-08-03 04:49:58
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.tiktok.com - site is not usable
|
priority-critical browser-fixme
|
<!-- @browser: Dragon 65.0.2 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:65.0) Gecko/20100101 Firefox/65.0 IceDragon/65.0.2 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/125298 -->
**URL**: https://www.tiktok.com/@xx_leahsmith/video/7260844167096454427
**Browser / Version**: Dragon 65.0.2
**Operating System**: Windows 10
**Tested Another Browser**: Yes Firefox
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
"Page not available
Sorry about that! Please try again later." - and it says it at any page on tiktok, even the home page.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/8/c7753979-2a3e-4b65-9f2e-c0422dc12054.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20190318120942</li><li>channel: default</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2023/8/9a8a3ed8-dbf9-448f-86e9-e65721abd991)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.tiktok.com - site is not usable - <!-- @browser: Dragon 65.0.2 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:65.0) Gecko/20100101 Firefox/65.0 IceDragon/65.0.2 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/125298 -->
**URL**: https://www.tiktok.com/@xx_leahsmith/video/7260844167096454427
**Browser / Version**: Dragon 65.0.2
**Operating System**: Windows 10
**Tested Another Browser**: Yes Firefox
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
"Page not available
Sorry about that! Please try again later." - and it says it at any page on tiktok, even the home page.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/8/c7753979-2a3e-4b65-9f2e-c0422dc12054.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20190318120942</li><li>channel: default</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2023/8/9a8a3ed8-dbf9-448f-86e9-e65721abd991)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
site is not usable url browser version dragon operating system windows tested another browser yes firefox problem type site is not usable description page not loading correctly steps to reproduce page not available sorry about that please try again later and it says it at any page on tiktok even the home page view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel default hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
20,278
| 26,912,278,024
|
IssuesEvent
|
2023-02-07 01:27:45
|
googleapis/google-cloud-node
|
https://api.github.com/repos/googleapis/google-cloud-node
|
opened
|
Nodejs 12 EOL: Upgrade these packages to stable when we update Node
|
type: process
|
These will be breaking changes, so upgrade these to stable:
- [ ] certificatemanager (https://github.com/googleapis/google-cloud-node/issues/3423)
- [ ] cloud life sciences (https://github.com/googleapis/google-cloud-node/issues/3542)
- [ ] nodejs-analytics-data (https://github.com/googleapis/google-cloud-node/issues/3540)
|
1.0
|
Nodejs 12 EOL: Upgrade these packages to stable when we update Node - These will be breaking changes, so upgrade these to stable:
- [ ] certificatemanager (https://github.com/googleapis/google-cloud-node/issues/3423)
- [ ] cloud life sciences (https://github.com/googleapis/google-cloud-node/issues/3542)
- [ ] nodejs-analytics-data (https://github.com/googleapis/google-cloud-node/issues/3540)
|
process
|
nodejs eol upgrade these packages to stable when we update node these will be breaking changes so upgrade these to stable certificatemanager cloud life sciences nodejs analytics data
| 1
|
390,559
| 11,545,324,180
|
IssuesEvent
|
2020-02-18 13:14:33
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
store-kr.uniqlo.com - site is not usable
|
browser-fenix engine-gecko priority-important
|
<!-- @browser: Firefox Mobile 73.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:73.0) Gecko/73.0 Firefox/73.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/48338 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://store-kr.uniqlo.com/
**Browser / Version**: Firefox Mobile 73.0
**Operating System**: Android
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: load webpage repeatedly
**Steps to Reproduce**:
this page have to redirect to "www.uniqlo.com", but not works.
This page work well on firefox 68.5.0
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
Submitted in the name of `@GyeongJun`
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
store-kr.uniqlo.com - site is not usable - <!-- @browser: Firefox Mobile 73.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:73.0) Gecko/73.0 Firefox/73.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/48338 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://store-kr.uniqlo.com/
**Browser / Version**: Firefox Mobile 73.0
**Operating System**: Android
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: load webpage repeatedly
**Steps to Reproduce**:
this page have to redirect to "www.uniqlo.com", but not works.
This page work well on firefox 68.5.0
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
Submitted in the name of `@GyeongJun`
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
store kr uniqlo com site is not usable url browser version firefox mobile operating system android tested another browser yes problem type site is not usable description load webpage repeatedly steps to reproduce this page have to redirect to but not works this page work well on firefox browser configuration none submitted in the name of gyeongjun from with ❤️
| 0
|
291,947
| 25,187,381,191
|
IssuesEvent
|
2022-11-11 19:31:03
|
davidjerleke/embla-carousel
|
https://api.github.com/repos/davidjerleke/embla-carousel
|
closed
|
Not working with Nextjs
|
stalled missing test case
|

Using any of the default examples, i have a hydration problem. How to fix it?
|
1.0
|
Not working with Nextjs - 
Using any of the default examples, i have a hydration problem. How to fix it?
|
non_process
|
not working with nextjs using any of the default examples i have a hydration problem how to fix it
| 0
|
503,392
| 14,590,860,794
|
IssuesEvent
|
2020-12-19 10:03:13
|
MPAS-Dev/MPAS-Analysis
|
https://api.github.com/repos/MPAS-Dev/MPAS-Analysis
|
closed
|
Improvement to delta OHC plot
|
enhancement priority
|
The delta OHC plot from MPAS-Analysis is extremely useful. But some of us tend to think in W/m2 (at TOA) rather than 1.e22 Joules (and I have a hard time converting between the two in my head). I'd like suggest adding slanted lines to the plot showing the conversion in W/m2 (total earth area) for easy comparions with netTOA in atmospheric plots.
Here is a mock-up of what such a plot might look like (I made up the numbers):

|
1.0
|
Improvement to delta OHC plot - The delta OHC plot from MPAS-Analysis is extremely useful. But some of us tend to think in W/m2 (at TOA) rather than 1.e22 Joules (and I have a hard time converting between the two in my head). I'd like suggest adding slanted lines to the plot showing the conversion in W/m2 (total earth area) for easy comparions with netTOA in atmospheric plots.
Here is a mock-up of what such a plot might look like (I made up the numbers):

|
non_process
|
improvement to delta ohc plot the delta ohc plot from mpas analysis is extremely useful but some of us tend to think in w at toa rather than joules and i have a hard time converting between the two in my head i d like suggest adding slanted lines to the plot showing the conversion in w total earth area for easy comparions with nettoa in atmospheric plots here is a mock up of what such a plot might look like i made up the numbers
| 0
|
575,566
| 17,035,676,288
|
IssuesEvent
|
2021-07-05 06:43:16
|
Mezzanine-UI/mezzanine
|
https://api.github.com/repos/Mezzanine-UI/mezzanine
|
closed
|
React Select `fullWIdth` props does not work
|
Priority: HIGH bug react
|
### Describe the bug
When using `Select` component the `fullWidth` does not work. This is due to the root `<div />` of `Select` component does not have `width: 100%` in its style. And currently no way to put styles or css to the `<div />` element
### Steps To Reproduce
Steps to reproduce the behavior:
1. Implement `Select`
2. apply `fullWidth` like `<Select fullWidth={true} />`
3. see the component will not be full width of its parent
### Expected behavior
Be full width of its parent
### Screenshots
N/A
|
1.0
|
React Select `fullWIdth` props does not work - ### Describe the bug
When using `Select` component the `fullWidth` does not work. This is due to the root `<div />` of `Select` component does not have `width: 100%` in its style. And currently no way to put styles or css to the `<div />` element
### Steps To Reproduce
Steps to reproduce the behavior:
1. Implement `Select`
2. apply `fullWidth` like `<Select fullWidth={true} />`
3. see the component will not be full width of its parent
### Expected behavior
Be full width of its parent
### Screenshots
N/A
|
non_process
|
react select fullwidth props does not work describe the bug when using select component the fullwidth does not work this is due to the root of select component does not have width in its style and currently no way to put styles or css to the element steps to reproduce steps to reproduce the behavior implement select apply fullwidth like see the component will not be full width of its parent expected behavior be full width of its parent screenshots n a
| 0
|
831,901
| 32,064,483,314
|
IssuesEvent
|
2023-09-25 00:58:47
|
FRC-1294/frc2024
|
https://api.github.com/repos/FRC-1294/frc2024
|
closed
|
Attendance
|
Board Executive High Priority OffSeason
|
We have a working Google form. We are going to test it in the August off-season meetings and officially use it when the preseason starts in September.
|
1.0
|
Attendance - We have a working Google form. We are going to test it in the August off-season meetings and officially use it when the preseason starts in September.
|
non_process
|
attendance we have a working google form we are going to test it in the august off season meetings and officially use it when the preseason starts in september
| 0
|
11,077
| 13,912,982,492
|
IssuesEvent
|
2020-10-20 19:43:36
|
GoogleCloudPlatform/java-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/java-docs-samples
|
opened
|
functions/concepts/env-vars has illegal reflective operations (junit)
|
api: cloudfunctions priority: p2 type: process
|
Found in #4036
```
------------------------------------------------------------
- testing functions/concepts/env-vars
------------------------------------------------------------
[ERROR] WARNING: An illegal reflective access operation has occurred
[ERROR] WARNING: Illegal reflective access by org.junit.contrib.java.lang.system.EnvironmentVariables (file:/root/.m2/repository/com/github/stefanbirkner/system-rules/1.19.0/system-rules-1.19.0.jar) to field java.util.Collections$UnmodifiableMap.m
[ERROR] WARNING: Please consider reporting this to the maintainers of org.junit.contrib.java.lang.system.EnvironmentVariables
[ERROR] WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operations
[ERROR] WARNING: All illegal access operations will be denied in a future release
Testing completed.
```
|
1.0
|
functions/concepts/env-vars has illegal reflective operations (junit) - Found in #4036
```
------------------------------------------------------------
- testing functions/concepts/env-vars
------------------------------------------------------------
[ERROR] WARNING: An illegal reflective access operation has occurred
[ERROR] WARNING: Illegal reflective access by org.junit.contrib.java.lang.system.EnvironmentVariables (file:/root/.m2/repository/com/github/stefanbirkner/system-rules/1.19.0/system-rules-1.19.0.jar) to field java.util.Collections$UnmodifiableMap.m
[ERROR] WARNING: Please consider reporting this to the maintainers of org.junit.contrib.java.lang.system.EnvironmentVariables
[ERROR] WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operations
[ERROR] WARNING: All illegal access operations will be denied in a future release
Testing completed.
```
|
process
|
functions concepts env vars has illegal reflective operations junit found in testing functions concepts env vars warning an illegal reflective access operation has occurred warning illegal reflective access by org junit contrib java lang system environmentvariables file root repository com github stefanbirkner system rules system rules jar to field java util collections unmodifiablemap m warning please consider reporting this to the maintainers of org junit contrib java lang system environmentvariables warning use illegal access warn to enable warnings of further illegal reflective access operations warning all illegal access operations will be denied in a future release testing completed
| 1
|
306,851
| 9,412,244,246
|
IssuesEvent
|
2019-04-10 03:08:42
|
CS2103-AY1819S2-W15-2/main
|
https://api.github.com/repos/CS2103-AY1819S2-W15-2/main
|
closed
|
Undo command does not recalculate budget
|
bug priority.High
|
**Describe the bug**
Undo command does not undo the action on budget.
**To Reproduce**
Steps to reproduce the behavior:
1. Add a budget for transport with amount $20, start date 05-04-2019 and end date 20-04-2019.
2. Add an expense for "bus" with cost $5 and category TRANSPORT with date 05-04-2019. This expense is reflected on the transport budget.
3. Undo
**Expected behavior**
The transport budget should undo its update, and have an amount of $0 after the undo command.
**Screenshots**
If applicable, add screenshots to help explain your problem

**Additional context**
Add any other context about the problem here.
<hr>
**Reported by:** @kev-inc
**Severity:** Medium
<sub>[original: nus-cs2103-AY1819S2/pe-dry-run#623]</sub>
|
1.0
|
Undo command does not recalculate budget - **Describe the bug**
Undo command does not undo the action on budget.
**To Reproduce**
Steps to reproduce the behavior:
1. Add a budget for transport with amount $20, start date 05-04-2019 and end date 20-04-2019.
2. Add an expense for "bus" with cost $5 and category TRANSPORT with date 05-04-2019. This expense is reflected on the transport budget.
3. Undo
**Expected behavior**
The transport budget should undo its update, and have an amount of $0 after the undo command.
**Screenshots**
If applicable, add screenshots to help explain your problem

**Additional context**
Add any other context about the problem here.
<hr>
**Reported by:** @kev-inc
**Severity:** Medium
<sub>[original: nus-cs2103-AY1819S2/pe-dry-run#623]</sub>
|
non_process
|
undo command does not recalculate budget describe the bug undo command does not undo the action on budget to reproduce steps to reproduce the behavior add a budget for transport with amount start date and end date add an expense for bus with cost and category transport with date this expense is reflected on the transport budget undo expected behavior the transport budget should undo its update and have an amount of after the undo command screenshots if applicable add screenshots to help explain your problem additional context add any other context about the problem here reported by kev inc severity medium
| 0
|
14,363
| 17,382,761,862
|
IssuesEvent
|
2021-08-01 03:07:34
|
gfx-rs/naga
|
https://api.github.com/repos/gfx-rs/naga
|
opened
|
Many redundant assignments in produced shaders
|
area: processing help wanted kind: refactor
|
#925 had to make the `Load` expressions emitted unconditionally, so that the ordering of load/store operations is preserved.
Unfortunately, it results it a lot of dummy assignments. Perhaps, we can figure out a simple heuristics to avoid those?
|
1.0
|
Many redundant assignments in produced shaders - #925 had to make the `Load` expressions emitted unconditionally, so that the ordering of load/store operations is preserved.
Unfortunately, it results it a lot of dummy assignments. Perhaps, we can figure out a simple heuristics to avoid those?
|
process
|
many redundant assignments in produced shaders had to make the load expressions emitted unconditionally so that the ordering of load store operations is preserved unfortunately it results it a lot of dummy assignments perhaps we can figure out a simple heuristics to avoid those
| 1
|
63,291
| 17,571,335,271
|
IssuesEvent
|
2021-08-14 19:09:11
|
pymc-devs/pymc3
|
https://api.github.com/repos/pymc-devs/pymc3
|
closed
|
Binder link is broken
|
defects beginner friendly help wanted
|
<img width="1168" alt="image" src="https://user-images.githubusercontent.com/674200/125451212-ecf21d1c-4b20-4583-b1e1-7efbaf1548b2.png">
I think we also need to link to the pymc-examples repo instead.
|
1.0
|
Binder link is broken - <img width="1168" alt="image" src="https://user-images.githubusercontent.com/674200/125451212-ecf21d1c-4b20-4583-b1e1-7efbaf1548b2.png">
I think we also need to link to the pymc-examples repo instead.
|
non_process
|
binder link is broken img width alt image src i think we also need to link to the pymc examples repo instead
| 0
|
20,099
| 5,987,766,652
|
IssuesEvent
|
2017-06-02 01:10:22
|
fire-eggs/FamilyLines
|
https://api.github.com/repos/fire-eggs/FamilyLines
|
opened
|
Make URLs 'active' in HTML export
|
CodePlex
|
<b>notonyournelly[CodePlex]</b> <br />Web URLs in sources or notes are not made 'active' when exporting to HTML.
This can be seen with the sample windsor family, when exporting to HTML with sources amp repositories. Several source 'names' are URLs, but have not been made 'active'.
A URL would be defined as text that starts with quothttp://quot or quothttps://quot. These should be turned into links, e.g. the text quothttp://gw.geneanet.org/rebusyquot
should be exported as the HTML tag:
lta href=quothttp://gw.geneanet.org/rebusyquotgthttp://gw.geneanet.org/rebusylt/agt
Scan source names, notes Scan people notes
|
1.0
|
Make URLs 'active' in HTML export - <b>notonyournelly[CodePlex]</b> <br />Web URLs in sources or notes are not made 'active' when exporting to HTML.
This can be seen with the sample windsor family, when exporting to HTML with sources amp repositories. Several source 'names' are URLs, but have not been made 'active'.
A URL would be defined as text that starts with quothttp://quot or quothttps://quot. These should be turned into links, e.g. the text quothttp://gw.geneanet.org/rebusyquot
should be exported as the HTML tag:
lta href=quothttp://gw.geneanet.org/rebusyquotgthttp://gw.geneanet.org/rebusylt/agt
Scan source names, notes Scan people notes
|
non_process
|
make urls active in html export notonyournelly web urls in sources or notes are not made active when exporting to html this can be seen with the sample windsor family when exporting to html with sources amp repositories several source names are urls but have not been made active a url would be defined as text that starts with quot or quot these should be turned into links e g the text quot should be exported as the html tag lta href quot scan source names notes scan people notes
| 0
|
22,125
| 30,669,395,448
|
IssuesEvent
|
2023-07-25 20:59:07
|
h4sh5/npm-auto-scanner
|
https://api.github.com/repos/h4sh5/npm-auto-scanner
|
opened
|
office-addin-dev-settings 2.0.8 has 1 guarddog issues
|
npm-silent-process-execution
|
```{"npm-silent-process-execution":[{"code":" const subprocess = childProcess.spawn(commandLine, [], {\n detached: true,\n shell: true,\n stdio: \"ignore\",\n windowsHide: false,\n });","location":"package/lib/process.js:30","message":"This package is silently executing another executable"}]}```
|
1.0
|
office-addin-dev-settings 2.0.8 has 1 guarddog issues - ```{"npm-silent-process-execution":[{"code":" const subprocess = childProcess.spawn(commandLine, [], {\n detached: true,\n shell: true,\n stdio: \"ignore\",\n windowsHide: false,\n });","location":"package/lib/process.js:30","message":"This package is silently executing another executable"}]}```
|
process
|
office addin dev settings has guarddog issues npm silent process execution n detached true n shell true n stdio ignore n windowshide false n location package lib process js message this package is silently executing another executable
| 1
|
77,941
| 14,933,922,116
|
IssuesEvent
|
2021-01-25 09:51:38
|
haproxy/haproxy
|
https://api.github.com/repos/haproxy/haproxy
|
closed
|
src/http_ana.c: untrusted value as argument suspected by coverity
|
type: code-report
|
```
4585 struct channel *res = &s->res;
1. tainted_return_value: Function htx_from_buf returns tainted data. [show details]
2. var_assign: Assigning: htx = htx_from_buf(&res->buf), which taints htx.
4586 struct htx *htx = htx_from_buf(&res->buf);
4587
3. Condition s->txn->status == -1, taking true branch.
4588 if (s->txn->status == -1)
4589 s->txn->status = reply->status;
4590 channel_htx_truncate(res, htx);
4591
CID 1444718 (#1 of 1): Untrusted value as argument (TAINTED_SCALAR)4. tainted_data: Passing tainted expression htx->size to http_reply_to_htx, which uses it as an offset. [show details]
Ensure that tainted values are properly sanitized, by checking that their values are within a permissible range.
4592 if (http_reply_to_htx(s, htx, reply) == -1)
4593 goto fail;
4594
```
|
1.0
|
src/http_ana.c: untrusted value as argument suspected by coverity - ```
4585 struct channel *res = &s->res;
1. tainted_return_value: Function htx_from_buf returns tainted data. [show details]
2. var_assign: Assigning: htx = htx_from_buf(&res->buf), which taints htx.
4586 struct htx *htx = htx_from_buf(&res->buf);
4587
3. Condition s->txn->status == -1, taking true branch.
4588 if (s->txn->status == -1)
4589 s->txn->status = reply->status;
4590 channel_htx_truncate(res, htx);
4591
CID 1444718 (#1 of 1): Untrusted value as argument (TAINTED_SCALAR)4. tainted_data: Passing tainted expression htx->size to http_reply_to_htx, which uses it as an offset. [show details]
Ensure that tainted values are properly sanitized, by checking that their values are within a permissible range.
4592 if (http_reply_to_htx(s, htx, reply) == -1)
4593 goto fail;
4594
```
|
non_process
|
src http ana c untrusted value as argument suspected by coverity struct channel res s res tainted return value function htx from buf returns tainted data var assign assigning htx htx from buf res buf which taints htx struct htx htx htx from buf res buf condition s txn status taking true branch if s txn status s txn status reply status channel htx truncate res htx cid of untrusted value as argument tainted scalar tainted data passing tainted expression htx size to http reply to htx which uses it as an offset ensure that tainted values are properly sanitized by checking that their values are within a permissible range if http reply to htx s htx reply goto fail
| 0
|
390,290
| 11,541,626,994
|
IssuesEvent
|
2020-02-18 04:45:47
|
ansible/awx
|
https://api.github.com/repos/ansible/awx
|
opened
|
Obey new Ansible sanity rules: parameter-list-no-elements, invalid-ansiblemodule-schema
|
component:awx_collection priority:high state:needs_devel type:task
|
##### ISSUE TYPE
- CI failure
active with Ansible `devel` branch
##### SUMMARY
The recently-merged PR https://github.com/ansible/ansible/pull/66386 fails our modules in the AWX collection with
```
plugins/modules/tower_job_launch.py:0:0: parameter-list-no-elements: Argument 'tags' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_job_template.py:0:0: parameter-list-no-elements: Argument 'credentials' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_notification.py:0:0: parameter-list-no-elements: Argument 'channels' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_notification.py:0:0: parameter-list-no-elements: Argument 'recipients' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_notification.py:0:0: parameter-list-no-elements: Argument 'rooms' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_notification.py:0:0: parameter-list-no-elements: Argument 'targets' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_notification.py:0:0: parameter-list-no-elements: Argument 'to_numbers' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'credential' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'credential_type' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'inventory' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'inventory_script' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'job_template' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'notification_template' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'organization' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'project' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'team' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'user' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'workflow' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_send.py:0:0: parameter-list-no-elements: Argument 'files' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_send.py:0:0: parameter-list-no-elements: Argument 'prevent' in argument_spec defines type as list but elements is not defined
```
And the PR https://github.com/ansible/ansible/pull/65747 fails our modules with
```
plugins/modules/tower_credential.py:0:0: invalid-ansiblemodule-schema: AnsibleModule.argument_spec.kind.choices: expected a list for dictionary value @ data['argument_spec']['kind']['choices']. Got dict_keys(['ssh', 'vault', 'net', 'scm', 'aws', 'vmware', 'satellite6', 'cloudforms', 'gce', 'azure_rm', 'openstack', 'rhv', 'insights', 'tower'])
plugins/modules/tower_credential_type.py:0:0: invalid-ansiblemodule-schema: AnsibleModule.argument_spec.kind.choices: expected a list for dictionary value @ data['argument_spec']['kind']['choices']. Got dict_keys(['ssh', 'vault', 'net', 'scm', 'cloud', 'insights'])
plugins/modules/tower_inventory_source.py:0:0: invalid-ansiblemodule-schema: AnsibleModule.argument_spec.source.choices: expected a list for dictionary value @ data['argument_spec']['source']['choices']. Got dict_keys(['file', 'scm', 'ec2', 'gce', 'azure', 'azure_rm', 'vmware', 'satellite6', 'cloudforms', 'openstack', 'rhv', 'tower', 'custom'])
```
|
1.0
|
Obey new Ansible sanity rules: parameter-list-no-elements, invalid-ansiblemodule-schema - ##### ISSUE TYPE
- CI failure
active with Ansible `devel` branch
##### SUMMARY
The recently-merged PR https://github.com/ansible/ansible/pull/66386 fails our modules in the AWX collection with
```
plugins/modules/tower_job_launch.py:0:0: parameter-list-no-elements: Argument 'tags' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_job_template.py:0:0: parameter-list-no-elements: Argument 'credentials' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_notification.py:0:0: parameter-list-no-elements: Argument 'channels' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_notification.py:0:0: parameter-list-no-elements: Argument 'recipients' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_notification.py:0:0: parameter-list-no-elements: Argument 'rooms' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_notification.py:0:0: parameter-list-no-elements: Argument 'targets' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_notification.py:0:0: parameter-list-no-elements: Argument 'to_numbers' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'credential' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'credential_type' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'inventory' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'inventory_script' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'job_template' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'notification_template' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'organization' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'project' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'team' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'user' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_receive.py:0:0: parameter-list-no-elements: Argument 'workflow' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_send.py:0:0: parameter-list-no-elements: Argument 'files' in argument_spec defines type as list but elements is not defined
plugins/modules/tower_send.py:0:0: parameter-list-no-elements: Argument 'prevent' in argument_spec defines type as list but elements is not defined
```
And the PR https://github.com/ansible/ansible/pull/65747 fails our modules with
```
plugins/modules/tower_credential.py:0:0: invalid-ansiblemodule-schema: AnsibleModule.argument_spec.kind.choices: expected a list for dictionary value @ data['argument_spec']['kind']['choices']. Got dict_keys(['ssh', 'vault', 'net', 'scm', 'aws', 'vmware', 'satellite6', 'cloudforms', 'gce', 'azure_rm', 'openstack', 'rhv', 'insights', 'tower'])
plugins/modules/tower_credential_type.py:0:0: invalid-ansiblemodule-schema: AnsibleModule.argument_spec.kind.choices: expected a list for dictionary value @ data['argument_spec']['kind']['choices']. Got dict_keys(['ssh', 'vault', 'net', 'scm', 'cloud', 'insights'])
plugins/modules/tower_inventory_source.py:0:0: invalid-ansiblemodule-schema: AnsibleModule.argument_spec.source.choices: expected a list for dictionary value @ data['argument_spec']['source']['choices']. Got dict_keys(['file', 'scm', 'ec2', 'gce', 'azure', 'azure_rm', 'vmware', 'satellite6', 'cloudforms', 'openstack', 'rhv', 'tower', 'custom'])
```
|
non_process
|
obey new ansible sanity rules parameter list no elements invalid ansiblemodule schema issue type ci failure active with ansible devel branch summary the recently merged pr fails our modules in the awx collection with plugins modules tower job launch py parameter list no elements argument tags in argument spec defines type as list but elements is not defined plugins modules tower job template py parameter list no elements argument credentials in argument spec defines type as list but elements is not defined plugins modules tower notification py parameter list no elements argument channels in argument spec defines type as list but elements is not defined plugins modules tower notification py parameter list no elements argument recipients in argument spec defines type as list but elements is not defined plugins modules tower notification py parameter list no elements argument rooms in argument spec defines type as list but elements is not defined plugins modules tower notification py parameter list no elements argument targets in argument spec defines type as list but elements is not defined plugins modules tower notification py parameter list no elements argument to numbers in argument spec defines type as list but elements is not defined plugins modules tower receive py parameter list no elements argument credential in argument spec defines type as list but elements is not defined plugins modules tower receive py parameter list no elements argument credential type in argument spec defines type as list but elements is not defined plugins modules tower receive py parameter list no elements argument inventory in argument spec defines type as list but elements is not defined plugins modules tower receive py parameter list no elements argument inventory script in argument spec defines type as list but elements is not defined plugins modules tower receive py parameter list no elements argument job template in argument spec defines type as list but elements is not defined plugins modules tower receive py parameter list no elements argument notification template in argument spec defines type as list but elements is not defined plugins modules tower receive py parameter list no elements argument organization in argument spec defines type as list but elements is not defined plugins modules tower receive py parameter list no elements argument project in argument spec defines type as list but elements is not defined plugins modules tower receive py parameter list no elements argument team in argument spec defines type as list but elements is not defined plugins modules tower receive py parameter list no elements argument user in argument spec defines type as list but elements is not defined plugins modules tower receive py parameter list no elements argument workflow in argument spec defines type as list but elements is not defined plugins modules tower send py parameter list no elements argument files in argument spec defines type as list but elements is not defined plugins modules tower send py parameter list no elements argument prevent in argument spec defines type as list but elements is not defined and the pr fails our modules with plugins modules tower credential py invalid ansiblemodule schema ansiblemodule argument spec kind choices expected a list for dictionary value data got dict keys plugins modules tower credential type py invalid ansiblemodule schema ansiblemodule argument spec kind choices expected a list for dictionary value data got dict keys plugins modules tower inventory source py invalid ansiblemodule schema ansiblemodule argument spec source choices expected a list for dictionary value data got dict keys
| 0
|
616,159
| 19,294,904,688
|
IssuesEvent
|
2021-12-12 12:28:15
|
bounswe/2021SpringGroup9
|
https://api.github.com/repos/bounswe/2021SpringGroup9
|
closed
|
Show posts on the map
|
enhancement status: in progress priority: critical difficulty: hard android postory
|
### Task: Show posts on the map on the discover page
### Type of task (new feature, writing tests, refactoring): New Feature
### Standards and rules to follow: None
### Expected result: The markers should show the correct locations for the posts.
### Additional notes: None.
**Deadline: 12.12.2021**
|
1.0
|
Show posts on the map - ### Task: Show posts on the map on the discover page
### Type of task (new feature, writing tests, refactoring): New Feature
### Standards and rules to follow: None
### Expected result: The markers should show the correct locations for the posts.
### Additional notes: None.
**Deadline: 12.12.2021**
|
non_process
|
show posts on the map task show posts on the map on the discover page type of task new feature writing tests refactoring new feature standards and rules to follow none expected result the markers should show the correct locations for the posts additional notes none deadline
| 0
|
19,206
| 25,339,013,139
|
IssuesEvent
|
2022-11-18 19:34:43
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
when setting up bazel on ubuntu running on ARM64, getting "bazel-real: cannot execute binary file: Exec format error"
|
type: support / not a bug (process) untriaged team-OSS
|
### Description of the bug:
downloaded bazel-4.2.3-installer-darwin-arm64.sh and bazel-5.3.2-installer-darwin-arm64.sh and followed [instructions]( https://bazel.build/install/ubuntu) using binary installer to set up.
Getting this error when running the installer bash script for both installer versions -
**_Uncompressing....../home/vscode/bin/bazel: line 220: /home/vscode/.bazel/bin/bazel-real: cannot execute binary file: Exec format error_**
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
Run this script on an ubuntu VM running on arm64 macbook. This is based on set up instructions [here](https://bazel.build/install/ubuntu#:~:text=Installing%20Bazel%20on%20Ubuntu%201%20Step%201%3A%20Add,...%203%20Step%203%3A%20Install%20a%20JDK%20%28optional%29) using the binary installer.
sudo apt install g++ unzip zip -y
sudo apt-get install openjdk-11-jdk -y
#wget https://github.com/bazelbuild/bazel/releases/download/5.3.2/bazel-5.3.2-installer-darwin-arm64.sh
#chmod +x bazel-5.3.2-installer-darwin-arm64.sh
#./bazel-5.3.2-installer-darwin-arm64.sh --user
wget https://github.com/bazelbuild/bazel/releases/download/4.2.3/bazel-4.2.3-installer-darwin-arm64.sh
chmod +x bazel-4.2.3-installer-darwin-arm64.sh
./bazel-4.2.3-installer-darwin-arm64.sh --user
export PATH="$PATH:$HOME/bin"
### Which operating system are you running Bazel on?
ubuntu
### What is the output of `bazel info release`?
doesn't run because it failed ruing set up
### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel.
_No response_
### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ?
_No response_
### Have you found anything relevant by searching the web?
No
### Any other information, logs, or outputs that you want to share?
complete command run output
```
songy@ubuntu:~$ ./bazel-5.3.2-installer-darwin-arm64.sh --user
Bazel installer
---------------
Bazel is bundled with software licensed under the GPLv2 with Classpath exception.
You can find the sources next to the installer on our release page:
https://github.com/bazelbuild/bazel/releases
# Binary package at HEAD (@bc087f49584a6a6)
- [Commit](https://github.com/bazelbuild/bazel/commit/bc087f49584a6a6)
Uncompressing....../home/songy/bin/bazel: line 220: /home/songy/.bazel/bin/bazel-real: cannot execute binary file: Exec format error
/home/songy/bin/bazel: line 220: /home/songy/.bazel/bin/bazel-real: Success
```
|
1.0
|
when setting up bazel on ubuntu running on ARM64, getting "bazel-real: cannot execute binary file: Exec format error" - ### Description of the bug:
downloaded bazel-4.2.3-installer-darwin-arm64.sh and bazel-5.3.2-installer-darwin-arm64.sh and followed [instructions]( https://bazel.build/install/ubuntu) using binary installer to set up.
Getting this error when running the installer bash script for both installer versions -
**_Uncompressing....../home/vscode/bin/bazel: line 220: /home/vscode/.bazel/bin/bazel-real: cannot execute binary file: Exec format error_**
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
Run this script on an ubuntu VM running on arm64 macbook. This is based on set up instructions [here](https://bazel.build/install/ubuntu#:~:text=Installing%20Bazel%20on%20Ubuntu%201%20Step%201%3A%20Add,...%203%20Step%203%3A%20Install%20a%20JDK%20%28optional%29) using the binary installer.
sudo apt install g++ unzip zip -y
sudo apt-get install openjdk-11-jdk -y
#wget https://github.com/bazelbuild/bazel/releases/download/5.3.2/bazel-5.3.2-installer-darwin-arm64.sh
#chmod +x bazel-5.3.2-installer-darwin-arm64.sh
#./bazel-5.3.2-installer-darwin-arm64.sh --user
wget https://github.com/bazelbuild/bazel/releases/download/4.2.3/bazel-4.2.3-installer-darwin-arm64.sh
chmod +x bazel-4.2.3-installer-darwin-arm64.sh
./bazel-4.2.3-installer-darwin-arm64.sh --user
export PATH="$PATH:$HOME/bin"
### Which operating system are you running Bazel on?
ubuntu
### What is the output of `bazel info release`?
doesn't run because it failed ruing set up
### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel.
_No response_
### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ?
_No response_
### Have you found anything relevant by searching the web?
No
### Any other information, logs, or outputs that you want to share?
complete command run output
```
songy@ubuntu:~$ ./bazel-5.3.2-installer-darwin-arm64.sh --user
Bazel installer
---------------
Bazel is bundled with software licensed under the GPLv2 with Classpath exception.
You can find the sources next to the installer on our release page:
https://github.com/bazelbuild/bazel/releases
# Binary package at HEAD (@bc087f49584a6a6)
- [Commit](https://github.com/bazelbuild/bazel/commit/bc087f49584a6a6)
Uncompressing....../home/songy/bin/bazel: line 220: /home/songy/.bazel/bin/bazel-real: cannot execute binary file: Exec format error
/home/songy/bin/bazel: line 220: /home/songy/.bazel/bin/bazel-real: Success
```
|
process
|
when setting up bazel on ubuntu running on getting bazel real cannot execute binary file exec format error description of the bug downloaded bazel installer darwin sh and bazel installer darwin sh and followed using binary installer to set up getting this error when running the installer bash script for both installer versions uncompressing home vscode bin bazel line home vscode bazel bin bazel real cannot execute binary file exec format error what s the simplest easiest way to reproduce this bug please provide a minimal example if possible run this script on an ubuntu vm running on macbook this is based on set up instructions using the binary installer sudo apt install g unzip zip y sudo apt get install openjdk jdk y wget chmod x bazel installer darwin sh bazel installer darwin sh user wget chmod x bazel installer darwin sh bazel installer darwin sh user export path path home bin which operating system are you running bazel on ubuntu what is the output of bazel info release doesn t run because it failed ruing set up if bazel info release returns development version or non git tell us how you built bazel no response what s the output of git remote get url origin git rev parse master git rev parse head no response have you found anything relevant by searching the web no any other information logs or outputs that you want to share complete command run output songy ubuntu bazel installer darwin sh user bazel installer bazel is bundled with software licensed under the with classpath exception you can find the sources next to the installer on our release page binary package at head uncompressing home songy bin bazel line home songy bazel bin bazel real cannot execute binary file exec format error home songy bin bazel line home songy bazel bin bazel real success
| 1
|
14,908
| 18,294,751,769
|
IssuesEvent
|
2021-10-05 19:15:18
|
vectordotdev/vector
|
https://api.github.com/repos/vectordotdev/vector
|
opened
|
Process events in batches
|
type: enhancement domain: performance domain: processing
|
As part of making Vector more performant and memory efficient we should switch to processing events in batches for sources that support receiving events in batches. This is a placeholder issue representing that work.
|
1.0
|
Process events in batches - As part of making Vector more performant and memory efficient we should switch to processing events in batches for sources that support receiving events in batches. This is a placeholder issue representing that work.
|
process
|
process events in batches as part of making vector more performant and memory efficient we should switch to processing events in batches for sources that support receiving events in batches this is a placeholder issue representing that work
| 1
|
20,011
| 26,485,430,087
|
IssuesEvent
|
2023-01-17 17:36:45
|
GoogleCloudPlatform/java-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/java-docs-samples
|
closed
|
Cloud Run Filesystem sample is not cleaning up services
|
type: process api: run samples
|
In the java-docs-samples-testing project, new `filesystem<UUID>` services are being created daily.
They appear to be coming from [ApplicationTests.java](https://github.com/GoogleCloudPlatform/java-docs-samples/blob/main/run/filesystem/src/test/java/com/example/filesystem/ApplicationTests.java#L63), which does have a [cleanup step](https://github.com/GoogleCloudPlatform/java-docs-samples/blob/2734187ea2214bb3e13f95923d25e89cde61d2f4/run/filesystem/src/test/java/com/example/filesystem/ApplicationTests.java#L123).
|
1.0
|
Cloud Run Filesystem sample is not cleaning up services - In the java-docs-samples-testing project, new `filesystem<UUID>` services are being created daily.
They appear to be coming from [ApplicationTests.java](https://github.com/GoogleCloudPlatform/java-docs-samples/blob/main/run/filesystem/src/test/java/com/example/filesystem/ApplicationTests.java#L63), which does have a [cleanup step](https://github.com/GoogleCloudPlatform/java-docs-samples/blob/2734187ea2214bb3e13f95923d25e89cde61d2f4/run/filesystem/src/test/java/com/example/filesystem/ApplicationTests.java#L123).
|
process
|
cloud run filesystem sample is not cleaning up services in the java docs samples testing project new filesystem services are being created daily they appear to be coming from which does have a
| 1
|
14,869
| 18,277,814,657
|
IssuesEvent
|
2021-10-04 21:09:27
|
GSA/EDX
|
https://api.github.com/repos/GSA/EDX
|
opened
|
Datamap
|
process
|
Datamap is an exploration of layers of organizational data:

* Staff
* Websites
* Programs
* Products
* PMO's - Product teams
* Focus Areas
* Divisions
* Agency (on Agency)
* Agencies
---
#### Opportunities for further research
* [ ] add Tasks as layer, atop Staff, to convey resourcing (is it possible to visualize a resource constraint? capacity time vs. backlog estimate time)
* highlight Staff who have Roles that relate to Services (requires modeling Services, using Roles)
* [ ] other visualizations
|
1.0
|
Datamap - Datamap is an exploration of layers of organizational data:

* Staff
* Websites
* Programs
* Products
* PMO's - Product teams
* Focus Areas
* Divisions
* Agency (on Agency)
* Agencies
---
#### Opportunities for further research
* [ ] add Tasks as layer, atop Staff, to convey resourcing (is it possible to visualize a resource constraint? capacity time vs. backlog estimate time)
* highlight Staff who have Roles that relate to Services (requires modeling Services, using Roles)
* [ ] other visualizations
|
process
|
datamap datamap is an exploration of layers of organizational data staff websites programs products pmo s product teams focus areas divisions agency on agency agencies opportunities for further research add tasks as layer atop staff to convey resourcing is it possible to visualize a resource constraint capacity time vs backlog estimate time highlight staff who have roles that relate to services requires modeling services using roles other visualizations
| 1
|
18,969
| 24,942,081,057
|
IssuesEvent
|
2022-10-31 19:52:09
|
googleapis/sphinx-docfx-yaml
|
https://api.github.com/repos/googleapis/sphinx-docfx-yaml
|
closed
|
Add "google-cloud-aiplatform[prediction]" as custom requirement for Kokoro script
|
type: process priority: p3
|
There is a new library in python-aiplatform that requires a special install. Filing the issue to keep track of this.
|
1.0
|
Add "google-cloud-aiplatform[prediction]" as custom requirement for Kokoro script - There is a new library in python-aiplatform that requires a special install. Filing the issue to keep track of this.
|
process
|
add google cloud aiplatform as custom requirement for kokoro script there is a new library in python aiplatform that requires a special install filing the issue to keep track of this
| 1
|
21,182
| 28,149,907,897
|
IssuesEvent
|
2023-04-02 22:40:32
|
AbdElAziz333/Pluto
|
https://api.github.com/repos/AbdElAziz333/Pluto
|
closed
|
Version 1.19.2-0.0.8 incompatible with Immersive Portals for Forge
|
bug in processing
|
[crash-2023-03-30_09.36.46-server.txt](https://github.com/AbdElAziz333/Pluto/files/11113362/crash-2023-03-30_09.36.46-server.txt)
The only mods I have installed are cloth config (req of Imm Portals), Imm Portals, and Pluto. The crash does not happen with Version 0.0.7
I'm using Forge 43.2.8 but I have tested it on many versions and the result is the same.
|
1.0
|
Version 1.19.2-0.0.8 incompatible with Immersive Portals for Forge - [crash-2023-03-30_09.36.46-server.txt](https://github.com/AbdElAziz333/Pluto/files/11113362/crash-2023-03-30_09.36.46-server.txt)
The only mods I have installed are cloth config (req of Imm Portals), Imm Portals, and Pluto. The crash does not happen with Version 0.0.7
I'm using Forge 43.2.8 but I have tested it on many versions and the result is the same.
|
process
|
version incompatible with immersive portals for forge the only mods i have installed are cloth config req of imm portals imm portals and pluto the crash does not happen with version i m using forge but i have tested it on many versions and the result is the same
| 1
|
14,900
| 18,291,580,324
|
IssuesEvent
|
2021-10-05 15:46:51
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[Android][manage apps] Need to update the text in the pop-up for the app upgrade screens
|
Bug P2 Android Process: Fixed Process: Tested QA Process: Tested dev
|
Current it is displaying same text for both scenarios.

**Force upgrade screen**
Pop-up text :-
"Please upgrade the app to continue."
Buttons:-
1) Ok
2) Cancel
**Optional upgrade screen**
Pop-up text :-
"A new version of this app is available. Do you want to update it now?"
Buttons:-
1) Yes
2) Skip
|
3.0
|
[Android][manage apps] Need to update the text in the pop-up for the app upgrade screens - Current it is displaying same text for both scenarios.

**Force upgrade screen**
Pop-up text :-
"Please upgrade the app to continue."
Buttons:-
1) Ok
2) Cancel
**Optional upgrade screen**
Pop-up text :-
"A new version of this app is available. Do you want to update it now?"
Buttons:-
1) Yes
2) Skip
|
process
|
need to update the text in the pop up for the app upgrade screens current it is displaying same text for both scenarios force upgrade screen pop up text please upgrade the app to continue buttons ok cancel optional upgrade screen pop up text a new version of this app is available do you want to update it now buttons yes skip
| 1
|
5,401
| 8,233,501,848
|
IssuesEvent
|
2018-09-08 01:42:09
|
aspnet/IISIntegration
|
https://api.github.com/repos/aspnet/IISIntegration
|
closed
|
Socket bind exception not captured in stdout log
|
cost: M diagnostics investigate out-of-process
|
If two apps fail to start due to a port collision (see https://github.com/aspnet/IISIntegration/issues/1124), I would expect to see the managed exception (shown below) in the stdout log for the app. However, for some reason this message is not captured in the stdout log. I have verified a normal exception thrown from `Main()` is captured, and I've also verified that any calls to `Console.WriteLine()` before the port collision are also captured.
I have verified this impacts both v1 and v2 out-of-process.
The event log does contain messages from IIS which make it clear the issue is port collision, so this issue is a lower priority, but it's puzzling.
```
$ dotnet run
crit: Microsoft.AspNetCore.Server.Kestrel[0]
Unable to start Kestrel.
System.IO.IOException: Failed to bind to address http://127.0.0.1:5000: address already in use. ---> Microsoft.AspNetCore.Connections.AddressInUseException: Only one usage of each socket address (protocol/network address/port) is normally permitted ---> System.Net.Sockets.SocketException: Only one usage of each socket address (protocol/network address/port) is normally permitted
at System.Net.Sockets.Socket.UpdateStatusAfterSocketErrorAndThrowException(SocketError error, String callerName)
at System.Net.Sockets.Socket.DoBind(EndPoint endPointSnapshot, SocketAddress socketAddress)
at System.Net.Sockets.Socket.Bind(EndPoint localEP)
at Microsoft.AspNetCore.Server.Kestrel.Transport.Sockets.SocketTransport.BindAsync()
--- End of inner exception stack trace ---
at Microsoft.AspNetCore.Server.Kestrel.Transport.Sockets.SocketTransport.BindAsync()
at Microsoft.AspNetCore.Server.Kestrel.Core.KestrelServer.<>c__DisplayClass22_0`1.<<StartAsync>g__OnBind|0>d.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at Microsoft.AspNetCore.Server.Kestrel.Core.Internal.AddressBinder.BindEndpointAsync(ListenOptions endpoint, AddressBindContext context)
--- End of inner exception stack trace ---
at Microsoft.AspNetCore.Server.Kestrel.Core.Internal.AddressBinder.BindEndpointAsync(ListenOptions endpoint, AddressBindContext context)
at Microsoft.AspNetCore.Server.Kestrel.Core.LocalhostListenOptions.BindAsync(AddressBindContext context)
at Microsoft.AspNetCore.Server.Kestrel.Core.Internal.AddressBinder.DefaultAddressStrategy.BindAsync(AddressBindContext context)
at Microsoft.AspNetCore.Server.Kestrel.Core.Internal.AddressBinder.BindAsync(IServerAddressesFeature addresses, KestrelServerOptions serverOptions, ILogger logger, Func`2 createBinding)
at Microsoft.AspNetCore.Server.Kestrel.Core.KestrelServer.StartAsync[TContext](IHttpApplication`1 application, CancellationToken cancellationToken)
```
|
1.0
|
Socket bind exception not captured in stdout log - If two apps fail to start due to a port collision (see https://github.com/aspnet/IISIntegration/issues/1124), I would expect to see the managed exception (shown below) in the stdout log for the app. However, for some reason this message is not captured in the stdout log. I have verified a normal exception thrown from `Main()` is captured, and I've also verified that any calls to `Console.WriteLine()` before the port collision are also captured.
I have verified this impacts both v1 and v2 out-of-process.
The event log does contain messages from IIS which make it clear the issue is port collision, so this issue is a lower priority, but it's puzzling.
```
$ dotnet run
crit: Microsoft.AspNetCore.Server.Kestrel[0]
Unable to start Kestrel.
System.IO.IOException: Failed to bind to address http://127.0.0.1:5000: address already in use. ---> Microsoft.AspNetCore.Connections.AddressInUseException: Only one usage of each socket address (protocol/network address/port) is normally permitted ---> System.Net.Sockets.SocketException: Only one usage of each socket address (protocol/network address/port) is normally permitted
at System.Net.Sockets.Socket.UpdateStatusAfterSocketErrorAndThrowException(SocketError error, String callerName)
at System.Net.Sockets.Socket.DoBind(EndPoint endPointSnapshot, SocketAddress socketAddress)
at System.Net.Sockets.Socket.Bind(EndPoint localEP)
at Microsoft.AspNetCore.Server.Kestrel.Transport.Sockets.SocketTransport.BindAsync()
--- End of inner exception stack trace ---
at Microsoft.AspNetCore.Server.Kestrel.Transport.Sockets.SocketTransport.BindAsync()
at Microsoft.AspNetCore.Server.Kestrel.Core.KestrelServer.<>c__DisplayClass22_0`1.<<StartAsync>g__OnBind|0>d.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at Microsoft.AspNetCore.Server.Kestrel.Core.Internal.AddressBinder.BindEndpointAsync(ListenOptions endpoint, AddressBindContext context)
--- End of inner exception stack trace ---
at Microsoft.AspNetCore.Server.Kestrel.Core.Internal.AddressBinder.BindEndpointAsync(ListenOptions endpoint, AddressBindContext context)
at Microsoft.AspNetCore.Server.Kestrel.Core.LocalhostListenOptions.BindAsync(AddressBindContext context)
at Microsoft.AspNetCore.Server.Kestrel.Core.Internal.AddressBinder.DefaultAddressStrategy.BindAsync(AddressBindContext context)
at Microsoft.AspNetCore.Server.Kestrel.Core.Internal.AddressBinder.BindAsync(IServerAddressesFeature addresses, KestrelServerOptions serverOptions, ILogger logger, Func`2 createBinding)
at Microsoft.AspNetCore.Server.Kestrel.Core.KestrelServer.StartAsync[TContext](IHttpApplication`1 application, CancellationToken cancellationToken)
```
|
process
|
socket bind exception not captured in stdout log if two apps fail to start due to a port collision see i would expect to see the managed exception shown below in the stdout log for the app however for some reason this message is not captured in the stdout log i have verified a normal exception thrown from main is captured and i ve also verified that any calls to console writeline before the port collision are also captured i have verified this impacts both and out of process the event log does contain messages from iis which make it clear the issue is port collision so this issue is a lower priority but it s puzzling dotnet run crit microsoft aspnetcore server kestrel unable to start kestrel system io ioexception failed to bind to address address already in use microsoft aspnetcore connections addressinuseexception only one usage of each socket address protocol network address port is normally permitted system net sockets socketexception only one usage of each socket address protocol network address port is normally permitted at system net sockets socket updatestatusaftersocketerrorandthrowexception socketerror error string callername at system net sockets socket dobind endpoint endpointsnapshot socketaddress socketaddress at system net sockets socket bind endpoint localep at microsoft aspnetcore server kestrel transport sockets sockettransport bindasync end of inner exception stack trace at microsoft aspnetcore server kestrel transport sockets sockettransport bindasync at microsoft aspnetcore server kestrel core kestrelserver c g onbind d movenext end of stack trace from previous location where exception was thrown at microsoft aspnetcore server kestrel core internal addressbinder bindendpointasync listenoptions endpoint addressbindcontext context end of inner exception stack trace at microsoft aspnetcore server kestrel core internal addressbinder bindendpointasync listenoptions endpoint addressbindcontext context at microsoft aspnetcore server kestrel core localhostlistenoptions bindasync addressbindcontext context at microsoft aspnetcore server kestrel core internal addressbinder defaultaddressstrategy bindasync addressbindcontext context at microsoft aspnetcore server kestrel core internal addressbinder bindasync iserveraddressesfeature addresses kestrelserveroptions serveroptions ilogger logger func createbinding at microsoft aspnetcore server kestrel core kestrelserver startasync ihttpapplication application cancellationtoken cancellationtoken
| 1
|
82,445
| 10,253,902,769
|
IssuesEvent
|
2019-08-21 12:26:33
|
geocollections/sarv-edit
|
https://api.github.com/repos/geocollections/sarv-edit
|
closed
|
Test reference workspace with combined list and detail views
|
Critical: level 1 Difficulty: medium design enhancement
|
In some workflows it is desirable to have both list view and detailed view visible at the same time, so that user avoids clicking back and forth between the two full window views and doing a lot of scrolling in the list view. The filtered-sorted list would remain visible on the left, showing only 1-2 data fields (maybe only reference). If screen space is limited, this left panel could be hidden by vertical slider, or partly closed by user. Both panels should have independent vertical scrolling.
|
1.0
|
Test reference workspace with combined list and detail views - In some workflows it is desirable to have both list view and detailed view visible at the same time, so that user avoids clicking back and forth between the two full window views and doing a lot of scrolling in the list view. The filtered-sorted list would remain visible on the left, showing only 1-2 data fields (maybe only reference). If screen space is limited, this left panel could be hidden by vertical slider, or partly closed by user. Both panels should have independent vertical scrolling.
|
non_process
|
test reference workspace with combined list and detail views in some workflows it is desirable to have both list view and detailed view visible at the same time so that user avoids clicking back and forth between the two full window views and doing a lot of scrolling in the list view the filtered sorted list would remain visible on the left showing only data fields maybe only reference if screen space is limited this left panel could be hidden by vertical slider or partly closed by user both panels should have independent vertical scrolling
| 0
|
203,785
| 15,889,246,982
|
IssuesEvent
|
2021-04-10 10:41:11
|
LDSSA/wiki
|
https://api.github.com/repos/LDSSA/wiki
|
closed
|
The links to "disciplinary action" are broken
|
Documentation AOR wiki-github-page
|
On [this section](https://ldssa.github.io/wiki/Students/Exposing-concerns/) there are a couple of links that go to empty old wiki sections.
|
1.0
|
The links to "disciplinary action" are broken - On [this section](https://ldssa.github.io/wiki/Students/Exposing-concerns/) there are a couple of links that go to empty old wiki sections.
|
non_process
|
the links to disciplinary action are broken on there are a couple of links that go to empty old wiki sections
| 0
|
16,138
| 20,403,508,838
|
IssuesEvent
|
2022-02-23 00:41:42
|
vortexntnu/Vortex-CV
|
https://api.github.com/repos/vortexntnu/Vortex-CV
|
opened
|
Transform computed pose into the world frame (Pointcloud Processing)
|
enhancement high priority Pointcloud Processing Estimation
|
**Time estimate:** 10 hours
**Description of task:**
The Vision-EKF needs object poses in the odom (world) frame to be able to function properly. The output from the pointcloud processing node needs to be transformed from the local camera frame into the world frame. The complete TF is going to be a complex of both static and dynamic TFs.
|
1.0
|
Transform computed pose into the world frame (Pointcloud Processing) - **Time estimate:** 10 hours
**Description of task:**
The Vision-EKF needs object poses in the odom (world) frame to be able to function properly. The output from the pointcloud processing node needs to be transformed from the local camera frame into the world frame. The complete TF is going to be a complex of both static and dynamic TFs.
|
process
|
transform computed pose into the world frame pointcloud processing time estimate hours description of task the vision ekf needs object poses in the odom world frame to be able to function properly the output from the pointcloud processing node needs to be transformed from the local camera frame into the world frame the complete tf is going to be a complex of both static and dynamic tfs
| 1
|
14,276
| 17,242,752,882
|
IssuesEvent
|
2021-07-21 02:37:00
|
bisq-network/bisq
|
https://api.github.com/repos/bisq-network/bisq
|
closed
|
Delayed payout transaction is missing
|
in:trade-process was:dropped
|
<!--
SUPPORT REQUESTS: This is for reporting bugs in the Bisq app.
If you have a support request, please join #support on Bisq's
Keybase team at https://keybase.io/team/Bisq
-->
### Description
Bisq failed to create the delayed payout transaction needed to send funds to Arbitration.
<!-- brief description of the bug -->
#### Version
1.4.2
<!-- commit id or version number -->
### Steps to reproduce
Unable to reproduce
<!--if you can reliably reproduce the bug, list the steps here -->
### Expected behaviour
Bisq should create a delayed payout transaction sending funds to the Arbitrator address along with the deposit transaction.
<!--description of the expected behavior -->
### Actual behaviour
Bisq failed to create the delayed payout transaction needed to send funds to Arbitration.
<!-- explain what happened instead of the expected behaviour -->
### Screenshots
<img width="957" alt="btcBuyerTaker-errorMessage" src="https://user-images.githubusercontent.com/44140017/98450377-43890000-2134-11eb-9c97-9fdc2d29f249.PNG">
<img width="960" alt="btcBuyerTaker-tradeDetails" src="https://user-images.githubusercontent.com/44140017/98450359-15a3bb80-2134-11eb-9553-719271f2c3aa.PNG">
<!--Screenshots if gui related, drag and drop to add to the issue -->
#### Device or machine
Windows 10
<!-- device/machine used, operating system -->
|
1.0
|
Delayed payout transaction is missing - <!--
SUPPORT REQUESTS: This is for reporting bugs in the Bisq app.
If you have a support request, please join #support on Bisq's
Keybase team at https://keybase.io/team/Bisq
-->
### Description
Bisq failed to create the delayed payout transaction needed to send funds to Arbitration.
<!-- brief description of the bug -->
#### Version
1.4.2
<!-- commit id or version number -->
### Steps to reproduce
Unable to reproduce
<!--if you can reliably reproduce the bug, list the steps here -->
### Expected behaviour
Bisq should create a delayed payout transaction sending funds to the Arbitrator address along with the deposit transaction.
<!--description of the expected behavior -->
### Actual behaviour
Bisq failed to create the delayed payout transaction needed to send funds to Arbitration.
<!-- explain what happened instead of the expected behaviour -->
### Screenshots
<img width="957" alt="btcBuyerTaker-errorMessage" src="https://user-images.githubusercontent.com/44140017/98450377-43890000-2134-11eb-9c97-9fdc2d29f249.PNG">
<img width="960" alt="btcBuyerTaker-tradeDetails" src="https://user-images.githubusercontent.com/44140017/98450359-15a3bb80-2134-11eb-9553-719271f2c3aa.PNG">
<!--Screenshots if gui related, drag and drop to add to the issue -->
#### Device or machine
Windows 10
<!-- device/machine used, operating system -->
|
process
|
delayed payout transaction is missing support requests this is for reporting bugs in the bisq app if you have a support request please join support on bisq s keybase team at description bisq failed to create the delayed payout transaction needed to send funds to arbitration version steps to reproduce unable to reproduce expected behaviour bisq should create a delayed payout transaction sending funds to the arbitrator address along with the deposit transaction actual behaviour bisq failed to create the delayed payout transaction needed to send funds to arbitration screenshots img width alt btcbuyertaker errormessage src img width alt btcbuyertaker tradedetails src device or machine windows
| 1
|
21,547
| 29,865,346,995
|
IssuesEvent
|
2023-06-20 02:59:08
|
cncf/tag-security
|
https://api.github.com/repos/cncf/tag-security
|
closed
|
[SecAssess WG] Provide some consistency across reviews
|
help wanted assessment-process suggestion inactive
|
This issue was created from results of the Security Assessment Improvement Working Group (https://github.com/cncf/sig-security/issues/167#issuecomment-714514142).
# Provide additional consistency across security assessment reviews
## Premise
- Auditing imperative systems is more of an art than a science
- Threat modeling is important for projects, we should figure out how to help them be consistent with it
- Different projects have a variety in the level of detail for different sections and emphasis on different aspects
## Ideas
- Make adjustments to assessment doc structure
- Give recommendations of word length for sections
- Map assessment findings to MIRTE ATTACK or similar
- Use an existing assessment framework + template - can be tweaked for CNCF but start with that so there is a consistent process and vocabulary
- Provide qualitative ranking for projects i.e. scoring between 1 - 10
- Provide ideas/guidelines for how one should perform a review
## Logistics
- [ ] Contributors (For multiple contributors, 1 lead to coordinate)
- Placeholder_2
- [ ] SIG-Representative
|
1.0
|
[SecAssess WG] Provide some consistency across reviews - This issue was created from results of the Security Assessment Improvement Working Group (https://github.com/cncf/sig-security/issues/167#issuecomment-714514142).
# Provide additional consistency across security assessment reviews
## Premise
- Auditing imperative systems is more of an art than a science
- Threat modeling is important for projects, we should figure out how to help them be consistent with it
- Different projects have a variety in the level of detail for different sections and emphasis on different aspects
## Ideas
- Make adjustments to assessment doc structure
- Give recommendations of word length for sections
- Map assessment findings to MIRTE ATTACK or similar
- Use an existing assessment framework + template - can be tweaked for CNCF but start with that so there is a consistent process and vocabulary
- Provide qualitative ranking for projects i.e. scoring between 1 - 10
- Provide ideas/guidelines for how one should perform a review
## Logistics
- [ ] Contributors (For multiple contributors, 1 lead to coordinate)
- Placeholder_2
- [ ] SIG-Representative
|
process
|
provide some consistency across reviews this issue was created from results of the security assessment improvement working group provide additional consistency across security assessment reviews premise auditing imperative systems is more of an art than a science threat modeling is important for projects we should figure out how to help them be consistent with it different projects have a variety in the level of detail for different sections and emphasis on different aspects ideas make adjustments to assessment doc structure give recommendations of word length for sections map assessment findings to mirte attack or similar use an existing assessment framework template can be tweaked for cncf but start with that so there is a consistent process and vocabulary provide qualitative ranking for projects i e scoring between provide ideas guidelines for how one should perform a review logistics contributors for multiple contributors lead to coordinate placeholder sig representative
| 1
|
11,118
| 13,957,683,638
|
IssuesEvent
|
2020-10-24 08:08:05
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
SK: Ongoing harvesting
|
Geoportal Harvesting process SK - Slovakia
|
Dear colleagues,
in connection to the harvesting of the national catalogues and the Monitoring deadline on 15th of December, we would like to check, how will be evaluated the harvests still in the pipe on EC Geoportal harvester?
We have started the session on Wednesday 11.12.2019 20:53 (INSPIRE-da77b119-9d6e-11e7-b5a7-52540023a883), but so far, the harvesting still seems to be under the processing.
Thanks.
Martin
|
1.0
|
SK: Ongoing harvesting - Dear colleagues,
in connection to the harvesting of the national catalogues and the Monitoring deadline on 15th of December, we would like to check, how will be evaluated the harvests still in the pipe on EC Geoportal harvester?
We have started the session on Wednesday 11.12.2019 20:53 (INSPIRE-da77b119-9d6e-11e7-b5a7-52540023a883), but so far, the harvesting still seems to be under the processing.
Thanks.
Martin
|
process
|
sk ongoing harvesting dear colleagues in connection to the harvesting of the national catalogues and the monitoring deadline on of december we would like to check how will be evaluated the harvests still in the pipe on ec geoportal harvester we have started the session on wednesday inspire but so far the harvesting still seems to be under the processing thanks martin
| 1
|
789,961
| 27,811,043,825
|
IssuesEvent
|
2023-03-18 05:23:23
|
AY2223S2-CS2113-T15-4/tp
|
https://api.github.com/repos/AY2223S2-CS2113-T15-4/tp
|
closed
|
Follow the user guide input/output format more closely for "update"
|
type.Enhancement priority.High
|
* [ ] Update the input format of `update` so that users use a query instead of an index number to get the cards.
* [ ] Write developer guide.
|
1.0
|
Follow the user guide input/output format more closely for "update" - * [ ] Update the input format of `update` so that users use a query instead of an index number to get the cards.
* [ ] Write developer guide.
|
non_process
|
follow the user guide input output format more closely for update update the input format of update so that users use a query instead of an index number to get the cards write developer guide
| 0
|
7,002
| 3,933,932,686
|
IssuesEvent
|
2016-04-25 20:49:29
|
jens-maus/yam
|
https://api.github.com/repos/jens-maus/yam
|
closed
|
Warn about incompatible filetype when importing mails
|
#trivial @normal enhancement fixed GUI nightly build
|
**Originally by _mario@cattaneo.de_ on 2011-01-22 19:47:12 +0100**
___
Try to import eg. S:Startup-Sequence (I know, you have to change the filter to do so, but this is just an example ;)).
YAM will silently ignore that file, ie. do nothing. Would be nice, if there was a requester informing the user that something went wrong (incompatiible/corrupted...) with the selected file.
|
1.0
|
Warn about incompatible filetype when importing mails - **Originally by _mario@cattaneo.de_ on 2011-01-22 19:47:12 +0100**
___
Try to import eg. S:Startup-Sequence (I know, you have to change the filter to do so, but this is just an example ;)).
YAM will silently ignore that file, ie. do nothing. Would be nice, if there was a requester informing the user that something went wrong (incompatiible/corrupted...) with the selected file.
|
non_process
|
warn about incompatible filetype when importing mails originally by mario cattaneo de on try to import eg s startup sequence i know you have to change the filter to do so but this is just an example yam will silently ignore that file ie do nothing would be nice if there was a requester informing the user that something went wrong incompatiible corrupted with the selected file
| 0
|
158,077
| 20,007,575,148
|
IssuesEvent
|
2022-02-01 00:02:02
|
timf-app-sandbox/t1
|
https://api.github.com/repos/timf-app-sandbox/t1
|
opened
|
CVE-2021-3918 (High) detected in json-schema-0.2.3.tgz
|
security vulnerability
|
## CVE-2021-3918 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p></summary>
<p>JSON Schema validation and specifications</p>
<p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/npm/node_modules/request/node_modules/http-signature/node_modules/jsprim/node_modules/json-schema/package.json,/node_modules/json-schema/package.json</p>
<p>
Dependency Hierarchy:
- grunt-npm-install-0.3.1.tgz (Root Library)
- npm-3.10.10.tgz
- request-2.75.0.tgz
- http-signature-1.1.1.tgz
- jsprim-1.3.1.tgz
- :x: **json-schema-0.2.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/timf-app-sandbox/t1/commit/ade63e4d4d7f78775fe064056f81c5da1b53aecf">ade63e4d4d7f78775fe064056f81c5da1b53aecf</a></p>
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution')
<p>Publish Date: 2021-11-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918>CVE-2021-3918</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-3918">https://nvd.nist.gov/vuln/detail/CVE-2021-3918</a></p>
<p>Release Date: 2021-11-13</p>
<p>Fix Resolution: json-schema - 0.4.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"json-schema","packageVersion":"0.2.3","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;http-signature:1.1.1;jsprim:1.3.1;json-schema:0.2.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"json-schema - 0.4.0","isBinary":false}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2021-3918","vulnerabilityDetails":"json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes (\u0027Prototype Pollution\u0027)","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2021-3918 (High) detected in json-schema-0.2.3.tgz - ## CVE-2021-3918 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p></summary>
<p>JSON Schema validation and specifications</p>
<p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/npm/node_modules/request/node_modules/http-signature/node_modules/jsprim/node_modules/json-schema/package.json,/node_modules/json-schema/package.json</p>
<p>
Dependency Hierarchy:
- grunt-npm-install-0.3.1.tgz (Root Library)
- npm-3.10.10.tgz
- request-2.75.0.tgz
- http-signature-1.1.1.tgz
- jsprim-1.3.1.tgz
- :x: **json-schema-0.2.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/timf-app-sandbox/t1/commit/ade63e4d4d7f78775fe064056f81c5da1b53aecf">ade63e4d4d7f78775fe064056f81c5da1b53aecf</a></p>
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution')
<p>Publish Date: 2021-11-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918>CVE-2021-3918</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-3918">https://nvd.nist.gov/vuln/detail/CVE-2021-3918</a></p>
<p>Release Date: 2021-11-13</p>
<p>Fix Resolution: json-schema - 0.4.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"json-schema","packageVersion":"0.2.3","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;http-signature:1.1.1;jsprim:1.3.1;json-schema:0.2.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"json-schema - 0.4.0","isBinary":false}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2021-3918","vulnerabilityDetails":"json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes (\u0027Prototype Pollution\u0027)","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in json schema tgz cve high severity vulnerability vulnerable library json schema tgz json schema validation and specifications library home page a href path to dependency file package json path to vulnerable library node modules npm node modules request node modules http signature node modules jsprim node modules json schema package json node modules json schema package json dependency hierarchy grunt npm install tgz root library npm tgz request tgz http signature tgz jsprim tgz x json schema tgz vulnerable library found in head commit a href found in base branch dev vulnerability details json schema is vulnerable to improperly controlled modification of object prototype attributes prototype pollution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution json schema isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree grunt npm install npm request http signature jsprim json schema isminimumfixversionavailable true minimumfixversion json schema isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails json schema is vulnerable to improperly controlled modification of object prototype attributes pollution vulnerabilityurl
| 0
|
210,583
| 16,375,276,123
|
IssuesEvent
|
2021-05-16 00:39:43
|
RespiraWorks/Ventilator
|
https://api.github.com/repos/RespiraWorks/Ventilator
|
opened
|
RW-SYS-148 Comments carried over from Valispace
|
Documentation Question Systems
|
Relates to #803
Comments from Valispace:
if the test you are envisioning involve shorting the inhalation and exhalation limbs and plugging the output port while running the blower at max capacity maybe we can source it based on the Pb560 leak detection accuracy to something significantly less than 3 lpm +/- 20% of set pressure.
open
0 responses
Created 12 months ago by RespiraWorks Team 6 Reader
|
1.0
|
RW-SYS-148 Comments carried over from Valispace - Relates to #803
Comments from Valispace:
if the test you are envisioning involve shorting the inhalation and exhalation limbs and plugging the output port while running the blower at max capacity maybe we can source it based on the Pb560 leak detection accuracy to something significantly less than 3 lpm +/- 20% of set pressure.
open
0 responses
Created 12 months ago by RespiraWorks Team 6 Reader
|
non_process
|
rw sys comments carried over from valispace relates to comments from valispace if the test you are envisioning involve shorting the inhalation and exhalation limbs and plugging the output port while running the blower at max capacity maybe we can source it based on the leak detection accuracy to something significantly less than lpm of set pressure open responses created months ago by respiraworks team reader
| 0
|
61,304
| 14,968,661,907
|
IssuesEvent
|
2021-01-27 17:07:15
|
sandboxie-plus/Sandboxie
|
https://api.github.com/repos/sandboxie-plus/Sandboxie
|
closed
|
Problems with modal windows
|
fixed in next build
|
1) Some modal windows block main window but not shown itself:
<details>
<summary>Show spoiler (gif with problem visualization)</summary>

</details>
2) If enabled "Always on Top" all modal windows showed behind main window:
<details>
<summary>Show spoiler (gif with problem visualization)</summary>

</details>
SBIE version is 0.5.5 / 5.46.4 (64 bit)
|
1.0
|
Problems with modal windows - 1) Some modal windows block main window but not shown itself:
<details>
<summary>Show spoiler (gif with problem visualization)</summary>

</details>
2) If enabled "Always on Top" all modal windows showed behind main window:
<details>
<summary>Show spoiler (gif with problem visualization)</summary>

</details>
SBIE version is 0.5.5 / 5.46.4 (64 bit)
|
non_process
|
problems with modal windows some modal windows block main window but not shown itself show spoiler gif with problem visualization if enabled always on top all modal windows showed behind main window show spoiler gif with problem visualization sbie version is bit
| 0
|
7,539
| 10,617,857,672
|
IssuesEvent
|
2019-10-12 22:31:35
|
testingrequired/tf
|
https://api.github.com/repos/testingrequired/tf
|
closed
|
Release document
|
process
|
- Run `npm run verify`
- Run using `tf-example` and local tar file
- **tf:** `npm run pack-preview`
- **tf-example:** `npm i -D testingrequired-tf-VERSION.tgz`
- **tf-example:** `npm run test`
- Bump version e.g. `0.0.21`
- Run `npm i` to bump package lock
- Create version git tag
- Push
- Publish to `@types`
|
1.0
|
Release document - - Run `npm run verify`
- Run using `tf-example` and local tar file
- **tf:** `npm run pack-preview`
- **tf-example:** `npm i -D testingrequired-tf-VERSION.tgz`
- **tf-example:** `npm run test`
- Bump version e.g. `0.0.21`
- Run `npm i` to bump package lock
- Create version git tag
- Push
- Publish to `@types`
|
process
|
release document run npm run verify run using tf example and local tar file tf npm run pack preview tf example npm i d testingrequired tf version tgz tf example npm run test bump version e g run npm i to bump package lock create version git tag push publish to types
| 1
|
9,312
| 12,323,624,857
|
IssuesEvent
|
2020-05-13 12:29:36
|
googleapis/python-bigquery
|
https://api.github.com/repos/googleapis/python-bigquery
|
opened
|
Remove workarounds for the BQ Storage API issue with small result sets
|
status: blocked type: process
|
Context: https://github.com/googleapis/python-bigquery/pull/55#discussion_r423382586
The BigQuery Storage API has an issue with fetching small result sets, thus the BigQuery library implements a fallback to `tabledata.list` API. Since the fix for that issue on the backend will be rolled out soon, the fallback will not be necessary anymore and should be removed.
|
1.0
|
Remove workarounds for the BQ Storage API issue with small result sets - Context: https://github.com/googleapis/python-bigquery/pull/55#discussion_r423382586
The BigQuery Storage API has an issue with fetching small result sets, thus the BigQuery library implements a fallback to `tabledata.list` API. Since the fix for that issue on the backend will be rolled out soon, the fallback will not be necessary anymore and should be removed.
|
process
|
remove workarounds for the bq storage api issue with small result sets context the bigquery storage api has an issue with fetching small result sets thus the bigquery library implements a fallback to tabledata list api since the fix for that issue on the backend will be rolled out soon the fallback will not be necessary anymore and should be removed
| 1
|
8,066
| 11,242,660,294
|
IssuesEvent
|
2020-01-10 00:03:26
|
NottingHack/hms2
|
https://api.github.com/repos/NottingHack/hms2
|
closed
|
member.approval audit
|
2.1 Process suggestion
|
This requires #296 in place but then we can
- Any not yet reviewed and
- outstanding after 48 hours notify team.membership
- outstanding after 96 hours notify team.trustees
|
1.0
|
member.approval audit - This requires #296 in place but then we can
- Any not yet reviewed and
- outstanding after 48 hours notify team.membership
- outstanding after 96 hours notify team.trustees
|
process
|
member approval audit this requires in place but then we can any not yet reviewed and outstanding after hours notify team membership outstanding after hours notify team trustees
| 1
|
316,587
| 27,168,409,394
|
IssuesEvent
|
2023-02-17 17:08:14
|
MPMG-DCC-UFMG/F01
|
https://api.github.com/repos/MPMG-DCC-UFMG/F01
|
closed
|
Teste de generalizacao para a tag Receitas - Dados das Receitas - Brumadinho
|
generalization test development template - ABO (21) tag - Receitas subtag - Dados das receitas
|
DoD: Realizar o teste de Generalização do validador da tag Receitas - Dados das Receitas para o Município de Brumadinho.
|
1.0
|
Teste de generalizacao para a tag Receitas - Dados das Receitas - Brumadinho - DoD: Realizar o teste de Generalização do validador da tag Receitas - Dados das Receitas para o Município de Brumadinho.
|
non_process
|
teste de generalizacao para a tag receitas dados das receitas brumadinho dod realizar o teste de generalização do validador da tag receitas dados das receitas para o município de brumadinho
| 0
|
75,668
| 3,470,756,976
|
IssuesEvent
|
2015-12-23 10:44:29
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
opened
|
e2e flake: Services should work after restarting apiserver
|
kind/flake priority/P1 team/CSI
|
http://kubekins.dls.corp.google.com/view/Critical%20Builds/job/kubernetes-e2e-gce/10009/
```
Services should work after restarting apiserver [Disruptive]
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/service.go:351 Dec 23 00:34:07.616: error while waiting for apiserver up: waiting for apiserver timed out
```
@kubernetes/goog-csi
|
1.0
|
e2e flake: Services should work after restarting apiserver - http://kubekins.dls.corp.google.com/view/Critical%20Builds/job/kubernetes-e2e-gce/10009/
```
Services should work after restarting apiserver [Disruptive]
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/service.go:351 Dec 23 00:34:07.616: error while waiting for apiserver up: waiting for apiserver timed out
```
@kubernetes/goog-csi
|
non_process
|
flake services should work after restarting apiserver services should work after restarting apiserver go src io kubernetes output dockerized go src io kubernetes test service go dec error while waiting for apiserver up waiting for apiserver timed out kubernetes goog csi
| 0
|
141,147
| 5,429,988,038
|
IssuesEvent
|
2017-03-03 19:54:40
|
Angry-Pixel/The-Betweenlands
|
https://api.github.com/repos/Angry-Pixel/The-Betweenlands
|
closed
|
Improve first person decay renderer
|
1.10 High Priority Incompatibility
|
Currently overrides the hand model which causes issues with Optifine and likely also other mods that modify the player model
|
1.0
|
Improve first person decay renderer - Currently overrides the hand model which causes issues with Optifine and likely also other mods that modify the player model
|
non_process
|
improve first person decay renderer currently overrides the hand model which causes issues with optifine and likely also other mods that modify the player model
| 0
|
15,501
| 19,703,263,239
|
IssuesEvent
|
2022-01-12 18:52:05
|
googleapis/java-analytics-data
|
https://api.github.com/repos/googleapis/java-analytics-data
|
opened
|
Your .repo-metadata.json file has a problem 🤒
|
type: process repo-metadata: lint
|
You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* release_level must be equal to one of the allowed values in .repo-metadata.json
* api_shortname 'analytics-data' invalid in .repo-metadata.json
☝️ Once you correct these problems, you can close this issue.
Reach out to **go/github-automation** if you have any questions.
|
1.0
|
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* release_level must be equal to one of the allowed values in .repo-metadata.json
* api_shortname 'analytics-data' invalid in .repo-metadata.json
☝️ Once you correct these problems, you can close this issue.
Reach out to **go/github-automation** if you have any questions.
|
process
|
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 release level must be equal to one of the allowed values in repo metadata json api shortname analytics data invalid in repo metadata json ☝️ once you correct these problems you can close this issue reach out to go github automation if you have any questions
| 1
|
21,674
| 30,120,032,075
|
IssuesEvent
|
2023-06-30 14:29:03
|
microsoft/vscode
|
https://api.github.com/repos/microsoft/vscode
|
closed
|
Getting terminal revive environments is currently slow
|
bug perf terminal-process
|
This has many async calls, some could probably be combined and made parallel
https://github.com/microsoft/vscode/blob/6314fae1a37738b6555c8748af9ed000606f1788/src/vs/workbench/contrib/terminal/electron-sandbox/localTerminalBackend.ts#L281-L286
|
1.0
|
Getting terminal revive environments is currently slow - This has many async calls, some could probably be combined and made parallel
https://github.com/microsoft/vscode/blob/6314fae1a37738b6555c8748af9ed000606f1788/src/vs/workbench/contrib/terminal/electron-sandbox/localTerminalBackend.ts#L281-L286
|
process
|
getting terminal revive environments is currently slow this has many async calls some could probably be combined and made parallel
| 1
|
80,303
| 3,560,610,290
|
IssuesEvent
|
2016-01-23 05:59:56
|
onyxfish/agate
|
https://api.github.com/repos/onyxfish/agate
|
closed
|
Integers are being rendered with an unnecessary decimal point (Table.to_csv)
|
bug priority-normal
|
```
text,date,integer,boolean,float,datetime,empty_column,
Chicago Reader,1971-01-01,40.0,True,1.0,1971-01-01 04:14:00,,
Chicago Sun-Times,1948-01-01,63.0,True,1.27,1948-01-01 14:57:13,,Extra data beyond headers will be trimmed
Chicago Tribune,1920-01-01,164.0,False,41800000.01,1920-01-01,,
This row has blanks,,,,,,,
Unicode! Σ,,,,,,,
```
|
1.0
|
Integers are being rendered with an unnecessary decimal point (Table.to_csv) - ```
text,date,integer,boolean,float,datetime,empty_column,
Chicago Reader,1971-01-01,40.0,True,1.0,1971-01-01 04:14:00,,
Chicago Sun-Times,1948-01-01,63.0,True,1.27,1948-01-01 14:57:13,,Extra data beyond headers will be trimmed
Chicago Tribune,1920-01-01,164.0,False,41800000.01,1920-01-01,,
This row has blanks,,,,,,,
Unicode! Σ,,,,,,,
```
|
non_process
|
integers are being rendered with an unnecessary decimal point table to csv text date integer boolean float datetime empty column chicago reader true chicago sun times true extra data beyond headers will be trimmed chicago tribune false this row has blanks unicode σ
| 0
|
256,208
| 8,127,034,449
|
IssuesEvent
|
2018-08-17 06:16:27
|
aowen87/BAR
|
https://api.github.com/repos/aowen87/BAR
|
closed
|
Ultra Wrapper Documentation
|
Expected Use: 3 - Occasional Feature Impact: 3 - Medium Priority: Normal Support Group: DOE/ASC
|
Hank Shay requests documentation on the ultra wrapper, specifically which subsets of ultra are implemented.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 174
Status: Resolved
Project: VisIt
Tracker: Feature
Priority: Normal
Subject: Ultra Wrapper Documentation
Assigned to: Kathleen Biagas
Category:
Target version: 2.1
Author: Cyrus Harrison
Start: 06/22/2010
Due date:
% Done: 0
Estimated time:
Created: 06/22/2010 01:13 pm
Updated: 09/01/2010 11:51 am
Likelihood:
Severity:
Found in version:
Impact: 3 - Medium
Expected Use: 3 - Occasional
OS: All
Support Group: DOE/ASC
Description:
Hank Shay requests documentation on the ultra wrapper, specifically which subsets of ultra are implemented.
Comments:
Assignment from LLNL VisIt 2.1 Release Meeting
Added to VisIt's help.SVN revision 12353.
|
1.0
|
Ultra Wrapper Documentation - Hank Shay requests documentation on the ultra wrapper, specifically which subsets of ultra are implemented.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 174
Status: Resolved
Project: VisIt
Tracker: Feature
Priority: Normal
Subject: Ultra Wrapper Documentation
Assigned to: Kathleen Biagas
Category:
Target version: 2.1
Author: Cyrus Harrison
Start: 06/22/2010
Due date:
% Done: 0
Estimated time:
Created: 06/22/2010 01:13 pm
Updated: 09/01/2010 11:51 am
Likelihood:
Severity:
Found in version:
Impact: 3 - Medium
Expected Use: 3 - Occasional
OS: All
Support Group: DOE/ASC
Description:
Hank Shay requests documentation on the ultra wrapper, specifically which subsets of ultra are implemented.
Comments:
Assignment from LLNL VisIt 2.1 Release Meeting
Added to VisIt's help.SVN revision 12353.
|
non_process
|
ultra wrapper documentation hank shay requests documentation on the ultra wrapper specifically which subsets of ultra are implemented redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker feature priority normal subject ultra wrapper documentation assigned to kathleen biagas category target version author cyrus harrison start due date done estimated time created pm updated am likelihood severity found in version impact medium expected use occasional os all support group doe asc description hank shay requests documentation on the ultra wrapper specifically which subsets of ultra are implemented comments assignment from llnl visit release meeting added to visit s help svn revision
| 0
|
21,532
| 29,828,452,941
|
IssuesEvent
|
2023-06-18 00:42:02
|
devssa/onde-codar-em-salvador
|
https://api.github.com/repos/devssa/onde-codar-em-salvador
|
closed
|
[Hibrido / Belo Horizonte, Minas Gerais, Brazil] Office 365 Developer na Coodesh
|
SALVADOR FRONT-END PJ JAVASCRIPT FULL-STACK HTML REQUISITOS PROCESSOS GITHUB Sketch ADOBE XD UMA SHAREPOINT AUTOMAÇÃO DE PROCESSOS ALOCADO Stale
|
## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/jobs/office-365-developer-194401047?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>O TSX Group está em busca de Office 365 Developer com experiência em Sharepoint e design front-end para se juntar à nossa equipe.</p>
<p>A especialização das Unidades de Negócio do TSX Group constitui o cerne do diferencial competitivo de suas atribuições empresariais com a geração de valor para clientes, parceiros e sociedade a partir de uma atuação multidisciplinar e integrada. 30 anos de experiência orientam processos lógicos e maduros para soluções e performance de resultados com alto grau de maturidade no tratamento sistêmico de projetos.</p>
<p>O desafio é estruturar a intranet da empresa, sendo responsável por criar toda a estrutura da intranet, cuidar da gestão do Sharepoint e também da parte de design frontend.</p>
## TSX Group:
<p>A especialização das Unidades de Negócio do TSX Group constitui o cerne do diferencial competitivo de suas atribuições empresariais com a geração de valor para clientes, parceiros e sociedade a partir de uma atuação multidisciplinar e integrada.</p>
<p><strong>30 anos de experiência</strong> orientam processos lógicos e maduros para soluções e performance de resultados com alto grau de maturidade no tratamento sistêmico de projetos.</p><a href='https://coodesh.com/companies/tsx-group'>Veja mais no site</a>
## Habilidades:
- SharePoint
- HTML
- CSS
- Javascript
- Adobe XD
## Local:
Belo Horizonte, Minas Gerais, Brazil
## Requisitos:
- Experiência comprovada em desenvolvimento de intranet usando ferramentas do Office 365, em especial o Sharepoint;
- Conhecimento sólido em tecnologias front-end, como HTML, CSS, JavaScript;
- Familiaridade com ferramentas de design, como Adobe XD ou Sketch;
- Experiência em automação de processos usando o Microsoft Power Automate ou outras ferramentas de automação;
- Habilidades de comunicação e trabalho em equipe;
- Criatividade e proatividade são essenciais.
## Diferenciais:
- Experiência em customização e desenvolvimento de soluções usando Sharepoint Framework (SPFx).
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Office 365 Developer na TSX Group](https://coodesh.com/jobs/office-365-developer-194401047?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Alocado
#### Regime
PJ
#### Categoria
Full-Stack
|
2.0
|
[Hibrido / Belo Horizonte, Minas Gerais, Brazil] Office 365 Developer na Coodesh - ## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/jobs/office-365-developer-194401047?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>O TSX Group está em busca de Office 365 Developer com experiência em Sharepoint e design front-end para se juntar à nossa equipe.</p>
<p>A especialização das Unidades de Negócio do TSX Group constitui o cerne do diferencial competitivo de suas atribuições empresariais com a geração de valor para clientes, parceiros e sociedade a partir de uma atuação multidisciplinar e integrada. 30 anos de experiência orientam processos lógicos e maduros para soluções e performance de resultados com alto grau de maturidade no tratamento sistêmico de projetos.</p>
<p>O desafio é estruturar a intranet da empresa, sendo responsável por criar toda a estrutura da intranet, cuidar da gestão do Sharepoint e também da parte de design frontend.</p>
## TSX Group:
<p>A especialização das Unidades de Negócio do TSX Group constitui o cerne do diferencial competitivo de suas atribuições empresariais com a geração de valor para clientes, parceiros e sociedade a partir de uma atuação multidisciplinar e integrada.</p>
<p><strong>30 anos de experiência</strong> orientam processos lógicos e maduros para soluções e performance de resultados com alto grau de maturidade no tratamento sistêmico de projetos.</p><a href='https://coodesh.com/companies/tsx-group'>Veja mais no site</a>
## Habilidades:
- SharePoint
- HTML
- CSS
- Javascript
- Adobe XD
## Local:
Belo Horizonte, Minas Gerais, Brazil
## Requisitos:
- Experiência comprovada em desenvolvimento de intranet usando ferramentas do Office 365, em especial o Sharepoint;
- Conhecimento sólido em tecnologias front-end, como HTML, CSS, JavaScript;
- Familiaridade com ferramentas de design, como Adobe XD ou Sketch;
- Experiência em automação de processos usando o Microsoft Power Automate ou outras ferramentas de automação;
- Habilidades de comunicação e trabalho em equipe;
- Criatividade e proatividade são essenciais.
## Diferenciais:
- Experiência em customização e desenvolvimento de soluções usando Sharepoint Framework (SPFx).
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Office 365 Developer na TSX Group](https://coodesh.com/jobs/office-365-developer-194401047?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Alocado
#### Regime
PJ
#### Categoria
Full-Stack
|
process
|
office developer na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 o tsx group está em busca de office developer com experiência em sharepoint e design front end para se juntar à nossa equipe a especialização das unidades de negócio do tsx group constitui o cerne do diferencial competitivo de suas atribuições empresariais com a geração de valor para clientes parceiros e sociedade a partir de uma atuação multidisciplinar e integrada anos de experiência orientam processos lógicos e maduros para soluções e performance de resultados com alto grau de maturidade no tratamento sistêmico de projetos o desafio é estruturar a intranet da empresa sendo responsável por criar toda a estrutura da intranet cuidar da gestão do sharepoint e também da parte de design frontend tsx group a especialização das unidades de negócio do tsx group constitui o cerne do diferencial competitivo de suas atribuições empresariais com a geração de valor para clientes parceiros e sociedade a partir de uma atuação multidisciplinar e integrada anos de experiência orientam processos lógicos e maduros para soluções e performance de resultados com alto grau de maturidade no tratamento sistêmico de projetos habilidades sharepoint html css javascript adobe xd local belo horizonte minas gerais brazil requisitos experiência comprovada em desenvolvimento de intranet usando ferramentas do office em especial o sharepoint conhecimento sólido em tecnologias front end como html css javascript familiaridade com ferramentas de design como adobe xd ou sketch experiência em automação de processos usando o microsoft power automate ou outras ferramentas de automação habilidades de comunicação e trabalho em equipe criatividade e proatividade são essenciais diferenciais experiência em customização e desenvolvimento de soluções usando sharepoint framework spfx como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação alocado regime pj categoria full stack
| 1
|
100,549
| 11,198,788,853
|
IssuesEvent
|
2020-01-03 16:50:50
|
18F/before-you-ship
|
https://api.github.com/repos/18F/before-you-ship
|
closed
|
Instructions on how to go between ATO types
|
Feedback Needed epic: ATO Improvements epic: documentation
|
## Problem Statement:
- The 1 Year is inclusive of all the ATO steps.
- Also
- additional sections of the SSP must be completed
- additional controls are assessed
- a more thorough penetration test is done which may reveal other vulnerabilities
- all need to be resolved before the ATO is granted
## Background information:
This could use more detail, as we've had the question of how to go from a 90-day to a 1-year/3-year come up several times recently.
## Actions to take:
- [x] Understand the ATO types
- [ ] Figure out the roles and responsibilities
- [ ] Set up conversation with GSA IT Security
## Acceptance criteria:
- [ ] Documentation about how to go between is written
## Supporting Documentation:
[The Types of ATO page](https://pages.18f.gov/before-you-ship/ato/types/#gsa-lato)
## Related Issues:
|
1.0
|
Instructions on how to go between ATO types - ## Problem Statement:
- The 1 Year is inclusive of all the ATO steps.
- Also
- additional sections of the SSP must be completed
- additional controls are assessed
- a more thorough penetration test is done which may reveal other vulnerabilities
- all need to be resolved before the ATO is granted
## Background information:
This could use more detail, as we've had the question of how to go from a 90-day to a 1-year/3-year come up several times recently.
## Actions to take:
- [x] Understand the ATO types
- [ ] Figure out the roles and responsibilities
- [ ] Set up conversation with GSA IT Security
## Acceptance criteria:
- [ ] Documentation about how to go between is written
## Supporting Documentation:
[The Types of ATO page](https://pages.18f.gov/before-you-ship/ato/types/#gsa-lato)
## Related Issues:
|
non_process
|
instructions on how to go between ato types problem statement the year is inclusive of all the ato steps also additional sections of the ssp must be completed additional controls are assessed a more thorough penetration test is done which may reveal other vulnerabilities all need to be resolved before the ato is granted background information this could use more detail as we ve had the question of how to go from a day to a year year come up several times recently actions to take understand the ato types figure out the roles and responsibilities set up conversation with gsa it security acceptance criteria documentation about how to go between is written supporting documentation related issues
| 0
|
13,265
| 15,730,852,536
|
IssuesEvent
|
2021-03-29 16:23:01
|
kubernetes/minikube
|
https://api.github.com/repos/kubernetes/minikube
|
closed
|
Bump metrics-server to v0.4.2
|
area/addons kind/process
|
Minikube has [metrics-server](https://github.com/kubernetes-sigs/metrics-server) addons.
Its addon has v0.2.1 image tag. It's very old, because it was published by 2017.
I'll send PR to bump metrics-server addon to v0.4.2(latest).
https://github.com/kubernetes-sigs/metrics-server/releases/tag/v0.4.2
/area addons
/kind process
|
1.0
|
Bump metrics-server to v0.4.2 - Minikube has [metrics-server](https://github.com/kubernetes-sigs/metrics-server) addons.
Its addon has v0.2.1 image tag. It's very old, because it was published by 2017.
I'll send PR to bump metrics-server addon to v0.4.2(latest).
https://github.com/kubernetes-sigs/metrics-server/releases/tag/v0.4.2
/area addons
/kind process
|
process
|
bump metrics server to minikube has addons its addon has image tag it s very old because it was published by i ll send pr to bump metrics server addon to latest area addons kind process
| 1
|
94,383
| 3,925,156,907
|
IssuesEvent
|
2016-04-22 17:52:53
|
chrisblakley/Nebula
|
https://api.github.com/repos/chrisblakley/Nebula
|
opened
|
Re-check GA user timings for load time
|
Bug Frontend (Script) Low Priority
|
Not sure these are adding up properly. Seeing numbers in the 2 millions for perceived load.
|
1.0
|
Re-check GA user timings for load time - Not sure these are adding up properly. Seeing numbers in the 2 millions for perceived load.
|
non_process
|
re check ga user timings for load time not sure these are adding up properly seeing numbers in the millions for perceived load
| 0
|
223,033
| 17,535,814,899
|
IssuesEvent
|
2021-08-12 06:14:28
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
roachtest: backup/nodeShutdown/coordinator/n4cpu4 failed
|
C-test-failure O-robot O-roachtest release-blocker branch-release-21.1
|
roachtest.backup/nodeShutdown/coordinator/n4cpu4 [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3290967&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3290967&tab=artifacts#/backup/nodeShutdown/coordinator/n4cpu4) on release-21.1 @ [c4d0e7baee3925541eed599ae771abb95c97732b](https://github.com/cockroachdb/cockroach/commits/c4d0e7baee3925541eed599ae771abb95c97732b):
```
The test failed on branch=release-21.1, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/backup/nodeShutdown/coordinator/n4cpu4/run_1
jobs.go:131,backup.go:132,test_runner.go:733: unexpectedly found job 683797475003269122 in state failed
(1) attached stack trace
-- stack trace:
| main.jobSurvivesNodeShutdown.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/jobs.go:79
| main.(*monitor).Go.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2666
| golang.org/x/sync/errgroup.(*Group).Go.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup/errgroup.go:57
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1371
Wraps: (2) unexpectedly found job 683797475003269122 in state failed
Error types: (1) *withstack.withStack (2) *errutil.leafError
cluster.go:1667,context.go:89,cluster.go:1656,test_runner.go:820: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-3290967-1628748601-01-n4cpu4 --oneshot --ignore-empty-nodes: exit status 1 2: dead (exit status 137)
4: 10659
3: 10607
1: 11767
Error: UNCLASSIFIED_PROBLEM: 2: dead (exit status 137)
(1) UNCLASSIFIED_PROBLEM
Wraps: (2) attached stack trace
-- stack trace:
| main.glob..func14
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1147
| main.wrap.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:271
| github.com/spf13/cobra.(*Command).execute
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:830
| github.com/spf13/cobra.(*Command).ExecuteC
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:914
| github.com/spf13/cobra.(*Command).Execute
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:864
| main.main
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1889
| runtime.main
| /usr/local/go/src/runtime/proc.go:225
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1371
Wraps: (3) 2: dead (exit status 137)
Error types: (1) errors.Unclassified (2) *withstack.withStack (3) *errutil.leafError
```
<details><summary>Reproduce</summary>
<p>
<p>To reproduce, try:
```bash
# From https://go.crdb.dev/p/roachstress, perhaps edited lightly.
caffeinate ./roachstress.sh backup/nodeShutdown/coordinator/n4cpu4
```
</p>
</p>
</details>
/cc @cockroachdb/bulk-io
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*backup/nodeShutdown/coordinator/n4cpu4.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
2.0
|
roachtest: backup/nodeShutdown/coordinator/n4cpu4 failed - roachtest.backup/nodeShutdown/coordinator/n4cpu4 [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3290967&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3290967&tab=artifacts#/backup/nodeShutdown/coordinator/n4cpu4) on release-21.1 @ [c4d0e7baee3925541eed599ae771abb95c97732b](https://github.com/cockroachdb/cockroach/commits/c4d0e7baee3925541eed599ae771abb95c97732b):
```
The test failed on branch=release-21.1, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/backup/nodeShutdown/coordinator/n4cpu4/run_1
jobs.go:131,backup.go:132,test_runner.go:733: unexpectedly found job 683797475003269122 in state failed
(1) attached stack trace
-- stack trace:
| main.jobSurvivesNodeShutdown.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/jobs.go:79
| main.(*monitor).Go.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2666
| golang.org/x/sync/errgroup.(*Group).Go.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup/errgroup.go:57
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1371
Wraps: (2) unexpectedly found job 683797475003269122 in state failed
Error types: (1) *withstack.withStack (2) *errutil.leafError
cluster.go:1667,context.go:89,cluster.go:1656,test_runner.go:820: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-3290967-1628748601-01-n4cpu4 --oneshot --ignore-empty-nodes: exit status 1 2: dead (exit status 137)
4: 10659
3: 10607
1: 11767
Error: UNCLASSIFIED_PROBLEM: 2: dead (exit status 137)
(1) UNCLASSIFIED_PROBLEM
Wraps: (2) attached stack trace
-- stack trace:
| main.glob..func14
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1147
| main.wrap.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:271
| github.com/spf13/cobra.(*Command).execute
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:830
| github.com/spf13/cobra.(*Command).ExecuteC
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:914
| github.com/spf13/cobra.(*Command).Execute
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:864
| main.main
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1889
| runtime.main
| /usr/local/go/src/runtime/proc.go:225
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1371
Wraps: (3) 2: dead (exit status 137)
Error types: (1) errors.Unclassified (2) *withstack.withStack (3) *errutil.leafError
```
<details><summary>Reproduce</summary>
<p>
<p>To reproduce, try:
```bash
# From https://go.crdb.dev/p/roachstress, perhaps edited lightly.
caffeinate ./roachstress.sh backup/nodeShutdown/coordinator/n4cpu4
```
</p>
</p>
</details>
/cc @cockroachdb/bulk-io
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*backup/nodeShutdown/coordinator/n4cpu4.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
non_process
|
roachtest backup nodeshutdown coordinator failed roachtest backup nodeshutdown coordinator with on release the test failed on branch release cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts backup nodeshutdown coordinator run jobs go backup go test runner go unexpectedly found job in state failed attached stack trace stack trace main jobsurvivesnodeshutdown home agent work go src github com cockroachdb cockroach pkg cmd roachtest jobs go main monitor go home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go golang org x sync errgroup group go home agent work go src github com cockroachdb cockroach vendor golang org x sync errgroup errgroup go runtime goexit usr local go src runtime asm s wraps unexpectedly found job in state failed error types withstack withstack errutil leaferror cluster go context go cluster go test runner go dead node detection home agent work go src github com cockroachdb cockroach bin roachprod monitor teamcity oneshot ignore empty nodes exit status dead exit status error unclassified problem dead exit status unclassified problem wraps attached stack trace stack trace main glob home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go main wrap home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cobra command executec home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go main main home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps dead exit status error types errors unclassified withstack withstack errutil leaferror reproduce to reproduce try bash from perhaps edited lightly caffeinate roachstress sh backup nodeshutdown coordinator cc cockroachdb bulk io
| 0
|
252,565
| 19,031,148,465
|
IssuesEvent
|
2021-11-24 10:48:26
|
openshift/okd
|
https://api.github.com/repos/openshift/okd
|
reopened
|
OKD on OpenStack docs refer to RHOSP only
|
documentation kind/documentation
|
Hello, I'm not quite sure if this is the right place to ask, but I'm bugged with something. I have an OpenStack (Mitaka release) cluster, currently running OKD 3.11. I would like to updagre to OKD 4.6 but the documentation only says of RHOSP. So I'm not sure, can I OKD 4.6 onto Mitaka or a newer release or do I need RHOSP?
|
2.0
|
OKD on OpenStack docs refer to RHOSP only - Hello, I'm not quite sure if this is the right place to ask, but I'm bugged with something. I have an OpenStack (Mitaka release) cluster, currently running OKD 3.11. I would like to updagre to OKD 4.6 but the documentation only says of RHOSP. So I'm not sure, can I OKD 4.6 onto Mitaka or a newer release or do I need RHOSP?
|
non_process
|
okd on openstack docs refer to rhosp only hello i m not quite sure if this is the right place to ask but i m bugged with something i have an openstack mitaka release cluster currently running okd i would like to updagre to okd but the documentation only says of rhosp so i m not sure can i okd onto mitaka or a newer release or do i need rhosp
| 0
|
229,979
| 17,597,575,092
|
IssuesEvent
|
2021-08-17 07:45:53
|
appnaodo-3/create-component
|
https://api.github.com/repos/appnaodo-3/create-component
|
opened
|
Slide of image and some requirement to deploy and checking
|
documentation
|
As a developer, some time i meet the slider of images, it is same that slack.
Requirement:
- Browser: chrome, safari of pc and mobile
- Automatic testing and report it to email
|
1.0
|
Slide of image and some requirement to deploy and checking - As a developer, some time i meet the slider of images, it is same that slack.
Requirement:
- Browser: chrome, safari of pc and mobile
- Automatic testing and report it to email
|
non_process
|
slide of image and some requirement to deploy and checking as a developer some time i meet the slider of images it is same that slack requirement browser chrome safari of pc and mobile automatic testing and report it to email
| 0
|
430,146
| 30,118,767,039
|
IssuesEvent
|
2023-06-30 13:38:09
|
timescale/docs
|
https://api.github.com/repos/timescale/docs
|
closed
|
[Feedback] Page: /use-timescale/latest/ingest-data/ingest-prometheus/
|
documentation community feedback
|
### Is it easy to find the information you need?
No
### Are the instructions clear?
No
### How could we improve the Timescale documentation site?
uh this has no details on how to get data scraped from a prometheus compatible endpoint into timescaledb. it just says what prometheus is
|
1.0
|
[Feedback] Page: /use-timescale/latest/ingest-data/ingest-prometheus/ - ### Is it easy to find the information you need?
No
### Are the instructions clear?
No
### How could we improve the Timescale documentation site?
uh this has no details on how to get data scraped from a prometheus compatible endpoint into timescaledb. it just says what prometheus is
|
non_process
|
page use timescale latest ingest data ingest prometheus is it easy to find the information you need no are the instructions clear no how could we improve the timescale documentation site uh this has no details on how to get data scraped from a prometheus compatible endpoint into timescaledb it just says what prometheus is
| 0
|
19,901
| 26,354,981,545
|
IssuesEvent
|
2023-01-11 09:01:58
|
googleapis/google-cloud-php
|
https://api.github.com/repos/googleapis/google-cloud-php
|
opened
|
Your .repo-metadata.json files have a problem 🤒
|
type: process repo-metadata: lint
|
You have a problem with your .repo-metadata.json files:
Result of scan 📈:
* release_level must be equal to one of the allowed values in AccessApproval/.repo-metadata.json
* api_shortname field missing from AccessApproval/.repo-metadata.json
* release_level must be equal to one of the allowed values in AccessContextManager/.repo-metadata.json
* api_shortname field missing from AccessContextManager/.repo-metadata.json
* release_level must be equal to one of the allowed values in AiPlatform/.repo-metadata.json
* api_shortname field missing from AiPlatform/.repo-metadata.json
* release_level must be equal to one of the allowed values in AnalyticsAdmin/.repo-metadata.json
* api_shortname field missing from AnalyticsAdmin/.repo-metadata.json
* release_level must be equal to one of the allowed values in AnalyticsData/.repo-metadata.json
* api_shortname field missing from AnalyticsData/.repo-metadata.json
* release_level must be equal to one of the allowed values in ApiGateway/.repo-metadata.json
* api_shortname field missing from ApiGateway/.repo-metadata.json
* release_level must be equal to one of the allowed values in ApiKeys/.repo-metadata.json
* api_shortname field missing from ApiKeys/.repo-metadata.json
* release_level must be equal to one of the allowed values in ApigeeConnect/.repo-metadata.json
* api_shortname field missing from ApigeeConnect/.repo-metadata.json
* release_level must be equal to one of the allowed values in ApigeeRegistry/.repo-metadata.json
* api_shortname field missing from ApigeeRegistry/.repo-metadata.json
* release_level must be equal to one of the allowed values in AppEngineAdmin/.repo-metadata.json
* api_shortname field missing from AppEngineAdmin/.repo-metadata.json
* release_level must be equal to one of the allowed values in ArtifactRegistry/.repo-metadata.json
* api_shortname field missing from ArtifactRegistry/.repo-metadata.json
* release_level must be equal to one of the allowed values in Asset/.repo-metadata.json
* api_shortname field missing from Asset/.repo-metadata.json
* release_level must be equal to one of the allowed values in AssuredWorkloads/.repo-metadata.json
* api_shortname field missing from AssuredWorkloads/.repo-metadata.json
* release_level must be equal to one of the allowed values in AutoMl/.repo-metadata.json
* api_shortname field missing from AutoMl/.repo-metadata.json
* release_level must be equal to one of the allowed values in BareMetalSolution/.repo-metadata.json
* api_shortname field missing from BareMetalSolution/.repo-metadata.json
* release_level must be equal to one of the allowed values in Batch/.repo-metadata.json
* api_shortname field missing from Batch/.repo-metadata.json
* release_level must be equal to one of the allowed values in BeyondCorpAppConnections/.repo-metadata.json
* api_shortname field missing from BeyondCorpAppConnections/.repo-metadata.json
* release_level must be equal to one of the allowed values in BeyondCorpAppConnectors/.repo-metadata.json
* api_shortname field missing from BeyondCorpAppConnectors/.repo-metadata.json
* release_level must be equal to one of the allowed values in BeyondCorpAppGateways/.repo-metadata.json
* api_shortname field missing from BeyondCorpAppGateways/.repo-metadata.json
* release_level must be equal to one of the allowed values in BeyondCorpClientConnectorServices/.repo-metadata.json
* api_shortname field missing from BeyondCorpClientConnectorServices/.repo-metadata.json
* release_level must be equal to one of the allowed values in BeyondCorpClientGateways/.repo-metadata.json
* api_shortname field missing from BeyondCorpClientGateways/.repo-metadata.json
* release_level must be equal to one of the allowed values in BigQuery/.repo-metadata.json
* api_shortname field missing from BigQuery/.repo-metadata.json
* release_level must be equal to one of the allowed values in BigQueryAnalyticsHub/.repo-metadata.json
* api_shortname field missing from BigQueryAnalyticsHub/.repo-metadata.json
* release_level must be equal to one of the allowed values in BigQueryConnection/.repo-metadata.json
* api_shortname field missing from BigQueryConnection/.repo-metadata.json
* release_level must be equal to one of the allowed values in BigQueryDataExchange/.repo-metadata.json
* api_shortname field missing from BigQueryDataExchange/.repo-metadata.json
* release_level must be equal to one of the allowed values in BigQueryDataPolicies/.repo-metadata.json
* api_shortname field missing from BigQueryDataPolicies/.repo-metadata.json
* release_level must be equal to one of the allowed values in BigQueryDataTransfer/.repo-metadata.json
* api_shortname field missing from BigQueryDataTransfer/.repo-metadata.json
* release_level must be equal to one of the allowed values in BigQueryMigration/.repo-metadata.json
* api_shortname field missing from BigQueryMigration/.repo-metadata.json
* release_level must be equal to one of the allowed values in BigQueryReservation/.repo-metadata.json
* api_shortname field missing from BigQueryReservation/.repo-metadata.json
* release_level must be equal to one of the allowed values in BigQueryStorage/.repo-metadata.json
* api_shortname field missing from BigQueryStorage/.repo-metadata.json
* release_level must be equal to one of the allowed values in Bigtable/.repo-metadata.json
* api_shortname field missing from Bigtable/.repo-metadata.json
* release_level must be equal to one of the allowed values in Billing/.repo-metadata.json
* api_shortname field missing from Billing/.repo-metadata.json
* release_level must be equal to one of the allowed values in BillingBudgets/.repo-metadata.json
* api_shortname field missing from BillingBudgets/.repo-metadata.json
* release_level must be equal to one of the allowed values in BinaryAuthorization/.repo-metadata.json
* api_shortname field missing from BinaryAuthorization/.repo-metadata.json
* release_level must be equal to one of the allowed values in Build/.repo-metadata.json
* api_shortname field missing from Build/.repo-metadata.json
* release_level must be equal to one of the allowed values in CertificateManager/.repo-metadata.json
* api_shortname field missing from CertificateManager/.repo-metadata.json
* release_level must be equal to one of the allowed values in Channel/.repo-metadata.json
* api_shortname field missing from Channel/.repo-metadata.json
* release_level must be equal to one of the allowed values in CommonProtos/.repo-metadata.json
* release_level must be equal to one of the allowed values in Compute/.repo-metadata.json
* api_shortname field missing from Compute/.repo-metadata.json
* release_level must be equal to one of the allowed values in ContactCenterInsights/.repo-metadata.json
* api_shortname field missing from ContactCenterInsights/.repo-metadata.json
* release_level must be equal to one of the allowed values in Container/.repo-metadata.json
* api_shortname field missing from Container/.repo-metadata.json
* release_level must be equal to one of the allowed values in ContainerAnalysis/.repo-metadata.json
* api_shortname field missing from ContainerAnalysis/.repo-metadata.json
* release_level must be equal to one of the allowed values in Core/.repo-metadata.json
* release_level must be equal to one of the allowed values in DataCatalog/.repo-metadata.json
* api_shortname field missing from DataCatalog/.repo-metadata.json
* release_level must be equal to one of the allowed values in DataFusion/.repo-metadata.json
* api_shortname field missing from DataFusion/.repo-metadata.json
* release_level must be equal to one of the allowed values in DataLabeling/.repo-metadata.json
* api_shortname field missing from DataLabeling/.repo-metadata.json
* release_level must be equal to one of the allowed values in Dataflow/.repo-metadata.json
* api_shortname field missing from Dataflow/.repo-metadata.json
* release_level must be equal to one of the allowed values in Dataform/.repo-metadata.json
* api_shortname field missing from Dataform/.repo-metadata.json
* release_level must be equal to one of the allowed values in Dataplex/.repo-metadata.json
* api_shortname field missing from Dataplex/.repo-metadata.json
* release_level must be equal to one of the allowed values in Dataproc/.repo-metadata.json
* api_shortname field missing from Dataproc/.repo-metadata.json
* release_level must be equal to one of the allowed values in DataprocMetastore/.repo-metadata.json
* api_shortname field missing from DataprocMetastore/.repo-metadata.json
* release_level must be equal to one of the allowed values in Datastore/.repo-metadata.json
* api_shortname field missing from Datastore/.repo-metadata.json
* release_level must be equal to one of the allowed values in DatastoreAdmin/.repo-metadata.json
* api_shortname field missing from DatastoreAdmin/.repo-metadata.json
* release_level must be equal to one of the allowed values in Datastream/.repo-metadata.json
* api_shortname field missing from Datastream/.repo-metadata.json
* release_level must be equal to one of the allowed values in Debugger/.repo-metadata.json
* api_shortname field missing from Debugger/.repo-metadata.json
* release_level must be equal to one of the allowed values in Deploy/.repo-metadata.json
* api_shortname field missing from Deploy/.repo-metadata.json
* release_level must be equal to one of the allowed values in Dialogflow/.repo-metadata.json
* api_shortname field missing from Dialogflow/.repo-metadata.json
* release_level must be equal to one of the allowed values in Dlp/.repo-metadata.json
* api_shortname field missing from Dlp/.repo-metadata.json
* release_level must be equal to one of the allowed values in Dms/.repo-metadata.json
* api_shortname field missing from Dms/.repo-metadata.json
* release_level must be equal to one of the allowed values in DocumentAi/.repo-metadata.json
* api_shortname field missing from DocumentAi/.repo-metadata.json
* release_level must be equal to one of the allowed values in Domains/.repo-metadata.json
* api_shortname field missing from Domains/.repo-metadata.json
* release_level must be equal to one of the allowed values in ErrorReporting/.repo-metadata.json
* api_shortname field missing from ErrorReporting/.repo-metadata.json
* release_level must be equal to one of the allowed values in EssentialContacts/.repo-metadata.json
* api_shortname field missing from EssentialContacts/.repo-metadata.json
* release_level must be equal to one of the allowed values in Eventarc/.repo-metadata.json
* api_shortname field missing from Eventarc/.repo-metadata.json
* release_level must be equal to one of the allowed values in EventarcPublishing/.repo-metadata.json
* api_shortname field missing from EventarcPublishing/.repo-metadata.json
* release_level must be equal to one of the allowed values in Filestore/.repo-metadata.json
* api_shortname field missing from Filestore/.repo-metadata.json
* release_level must be equal to one of the allowed values in Firestore/.repo-metadata.json
* api_shortname field missing from Firestore/.repo-metadata.json
* release_level must be equal to one of the allowed values in Functions/.repo-metadata.json
* api_shortname field missing from Functions/.repo-metadata.json
* release_level must be equal to one of the allowed values in GSuiteAddOns/.repo-metadata.json
* api_shortname field missing from GSuiteAddOns/.repo-metadata.json
* release_level must be equal to one of the allowed values in Gaming/.repo-metadata.json
* api_shortname field missing from Gaming/.repo-metadata.json
* release_level must be equal to one of the allowed values in GkeBackup/.repo-metadata.json
* api_shortname field missing from GkeBackup/.repo-metadata.json
* release_level must be equal to one of the allowed values in GkeConnectGateway/.repo-metadata.json
* api_shortname field missing from GkeConnectGateway/.repo-metadata.json
* release_level must be equal to one of the allowed values in GkeHub/.repo-metadata.json
* api_shortname field missing from GkeHub/.repo-metadata.json
* release_level must be equal to one of the allowed values in GkeMultiCloud/.repo-metadata.json
* api_shortname field missing from GkeMultiCloud/.repo-metadata.json
* release_level must be equal to one of the allowed values in Grafeas/.repo-metadata.json
* api_shortname field missing from Grafeas/.repo-metadata.json
* release_level must be equal to one of the allowed values in Iam/.repo-metadata.json
* api_shortname field missing from Iam/.repo-metadata.json
* release_level must be equal to one of the allowed values in IamCredentials/.repo-metadata.json
* api_shortname field missing from IamCredentials/.repo-metadata.json
* release_level must be equal to one of the allowed values in Iap/.repo-metadata.json
* api_shortname field missing from Iap/.repo-metadata.json
* release_level must be equal to one of the allowed values in Ids/.repo-metadata.json
* api_shortname field missing from Ids/.repo-metadata.json
* release_level must be equal to one of the allowed values in Iot/.repo-metadata.json
* api_shortname field missing from Iot/.repo-metadata.json
* release_level must be equal to one of the allowed values in Kms/.repo-metadata.json
* api_shortname field missing from Kms/.repo-metadata.json
* release_level must be equal to one of the allowed values in Language/.repo-metadata.json
* api_shortname field missing from Language/.repo-metadata.json
* release_level must be equal to one of the allowed values in LifeSciences/.repo-metadata.json
* api_shortname field missing from LifeSciences/.repo-metadata.json
* release_level must be equal to one of the allowed values in Logging/.repo-metadata.json
* api_shortname field missing from Logging/.repo-metadata.json
* release_level must be equal to one of the allowed values in LongRunning/.repo-metadata.json
* api_shortname field missing from LongRunning/.repo-metadata.json
* release_level must be equal to one of the allowed values in ManagedIdentities/.repo-metadata.json
* api_shortname field missing from ManagedIdentities/.repo-metadata.json
* release_level must be equal to one of the allowed values in MediaTranslation/.repo-metadata.json
* api_shortname field missing from MediaTranslation/.repo-metadata.json
* release_level must be equal to one of the allowed values in Memcache/.repo-metadata.json
* api_shortname field missing from Memcache/.repo-metadata.json
* release_level must be equal to one of the allowed values in Monitoring/.repo-metadata.json
* api_shortname field missing from Monitoring/.repo-metadata.json
* release_level must be equal to one of the allowed values in NetworkConnectivity/.repo-metadata.json
* api_shortname field missing from NetworkConnectivity/.repo-metadata.json
* release_level must be equal to one of the allowed values in NetworkManagement/.repo-metadata.json
* api_shortname field missing from NetworkManagement/.repo-metadata.json
* release_level must be equal to one of the allowed values in NetworkSecurity/.repo-metadata.json
* api_shortname field missing from NetworkSecurity/.repo-metadata.json
* release_level must be equal to one of the allowed values in Notebooks/.repo-metadata.json
* api_shortname field missing from Notebooks/.repo-metadata.json
* release_level must be equal to one of the allowed values in Optimization/.repo-metadata.json
* api_shortname field missing from Optimization/.repo-metadata.json
* release_level must be equal to one of the allowed values in OrchestrationAirflow/.repo-metadata.json
* api_shortname field missing from OrchestrationAirflow/.repo-metadata.json
* release_level must be equal to one of the allowed values in OrgPolicy/.repo-metadata.json
* api_shortname field missing from OrgPolicy/.repo-metadata.json
* release_level must be equal to one of the allowed values in OsConfig/.repo-metadata.json
* api_shortname field missing from OsConfig/.repo-metadata.json
* release_level must be equal to one of the allowed values in OsLogin/.repo-metadata.json
* api_shortname field missing from OsLogin/.repo-metadata.json
* release_level must be equal to one of the allowed values in PolicyTroubleshooter/.repo-metadata.json
* api_shortname field missing from PolicyTroubleshooter/.repo-metadata.json
* release_level must be equal to one of the allowed values in PrivateCatalog/.repo-metadata.json
* api_shortname field missing from PrivateCatalog/.repo-metadata.json
* release_level must be equal to one of the allowed values in Profiler/.repo-metadata.json
* api_shortname field missing from Profiler/.repo-metadata.json
* release_level must be equal to one of the allowed values in PubSub/.repo-metadata.json
* api_shortname field missing from PubSub/.repo-metadata.json
* release_level must be equal to one of the allowed values in RecaptchaEnterprise/.repo-metadata.json
* api_shortname field missing from RecaptchaEnterprise/.repo-metadata.json
* release_level must be equal to one of the allowed values in RecommendationEngine/.repo-metadata.json
* api_shortname field missing from RecommendationEngine/.repo-metadata.json
* release_level must be equal to one of the allowed values in Recommender/.repo-metadata.json
* api_shortname field missing from Recommender/.repo-metadata.json
* release_level must be equal to one of the allowed values in Redis/.repo-metadata.json
* api_shortname field missing from Redis/.repo-metadata.json
* release_level must be equal to one of the allowed values in ResourceManager/.repo-metadata.json
* api_shortname field missing from ResourceManager/.repo-metadata.json
* release_level must be equal to one of the allowed values in ResourceSettings/.repo-metadata.json
* api_shortname field missing from ResourceSettings/.repo-metadata.json
* release_level must be equal to one of the allowed values in Retail/.repo-metadata.json
* api_shortname field missing from Retail/.repo-metadata.json
* release_level must be equal to one of the allowed values in Run/.repo-metadata.json
* api_shortname field missing from Run/.repo-metadata.json
* release_level must be equal to one of the allowed values in Scheduler/.repo-metadata.json
* api_shortname field missing from Scheduler/.repo-metadata.json
* release_level must be equal to one of the allowed values in SecretManager/.repo-metadata.json
* api_shortname field missing from SecretManager/.repo-metadata.json
* release_level must be equal to one of the allowed values in SecurityCenter/.repo-metadata.json
* api_shortname field missing from SecurityCenter/.repo-metadata.json
* release_level must be equal to one of the allowed values in SecurityPrivateCa/.repo-metadata.json
* api_shortname field missing from SecurityPrivateCa/.repo-metadata.json
* release_level must be equal to one of the allowed values in SecurityPublicCA/.repo-metadata.json
* api_shortname field missing from SecurityPublicCA/.repo-metadata.json
* release_level must be equal to one of the allowed values in ServiceControl/.repo-metadata.json
* api_shortname field missing from ServiceControl/.repo-metadata.json
* release_level must be equal to one of the allowed values in ServiceDirectory/.repo-metadata.json
* api_shortname field missing from ServiceDirectory/.repo-metadata.json
* release_level must be equal to one of the allowed values in ServiceManagement/.repo-metadata.json
* api_shortname field missing from ServiceManagement/.repo-metadata.json
* release_level must be equal to one of the allowed values in ServiceUsage/.repo-metadata.json
* api_shortname field missing from ServiceUsage/.repo-metadata.json
* release_level must be equal to one of the allowed values in Shell/.repo-metadata.json
* api_shortname field missing from Shell/.repo-metadata.json
* release_level must be equal to one of the allowed values in Spanner/.repo-metadata.json
* api_shortname field missing from Spanner/.repo-metadata.json
* release_level must be equal to one of the allowed values in Speech/.repo-metadata.json
* api_shortname field missing from Speech/.repo-metadata.json
* release_level must be equal to one of the allowed values in SqlAdmin/.repo-metadata.json
* api_shortname field missing from SqlAdmin/.repo-metadata.json
* release_level must be equal to one of the allowed values in Storage/.repo-metadata.json
* api_shortname field missing from Storage/.repo-metadata.json
* release_level must be equal to one of the allowed values in StorageTransfer/.repo-metadata.json
* api_shortname field missing from StorageTransfer/.repo-metadata.json
* release_level must be equal to one of the allowed values in Talent/.repo-metadata.json
* api_shortname field missing from Talent/.repo-metadata.json
* release_level must be equal to one of the allowed values in Tasks/.repo-metadata.json
* api_shortname field missing from Tasks/.repo-metadata.json
* release_level must be equal to one of the allowed values in TextToSpeech/.repo-metadata.json
* api_shortname field missing from TextToSpeech/.repo-metadata.json
* release_level must be equal to one of the allowed values in Tpu/.repo-metadata.json
* api_shortname field missing from Tpu/.repo-metadata.json
* release_level must be equal to one of the allowed values in Trace/.repo-metadata.json
* api_shortname field missing from Trace/.repo-metadata.json
* release_level must be equal to one of the allowed values in Translate/.repo-metadata.json
* api_shortname field missing from Translate/.repo-metadata.json
* release_level must be equal to one of the allowed values in VideoIntelligence/.repo-metadata.json
* api_shortname field missing from VideoIntelligence/.repo-metadata.json
* release_level must be equal to one of the allowed values in VideoLiveStream/.repo-metadata.json
* api_shortname field missing from VideoLiveStream/.repo-metadata.json
* release_level must be equal to one of the allowed values in VideoStitcher/.repo-metadata.json
* api_shortname field missing from VideoStitcher/.repo-metadata.json
* release_level must be equal to one of the allowed values in VideoTranscoder/.repo-metadata.json
* api_shortname field missing from VideoTranscoder/.repo-metadata.json
* release_level must be equal to one of the allowed values in Vision/.repo-metadata.json
* api_shortname field missing from Vision/.repo-metadata.json
* release_level must be equal to one of the allowed values in VmMigration/.repo-metadata.json
* api_shortname field missing from VmMigration/.repo-metadata.json
* release_level must be equal to one of the allowed values in VmwareEngine/.repo-metadata.json
* api_shortname field missing from VmwareEngine/.repo-metadata.json
* release_level must be equal to one of the allowed values in VpcAccess/.repo-metadata.json
* api_shortname field missing from VpcAccess/.repo-metadata.json
* release_level must be equal to one of the allowed values in WebRisk/.repo-metadata.json
* api_shortname field missing from WebRisk/.repo-metadata.json
* release_level must be equal to one of the allowed values in WebSecurityScanner/.repo-metadata.json
* api_shortname field missing from WebSecurityScanner/.repo-metadata.json
* release_level must be equal to one of the allowed values in Workflows/.repo-metadata.json
* api_shortname field missing from Workflows/.repo-metadata.json
* client_documentation must match pattern "^https://.*" in dev/tests/fixtures/component/Vision/.repo-metadata.json
* release_level must be equal to one of the allowed values in dev/tests/fixtures/component/Vision/.repo-metadata.json
* api_shortname field missing from dev/tests/fixtures/component/Vision/.repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
1.0
|
Your .repo-metadata.json files have a problem 🤒 - You have a problem with your .repo-metadata.json files:
Result of scan 📈:
* release_level must be equal to one of the allowed values in AccessApproval/.repo-metadata.json
* api_shortname field missing from AccessApproval/.repo-metadata.json
* release_level must be equal to one of the allowed values in AccessContextManager/.repo-metadata.json
* api_shortname field missing from AccessContextManager/.repo-metadata.json
* release_level must be equal to one of the allowed values in AiPlatform/.repo-metadata.json
* api_shortname field missing from AiPlatform/.repo-metadata.json
* release_level must be equal to one of the allowed values in AnalyticsAdmin/.repo-metadata.json
* api_shortname field missing from AnalyticsAdmin/.repo-metadata.json
* release_level must be equal to one of the allowed values in AnalyticsData/.repo-metadata.json
* api_shortname field missing from AnalyticsData/.repo-metadata.json
* release_level must be equal to one of the allowed values in ApiGateway/.repo-metadata.json
* api_shortname field missing from ApiGateway/.repo-metadata.json
* release_level must be equal to one of the allowed values in ApiKeys/.repo-metadata.json
* api_shortname field missing from ApiKeys/.repo-metadata.json
* release_level must be equal to one of the allowed values in ApigeeConnect/.repo-metadata.json
* api_shortname field missing from ApigeeConnect/.repo-metadata.json
* release_level must be equal to one of the allowed values in ApigeeRegistry/.repo-metadata.json
* api_shortname field missing from ApigeeRegistry/.repo-metadata.json
* release_level must be equal to one of the allowed values in AppEngineAdmin/.repo-metadata.json
* api_shortname field missing from AppEngineAdmin/.repo-metadata.json
* release_level must be equal to one of the allowed values in ArtifactRegistry/.repo-metadata.json
* api_shortname field missing from ArtifactRegistry/.repo-metadata.json
* release_level must be equal to one of the allowed values in Asset/.repo-metadata.json
* api_shortname field missing from Asset/.repo-metadata.json
* release_level must be equal to one of the allowed values in AssuredWorkloads/.repo-metadata.json
* api_shortname field missing from AssuredWorkloads/.repo-metadata.json
* release_level must be equal to one of the allowed values in AutoMl/.repo-metadata.json
* api_shortname field missing from AutoMl/.repo-metadata.json
* release_level must be equal to one of the allowed values in BareMetalSolution/.repo-metadata.json
* api_shortname field missing from BareMetalSolution/.repo-metadata.json
* release_level must be equal to one of the allowed values in Batch/.repo-metadata.json
* api_shortname field missing from Batch/.repo-metadata.json
* release_level must be equal to one of the allowed values in BeyondCorpAppConnections/.repo-metadata.json
* api_shortname field missing from BeyondCorpAppConnections/.repo-metadata.json
* release_level must be equal to one of the allowed values in BeyondCorpAppConnectors/.repo-metadata.json
* api_shortname field missing from BeyondCorpAppConnectors/.repo-metadata.json
* release_level must be equal to one of the allowed values in BeyondCorpAppGateways/.repo-metadata.json
* api_shortname field missing from BeyondCorpAppGateways/.repo-metadata.json
* release_level must be equal to one of the allowed values in BeyondCorpClientConnectorServices/.repo-metadata.json
* api_shortname field missing from BeyondCorpClientConnectorServices/.repo-metadata.json
* release_level must be equal to one of the allowed values in BeyondCorpClientGateways/.repo-metadata.json
* api_shortname field missing from BeyondCorpClientGateways/.repo-metadata.json
* release_level must be equal to one of the allowed values in BigQuery/.repo-metadata.json
* api_shortname field missing from BigQuery/.repo-metadata.json
* release_level must be equal to one of the allowed values in BigQueryAnalyticsHub/.repo-metadata.json
* api_shortname field missing from BigQueryAnalyticsHub/.repo-metadata.json
* release_level must be equal to one of the allowed values in BigQueryConnection/.repo-metadata.json
* api_shortname field missing from BigQueryConnection/.repo-metadata.json
* release_level must be equal to one of the allowed values in BigQueryDataExchange/.repo-metadata.json
* api_shortname field missing from BigQueryDataExchange/.repo-metadata.json
* release_level must be equal to one of the allowed values in BigQueryDataPolicies/.repo-metadata.json
* api_shortname field missing from BigQueryDataPolicies/.repo-metadata.json
* release_level must be equal to one of the allowed values in BigQueryDataTransfer/.repo-metadata.json
* api_shortname field missing from BigQueryDataTransfer/.repo-metadata.json
* release_level must be equal to one of the allowed values in BigQueryMigration/.repo-metadata.json
* api_shortname field missing from BigQueryMigration/.repo-metadata.json
* release_level must be equal to one of the allowed values in BigQueryReservation/.repo-metadata.json
* api_shortname field missing from BigQueryReservation/.repo-metadata.json
* release_level must be equal to one of the allowed values in BigQueryStorage/.repo-metadata.json
* api_shortname field missing from BigQueryStorage/.repo-metadata.json
* release_level must be equal to one of the allowed values in Bigtable/.repo-metadata.json
* api_shortname field missing from Bigtable/.repo-metadata.json
* release_level must be equal to one of the allowed values in Billing/.repo-metadata.json
* api_shortname field missing from Billing/.repo-metadata.json
* release_level must be equal to one of the allowed values in BillingBudgets/.repo-metadata.json
* api_shortname field missing from BillingBudgets/.repo-metadata.json
* release_level must be equal to one of the allowed values in BinaryAuthorization/.repo-metadata.json
* api_shortname field missing from BinaryAuthorization/.repo-metadata.json
* release_level must be equal to one of the allowed values in Build/.repo-metadata.json
* api_shortname field missing from Build/.repo-metadata.json
* release_level must be equal to one of the allowed values in CertificateManager/.repo-metadata.json
* api_shortname field missing from CertificateManager/.repo-metadata.json
* release_level must be equal to one of the allowed values in Channel/.repo-metadata.json
* api_shortname field missing from Channel/.repo-metadata.json
* release_level must be equal to one of the allowed values in CommonProtos/.repo-metadata.json
* release_level must be equal to one of the allowed values in Compute/.repo-metadata.json
* api_shortname field missing from Compute/.repo-metadata.json
* release_level must be equal to one of the allowed values in ContactCenterInsights/.repo-metadata.json
* api_shortname field missing from ContactCenterInsights/.repo-metadata.json
* release_level must be equal to one of the allowed values in Container/.repo-metadata.json
* api_shortname field missing from Container/.repo-metadata.json
* release_level must be equal to one of the allowed values in ContainerAnalysis/.repo-metadata.json
* api_shortname field missing from ContainerAnalysis/.repo-metadata.json
* release_level must be equal to one of the allowed values in Core/.repo-metadata.json
* release_level must be equal to one of the allowed values in DataCatalog/.repo-metadata.json
* api_shortname field missing from DataCatalog/.repo-metadata.json
* release_level must be equal to one of the allowed values in DataFusion/.repo-metadata.json
* api_shortname field missing from DataFusion/.repo-metadata.json
* release_level must be equal to one of the allowed values in DataLabeling/.repo-metadata.json
* api_shortname field missing from DataLabeling/.repo-metadata.json
* release_level must be equal to one of the allowed values in Dataflow/.repo-metadata.json
* api_shortname field missing from Dataflow/.repo-metadata.json
* release_level must be equal to one of the allowed values in Dataform/.repo-metadata.json
* api_shortname field missing from Dataform/.repo-metadata.json
* release_level must be equal to one of the allowed values in Dataplex/.repo-metadata.json
* api_shortname field missing from Dataplex/.repo-metadata.json
* release_level must be equal to one of the allowed values in Dataproc/.repo-metadata.json
* api_shortname field missing from Dataproc/.repo-metadata.json
* release_level must be equal to one of the allowed values in DataprocMetastore/.repo-metadata.json
* api_shortname field missing from DataprocMetastore/.repo-metadata.json
* release_level must be equal to one of the allowed values in Datastore/.repo-metadata.json
* api_shortname field missing from Datastore/.repo-metadata.json
* release_level must be equal to one of the allowed values in DatastoreAdmin/.repo-metadata.json
* api_shortname field missing from DatastoreAdmin/.repo-metadata.json
* release_level must be equal to one of the allowed values in Datastream/.repo-metadata.json
* api_shortname field missing from Datastream/.repo-metadata.json
* release_level must be equal to one of the allowed values in Debugger/.repo-metadata.json
* api_shortname field missing from Debugger/.repo-metadata.json
* release_level must be equal to one of the allowed values in Deploy/.repo-metadata.json
* api_shortname field missing from Deploy/.repo-metadata.json
* release_level must be equal to one of the allowed values in Dialogflow/.repo-metadata.json
* api_shortname field missing from Dialogflow/.repo-metadata.json
* release_level must be equal to one of the allowed values in Dlp/.repo-metadata.json
* api_shortname field missing from Dlp/.repo-metadata.json
* release_level must be equal to one of the allowed values in Dms/.repo-metadata.json
* api_shortname field missing from Dms/.repo-metadata.json
* release_level must be equal to one of the allowed values in DocumentAi/.repo-metadata.json
* api_shortname field missing from DocumentAi/.repo-metadata.json
* release_level must be equal to one of the allowed values in Domains/.repo-metadata.json
* api_shortname field missing from Domains/.repo-metadata.json
* release_level must be equal to one of the allowed values in ErrorReporting/.repo-metadata.json
* api_shortname field missing from ErrorReporting/.repo-metadata.json
* release_level must be equal to one of the allowed values in EssentialContacts/.repo-metadata.json
* api_shortname field missing from EssentialContacts/.repo-metadata.json
* release_level must be equal to one of the allowed values in Eventarc/.repo-metadata.json
* api_shortname field missing from Eventarc/.repo-metadata.json
* release_level must be equal to one of the allowed values in EventarcPublishing/.repo-metadata.json
* api_shortname field missing from EventarcPublishing/.repo-metadata.json
* release_level must be equal to one of the allowed values in Filestore/.repo-metadata.json
* api_shortname field missing from Filestore/.repo-metadata.json
* release_level must be equal to one of the allowed values in Firestore/.repo-metadata.json
* api_shortname field missing from Firestore/.repo-metadata.json
* release_level must be equal to one of the allowed values in Functions/.repo-metadata.json
* api_shortname field missing from Functions/.repo-metadata.json
* release_level must be equal to one of the allowed values in GSuiteAddOns/.repo-metadata.json
* api_shortname field missing from GSuiteAddOns/.repo-metadata.json
* release_level must be equal to one of the allowed values in Gaming/.repo-metadata.json
* api_shortname field missing from Gaming/.repo-metadata.json
* release_level must be equal to one of the allowed values in GkeBackup/.repo-metadata.json
* api_shortname field missing from GkeBackup/.repo-metadata.json
* release_level must be equal to one of the allowed values in GkeConnectGateway/.repo-metadata.json
* api_shortname field missing from GkeConnectGateway/.repo-metadata.json
* release_level must be equal to one of the allowed values in GkeHub/.repo-metadata.json
* api_shortname field missing from GkeHub/.repo-metadata.json
* release_level must be equal to one of the allowed values in GkeMultiCloud/.repo-metadata.json
* api_shortname field missing from GkeMultiCloud/.repo-metadata.json
* release_level must be equal to one of the allowed values in Grafeas/.repo-metadata.json
* api_shortname field missing from Grafeas/.repo-metadata.json
* release_level must be equal to one of the allowed values in Iam/.repo-metadata.json
* api_shortname field missing from Iam/.repo-metadata.json
* release_level must be equal to one of the allowed values in IamCredentials/.repo-metadata.json
* api_shortname field missing from IamCredentials/.repo-metadata.json
* release_level must be equal to one of the allowed values in Iap/.repo-metadata.json
* api_shortname field missing from Iap/.repo-metadata.json
* release_level must be equal to one of the allowed values in Ids/.repo-metadata.json
* api_shortname field missing from Ids/.repo-metadata.json
* release_level must be equal to one of the allowed values in Iot/.repo-metadata.json
* api_shortname field missing from Iot/.repo-metadata.json
* release_level must be equal to one of the allowed values in Kms/.repo-metadata.json
* api_shortname field missing from Kms/.repo-metadata.json
* release_level must be equal to one of the allowed values in Language/.repo-metadata.json
* api_shortname field missing from Language/.repo-metadata.json
* release_level must be equal to one of the allowed values in LifeSciences/.repo-metadata.json
* api_shortname field missing from LifeSciences/.repo-metadata.json
* release_level must be equal to one of the allowed values in Logging/.repo-metadata.json
* api_shortname field missing from Logging/.repo-metadata.json
* release_level must be equal to one of the allowed values in LongRunning/.repo-metadata.json
* api_shortname field missing from LongRunning/.repo-metadata.json
* release_level must be equal to one of the allowed values in ManagedIdentities/.repo-metadata.json
* api_shortname field missing from ManagedIdentities/.repo-metadata.json
* release_level must be equal to one of the allowed values in MediaTranslation/.repo-metadata.json
* api_shortname field missing from MediaTranslation/.repo-metadata.json
* release_level must be equal to one of the allowed values in Memcache/.repo-metadata.json
* api_shortname field missing from Memcache/.repo-metadata.json
* release_level must be equal to one of the allowed values in Monitoring/.repo-metadata.json
* api_shortname field missing from Monitoring/.repo-metadata.json
* release_level must be equal to one of the allowed values in NetworkConnectivity/.repo-metadata.json
* api_shortname field missing from NetworkConnectivity/.repo-metadata.json
* release_level must be equal to one of the allowed values in NetworkManagement/.repo-metadata.json
* api_shortname field missing from NetworkManagement/.repo-metadata.json
* release_level must be equal to one of the allowed values in NetworkSecurity/.repo-metadata.json
* api_shortname field missing from NetworkSecurity/.repo-metadata.json
* release_level must be equal to one of the allowed values in Notebooks/.repo-metadata.json
* api_shortname field missing from Notebooks/.repo-metadata.json
* release_level must be equal to one of the allowed values in Optimization/.repo-metadata.json
* api_shortname field missing from Optimization/.repo-metadata.json
* release_level must be equal to one of the allowed values in OrchestrationAirflow/.repo-metadata.json
* api_shortname field missing from OrchestrationAirflow/.repo-metadata.json
* release_level must be equal to one of the allowed values in OrgPolicy/.repo-metadata.json
* api_shortname field missing from OrgPolicy/.repo-metadata.json
* release_level must be equal to one of the allowed values in OsConfig/.repo-metadata.json
* api_shortname field missing from OsConfig/.repo-metadata.json
* release_level must be equal to one of the allowed values in OsLogin/.repo-metadata.json
* api_shortname field missing from OsLogin/.repo-metadata.json
* release_level must be equal to one of the allowed values in PolicyTroubleshooter/.repo-metadata.json
* api_shortname field missing from PolicyTroubleshooter/.repo-metadata.json
* release_level must be equal to one of the allowed values in PrivateCatalog/.repo-metadata.json
* api_shortname field missing from PrivateCatalog/.repo-metadata.json
* release_level must be equal to one of the allowed values in Profiler/.repo-metadata.json
* api_shortname field missing from Profiler/.repo-metadata.json
* release_level must be equal to one of the allowed values in PubSub/.repo-metadata.json
* api_shortname field missing from PubSub/.repo-metadata.json
* release_level must be equal to one of the allowed values in RecaptchaEnterprise/.repo-metadata.json
* api_shortname field missing from RecaptchaEnterprise/.repo-metadata.json
* release_level must be equal to one of the allowed values in RecommendationEngine/.repo-metadata.json
* api_shortname field missing from RecommendationEngine/.repo-metadata.json
* release_level must be equal to one of the allowed values in Recommender/.repo-metadata.json
* api_shortname field missing from Recommender/.repo-metadata.json
* release_level must be equal to one of the allowed values in Redis/.repo-metadata.json
* api_shortname field missing from Redis/.repo-metadata.json
* release_level must be equal to one of the allowed values in ResourceManager/.repo-metadata.json
* api_shortname field missing from ResourceManager/.repo-metadata.json
* release_level must be equal to one of the allowed values in ResourceSettings/.repo-metadata.json
* api_shortname field missing from ResourceSettings/.repo-metadata.json
* release_level must be equal to one of the allowed values in Retail/.repo-metadata.json
* api_shortname field missing from Retail/.repo-metadata.json
* release_level must be equal to one of the allowed values in Run/.repo-metadata.json
* api_shortname field missing from Run/.repo-metadata.json
* release_level must be equal to one of the allowed values in Scheduler/.repo-metadata.json
* api_shortname field missing from Scheduler/.repo-metadata.json
* release_level must be equal to one of the allowed values in SecretManager/.repo-metadata.json
* api_shortname field missing from SecretManager/.repo-metadata.json
* release_level must be equal to one of the allowed values in SecurityCenter/.repo-metadata.json
* api_shortname field missing from SecurityCenter/.repo-metadata.json
* release_level must be equal to one of the allowed values in SecurityPrivateCa/.repo-metadata.json
* api_shortname field missing from SecurityPrivateCa/.repo-metadata.json
* release_level must be equal to one of the allowed values in SecurityPublicCA/.repo-metadata.json
* api_shortname field missing from SecurityPublicCA/.repo-metadata.json
* release_level must be equal to one of the allowed values in ServiceControl/.repo-metadata.json
* api_shortname field missing from ServiceControl/.repo-metadata.json
* release_level must be equal to one of the allowed values in ServiceDirectory/.repo-metadata.json
* api_shortname field missing from ServiceDirectory/.repo-metadata.json
* release_level must be equal to one of the allowed values in ServiceManagement/.repo-metadata.json
* api_shortname field missing from ServiceManagement/.repo-metadata.json
* release_level must be equal to one of the allowed values in ServiceUsage/.repo-metadata.json
* api_shortname field missing from ServiceUsage/.repo-metadata.json
* release_level must be equal to one of the allowed values in Shell/.repo-metadata.json
* api_shortname field missing from Shell/.repo-metadata.json
* release_level must be equal to one of the allowed values in Spanner/.repo-metadata.json
* api_shortname field missing from Spanner/.repo-metadata.json
* release_level must be equal to one of the allowed values in Speech/.repo-metadata.json
* api_shortname field missing from Speech/.repo-metadata.json
* release_level must be equal to one of the allowed values in SqlAdmin/.repo-metadata.json
* api_shortname field missing from SqlAdmin/.repo-metadata.json
* release_level must be equal to one of the allowed values in Storage/.repo-metadata.json
* api_shortname field missing from Storage/.repo-metadata.json
* release_level must be equal to one of the allowed values in StorageTransfer/.repo-metadata.json
* api_shortname field missing from StorageTransfer/.repo-metadata.json
* release_level must be equal to one of the allowed values in Talent/.repo-metadata.json
* api_shortname field missing from Talent/.repo-metadata.json
* release_level must be equal to one of the allowed values in Tasks/.repo-metadata.json
* api_shortname field missing from Tasks/.repo-metadata.json
* release_level must be equal to one of the allowed values in TextToSpeech/.repo-metadata.json
* api_shortname field missing from TextToSpeech/.repo-metadata.json
* release_level must be equal to one of the allowed values in Tpu/.repo-metadata.json
* api_shortname field missing from Tpu/.repo-metadata.json
* release_level must be equal to one of the allowed values in Trace/.repo-metadata.json
* api_shortname field missing from Trace/.repo-metadata.json
* release_level must be equal to one of the allowed values in Translate/.repo-metadata.json
* api_shortname field missing from Translate/.repo-metadata.json
* release_level must be equal to one of the allowed values in VideoIntelligence/.repo-metadata.json
* api_shortname field missing from VideoIntelligence/.repo-metadata.json
* release_level must be equal to one of the allowed values in VideoLiveStream/.repo-metadata.json
* api_shortname field missing from VideoLiveStream/.repo-metadata.json
* release_level must be equal to one of the allowed values in VideoStitcher/.repo-metadata.json
* api_shortname field missing from VideoStitcher/.repo-metadata.json
* release_level must be equal to one of the allowed values in VideoTranscoder/.repo-metadata.json
* api_shortname field missing from VideoTranscoder/.repo-metadata.json
* release_level must be equal to one of the allowed values in Vision/.repo-metadata.json
* api_shortname field missing from Vision/.repo-metadata.json
* release_level must be equal to one of the allowed values in VmMigration/.repo-metadata.json
* api_shortname field missing from VmMigration/.repo-metadata.json
* release_level must be equal to one of the allowed values in VmwareEngine/.repo-metadata.json
* api_shortname field missing from VmwareEngine/.repo-metadata.json
* release_level must be equal to one of the allowed values in VpcAccess/.repo-metadata.json
* api_shortname field missing from VpcAccess/.repo-metadata.json
* release_level must be equal to one of the allowed values in WebRisk/.repo-metadata.json
* api_shortname field missing from WebRisk/.repo-metadata.json
* release_level must be equal to one of the allowed values in WebSecurityScanner/.repo-metadata.json
* api_shortname field missing from WebSecurityScanner/.repo-metadata.json
* release_level must be equal to one of the allowed values in Workflows/.repo-metadata.json
* api_shortname field missing from Workflows/.repo-metadata.json
* client_documentation must match pattern "^https://.*" in dev/tests/fixtures/component/Vision/.repo-metadata.json
* release_level must be equal to one of the allowed values in dev/tests/fixtures/component/Vision/.repo-metadata.json
* api_shortname field missing from dev/tests/fixtures/component/Vision/.repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
process
|
your repo metadata json files have a problem 🤒 you have a problem with your repo metadata json files result of scan 📈 release level must be equal to one of the allowed values in accessapproval repo metadata json api shortname field missing from accessapproval repo metadata json release level must be equal to one of the allowed values in accesscontextmanager repo metadata json api shortname field missing from accesscontextmanager repo metadata json release level must be equal to one of the allowed values in aiplatform repo metadata json api shortname field missing from aiplatform repo metadata json release level must be equal to one of the allowed values in analyticsadmin repo metadata json api shortname field missing from analyticsadmin repo metadata json release level must be equal to one of the allowed values in analyticsdata repo metadata json api shortname field missing from analyticsdata repo metadata json release level must be equal to one of the allowed values in apigateway repo metadata json api shortname field missing from apigateway repo metadata json release level must be equal to one of the allowed values in apikeys repo metadata json api shortname field missing from apikeys repo metadata json release level must be equal to one of the allowed values in apigeeconnect repo metadata json api shortname field missing from apigeeconnect repo metadata json release level must be equal to one of the allowed values in apigeeregistry repo metadata json api shortname field missing from apigeeregistry repo metadata json release level must be equal to one of the allowed values in appengineadmin repo metadata json api shortname field missing from appengineadmin repo metadata json release level must be equal to one of the allowed values in artifactregistry repo metadata json api shortname field missing from artifactregistry repo metadata json release level must be equal to one of the allowed values in asset repo metadata json api shortname field missing from asset repo metadata json release level must be equal to one of the allowed values in assuredworkloads repo metadata json api shortname field missing from assuredworkloads repo metadata json release level must be equal to one of the allowed values in automl repo metadata json api shortname field missing from automl repo metadata json release level must be equal to one of the allowed values in baremetalsolution repo metadata json api shortname field missing from baremetalsolution repo metadata json release level must be equal to one of the allowed values in batch repo metadata json api shortname field missing from batch repo metadata json release level must be equal to one of the allowed values in beyondcorpappconnections repo metadata json api shortname field missing from beyondcorpappconnections repo metadata json release level must be equal to one of the allowed values in beyondcorpappconnectors repo metadata json api shortname field missing from beyondcorpappconnectors repo metadata json release level must be equal to one of the allowed values in beyondcorpappgateways repo metadata json api shortname field missing from beyondcorpappgateways repo metadata json release level must be equal to one of the allowed values in beyondcorpclientconnectorservices repo metadata json api shortname field missing from beyondcorpclientconnectorservices repo metadata json release level must be equal to one of the allowed values in beyondcorpclientgateways repo metadata json api shortname field missing from beyondcorpclientgateways repo metadata json release level must be equal to one of the allowed values in bigquery repo metadata json api shortname field missing from bigquery repo metadata json release level must be equal to one of the allowed values in bigqueryanalyticshub repo metadata json api shortname field missing from bigqueryanalyticshub repo metadata json release level must be equal to one of the allowed values in bigqueryconnection repo metadata json api shortname field missing from bigqueryconnection repo metadata json release level must be equal to one of the allowed values in bigquerydataexchange repo metadata json api shortname field missing from bigquerydataexchange repo metadata json release level must be equal to one of the allowed values in bigquerydatapolicies repo metadata json api shortname field missing from bigquerydatapolicies repo metadata json release level must be equal to one of the allowed values in bigquerydatatransfer repo metadata json api shortname field missing from bigquerydatatransfer repo metadata json release level must be equal to one of the allowed values in bigquerymigration repo metadata json api shortname field missing from bigquerymigration repo metadata json release level must be equal to one of the allowed values in bigqueryreservation repo metadata json api shortname field missing from bigqueryreservation repo metadata json release level must be equal to one of the allowed values in bigquerystorage repo metadata json api shortname field missing from bigquerystorage repo metadata json release level must be equal to one of the allowed values in bigtable repo metadata json api shortname field missing from bigtable repo metadata json release level must be equal to one of the allowed values in billing repo metadata json api shortname field missing from billing repo metadata json release level must be equal to one of the allowed values in billingbudgets repo metadata json api shortname field missing from billingbudgets repo metadata json release level must be equal to one of the allowed values in binaryauthorization repo metadata json api shortname field missing from binaryauthorization repo metadata json release level must be equal to one of the allowed values in build repo metadata json api shortname field missing from build repo metadata json release level must be equal to one of the allowed values in certificatemanager repo metadata json api shortname field missing from certificatemanager repo metadata json release level must be equal to one of the allowed values in channel repo metadata json api shortname field missing from channel repo metadata json release level must be equal to one of the allowed values in commonprotos repo metadata json release level must be equal to one of the allowed values in compute repo metadata json api shortname field missing from compute repo metadata json release level must be equal to one of the allowed values in contactcenterinsights repo metadata json api shortname field missing from contactcenterinsights repo metadata json release level must be equal to one of the allowed values in container repo metadata json api shortname field missing from container repo metadata json release level must be equal to one of the allowed values in containeranalysis repo metadata json api shortname field missing from containeranalysis repo metadata json release level must be equal to one of the allowed values in core repo metadata json release level must be equal to one of the allowed values in datacatalog repo metadata json api shortname field missing from datacatalog repo metadata json release level must be equal to one of the allowed values in datafusion repo metadata json api shortname field missing from datafusion repo metadata json release level must be equal to one of the allowed values in datalabeling repo metadata json api shortname field missing from datalabeling repo metadata json release level must be equal to one of the allowed values in dataflow repo metadata json api shortname field missing from dataflow repo metadata json release level must be equal to one of the allowed values in dataform repo metadata json api shortname field missing from dataform repo metadata json release level must be equal to one of the allowed values in dataplex repo metadata json api shortname field missing from dataplex repo metadata json release level must be equal to one of the allowed values in dataproc repo metadata json api shortname field missing from dataproc repo metadata json release level must be equal to one of the allowed values in dataprocmetastore repo metadata json api shortname field missing from dataprocmetastore repo metadata json release level must be equal to one of the allowed values in datastore repo metadata json api shortname field missing from datastore repo metadata json release level must be equal to one of the allowed values in datastoreadmin repo metadata json api shortname field missing from datastoreadmin repo metadata json release level must be equal to one of the allowed values in datastream repo metadata json api shortname field missing from datastream repo metadata json release level must be equal to one of the allowed values in debugger repo metadata json api shortname field missing from debugger repo metadata json release level must be equal to one of the allowed values in deploy repo metadata json api shortname field missing from deploy repo metadata json release level must be equal to one of the allowed values in dialogflow repo metadata json api shortname field missing from dialogflow repo metadata json release level must be equal to one of the allowed values in dlp repo metadata json api shortname field missing from dlp repo metadata json release level must be equal to one of the allowed values in dms repo metadata json api shortname field missing from dms repo metadata json release level must be equal to one of the allowed values in documentai repo metadata json api shortname field missing from documentai repo metadata json release level must be equal to one of the allowed values in domains repo metadata json api shortname field missing from domains repo metadata json release level must be equal to one of the allowed values in errorreporting repo metadata json api shortname field missing from errorreporting repo metadata json release level must be equal to one of the allowed values in essentialcontacts repo metadata json api shortname field missing from essentialcontacts repo metadata json release level must be equal to one of the allowed values in eventarc repo metadata json api shortname field missing from eventarc repo metadata json release level must be equal to one of the allowed values in eventarcpublishing repo metadata json api shortname field missing from eventarcpublishing repo metadata json release level must be equal to one of the allowed values in filestore repo metadata json api shortname field missing from filestore repo metadata json release level must be equal to one of the allowed values in firestore repo metadata json api shortname field missing from firestore repo metadata json release level must be equal to one of the allowed values in functions repo metadata json api shortname field missing from functions repo metadata json release level must be equal to one of the allowed values in gsuiteaddons repo metadata json api shortname field missing from gsuiteaddons repo metadata json release level must be equal to one of the allowed values in gaming repo metadata json api shortname field missing from gaming repo metadata json release level must be equal to one of the allowed values in gkebackup repo metadata json api shortname field missing from gkebackup repo metadata json release level must be equal to one of the allowed values in gkeconnectgateway repo metadata json api shortname field missing from gkeconnectgateway repo metadata json release level must be equal to one of the allowed values in gkehub repo metadata json api shortname field missing from gkehub repo metadata json release level must be equal to one of the allowed values in gkemulticloud repo metadata json api shortname field missing from gkemulticloud repo metadata json release level must be equal to one of the allowed values in grafeas repo metadata json api shortname field missing from grafeas repo metadata json release level must be equal to one of the allowed values in iam repo metadata json api shortname field missing from iam repo metadata json release level must be equal to one of the allowed values in iamcredentials repo metadata json api shortname field missing from iamcredentials repo metadata json release level must be equal to one of the allowed values in iap repo metadata json api shortname field missing from iap repo metadata json release level must be equal to one of the allowed values in ids repo metadata json api shortname field missing from ids repo metadata json release level must be equal to one of the allowed values in iot repo metadata json api shortname field missing from iot repo metadata json release level must be equal to one of the allowed values in kms repo metadata json api shortname field missing from kms repo metadata json release level must be equal to one of the allowed values in language repo metadata json api shortname field missing from language repo metadata json release level must be equal to one of the allowed values in lifesciences repo metadata json api shortname field missing from lifesciences repo metadata json release level must be equal to one of the allowed values in logging repo metadata json api shortname field missing from logging repo metadata json release level must be equal to one of the allowed values in longrunning repo metadata json api shortname field missing from longrunning repo metadata json release level must be equal to one of the allowed values in managedidentities repo metadata json api shortname field missing from managedidentities repo metadata json release level must be equal to one of the allowed values in mediatranslation repo metadata json api shortname field missing from mediatranslation repo metadata json release level must be equal to one of the allowed values in memcache repo metadata json api shortname field missing from memcache repo metadata json release level must be equal to one of the allowed values in monitoring repo metadata json api shortname field missing from monitoring repo metadata json release level must be equal to one of the allowed values in networkconnectivity repo metadata json api shortname field missing from networkconnectivity repo metadata json release level must be equal to one of the allowed values in networkmanagement repo metadata json api shortname field missing from networkmanagement repo metadata json release level must be equal to one of the allowed values in networksecurity repo metadata json api shortname field missing from networksecurity repo metadata json release level must be equal to one of the allowed values in notebooks repo metadata json api shortname field missing from notebooks repo metadata json release level must be equal to one of the allowed values in optimization repo metadata json api shortname field missing from optimization repo metadata json release level must be equal to one of the allowed values in orchestrationairflow repo metadata json api shortname field missing from orchestrationairflow repo metadata json release level must be equal to one of the allowed values in orgpolicy repo metadata json api shortname field missing from orgpolicy repo metadata json release level must be equal to one of the allowed values in osconfig repo metadata json api shortname field missing from osconfig repo metadata json release level must be equal to one of the allowed values in oslogin repo metadata json api shortname field missing from oslogin repo metadata json release level must be equal to one of the allowed values in policytroubleshooter repo metadata json api shortname field missing from policytroubleshooter repo metadata json release level must be equal to one of the allowed values in privatecatalog repo metadata json api shortname field missing from privatecatalog repo metadata json release level must be equal to one of the allowed values in profiler repo metadata json api shortname field missing from profiler repo metadata json release level must be equal to one of the allowed values in pubsub repo metadata json api shortname field missing from pubsub repo metadata json release level must be equal to one of the allowed values in recaptchaenterprise repo metadata json api shortname field missing from recaptchaenterprise repo metadata json release level must be equal to one of the allowed values in recommendationengine repo metadata json api shortname field missing from recommendationengine repo metadata json release level must be equal to one of the allowed values in recommender repo metadata json api shortname field missing from recommender repo metadata json release level must be equal to one of the allowed values in redis repo metadata json api shortname field missing from redis repo metadata json release level must be equal to one of the allowed values in resourcemanager repo metadata json api shortname field missing from resourcemanager repo metadata json release level must be equal to one of the allowed values in resourcesettings repo metadata json api shortname field missing from resourcesettings repo metadata json release level must be equal to one of the allowed values in retail repo metadata json api shortname field missing from retail repo metadata json release level must be equal to one of the allowed values in run repo metadata json api shortname field missing from run repo metadata json release level must be equal to one of the allowed values in scheduler repo metadata json api shortname field missing from scheduler repo metadata json release level must be equal to one of the allowed values in secretmanager repo metadata json api shortname field missing from secretmanager repo metadata json release level must be equal to one of the allowed values in securitycenter repo metadata json api shortname field missing from securitycenter repo metadata json release level must be equal to one of the allowed values in securityprivateca repo metadata json api shortname field missing from securityprivateca repo metadata json release level must be equal to one of the allowed values in securitypublicca repo metadata json api shortname field missing from securitypublicca repo metadata json release level must be equal to one of the allowed values in servicecontrol repo metadata json api shortname field missing from servicecontrol repo metadata json release level must be equal to one of the allowed values in servicedirectory repo metadata json api shortname field missing from servicedirectory repo metadata json release level must be equal to one of the allowed values in servicemanagement repo metadata json api shortname field missing from servicemanagement repo metadata json release level must be equal to one of the allowed values in serviceusage repo metadata json api shortname field missing from serviceusage repo metadata json release level must be equal to one of the allowed values in shell repo metadata json api shortname field missing from shell repo metadata json release level must be equal to one of the allowed values in spanner repo metadata json api shortname field missing from spanner repo metadata json release level must be equal to one of the allowed values in speech repo metadata json api shortname field missing from speech repo metadata json release level must be equal to one of the allowed values in sqladmin repo metadata json api shortname field missing from sqladmin repo metadata json release level must be equal to one of the allowed values in storage repo metadata json api shortname field missing from storage repo metadata json release level must be equal to one of the allowed values in storagetransfer repo metadata json api shortname field missing from storagetransfer repo metadata json release level must be equal to one of the allowed values in talent repo metadata json api shortname field missing from talent repo metadata json release level must be equal to one of the allowed values in tasks repo metadata json api shortname field missing from tasks repo metadata json release level must be equal to one of the allowed values in texttospeech repo metadata json api shortname field missing from texttospeech repo metadata json release level must be equal to one of the allowed values in tpu repo metadata json api shortname field missing from tpu repo metadata json release level must be equal to one of the allowed values in trace repo metadata json api shortname field missing from trace repo metadata json release level must be equal to one of the allowed values in translate repo metadata json api shortname field missing from translate repo metadata json release level must be equal to one of the allowed values in videointelligence repo metadata json api shortname field missing from videointelligence repo metadata json release level must be equal to one of the allowed values in videolivestream repo metadata json api shortname field missing from videolivestream repo metadata json release level must be equal to one of the allowed values in videostitcher repo metadata json api shortname field missing from videostitcher repo metadata json release level must be equal to one of the allowed values in videotranscoder repo metadata json api shortname field missing from videotranscoder repo metadata json release level must be equal to one of the allowed values in vision repo metadata json api shortname field missing from vision repo metadata json release level must be equal to one of the allowed values in vmmigration repo metadata json api shortname field missing from vmmigration repo metadata json release level must be equal to one of the allowed values in vmwareengine repo metadata json api shortname field missing from vmwareengine repo metadata json release level must be equal to one of the allowed values in vpcaccess repo metadata json api shortname field missing from vpcaccess repo metadata json release level must be equal to one of the allowed values in webrisk repo metadata json api shortname field missing from webrisk repo metadata json release level must be equal to one of the allowed values in websecurityscanner repo metadata json api shortname field missing from websecurityscanner repo metadata json release level must be equal to one of the allowed values in workflows repo metadata json api shortname field missing from workflows repo metadata json client documentation must match pattern in dev tests fixtures component vision repo metadata json release level must be equal to one of the allowed values in dev tests fixtures component vision repo metadata json api shortname field missing from dev tests fixtures component vision repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
| 1
|
12,208
| 14,742,806,321
|
IssuesEvent
|
2021-01-07 12:55:46
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
Missing Emailed Invoice
|
anc-process anp-1.5 ant-child/secondary ant-enhancement ant-support
|
In GitLab by @kdjstudios on Jun 6, 2019, 12:17
**Submitted by:** "Arianna Screen" <arianna.screen@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2019-06-06-33233/conversation
**Server:** Internal
**Client/Site:** 056
**Account:** A06295
**Issue:**
It has been brought to my attention that one of our clients is not getting her emailed invoices.
The account, 056-A06295, is set to have the invoice emailed.
Client has said that for quite a few months she has not received an email from AnswerNet except for the personally typed emails from me.
I checked the Site: Santa Rosa, CA, Billing Cycle: 05/28/2019 Master, E-mailed:125 invoices and her email is not listed even though it is in SAB.
I don’t have previous Master E-mails to verify how long she has not been receiving her invoices.
|
1.0
|
Missing Emailed Invoice - In GitLab by @kdjstudios on Jun 6, 2019, 12:17
**Submitted by:** "Arianna Screen" <arianna.screen@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2019-06-06-33233/conversation
**Server:** Internal
**Client/Site:** 056
**Account:** A06295
**Issue:**
It has been brought to my attention that one of our clients is not getting her emailed invoices.
The account, 056-A06295, is set to have the invoice emailed.
Client has said that for quite a few months she has not received an email from AnswerNet except for the personally typed emails from me.
I checked the Site: Santa Rosa, CA, Billing Cycle: 05/28/2019 Master, E-mailed:125 invoices and her email is not listed even though it is in SAB.
I don’t have previous Master E-mails to verify how long she has not been receiving her invoices.
|
process
|
missing emailed invoice in gitlab by kdjstudios on jun submitted by arianna screen helpdesk server internal client site account issue it has been brought to my attention that one of our clients is not getting her emailed invoices the account is set to have the invoice emailed client has said that for quite a few months she has not received an email from answernet except for the personally typed emails from me i checked the site santa rosa ca billing cycle master e mailed invoices and her email is not listed even though it is in sab i don’t have previous master e mails to verify how long she has not been receiving her invoices
| 1
|
2,661
| 5,435,724,176
|
IssuesEvent
|
2017-03-05 19:24:32
|
worcester-esu/wesu-website
|
https://api.github.com/repos/worcester-esu/wesu-website
|
closed
|
Deployment should use compression
|
enhancement process
|
Currently, deployment uploads all files individually without compression. It should instead upload an archive and extract it on the web server.
|
1.0
|
Deployment should use compression - Currently, deployment uploads all files individually without compression. It should instead upload an archive and extract it on the web server.
|
process
|
deployment should use compression currently deployment uploads all files individually without compression it should instead upload an archive and extract it on the web server
| 1
|
6,360
| 9,416,130,440
|
IssuesEvent
|
2019-04-10 14:05:21
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
Dashboard plots fail: Value must be an integer
|
Bug Priority/P2 Query Processor
|
- Chrome 73.0
- OSX 10.14
- postgres 11.2
- Metabase 32.0
- Metabase docker
- Metabase internal database H2
On the summary dashboard for very large tables (>25M rows), I'm seeing "There was a problem displaying this chart" for nearly every plot that ought to display. On dashboards for smaller tables (when DB load is low), plots display normally. I looked at the server logs, and the only clue I can find is this:
```
Mar 29 19:02:14 DEBUG metabase.middleware.log :: POST /api/dataset 400 (1 ms) (0 DB calls).
{:errors {:database "value must be an integer."}}
```
My theory is that the DB query is timing out but not bubbling the exception up in a transparent manner.
|
1.0
|
Dashboard plots fail: Value must be an integer - - Chrome 73.0
- OSX 10.14
- postgres 11.2
- Metabase 32.0
- Metabase docker
- Metabase internal database H2
On the summary dashboard for very large tables (>25M rows), I'm seeing "There was a problem displaying this chart" for nearly every plot that ought to display. On dashboards for smaller tables (when DB load is low), plots display normally. I looked at the server logs, and the only clue I can find is this:
```
Mar 29 19:02:14 DEBUG metabase.middleware.log :: POST /api/dataset 400 (1 ms) (0 DB calls).
{:errors {:database "value must be an integer."}}
```
My theory is that the DB query is timing out but not bubbling the exception up in a transparent manner.
|
process
|
dashboard plots fail value must be an integer chrome osx postgres metabase metabase docker metabase internal database on the summary dashboard for very large tables rows i m seeing there was a problem displaying this chart for nearly every plot that ought to display on dashboards for smaller tables when db load is low plots display normally i looked at the server logs and the only clue i can find is this mar debug metabase middleware log post api dataset ms db calls errors database value must be an integer my theory is that the db query is timing out but not bubbling the exception up in a transparent manner
| 1
|
455,287
| 13,123,815,690
|
IssuesEvent
|
2020-08-06 01:47:34
|
w3c/w3process
|
https://api.github.com/repos/w3c/w3process
|
opened
|
Tooling Requirements
|
P2021: Priority
|
Tooling across WGs is fragmenting severely, and is sometimes resulting in inaccessible WG operations. We should include in our Process some of the core requirements, e.g. archived for posterity, accessible to participants with disabilities and in different countries, documented so that newcomers and observers can find discussions, etc. (And put in a separate policy document details / best practices / Team support status / etc. that can be maintained by the systeam to reflect the current state of the world.)
Some initial thoughts on this topic have been drafted into https://www.w3.org/wiki/W3C_Tooling_Policy https://www.w3.org/wiki/Tooling_Considerations and https://lists.w3.org/Archives/Member/w3c-ac-forum/2020AprJun/0137.html
|
1.0
|
Tooling Requirements - Tooling across WGs is fragmenting severely, and is sometimes resulting in inaccessible WG operations. We should include in our Process some of the core requirements, e.g. archived for posterity, accessible to participants with disabilities and in different countries, documented so that newcomers and observers can find discussions, etc. (And put in a separate policy document details / best practices / Team support status / etc. that can be maintained by the systeam to reflect the current state of the world.)
Some initial thoughts on this topic have been drafted into https://www.w3.org/wiki/W3C_Tooling_Policy https://www.w3.org/wiki/Tooling_Considerations and https://lists.w3.org/Archives/Member/w3c-ac-forum/2020AprJun/0137.html
|
non_process
|
tooling requirements tooling across wgs is fragmenting severely and is sometimes resulting in inaccessible wg operations we should include in our process some of the core requirements e g archived for posterity accessible to participants with disabilities and in different countries documented so that newcomers and observers can find discussions etc and put in a separate policy document details best practices team support status etc that can be maintained by the systeam to reflect the current state of the world some initial thoughts on this topic have been drafted into and
| 0
|
15,609
| 19,730,329,523
|
IssuesEvent
|
2022-01-14 01:12:20
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Script used to create Automation RunAs Accounts
|
automation/svc triaged cxp awaiting-product-team-response docs-experience process-automation/subsvc Pri1
|
Hello, the article itself is great and was a good read. I'd like to know if it's possible to be given the script that's used on the backend to create automation RunAs accounts. They often expire without me knowing and i'd like to add the script into a runbook so it deploys every few months or so. Thanks
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 8721e209-24ce-2170-6caa-ed12a7060080
* Version Independent ID: ac13f91d-460c-cbe9-4778-50d20765b252
* Content: [Azure Automation account authentication overview](https://docs.microsoft.com/en-us/azure/automation/automation-security-overview)
* Content Source: [articles/automation/automation-security-overview.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-security-overview.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SGSneha
* Microsoft Alias: **v-ssudhir**
|
1.0
|
Script used to create Automation RunAs Accounts -
Hello, the article itself is great and was a good read. I'd like to know if it's possible to be given the script that's used on the backend to create automation RunAs accounts. They often expire without me knowing and i'd like to add the script into a runbook so it deploys every few months or so. Thanks
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 8721e209-24ce-2170-6caa-ed12a7060080
* Version Independent ID: ac13f91d-460c-cbe9-4778-50d20765b252
* Content: [Azure Automation account authentication overview](https://docs.microsoft.com/en-us/azure/automation/automation-security-overview)
* Content Source: [articles/automation/automation-security-overview.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-security-overview.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SGSneha
* Microsoft Alias: **v-ssudhir**
|
process
|
script used to create automation runas accounts hello the article itself is great and was a good read i d like to know if it s possible to be given the script that s used on the backend to create automation runas accounts they often expire without me knowing and i d like to add the script into a runbook so it deploys every few months or so thanks document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login sgsneha microsoft alias v ssudhir
| 1
|
7,559
| 10,679,650,958
|
IssuesEvent
|
2019-10-21 19:43:15
|
ECP-WarpX/WarpX
|
https://api.github.com/repos/ECP-WarpX/WarpX
|
closed
|
Slice for back-transformed data in boosted-frame simulations
|
post-processing
|
Generate slice diagnostics to obtained back-transformed data on a reduced domain (1D, 2D slice or 3D reduced box).
See
https://github.com/ECP-WarpX/WarpX/pull/197
|
1.0
|
Slice for back-transformed data in boosted-frame simulations - Generate slice diagnostics to obtained back-transformed data on a reduced domain (1D, 2D slice or 3D reduced box).
See
https://github.com/ECP-WarpX/WarpX/pull/197
|
process
|
slice for back transformed data in boosted frame simulations generate slice diagnostics to obtained back transformed data on a reduced domain slice or reduced box see
| 1
|
80,619
| 7,752,321,821
|
IssuesEvent
|
2018-05-30 19:54:52
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
github.com/cockroachdb/cockroach/pkg/storage: TestReplicateQueueRebalance failed under stress
|
A-core-kv C-test-failure O-robot
|
SHA: https://github.com/cockroachdb/cockroach/commits/89690e007f3f4bab973a107754f471d140d5373f
Parameters:
```
TAGS=
GOFLAGS=-race
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=686673&tab=buildLog
```
I180530 08:03:03.273889 137577 storage/replica_raftstorage.go:730 [n4,s4,r10/?:{-}] applying preemptive snapshot at index 147 (id=25e68233, encoded size=79301, 1 rocksdb batches, 50 log entries)
I180530 08:03:03.277311 134834 storage/replica_proposal.go:202 [n1,s1,r24/1:/{Table/53-Max}] new range lease repl=(n1,s1):1 seq=3 start=1527667357.762478092,0 epo=1 pro=1527667366.622400772,0 following repl=(n1,s1):1 seq=3 start=1527667357.762478092,0 epo=1 pro=1527667366.622400772,0
I180530 08:03:03.365237 137497 storage/replica.go:3320 [n5,s5,r13/3:/Table/1{6-7}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n4,s4):4 (n2,s2):2 (n5,s5):3] next=5
I180530 08:03:03.437939 137577 storage/replica_raftstorage.go:736 [n4,s4,r10/?:/Table/1{3-4}] applied preemptive snapshot in 164ms [clear=0ms batch=0ms entries=161ms commit=2ms]
W180530 08:03:03.454242 137521 storage/replica.go:3131 [n5,s5,r13/3:/Table/1{6-7}] context canceled before proposing: 1 HeartbeatTxn
I180530 08:03:03.471034 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.471502 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.481953 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.484759 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.486223 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.487058 137515 storage/replica_command.go:1777 [replicate,n1,s1,r10/1:/Table/1{3-4}] change replicas (ADD_REPLICA (n4,s4):4): read existing descriptor r10:/Table/1{3-4} [(n1,s1):1, (n3,s3):2, (n5,s5):3, next=4]
I180530 08:03:03.490659 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.491741 136462 storage/store.go:3508 [n1,s1,r13/1:/Table/1{6-7}] added to replica GC queue (peer suggestion)
I180530 08:03:03.503804 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.512252 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.512759 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.513974 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.514300 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.520958 137558 storage/raft_transport.go:459 [n3] raft transport stream to node 5 established
I180530 08:03:03.540034 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.553581 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.553891 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.564197 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.569293 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.574537 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.578505 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.588426 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.590611 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.594118 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.601265 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.606347 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.619177 137457 storage/replica_raftstorage.go:520 [replicate,n5,s5,r2/3:/System/{-NodeLive…}] generated preemptive snapshot 5c19a28a at index 35
I180530 08:03:03.629308 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.631829 137563 storage/store.go:2496 [replicaGC,n1,s1,r13/1:/Table/1{6-7}] removing replica
I180530 08:03:03.635390 137563 storage/replica.go:829 [replicaGC,n1,s1,r13/1:/Table/1{6-7}] removed 8 (0+8) keys in 2ms [clear=0ms commit=2ms]
I180530 08:03:03.646053 134647 storage/replicate_queue_test.go:94 not balanced: [24 16 14 10 14]
I180530 08:03:03.656411 137457 storage/store_snapshot.go:605 [replicate,n5,s5,r2/3:/System/{-NodeLive…}] streamed snapshot to (n4,s4):?: kv pairs: 11, log entries: 16, rate-limit: 2.0 MiB/sec, 33ms
I180530 08:03:03.664356 137606 storage/replica_raftstorage.go:730 [n4,s4,r2/?:{-}] applying preemptive snapshot at index 35 (id=5c19a28a, encoded size=56945, 1 rocksdb batches, 16 log entries)
I180530 08:03:03.666032 134647 storage/replicate_queue_test.go:94 not balanced: [24 16 14 11 14]
I180530 08:03:03.731311 134647 storage/replicate_queue_test.go:94 not balanced: [24 16 14 11 14]
I180530 08:03:03.799007 134647 storage/replicate_queue_test.go:94 not balanced: [24 16 14 11 14]
I180530 08:03:03.861955 137606 storage/replica_raftstorage.go:736 [n4,s4,r2/?:/System/{-NodeLive…}] applied preemptive snapshot in 197ms [clear=0ms batch=0ms entries=168ms commit=3ms]
I180530 08:03:03.877066 137515 storage/replica.go:3320 [n1,s1,r10/1:/Table/1{3-4}] proposing ADD_REPLICA((n4,s4):4): updated=[(n1,s1):1 (n3,s3):2 (n5,s5):3 (n4,s4):4] next=5
I180530 08:03:03.917594 137457 storage/replica_command.go:1777 [replicate,n5,s5,r2/3:/System/{-NodeLive…}] change replicas (ADD_REPLICA (n4,s4):4): read existing descriptor r2:/System/{-NodeLiveness} [(n1,s1):1, (n3,s3):2, (n5,s5):3, next=4]
I180530 08:03:03.933851 134647 storage/replicate_queue_test.go:94 not balanced: [24 16 14 11 14]
I180530 08:03:03.988124 137661 storage/replica_raftstorage.go:520 [replicate,n1,s1,r1/1:/{Min-System/}] generated preemptive snapshot 7d835994 at index 177
I180530 08:03:04.001521 136424 storage/replica_proposal.go:202 [n5,s5,r10/3:/Table/1{3-4}] new range lease repl=(n5,s5):3 seq=4 start=1527667383.923595251,0 epo=1 pro=1527667383.923610141,0 following repl=(n1,s1):1 seq=3 start=1527667357.762478092,0 epo=1 pro=1527667366.995510910,0
I180530 08:03:04.100725 137661 storage/store_snapshot.go:605 [replicate,n1,s1,r1/1:/{Min-System/}] streamed snapshot to (n5,s5):?: kv pairs: 114, log entries: 58, rate-limit: 2.0 MiB/sec, 111ms
I180530 08:03:04.108915 137623 storage/replica_raftstorage.go:730 [n5,s5,r1/?:{-}] applying preemptive snapshot at index 177 (id=7d835994, encoded size=16094, 1 rocksdb batches, 58 log entries)
I180530 08:03:04.193570 137623 storage/replica_raftstorage.go:736 [n5,s5,r1/?:/{Min-System/}] applied preemptive snapshot in 84ms [clear=0ms batch=0ms entries=83ms commit=1ms]
I180530 08:03:04.205800 134647 storage/replicate_queue_test.go:94 not balanced: [24 16 14 11 15]
I180530 08:03:04.468945 137661 storage/replica_command.go:1777 [replicate,n1,s1,r1/1:/{Min-System/}] change replicas (ADD_REPLICA (n5,s5):4): read existing descriptor r1:/{Min-System/} [(n1,s1):1, (n4,s4):2, (n3,s3):3, next=4]
I180530 08:03:04.843957 137740 storage/replica.go:3320 [n1,s1,r1/1:/{Min-System/}] proposing ADD_REPLICA((n5,s5):4): updated=[(n1,s1):1 (n4,s4):2 (n3,s3):3 (n5,s5):4] next=5
I180530 08:03:04.846363 134647 storage/replicate_queue_test.go:94 not balanced: [24 16 14 11 15]
I180530 08:03:04.858516 137692 storage/replica.go:3320 [n5,s5,r2/3:/System/{-NodeLive…}] proposing ADD_REPLICA((n4,s4):4): updated=[(n1,s1):1 (n3,s3):2 (n5,s5):3 (n4,s4):4] next=5
I180530 08:03:04.939134 137634 storage/replica_command.go:1777 [replicate,n5,s5,r15/3:/Table/1{8-9}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r15:/Table/1{8-9} [(n1,s1):1, (n4,s4):2, (n5,s5):3, (n3,s3):4, next=5]
I180530 08:03:05.158183 137785 storage/replica_raftstorage.go:520 [replicate,n1,s1,r8/1:/Table/1{1-2}] generated preemptive snapshot eb16c8b1 at index 25
E180530 08:03:05.165417 137859 storage/queue.go:778 [replicate,n3,s3,r1/3:/{Min-System/}] no removable replicas from range that needs a removal: [raft progress unknown]
I180530 08:03:05.201642 137785 storage/store_snapshot.go:605 [replicate,n1,s1,r8/1:/Table/1{1-2}] streamed snapshot to (n3,s3):?: kv pairs: 11, log entries: 15, rate-limit: 2.0 MiB/sec, 35ms
I180530 08:03:05.205018 137862 storage/replica_raftstorage.go:730 [n3,s3,r8/?:{-}] applying preemptive snapshot at index 25 (id=eb16c8b1, encoded size=5859, 1 rocksdb batches, 15 log entries)
I180530 08:03:05.218754 137862 storage/replica_raftstorage.go:736 [n3,s3,r8/?:/Table/1{1-2}] applied preemptive snapshot in 13ms [clear=0ms batch=0ms entries=12ms commit=0ms]
I180530 08:03:05.234417 137785 storage/replica_command.go:1777 [replicate,n1,s1,r8/1:/Table/1{1-2}] change replicas (ADD_REPLICA (n3,s3):4): read existing descriptor r8:/Table/1{1-2} [(n1,s1):1, (n2,s2):2, (n4,s4):3, next=4]
I180530 08:03:05.605347 137785 storage/replica.go:3320 [n1,s1,r8/1:/Table/1{1-2}] proposing ADD_REPLICA((n3,s3):4): updated=[(n1,s1):1 (n2,s2):2 (n4,s4):3 (n3,s3):4] next=5
I180530 08:03:05.706603 137857 storage/replica_command.go:1777 [replicate,n2,s2,r14/3:/Table/1{7-8}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r14:/Table/1{7-8} [(n1,s1):1, (n4,s4):2, (n2,s2):3, (n3,s3):4, next=5]
I180530 08:03:05.727595 137801 storage/raft_transport.go:459 [n3] raft transport stream to node 4 established
I180530 08:03:05.746581 135477 storage/replica_proposal.go:202 [n2,s2,r8/2:/Table/1{1-2}] new range lease repl=(n2,s2):2 seq=4 start=1527667385.689873739,0 epo=1 pro=1527667385.689888920,0 following repl=(n1,s1):1 seq=3 start=1527667357.762478092,0 epo=1 pro=1527667367.179186503,0
I180530 08:03:05.809625 137868 storage/replica_raftstorage.go:520 [replicate,n1,s1,r20/1:/Table/{23-50}] generated preemptive snapshot 05341a94 at index 24
I180530 08:03:05.904355 137868 storage/store_snapshot.go:605 [replicate,n1,s1,r20/1:/Table/{23-50}] streamed snapshot to (n4,s4):?: kv pairs: 15, log entries: 14, rate-limit: 2.0 MiB/sec, 33ms
I180530 08:03:05.907673 137940 storage/replica_raftstorage.go:730 [n4,s4,r20/?:{-}] applying preemptive snapshot at index 24 (id=05341a94, encoded size=4934, 1 rocksdb batches, 14 log entries)
I180530 08:03:05.939097 134647 storage/replicate_queue_test.go:94 not balanced: [24 16 15 12 15]
I180530 08:03:05.966461 137940 storage/replica_raftstorage.go:736 [n4,s4,r20/?:/Table/{23-50}] applied preemptive snapshot in 58ms [clear=0ms batch=0ms entries=57ms commit=0ms]
I180530 08:03:06.068020 137868 storage/replica_command.go:1777 [replicate,n1,s1,r20/1:/Table/{23-50}] change replicas (ADD_REPLICA (n4,s4):4): read existing descriptor r20:/Table/{23-50} [(n1,s1):1, (n3,s3):2, (n5,s5):3, next=4]
I180530 08:03:06.315430 137950 storage/replica.go:3320 [n5,s5,r15/3:/Table/1{8-9}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n3,s3):4 (n4,s4):2 (n5,s5):3] next=5
I180530 08:03:06.411922 137951 storage/replica_command.go:1777 [replicate,n5,s5,r2/3:/System/{-NodeLive…}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r2:/System/{-NodeLiveness} [(n1,s1):1, (n3,s3):2, (n5,s5):3, (n4,s4):4, next=5]
I180530 08:03:06.468100 136462 storage/store.go:3508 [n1,s1,r15/1:/Table/1{8-9}] added to replica GC queue (peer suggestion)
I180530 08:03:06.865267 137954 storage/store.go:2496 [replicaGC,n1,s1,r15/1:/Table/1{8-9}] removing replica
I180530 08:03:06.872442 137941 storage/raft_transport.go:459 [n4] raft transport stream to node 3 established
I180530 08:03:06.899959 137954 storage/replica.go:829 [replicaGC,n1,s1,r15/1:/Table/1{8-9}] removed 8 (0+8) keys in 33ms [clear=0ms commit=33ms]
I180530 08:03:06.941305 138008 storage/replica.go:3320 [n1,s1,r20/1:/Table/{23-50}] proposing ADD_REPLICA((n4,s4):4): updated=[(n1,s1):1 (n3,s3):2 (n5,s5):3 (n4,s4):4] next=5
W180530 08:03:06.949385 138024 storage/intent_resolver.go:320 [n3,s3] failed to push during intent resolution: failed to push "change-replica" id=675eb1e3 key=/Local/Range/Table/23/RangeDescriptor rw=true pri=0.01733080 iso=SERIALIZABLE stat=PENDING epo=0 ts=1527667386.030283448,0 orig=1527667386.030283448,0 max=1527667386.030283448,0 wto=false rop=false seq=2
I180530 08:03:06.973530 134647 storage/replicate_queue_test.go:94 not balanced: [23 16 15 12 15]
I180530 08:03:07.200909 138073 storage/replica.go:3320 [n2,s2,r14/3:/Table/1{7-8}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n3,s3):4 (n4,s4):2 (n2,s2):3] next=5
I180530 08:03:07.329645 135883 storage/store.go:3508 [n1,s1,r14/1:/Table/1{7-8}] added to replica GC queue (peer suggestion)
I180530 08:03:07.338457 135737 storage/replica_proposal.go:202 [n3,s3,r20/2:/Table/{23-50}] new range lease repl=(n3,s3):2 seq=4 start=1527667387.099999044,0 epo=1 pro=1527667387.100014019,0 following repl=(n1,s1):1 seq=3 start=1527667357.762478092,0 epo=1 pro=1527667367.166295614,0
I180530 08:03:07.370467 138057 storage/replica_raftstorage.go:520 [replicate,n1,s1,r5/1:/System/ts{d-e}] generated preemptive snapshot 22be3a91 at index 44
I180530 08:03:07.451894 135359 server/status/runtime.go:219 [n2] runtime stats: 1.4 GiB RSS, 1158 goroutines, 77 MiB/29 MiB/138 MiB GO alloc/idle/total, 158 MiB/201 MiB CGO alloc/total, 1180.92cgo/sec, 1.39/0.12 %(u/s)time, 0.02 %gc (6x)
I180530 08:03:07.502056 138109 storage/replica_command.go:1777 [replicate,n3,s3,r1/3:/{Min-System/}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r1:/{Min-System/} [(n1,s1):1, (n4,s4):2, (n3,s3):3, (n5,s5):4, next=5]
I180530 08:03:07.652014 138076 storage/store.go:2496 [replicaGC,n1,s1,r14/1:/Table/1{7-8}] removing replica
I180530 08:03:07.684198 138091 storage/replica_command.go:1777 [replicate,n2,s2,r8/2:/Table/1{1-2}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r8:/Table/1{1-2} [(n1,s1):1, (n2,s2):2, (n4,s4):3, (n3,s3):4, next=5]
I180530 08:03:07.879883 134886 server/status/runtime.go:219 [n1] runtime stats: 1.4 GiB RSS, 1152 goroutines, 88 MiB/19 MiB/138 MiB GO alloc/idle/total, 157 MiB/201 MiB CGO alloc/total, 1372.68cgo/sec, 1.39/0.12 %(u/s)time, 0.02 %gc (6x)
I180530 08:03:07.919865 138057 storage/store_snapshot.go:605 [replicate,n1,s1,r5/1:/System/ts{d-e}] streamed snapshot to (n4,s4):?: kv pairs: 2167, log entries: 3, rate-limit: 2.0 MiB/sec, 548ms
I180530 08:03:07.924031 138076 storage/replica.go:829 [replicaGC,n1,s1,r14/1:/Table/1{7-8}] removed 8 (0+8) keys in 205ms [clear=0ms commit=205ms]
I180530 08:03:07.934630 138108 storage/replica_raftstorage.go:730 [n4,s4,r5/?:{-}] applying preemptive snapshot at index 44 (id=22be3a91, encoded size=351993, 1 rocksdb batches, 3 log entries)
I180530 08:03:07.991175 134647 storage/replicate_queue_test.go:94 not balanced: [22 16 15 13 15]
I180530 08:03:08.029788 138143 storage/replica.go:3320 [n5,s5,r2/3:/System/{-NodeLive…}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n4,s4):4 (n3,s3):2 (n5,s5):3] next=5
I180530 08:03:08.205764 138108 storage/replica_raftstorage.go:736 [n4,s4,r5/?:/System/ts{d-e}] applied preemptive snapshot in 271ms [clear=0ms batch=2ms entries=266ms commit=2ms]
I180530 08:03:08.265580 136462 storage/store.go:3508 [n1,s1,r2/1:/System/{-NodeLive…}] added to replica GC queue (peer suggestion)
I180530 08:03:08.282224 138057 storage/replica_command.go:1777 [replicate,n1,s1,r5/1:/System/ts{d-e}] change replicas (ADD_REPLICA (n4,s4):4): read existing descriptor r5:/System/ts{d-e} [(n1,s1):1, (n3,s3):2, (n2,s2):3, next=4]
I180530 08:03:08.288916 138228 storage/replica_command.go:1777 [replicate,n5,s5,r10/3:/Table/1{3-4}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r10:/Table/1{3-4} [(n1,s1):1, (n3,s3):2, (n5,s5):3, (n4,s4):4, next=5]
I180530 08:03:08.601014 135675 server/status/runtime.go:219 [n3] runtime stats: 1.5 GiB RSS, 1159 goroutines, 110 MiB/6.2 MiB/144 MiB GO alloc/idle/total, 158 MiB/203 MiB CGO alloc/total, 1373.17cgo/sec, 1.40/0.11 %(u/s)time, 0.02 %gc (6x)
I180530 08:03:08.812404 138213 storage/store.go:2496 [replicaGC,n1,s1,r2/1:/System/{-NodeLive…}] removing replica
I180530 08:03:08.814577 138213 storage/replica.go:829 [replicaGC,n1,s1,r2/1:/System/{-NodeLive…}] removed 8 (0+8) keys in 1ms [clear=0ms commit=1ms]
I180530 08:03:08.849996 138275 storage/replica.go:3320 [n3,s3,r1/3:/{Min-System/}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n5,s5):4 (n4,s4):2 (n3,s3):3] next=5
I180530 08:03:09.030224 134647 storage/replicate_queue_test.go:94 not balanced: [21 16 15 13 15]
I180530 08:03:09.084849 138294 storage/store.go:2496 [replicaGC,n1,s1,r1/1:/{Min-System/}] removing replica
I180530 08:03:09.087464 138294 storage/replica.go:829 [replicaGC,n1,s1,r1/1:/{Min-System/}] removed 35 (26+9) keys in 2ms [clear=0ms commit=1ms]
I180530 08:03:09.260541 138210 storage/replica_command.go:1777 [replicate,n3,s3,r20/2:/Table/{23-50}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r20:/Table/{23-50} [(n1,s1):1, (n3,s3):2, (n5,s5):3, (n4,s4):4, next=5]
I180530 08:03:09.330240 138057 storage/replica.go:3320 [n1,s1,r5/1:/System/ts{d-e}] proposing ADD_REPLICA((n4,s4):4): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3 (n4,s4):4] next=5
W180530 08:03:09.432827 135677 server/node.go:802 [n3,summaries] health alerts detected: {Alerts:[{StoreID:3 Category:METRICS Description:queue.replicate.process.failure Value:1}]}
I180530 08:03:09.516300 138240 storage/replica.go:3320 [n2,s2,r8/2:/Table/1{1-2}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n3,s3):4 (n2,s2):2 (n4,s4):3] next=5
I180530 08:03:09.688443 135639 storage/replica_proposal.go:202 [n3,s3,r5/2:/System/ts{d-e}] new range lease repl=(n3,s3):2 seq=4 start=1527667389.417811299,0 epo=1 pro=1527667389.417824987,0 following repl=(n1,s1):1 seq=3 start=1527667357.762478092,0 epo=1 pro=1527667366.661823335,0
I180530 08:03:09.749964 138364 storage/replica_raftstorage.go:520 [replicate,n2,s2,r8/2:/Table/1{1-2}] generated preemptive snapshot 7df7c51a at index 33
I180530 08:03:09.891483 135993 server/status/runtime.go:219 [n4] runtime stats: 1.5 GiB RSS, 1157 goroutines, 102 MiB/12 MiB/144 MiB GO alloc/idle/total, 158 MiB/204 MiB CGO alloc/total, 1346.95cgo/sec, 1.39/0.12 %(u/s)time, 0.01 %gc (5x)
I180530 08:03:09.901951 138364 storage/store_snapshot.go:605 [replicate,n2,s2,r8/2:/Table/1{1-2}] streamed snapshot to (n5,s5):?: kv pairs: 14, log entries: 23, rate-limit: 2.0 MiB/sec, 132ms
I180530 08:03:09.911719 135883 storage/store.go:3508 [n1,s1,r8/1:/Table/1{1-2}] added to replica GC queue (peer suggestion)
I180530 08:03:09.939996 138312 storage/store.go:2496 [replicaGC,n1,s1,r8/1:/Table/1{1-2}] removing replica
I180530 08:03:09.971866 138391 storage/replica_raftstorage.go:520 [replicate,n1,s1,r21/1:/Table/5{0-1}] generated preemptive snapshot 9d827d66 at index 24
I180530 08:03:09.990858 138367 storage/replica_raftstorage.go:730 [n5,s5,r8/?:{-}] applying preemptive snapshot at index 33 (id=7df7c51a, encoded size=8789, 1 rocksdb batches, 23 log entries)
I180530 08:03:10.006369 138367 storage/replica_raftstorage.go:736 [n5,s5,r8/?:/Table/1{1-2}] applied preemptive snapshot in 15ms [clear=0ms batch=0ms entries=13ms commit=1ms]
I180530 08:03:10.029771 138312 storage/replica.go:829 [replicaGC,n1,s1,r8/1:/Table/1{1-2}] removed 8 (0+8) keys in 81ms [clear=0ms commit=80ms]
I180530 08:03:10.046214 138391 storage/store_snapshot.go:605 [replicate,n1,s1,r21/1:/Table/5{0-1}] streamed snapshot to (n5,s5):?: kv pairs: 11, log entries: 14, rate-limit: 2.0 MiB/sec, 70ms
I180530 08:03:10.051795 134647 storage/replicate_queue_test.go:94 not balanced: [19 16 15 13 16]
I180530 08:03:10.079905 138370 storage/replica_raftstorage.go:730 [n5,s5,r21/?:{-}] applying preemptive snapshot at index 24 (id=9d827d66, encoded size=4614, 1 rocksdb batches, 14 log entries)
I180530 08:03:10.128625 138364 storage/replica_command.go:1777 [replicate,n2,s2,r8/2:/Table/1{1-2}] change replicas (ADD_REPLICA (n5,s5):5): read existing descriptor r8:/Table/1{1-2} [(n3,s3):4, (n2,s2):2, (n4,s4):3, next=5]
==================
WARNING: DATA RACE
Write at 0x00c42306c8c9 by goroutine 137:
github.com/cockroachdb/cockroach/pkg/roachpb.(*Transaction).Update()
/go/src/github.com/cockroachdb/cockroach/pkg/roachpb/data.go:932 +0xd6f
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).heartbeat()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1126 +0x973
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).heartbeatLoop()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:967 +0x3f8
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).updateState.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1333 +0x55
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:264 +0xf3
Previous read at 0x00c42306c8c9 by goroutine 661:
[failed to restore the stack]
Goroutine 137 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:259 +0x14c
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).updateState()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1331 +0x673
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:494 +0x37f
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).sendUsingSender()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:553 +0x106
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:973 +0x66e
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Send-fm()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:500 +0x96
github.com/cockroachdb/cockroach/pkg/internal/client.sendAndFill()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:468 +0x13d
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Run()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:500 +0xfd
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).changeReplicas.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1790 +0x535
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:516 +0x5e
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Exec()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:748 +0x128
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:515 +0x192
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).changeReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1765 +0xa6f
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).ChangeReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1670 +0xc9
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).removeReplica()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:591 +0xd9
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).processOneChange()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:386 +0x14d1
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).process()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:218 +0x28b
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processReplica()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:734 +0x313
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1.2()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:617 +0xdd
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:264 +0xf3
Goroutine 661 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:259 +0x14c
github.com/cockroachdb/cockroach/pkg/kv.GRPCTransportFactory()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/transport_race.go:61 +0x2f7
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendToReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1263 +0x14e
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendRPC()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:393 +0x2be
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendSingleRange()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:460 +0x265
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendPartialBatch()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1053 +0x3dd
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:738 +0x1a05
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:655 +0x565
github.com/cockroachdb/cockroach/pkg/internal/client.lookupRangeFwdScan()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/range_lookup.go:339 +0x4e7
github.com/cockroachdb/cockroach/pkg/internal/client.RangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/range_lookup.go:202 +0xb0e
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).RangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:299 +0x104
github.com/cockroachdb/cockroach/pkg/kv.(*RangeDescriptorCache).performRangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/range_cache.go:434 +0x42d
github.com/cockroachdb/cockroach/pkg/kv.(*RangeDescriptorCache).lookupRangeDescriptorInternal.func3()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/range_cache.go:310 +0x167
github.com/cockroachdb/cockroach/pkg/util/syncutil/singleflight.(*Group).doCall()
/go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/singleflight/singleflight.go:118 +0x4c
==================
==================
WARNING: DATA RACE
Write at 0x00c42306c8f8 by goroutine 137:
github.com/cockroachdb/cockroach/pkg/roachpb.(*Transaction).Update()
/go/src/github.com/cockroachdb/cockroach/pkg/roachpb/data.go:934 +0xe05
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).heartbeat()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1126 +0x973
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).heartbeatLoop()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:967 +0x3f8
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).updateState.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1333 +0x55
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:264 +0xf3
Previous read at 0x00c42306c8f8 by goroutine 661:
[failed to restore the stack]
Goroutine 137 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:259 +0x14c
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).updateState()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1331 +0x673
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:494 +0x37f
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).sendUsingSender()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:553 +0x106
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:973 +0x66e
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Send-fm()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:500 +0x96
github.com/cockroachdb/cockroach/pkg/internal/client.sendAndFill()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:468 +0x13d
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Run()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:500 +0xfd
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).changeReplicas.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1790 +0x535
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:516 +0x5e
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Exec()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:748 +0x128
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:515 +0x192
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).changeReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1765 +0xa6f
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).ChangeReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1670 +0xc9
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).removeReplica()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:591 +0xd9
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).processOneChange()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:386 +0x14d1
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).process()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:218 +0x28b
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processReplica()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:734 +0x313
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1.2()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:617 +0xdd
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:264 +0xf3
Goroutine 661 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:259 +0x14c
github.com/cockroachdb/cockroach/pkg/kv.GRPCTransportFactory()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/transport_race.go:61 +0x2f7
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendToReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1263 +0x14e
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendRPC()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:393 +0x2be
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendSingleRange()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:460 +0x265
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendPartialBatch()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1053 +0x3dd
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:738 +0x1a05
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:655 +0x565
github.com/cockroachdb/cockroach/pkg/internal/client.lookupRangeFwdScan()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/range_lookup.go:339 +0x4e7
github.com/cockroachdb/cockroach/pkg/internal/client.RangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/range_lookup.go:202 +0xb0e
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).RangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:299 +0x104
github.com/cockroachdb/cockroach/pkg/kv.(*RangeDescriptorCache).performRangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/range_cache.go:434 +0x42d
github.com/cockroachdb/cockroach/pkg/kv.(*RangeDescriptorCache).lookupRangeDescriptorInternal.func3()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/range_cache.go:310 +0x167
github.com/cockroachdb/cockroach/pkg/util/syncutil/singleflight.(*Group).doCall()
/go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/singleflight/singleflight.go:118 +0x4c
==================
==================
WARNING: DATA RACE
Write at 0x00c42306c870 by goroutine 137:
github.com/cockroachdb/cockroach/pkg/roachpb.(*Transaction).Update()
/go/src/github.com/cockroachdb/cockroach/pkg/util/hlc/timestamp.go:106 +0xbfc
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).heartbeat()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1126 +0x973
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).heartbeatLoop()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:967 +0x3f8
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).updateState.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1333 +0x55
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:264 +0xf3
Previous read at 0x00c42306c870 by goroutine 661:
[failed to restore the stack]
Goroutine 137 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:259 +0x14c
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).updateState()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1331 +0x673
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:494 +0x37f
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).sendUsingSender()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:553 +0x106
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:973 +0x66e
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Send-fm()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:500 +0x96
github.com/cockroachdb/cockroach/pkg/internal/client.sendAndFill()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:468 +0x13d
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Run()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:500 +0xfd
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).changeReplicas.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1790 +0x535
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:516 +0x5e
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Exec()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:748 +0x128
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:515 +0x192
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).changeReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1765 +0xa6f
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).ChangeReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1670 +0xc9
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).removeReplica()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:591 +0xd9
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).processOneChange()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:386 +0x14d1
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).process()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:218 +0x28b
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processReplica()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:734 +0x313
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1.2()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:617 +0xdd
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:264 +0xf3
Goroutine 661 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:259 +0x14c
github.com/cockroachdb/cockroach/pkg/kv.GRPCTransportFactory()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/transport_race.go:61 +0x2f7
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendToReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1263 +0x14e
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendRPC()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:393 +0x2be
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendSingleRange()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:460 +0x265
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendPartialBatch()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1053 +0x3dd
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:738 +0x1a05
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:655 +0x565
github.com/cockroachdb/cockroach/pkg/internal/client.lookupRangeFwdScan()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/range_lookup.go:339 +0x4e7
github.com/cockroachdb/cockroach/pkg/internal/client.RangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/range_lookup.go:202 +0xb0e
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).RangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:299 +0x104
github.com/cockroachdb/cockroach/pkg/kv.(*RangeDescriptorCache).performRangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/range_cache.go:434 +0x42d
github.com/cockroachdb/cockroach/pkg/kv.(*RangeDescriptorCache).lookupRangeDescriptorInternal.func3()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/range_cache.go:310 +0x167
github.com/cockroachdb/cockroach/pkg/util/syncutil/singleflight.(*Group).doCall()
/go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/singleflight/singleflight.go:118 +0x4c
==================
==================
WARNING: DATA RACE
Write at 0x00c42306c8b0 by goroutine 137:
github.com/cockroachdb/cockroach/pkg/roachpb.(*Transaction).UpdateObservedTimestamp()
/go/src/github.com/cockroachdb/cockroach/pkg/roachpb/data.go:1027 +0x100
github.com/cockroachdb/cockroach/pkg/roachpb.(*Transaction).Update()
/go/src/github.com/cockroachdb/cockroach/pkg/roachpb/data.go:949 +0x695
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).heartbeat()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1126 +0x973
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).heartbeatLoop()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:967 +0x3f8
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).updateState.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1333 +0x55
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:264 +0xf3
Previous read at 0x00c42306c8b0 by goroutine 661:
[failed to restore the stack]
Goroutine 137 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:259 +0x14c
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).updateState()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1331 +0x673
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:494 +0x37f
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).sendUsingSender()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:553 +0x106
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:973 +0x66e
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Send-fm()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:500 +0x96
github.com/cockroachdb/cockroach/pkg/internal/client.sendAndFill()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:468 +0x13d
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Run()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:500 +0xfd
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).changeReplicas.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1790 +0x535
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:516 +0x5e
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Exec()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:748 +0x128
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:515 +0x192
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).changeReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1765 +0xa6f
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).ChangeReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1670 +0xc9
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).removeReplica()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:591 +0xd9
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).processOneChange()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:386 +0x14d1
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).process()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:218 +0x28b
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processReplica()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:734 +0x313
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1.2()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:617 +0xdd
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:264 +0xf3
Goroutine 661 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:259 +0x14c
github.com/cockroachdb/cockroach/pkg/kv.GRPCTransportFactory()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/transport_race.go:61 +0x2f7
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendToReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1263 +0x14e
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendRPC()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:393 +0x2be
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendSingleRange()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:460 +0x265
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendPartialBatch()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1053 +0x3dd
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:738 +0x1a05
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:655 +0x565
github.com/cockroachdb/cockroach/pkg/internal/client.lookupRangeFwdScan()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/range_lookup.go:339 +0x4e7
github.com/cockroachdb/cockroach/pkg/internal/client.RangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/range_lookup.go:202 +0xb0e
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).RangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:299 +0x104
github.com/cockroachdb/cockroach/pkg/kv.(*RangeDescriptorCache).performRangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/range_cache.go:434 +0x42d
github.com/cockroachdb/cockroach/pkg/kv.(*RangeDescriptorCache).lookupRangeDescriptorInternal.func3()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/range_cache.go:310 +0x167
github.com/cockroachdb/cockroach/pkg/util/syncutil/singleflight.(*Group).doCall()
/go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/singleflight/singleflight.go:118 +0x4c
==================
==================
WARNING: DATA RACE
Write at 0x00c42306c8c8 by goroutine 137:
github.com/cockroachdb/cockroach/pkg/roachpb.(*Transaction).Update()
/go/src/github.com/cockroachdb/cockroach/pkg/roachpb/data.go:955 +0x724
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).heartbeat()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1126 +0x973
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).heartbeatLoop()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:967 +0x3f8
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).updateState.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1333 +0x55
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:264 +0xf3
Previous read at 0x00c42306c8c8 by goroutine 661:
[failed to restore the stack]
Goroutine 137 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:259 +0x14c
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).updateState()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1331 +0x673
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:494 +0x37f
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).sendUsingSender()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:553 +0x106
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:973 +0x66e
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Send-fm()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:500 +0x96
github.com/cockroachdb/cockroach/pkg/internal/client.sendAndFill()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:468 +0x13d
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Run()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:500 +0xfd
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).changeReplicas.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1790 +0x535
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:516 +0x5e
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Exec()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:748 +0x128
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:515 +0x192
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).changeReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1765 +0xa6f
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).ChangeReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1670 +0xc9
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).removeReplica()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:591 +0xd9
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).processOneChange()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:386 +0x14d1
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).process()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:218 +0x28b
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processReplica()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:734 +0x313
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1.2()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:617 +0xdd
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:264 +0xf3
Goroutine 661 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:259 +0x14c
github.com/cockroachdb/cockroach/pkg/kv.GRPCTransportFactory()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/transport_race.go:61 +0x2f7
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendToReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1263 +0x14e
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendRPC()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:393 +0x2be
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendSingleRange()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:460 +0x265
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendPartialBatch()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1053 +0x3dd
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:738 +0x1a05
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:655 +0x565
github.com/cockroachdb/cockroach/pkg/internal/client.lookupRangeFwdScan()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/range_lookup.go:339 +0x4e7
github.com/cockroachdb/cockroach/pkg/internal/client.RangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/range_lookup.go:202 +0xb0e
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).RangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:299 +0x104
github.com/cockroachdb/cockroach/pkg/kv.(*RangeDescriptorCache).performRangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/range_cache.go:434 +0x42d
github.com/cockroachdb/cockroach/pkg/kv.(*RangeDescriptorCache).lookupRangeDescriptorInternal.func3()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/range_cache.go:310 +0x167
github.com/cockroachdb/cockroach/pkg/util/syncutil/singleflight.(*Group).doCall()
/go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/singleflight/singleflight.go:118 +0x4c
==================
I180530 08:03:10.582713 138487 storage/replica.go:3320 [n5,s5,r10/3:/Table/1{3-4}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n4,s4):4 (n3,s3):2 (n5,s5):3] next=5
I180530 08:03:10.606967 138370 storage/replica_raftstorage.go:736 [n5,s5,r21/?:/Table/5{0-1}] applied preemptive snapshot in 527ms [clear=0ms batch=0ms entries=431ms commit=20ms]
I180530 08:03:10.616792 136293 server/status/runtime.go:219 [n5] runtime stats: 1.5 GiB RSS, 1153 goroutines, 69 MiB/43 MiB/144 MiB GO alloc/idle/total, 157 MiB/204 MiB CGO alloc/total, 1301.03cgo/sec, 1.39/0.11 %(u/s)time, 0.02 %gc (6x)
I180530 08:03:10.732813 138491 storage/replica.go:3320 [n3,s3,r20/2:/Table/{23-50}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n4,s4):4 (n3,s3):2 (n5,s5):3] next=5
I180530 08:03:10.779589 138391 storage/replica_command.go:1777 [replicate,n1,s1,r21/1:/Table/5{0-1}] change replicas (ADD_REPLICA (n5,s5):4): read existing descriptor r21:/Table/5{0-1} [(n1,s1):1, (n2,s2):2, (n4,s4):3, next=4]
W180530 08:03:10.858350 138514 storage/replica.go:3131 [n5,s5,r10/3:/Table/1{3-4}] context canceled before proposing: 1 HeartbeatTxn
I180530 08:03:10.963074 138477 storage/replica_raftstorage.go:520 [replicate,n5,s5,r13/3:/Table/1{6-7}] generated preemptive snapshot d39327b1 at index 33
I180530 08:03:10.980790 136462 storage/store.go:3508 [n1,s1,r10/1:/Table/1{3-4}] added to replica GC queue (peer suggestion)
I180530 08:03:10.981529 136462 storage/store.go:3508 [n1,s1,r10/1:/Table/1{3-4}] added to replica GC queue (peer suggestion)
I180530 08:03:11.000981 136451 storage/store.go:3508 [n1,s1,r20/1:/Table/{23-50}] added to replica GC queue (peer suggestion)
I180530 08:03:11.037791 136451 storage/store.go:3508 [n1,s1,r20/1:/Table/{23-50}] added to replica GC queue (peer suggestion)
I180530 08:03:11.089333 134647 storage/replicate_queue_test.go:94 not balanced: [19 16 15 13 17]
I180530 08:03:11.089912 136451 storage/store.go:3508 [n1,s1,r20/1:/Table/{23-50}] added to replica GC queue (peer suggestion)
I180530 08:03:11.099719 138477 storage/store_snapshot.go:605 [replicate,n5,s5,r13/3:/Table/1{6-7}] streamed snapshot to (n3,s3):?: kv pairs: 13, log entries: 23, rate-limit: 2.0 MiB/sec, 123ms
I180530 08:03:11.108045 138434 storage/replica_raftstorage.go:730 [n3,s3,r13/?:{-}] applying preemptive snapshot at index 33 (id=d39327b1, encoded size=7268, 1 rocksdb batches, 23 log entries)
I180530 08:03:11.159912 138534 storage/store.go:2496 [replicaGC,n1,s1,r10/1:/Table/1{3-4}] removing replica
I180530 08:03:11.196769 138534 storage/replica.go:829 [replicaGC,n1,s1,r10/1:/Table/1{3-4}] removed 457 (449+8) keys in 36ms [clear=0ms commit=35ms]
I180530 08:03:11.213849 138567 storage/replica_command.go:1777 [replicate,n3,s3,r5/2:/System/ts{d-e}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r5:/System/ts{d-e} [(n1,s1):1, (n3,s3):2, (n2,s2):3, (n4,s4):4, next=5]
I180530 08:03:11.250905 138570 storage/store.go:2496 [replicaGC,n1,s1,r20/1:/Table/{23-50}] removing replica
I180530 08:03:11.259926 138434 storage/replica_raftstorage.go:736 [n3,s3,r13/?:/Table/1{6-7}] applied preemptive snapshot in 152ms [clear=0ms batch=0ms entries=139ms commit=11ms]
I180530 08:03:11.260340 138570 storage/replica.go:829 [replicaGC,n1,s1,r20/1:/Table/{23-50}] removed 12 (4+8) keys in 2ms [clear=0ms commit=2ms]
I180530 08:03:11.301258 138477 storage/replica_command.go:1777 [replicate,n5,s5,r13/3:/Table/1{6-7}] change replicas (ADD_REPLICA (n3,s3):5): read existing descriptor r13:/Table/1{6-7} [(n4,s4):4, (n2,s2):2, (n5,s5):3, next=5]
I180530 08:03:11.773875 138576 storage/replica.go:3320 [n2,s2,r8/2:/Table/1{1-2}] proposing ADD_REPLICA((n5,s5):5): updated=[(n3,s3):4 (n2,s2):2 (n4,s4):3 (n5,s5):5] next=6
I180530 08:03:11.808726 138578 storage/replica.go:3320 [n1,s1,r21/1:/Table/5{0-1}] proposing ADD_REPLICA((n5,s5):4): updated=[(n1,s1):1 (n2,s2):2 (n4,s4):3 (n5,s5):4] next=5
I180530 08:03:11.832736 138619 storage/replica_command.go:1777 [replicate,n2,s2,r8/2:/Table/1{1-2}] change replicas (REMOVE_REPLICA (n5,s5):5): read existing descriptor r8:/Table/1{1-2} [(n3,s3):4, (n2,s2):2, (n4,s4):3, (n5,s5):5, next=6]
I180530 08:03:11.861635 138671 storage/replica_command.go:1777 [replicate,n1,s1,r21/1:/Table/5{0-1}] change replicas (REMOVE_REPLICA (n5,s5):4): read existing descriptor r21:/Table/5{0-1} [(n1,s1):1, (n2,s2):2, (n4,s4):3, (n5,s5):4, next=5]
I180530 08:03:12.111461 134647 storage/replicate_queue_test.go:94 not balanced: [17 16 16 13 17]
I180530 08:03:12.236652 138655 storage/replica.go:3320 [n5,s5,r13/3:/Table/1{6-7}] proposing ADD_REPLICA((n3,s3):5): updated=[(n4,s4):4 (n2,s2):2 (n5,s5):3 (n3,s3):5] next=6
I180530 08:03:12.517898 138771 storage/replica.go:3320 [n3,s3,r5/2:/System/ts{d-e}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n4,s4):4 (n3,s3):2 (n2,s2):3] next=5
I180530 08:03:12.656662 138719 storage/replica.go:3320 [n1,s1,r21/1:/Table/5{0-1}] proposing REMOVE_REPLICA((n5,s5):4): updated=[(n1,s1):1 (n2,s2):2 (n4,s4):3] next=5
I180530 08:03:12.728088 136601 storage/store.go:3508 [n5,s5,r21/4:/Table/5{0-1}] added to replica GC queue (peer suggestion)
I180530 08:03:12.744551 136601 storage/store.go:3508 [n5,s5,r21/4:/Table/5{0-1}] added to replica GC queue (peer suggestion)
I180530 08:03:12.922358 138686 storage/store.go:2496 [replicaGC,n5,s5,r21/4:/Table/5{0-1}] removing replica
I180530 08:03:12.943024 136042 storage/replica_proposal.go:202 [n4,s4,r13/4:/Table/1{6-7}] new range lease repl=(n4,s4):4 seq=5 start=1527667392.464660635,0 epo=1 pro=1527667392.464693705,0 following repl=(n5,s5):3 seq=4 start=1527667380.380046322,0 epo=1 pro=1527667380.380059659,0
I180530 08:03:12.966633 138686 storage/replica.go:829 [replicaGC,n5,s5,r21/4:/Table/5{0-1}] removed 8 (0+8) keys in 43ms [clear=0ms commit=43ms]
I180530 08:03:13.000377 136451 storage/store.go:3508 [n1,s1,r5/1:/System/ts{d-e}] added to replica GC queue (peer suggestion)
I180530 08:03:13.000762 136451 storage/store.go:3508 [n1,s1,r5/1:/System/ts{d-e}] added to replica GC queue (peer suggestion)
I180530 08:03:13.023378 138795 storage/store.go:2496 [replicaGC,n1,s1,r5/1:/System/ts{d-e}] removing replica
I180530 08:03:13.025688 138795 storage/replica.go:829 [replicaGC,n1,s1,r5/1:/System/ts{d-e}] removed 440 (431+9) keys in 1ms [clear=0ms commit=1ms]
I180530 08:03:13.129934 134647 storage/replicate_queue_test.go:94 not balanced: [16 16 16 13 16]
I180530 08:03:13.400499 138636 storage/replica.go:3320 [n2,s2,r8/2:/Table/1{1-2}] proposing REMOVE_REPLICA((n5,s5):5): updated=[(n3,s3):4 (n2,s2):2 (n4,s4):3] next=6
I180530 08:03:13.406799 138810 storage/replica_raftstorage.go:520 [replicate,n1,s1,r16/1:/Table/{19-20}] generated preemptive snapshot fc9b5125 at index 24
I180530 08:03:13.459965 138810 storage/store_snapshot.go:605 [replicate,n1,s1,r16/1:/Table/{19-20}] streamed snapshot to (n4,s4):?: kv pairs: 11, log entries: 14, rate-limit: 2.0 MiB/sec, 50ms
I180530 08:03:13.463457 138832 storage/replica_raftstorage.go:730 [n4,s4,r16/?:{-}] applying preemptive snapshot at index 24 (id=fc9b5125, encoded size=4802, 1 rocksdb batches, 14 log entries)
I180530 08:03:13.558236 137255 storage/store.go:3508 [n5,s5,r8/5:/Table/1{1-2}] added to replica GC queue (peer suggestion)
I180530 08:03:13.690396 138832 storage/replica_raftstorage.go:736 [n4,s4,r16/?:/Table/{19-20}] applied preemptive snapshot in 227ms [clear=0ms batch=0ms entries=129ms commit=1ms]
I180530 08:03:13.730213 138810 storage/replica_command.go:1777 [replicate,n1,s1,r16/1:/Table/{19-20}] change replicas (ADD_REPLICA (n4,s4):4): read existing descriptor r16:/Table/{19-20} [(n1,s1):1, (n5,s5):2, (n2,s2):3, next=4]
I180530 08:03:13.835079 138833 storage/store.go:2496 [replicaGC,n5,s5,r8/5:/Table/1{1-2}] removing replica
I180530 08:03:13.867510 138833 storage/replica.go:829 [replicaGC,n5,s5,r8/5:/Table/1{1-2}] removed 8 (0+8) keys in 31ms [clear=0ms commit=31ms]
I180530 08:03:14.131744 138949 util/stop/stopper.go:471 quiescing; tasks left:
1 [async] transport racer
1 [async] storage.replicate: processing replica
1 [async] kv.TxnCoordSender: heartbeat loop
W180530 08:03:14.143274 136367 storage/raft_transport.go:465 [n1] raft transport stream to node 5 failed: rpc error: code = Unavailable desc = transport is closing
I180530 08:03:14.145859 138949 util/stop/stopper.go:471 quiescing; tasks left:
1 [async] transport racer
1 [async] storage.replicate: processing replica
W180530 08:03:14.149388 134895 storage/node_liveness.go:465 [n1,hb] failed node liveness heartbeat: node unavailable; try another peer
W180530 08:03:14.149715 138810 internal/client/txn.go:526 [replicate,n1,s1,r16/1:/Table/{19-20}] failure aborting transaction: node unavailable; try another peer; abort caused by: log-range-event: failed to send RPC: sending to all 4 replicas failed; last error: {<nil> context canceled}
I180530 08:03:14.152797 135895 gossip/gossip.go:1307 [n4] node has connected to cluster via gossip
I180530 08:03:14.154434 135895 storage/stores.go:241 [n4] wrote 4 node addresses to persistent storage
I180530 08:03:14.155346 138953 util/stop/stopper.go:471 quiescing; tasks left:
1 node.Node: batch
W180530 08:03:14.164827 136561 storage/raft_transport.go:465 [n4] raft transport stream to node 1 failed: EOF
W180530 08:03:14.174801 136421 gossip/infostore.go:303 [n5] node unavailable; try another peer
W180530 08:03:14.183405 137487 storage/raft_transport.go:465 [n5] raft transport stream to node 3 failed: EOF
W180530 08:03:14.193702 137389 storage/raft_transport.go:465 [n3] raft transport stream to node 2 failed: EOF
W180530 08:03:14.193813 137219 storage/raft_transport.go:465 [n5] raft transport stream to node 4 failed: EOF
W180530 08:03:14.207193 135833 storage/raft_transport.go:465 [n1] raft transport stream to node 2 failed: rpc error: code = Canceled desc = grpc: the client connection is closing
E180530 08:03:14.211888 138810 storage/queue.go:778 [replicate,n1,s1,r16/1:/Table/{19-20}] change replicas of r16 failed: log-range-event: failed to send RPC: sending to all 4 replicas failed; last error: {<nil> context canceled}
W180530 08:03:14.195876 137801 storage/raft_transport.go:465 [n3] raft transport stream to node 4 failed: EOF
W180530 08:03:14.230633 137309 storage/raft_transport.go:465 [n2] raft transport stream to node 4 failed: rpc error: code = Canceled desc = context canceled
I180530 08:03:14.230886 138949 util/stop/stopper.go:471 quiescing; tasks left:
1 [async] transport racer
W180530 08:03:14.194806 136218 gossip/gossip.go:1293 [n5] no incoming or outgoing connections
W180530 08:03:14.216441 136269 storage/raft_transport.go:465 [n3] raft transport stream to node 1 failed: rpc error: code = Canceled desc = context canceled
W180530 08:03:14.217362 137558 storage/raft_transport.go:465 [n3] raft transport stream to node 5 failed: rpc error: code = Canceled desc = context canceled
W180530 08:03:14.221091 135836 storage/raft_transport.go:465 [n2] raft transport stream to node 1 failed: rpc error: code = Canceled desc = context canceled
W180530 08:03:14.226614 137105 storage/raft_transport.go:465 [n5] raft transport stream to node 2 failed: EOF
W180530 08:03:14.227960 136599 storage/raft_transport.go:465 [n5] raft transport stream to node 1 failed: EOF
W180530 08:03:14.228232 136556 storage/raft_transport.go:465 [n1] raft transport stream to node 4 failed: rpc error: code = Canceled desc = grpc: the client connection is closing
W180530 08:03:14.230332 137299 storage/raft_transport.go:465 [n2] raft transport stream to node 5 failed: rpc error: code = Canceled desc = context canceled
W180530 08:03:14.203037 136433 storage/raft_transport.go:465 [n1] raft transport stream to node 3 failed: rpc error: code = Canceled desc = grpc: the client connection is closing
W180530 08:03:14.231432 137310 storage/raft_transport.go:465 [n2] raft transport stream to node 3 failed: rpc error: code = Canceled desc = grpc: the client connection is closing
W180530 08:03:14.238059 137241 storage/raft_transport.go:465 [n4] raft transport stream to node 5 failed: rpc error: code = Canceled desc = grpc: the client connection is closing
W180530 08:03:14.240091 137459 storage/raft_transport.go:465 [n4] raft transport stream to node 2 failed: rpc error: code = Canceled desc = context canceled
W180530 08:03:14.240671 137941 storage/raft_transport.go:465 [n4] raft transport stream to node 3 failed: rpc error: code = Canceled desc = grpc: the client connection is closing
I180530 08:03:16.398033 134873 kv/transport_race.go:67 transport race promotion: ran 101 iterations on up to 2394 requests
testing.go:730: race detected during execution of test
```
|
1.0
|
github.com/cockroachdb/cockroach/pkg/storage: TestReplicateQueueRebalance failed under stress - SHA: https://github.com/cockroachdb/cockroach/commits/89690e007f3f4bab973a107754f471d140d5373f
Parameters:
```
TAGS=
GOFLAGS=-race
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=686673&tab=buildLog
```
I180530 08:03:03.273889 137577 storage/replica_raftstorage.go:730 [n4,s4,r10/?:{-}] applying preemptive snapshot at index 147 (id=25e68233, encoded size=79301, 1 rocksdb batches, 50 log entries)
I180530 08:03:03.277311 134834 storage/replica_proposal.go:202 [n1,s1,r24/1:/{Table/53-Max}] new range lease repl=(n1,s1):1 seq=3 start=1527667357.762478092,0 epo=1 pro=1527667366.622400772,0 following repl=(n1,s1):1 seq=3 start=1527667357.762478092,0 epo=1 pro=1527667366.622400772,0
I180530 08:03:03.365237 137497 storage/replica.go:3320 [n5,s5,r13/3:/Table/1{6-7}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n4,s4):4 (n2,s2):2 (n5,s5):3] next=5
I180530 08:03:03.437939 137577 storage/replica_raftstorage.go:736 [n4,s4,r10/?:/Table/1{3-4}] applied preemptive snapshot in 164ms [clear=0ms batch=0ms entries=161ms commit=2ms]
W180530 08:03:03.454242 137521 storage/replica.go:3131 [n5,s5,r13/3:/Table/1{6-7}] context canceled before proposing: 1 HeartbeatTxn
I180530 08:03:03.471034 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.471502 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.481953 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.484759 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.486223 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.487058 137515 storage/replica_command.go:1777 [replicate,n1,s1,r10/1:/Table/1{3-4}] change replicas (ADD_REPLICA (n4,s4):4): read existing descriptor r10:/Table/1{3-4} [(n1,s1):1, (n3,s3):2, (n5,s5):3, next=4]
I180530 08:03:03.490659 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.491741 136462 storage/store.go:3508 [n1,s1,r13/1:/Table/1{6-7}] added to replica GC queue (peer suggestion)
I180530 08:03:03.503804 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.512252 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.512759 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.513974 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.514300 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.520958 137558 storage/raft_transport.go:459 [n3] raft transport stream to node 5 established
I180530 08:03:03.540034 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.553581 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.553891 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.564197 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.569293 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.574537 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.578505 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.588426 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.590611 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.594118 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.601265 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.606347 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.619177 137457 storage/replica_raftstorage.go:520 [replicate,n5,s5,r2/3:/System/{-NodeLive…}] generated preemptive snapshot 5c19a28a at index 35
I180530 08:03:03.629308 134647 storage/replicate_queue_test.go:94 not balanced: [25 16 14 10 14]
I180530 08:03:03.631829 137563 storage/store.go:2496 [replicaGC,n1,s1,r13/1:/Table/1{6-7}] removing replica
I180530 08:03:03.635390 137563 storage/replica.go:829 [replicaGC,n1,s1,r13/1:/Table/1{6-7}] removed 8 (0+8) keys in 2ms [clear=0ms commit=2ms]
I180530 08:03:03.646053 134647 storage/replicate_queue_test.go:94 not balanced: [24 16 14 10 14]
I180530 08:03:03.656411 137457 storage/store_snapshot.go:605 [replicate,n5,s5,r2/3:/System/{-NodeLive…}] streamed snapshot to (n4,s4):?: kv pairs: 11, log entries: 16, rate-limit: 2.0 MiB/sec, 33ms
I180530 08:03:03.664356 137606 storage/replica_raftstorage.go:730 [n4,s4,r2/?:{-}] applying preemptive snapshot at index 35 (id=5c19a28a, encoded size=56945, 1 rocksdb batches, 16 log entries)
I180530 08:03:03.666032 134647 storage/replicate_queue_test.go:94 not balanced: [24 16 14 11 14]
I180530 08:03:03.731311 134647 storage/replicate_queue_test.go:94 not balanced: [24 16 14 11 14]
I180530 08:03:03.799007 134647 storage/replicate_queue_test.go:94 not balanced: [24 16 14 11 14]
I180530 08:03:03.861955 137606 storage/replica_raftstorage.go:736 [n4,s4,r2/?:/System/{-NodeLive…}] applied preemptive snapshot in 197ms [clear=0ms batch=0ms entries=168ms commit=3ms]
I180530 08:03:03.877066 137515 storage/replica.go:3320 [n1,s1,r10/1:/Table/1{3-4}] proposing ADD_REPLICA((n4,s4):4): updated=[(n1,s1):1 (n3,s3):2 (n5,s5):3 (n4,s4):4] next=5
I180530 08:03:03.917594 137457 storage/replica_command.go:1777 [replicate,n5,s5,r2/3:/System/{-NodeLive…}] change replicas (ADD_REPLICA (n4,s4):4): read existing descriptor r2:/System/{-NodeLiveness} [(n1,s1):1, (n3,s3):2, (n5,s5):3, next=4]
I180530 08:03:03.933851 134647 storage/replicate_queue_test.go:94 not balanced: [24 16 14 11 14]
I180530 08:03:03.988124 137661 storage/replica_raftstorage.go:520 [replicate,n1,s1,r1/1:/{Min-System/}] generated preemptive snapshot 7d835994 at index 177
I180530 08:03:04.001521 136424 storage/replica_proposal.go:202 [n5,s5,r10/3:/Table/1{3-4}] new range lease repl=(n5,s5):3 seq=4 start=1527667383.923595251,0 epo=1 pro=1527667383.923610141,0 following repl=(n1,s1):1 seq=3 start=1527667357.762478092,0 epo=1 pro=1527667366.995510910,0
I180530 08:03:04.100725 137661 storage/store_snapshot.go:605 [replicate,n1,s1,r1/1:/{Min-System/}] streamed snapshot to (n5,s5):?: kv pairs: 114, log entries: 58, rate-limit: 2.0 MiB/sec, 111ms
I180530 08:03:04.108915 137623 storage/replica_raftstorage.go:730 [n5,s5,r1/?:{-}] applying preemptive snapshot at index 177 (id=7d835994, encoded size=16094, 1 rocksdb batches, 58 log entries)
I180530 08:03:04.193570 137623 storage/replica_raftstorage.go:736 [n5,s5,r1/?:/{Min-System/}] applied preemptive snapshot in 84ms [clear=0ms batch=0ms entries=83ms commit=1ms]
I180530 08:03:04.205800 134647 storage/replicate_queue_test.go:94 not balanced: [24 16 14 11 15]
I180530 08:03:04.468945 137661 storage/replica_command.go:1777 [replicate,n1,s1,r1/1:/{Min-System/}] change replicas (ADD_REPLICA (n5,s5):4): read existing descriptor r1:/{Min-System/} [(n1,s1):1, (n4,s4):2, (n3,s3):3, next=4]
I180530 08:03:04.843957 137740 storage/replica.go:3320 [n1,s1,r1/1:/{Min-System/}] proposing ADD_REPLICA((n5,s5):4): updated=[(n1,s1):1 (n4,s4):2 (n3,s3):3 (n5,s5):4] next=5
I180530 08:03:04.846363 134647 storage/replicate_queue_test.go:94 not balanced: [24 16 14 11 15]
I180530 08:03:04.858516 137692 storage/replica.go:3320 [n5,s5,r2/3:/System/{-NodeLive…}] proposing ADD_REPLICA((n4,s4):4): updated=[(n1,s1):1 (n3,s3):2 (n5,s5):3 (n4,s4):4] next=5
I180530 08:03:04.939134 137634 storage/replica_command.go:1777 [replicate,n5,s5,r15/3:/Table/1{8-9}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r15:/Table/1{8-9} [(n1,s1):1, (n4,s4):2, (n5,s5):3, (n3,s3):4, next=5]
I180530 08:03:05.158183 137785 storage/replica_raftstorage.go:520 [replicate,n1,s1,r8/1:/Table/1{1-2}] generated preemptive snapshot eb16c8b1 at index 25
E180530 08:03:05.165417 137859 storage/queue.go:778 [replicate,n3,s3,r1/3:/{Min-System/}] no removable replicas from range that needs a removal: [raft progress unknown]
I180530 08:03:05.201642 137785 storage/store_snapshot.go:605 [replicate,n1,s1,r8/1:/Table/1{1-2}] streamed snapshot to (n3,s3):?: kv pairs: 11, log entries: 15, rate-limit: 2.0 MiB/sec, 35ms
I180530 08:03:05.205018 137862 storage/replica_raftstorage.go:730 [n3,s3,r8/?:{-}] applying preemptive snapshot at index 25 (id=eb16c8b1, encoded size=5859, 1 rocksdb batches, 15 log entries)
I180530 08:03:05.218754 137862 storage/replica_raftstorage.go:736 [n3,s3,r8/?:/Table/1{1-2}] applied preemptive snapshot in 13ms [clear=0ms batch=0ms entries=12ms commit=0ms]
I180530 08:03:05.234417 137785 storage/replica_command.go:1777 [replicate,n1,s1,r8/1:/Table/1{1-2}] change replicas (ADD_REPLICA (n3,s3):4): read existing descriptor r8:/Table/1{1-2} [(n1,s1):1, (n2,s2):2, (n4,s4):3, next=4]
I180530 08:03:05.605347 137785 storage/replica.go:3320 [n1,s1,r8/1:/Table/1{1-2}] proposing ADD_REPLICA((n3,s3):4): updated=[(n1,s1):1 (n2,s2):2 (n4,s4):3 (n3,s3):4] next=5
I180530 08:03:05.706603 137857 storage/replica_command.go:1777 [replicate,n2,s2,r14/3:/Table/1{7-8}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r14:/Table/1{7-8} [(n1,s1):1, (n4,s4):2, (n2,s2):3, (n3,s3):4, next=5]
I180530 08:03:05.727595 137801 storage/raft_transport.go:459 [n3] raft transport stream to node 4 established
I180530 08:03:05.746581 135477 storage/replica_proposal.go:202 [n2,s2,r8/2:/Table/1{1-2}] new range lease repl=(n2,s2):2 seq=4 start=1527667385.689873739,0 epo=1 pro=1527667385.689888920,0 following repl=(n1,s1):1 seq=3 start=1527667357.762478092,0 epo=1 pro=1527667367.179186503,0
I180530 08:03:05.809625 137868 storage/replica_raftstorage.go:520 [replicate,n1,s1,r20/1:/Table/{23-50}] generated preemptive snapshot 05341a94 at index 24
I180530 08:03:05.904355 137868 storage/store_snapshot.go:605 [replicate,n1,s1,r20/1:/Table/{23-50}] streamed snapshot to (n4,s4):?: kv pairs: 15, log entries: 14, rate-limit: 2.0 MiB/sec, 33ms
I180530 08:03:05.907673 137940 storage/replica_raftstorage.go:730 [n4,s4,r20/?:{-}] applying preemptive snapshot at index 24 (id=05341a94, encoded size=4934, 1 rocksdb batches, 14 log entries)
I180530 08:03:05.939097 134647 storage/replicate_queue_test.go:94 not balanced: [24 16 15 12 15]
I180530 08:03:05.966461 137940 storage/replica_raftstorage.go:736 [n4,s4,r20/?:/Table/{23-50}] applied preemptive snapshot in 58ms [clear=0ms batch=0ms entries=57ms commit=0ms]
I180530 08:03:06.068020 137868 storage/replica_command.go:1777 [replicate,n1,s1,r20/1:/Table/{23-50}] change replicas (ADD_REPLICA (n4,s4):4): read existing descriptor r20:/Table/{23-50} [(n1,s1):1, (n3,s3):2, (n5,s5):3, next=4]
I180530 08:03:06.315430 137950 storage/replica.go:3320 [n5,s5,r15/3:/Table/1{8-9}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n3,s3):4 (n4,s4):2 (n5,s5):3] next=5
I180530 08:03:06.411922 137951 storage/replica_command.go:1777 [replicate,n5,s5,r2/3:/System/{-NodeLive…}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r2:/System/{-NodeLiveness} [(n1,s1):1, (n3,s3):2, (n5,s5):3, (n4,s4):4, next=5]
I180530 08:03:06.468100 136462 storage/store.go:3508 [n1,s1,r15/1:/Table/1{8-9}] added to replica GC queue (peer suggestion)
I180530 08:03:06.865267 137954 storage/store.go:2496 [replicaGC,n1,s1,r15/1:/Table/1{8-9}] removing replica
I180530 08:03:06.872442 137941 storage/raft_transport.go:459 [n4] raft transport stream to node 3 established
I180530 08:03:06.899959 137954 storage/replica.go:829 [replicaGC,n1,s1,r15/1:/Table/1{8-9}] removed 8 (0+8) keys in 33ms [clear=0ms commit=33ms]
I180530 08:03:06.941305 138008 storage/replica.go:3320 [n1,s1,r20/1:/Table/{23-50}] proposing ADD_REPLICA((n4,s4):4): updated=[(n1,s1):1 (n3,s3):2 (n5,s5):3 (n4,s4):4] next=5
W180530 08:03:06.949385 138024 storage/intent_resolver.go:320 [n3,s3] failed to push during intent resolution: failed to push "change-replica" id=675eb1e3 key=/Local/Range/Table/23/RangeDescriptor rw=true pri=0.01733080 iso=SERIALIZABLE stat=PENDING epo=0 ts=1527667386.030283448,0 orig=1527667386.030283448,0 max=1527667386.030283448,0 wto=false rop=false seq=2
I180530 08:03:06.973530 134647 storage/replicate_queue_test.go:94 not balanced: [23 16 15 12 15]
I180530 08:03:07.200909 138073 storage/replica.go:3320 [n2,s2,r14/3:/Table/1{7-8}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n3,s3):4 (n4,s4):2 (n2,s2):3] next=5
I180530 08:03:07.329645 135883 storage/store.go:3508 [n1,s1,r14/1:/Table/1{7-8}] added to replica GC queue (peer suggestion)
I180530 08:03:07.338457 135737 storage/replica_proposal.go:202 [n3,s3,r20/2:/Table/{23-50}] new range lease repl=(n3,s3):2 seq=4 start=1527667387.099999044,0 epo=1 pro=1527667387.100014019,0 following repl=(n1,s1):1 seq=3 start=1527667357.762478092,0 epo=1 pro=1527667367.166295614,0
I180530 08:03:07.370467 138057 storage/replica_raftstorage.go:520 [replicate,n1,s1,r5/1:/System/ts{d-e}] generated preemptive snapshot 22be3a91 at index 44
I180530 08:03:07.451894 135359 server/status/runtime.go:219 [n2] runtime stats: 1.4 GiB RSS, 1158 goroutines, 77 MiB/29 MiB/138 MiB GO alloc/idle/total, 158 MiB/201 MiB CGO alloc/total, 1180.92cgo/sec, 1.39/0.12 %(u/s)time, 0.02 %gc (6x)
I180530 08:03:07.502056 138109 storage/replica_command.go:1777 [replicate,n3,s3,r1/3:/{Min-System/}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r1:/{Min-System/} [(n1,s1):1, (n4,s4):2, (n3,s3):3, (n5,s5):4, next=5]
I180530 08:03:07.652014 138076 storage/store.go:2496 [replicaGC,n1,s1,r14/1:/Table/1{7-8}] removing replica
I180530 08:03:07.684198 138091 storage/replica_command.go:1777 [replicate,n2,s2,r8/2:/Table/1{1-2}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r8:/Table/1{1-2} [(n1,s1):1, (n2,s2):2, (n4,s4):3, (n3,s3):4, next=5]
I180530 08:03:07.879883 134886 server/status/runtime.go:219 [n1] runtime stats: 1.4 GiB RSS, 1152 goroutines, 88 MiB/19 MiB/138 MiB GO alloc/idle/total, 157 MiB/201 MiB CGO alloc/total, 1372.68cgo/sec, 1.39/0.12 %(u/s)time, 0.02 %gc (6x)
I180530 08:03:07.919865 138057 storage/store_snapshot.go:605 [replicate,n1,s1,r5/1:/System/ts{d-e}] streamed snapshot to (n4,s4):?: kv pairs: 2167, log entries: 3, rate-limit: 2.0 MiB/sec, 548ms
I180530 08:03:07.924031 138076 storage/replica.go:829 [replicaGC,n1,s1,r14/1:/Table/1{7-8}] removed 8 (0+8) keys in 205ms [clear=0ms commit=205ms]
I180530 08:03:07.934630 138108 storage/replica_raftstorage.go:730 [n4,s4,r5/?:{-}] applying preemptive snapshot at index 44 (id=22be3a91, encoded size=351993, 1 rocksdb batches, 3 log entries)
I180530 08:03:07.991175 134647 storage/replicate_queue_test.go:94 not balanced: [22 16 15 13 15]
I180530 08:03:08.029788 138143 storage/replica.go:3320 [n5,s5,r2/3:/System/{-NodeLive…}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n4,s4):4 (n3,s3):2 (n5,s5):3] next=5
I180530 08:03:08.205764 138108 storage/replica_raftstorage.go:736 [n4,s4,r5/?:/System/ts{d-e}] applied preemptive snapshot in 271ms [clear=0ms batch=2ms entries=266ms commit=2ms]
I180530 08:03:08.265580 136462 storage/store.go:3508 [n1,s1,r2/1:/System/{-NodeLive…}] added to replica GC queue (peer suggestion)
I180530 08:03:08.282224 138057 storage/replica_command.go:1777 [replicate,n1,s1,r5/1:/System/ts{d-e}] change replicas (ADD_REPLICA (n4,s4):4): read existing descriptor r5:/System/ts{d-e} [(n1,s1):1, (n3,s3):2, (n2,s2):3, next=4]
I180530 08:03:08.288916 138228 storage/replica_command.go:1777 [replicate,n5,s5,r10/3:/Table/1{3-4}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r10:/Table/1{3-4} [(n1,s1):1, (n3,s3):2, (n5,s5):3, (n4,s4):4, next=5]
I180530 08:03:08.601014 135675 server/status/runtime.go:219 [n3] runtime stats: 1.5 GiB RSS, 1159 goroutines, 110 MiB/6.2 MiB/144 MiB GO alloc/idle/total, 158 MiB/203 MiB CGO alloc/total, 1373.17cgo/sec, 1.40/0.11 %(u/s)time, 0.02 %gc (6x)
I180530 08:03:08.812404 138213 storage/store.go:2496 [replicaGC,n1,s1,r2/1:/System/{-NodeLive…}] removing replica
I180530 08:03:08.814577 138213 storage/replica.go:829 [replicaGC,n1,s1,r2/1:/System/{-NodeLive…}] removed 8 (0+8) keys in 1ms [clear=0ms commit=1ms]
I180530 08:03:08.849996 138275 storage/replica.go:3320 [n3,s3,r1/3:/{Min-System/}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n5,s5):4 (n4,s4):2 (n3,s3):3] next=5
I180530 08:03:09.030224 134647 storage/replicate_queue_test.go:94 not balanced: [21 16 15 13 15]
I180530 08:03:09.084849 138294 storage/store.go:2496 [replicaGC,n1,s1,r1/1:/{Min-System/}] removing replica
I180530 08:03:09.087464 138294 storage/replica.go:829 [replicaGC,n1,s1,r1/1:/{Min-System/}] removed 35 (26+9) keys in 2ms [clear=0ms commit=1ms]
I180530 08:03:09.260541 138210 storage/replica_command.go:1777 [replicate,n3,s3,r20/2:/Table/{23-50}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r20:/Table/{23-50} [(n1,s1):1, (n3,s3):2, (n5,s5):3, (n4,s4):4, next=5]
I180530 08:03:09.330240 138057 storage/replica.go:3320 [n1,s1,r5/1:/System/ts{d-e}] proposing ADD_REPLICA((n4,s4):4): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3 (n4,s4):4] next=5
W180530 08:03:09.432827 135677 server/node.go:802 [n3,summaries] health alerts detected: {Alerts:[{StoreID:3 Category:METRICS Description:queue.replicate.process.failure Value:1}]}
I180530 08:03:09.516300 138240 storage/replica.go:3320 [n2,s2,r8/2:/Table/1{1-2}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n3,s3):4 (n2,s2):2 (n4,s4):3] next=5
I180530 08:03:09.688443 135639 storage/replica_proposal.go:202 [n3,s3,r5/2:/System/ts{d-e}] new range lease repl=(n3,s3):2 seq=4 start=1527667389.417811299,0 epo=1 pro=1527667389.417824987,0 following repl=(n1,s1):1 seq=3 start=1527667357.762478092,0 epo=1 pro=1527667366.661823335,0
I180530 08:03:09.749964 138364 storage/replica_raftstorage.go:520 [replicate,n2,s2,r8/2:/Table/1{1-2}] generated preemptive snapshot 7df7c51a at index 33
I180530 08:03:09.891483 135993 server/status/runtime.go:219 [n4] runtime stats: 1.5 GiB RSS, 1157 goroutines, 102 MiB/12 MiB/144 MiB GO alloc/idle/total, 158 MiB/204 MiB CGO alloc/total, 1346.95cgo/sec, 1.39/0.12 %(u/s)time, 0.01 %gc (5x)
I180530 08:03:09.901951 138364 storage/store_snapshot.go:605 [replicate,n2,s2,r8/2:/Table/1{1-2}] streamed snapshot to (n5,s5):?: kv pairs: 14, log entries: 23, rate-limit: 2.0 MiB/sec, 132ms
I180530 08:03:09.911719 135883 storage/store.go:3508 [n1,s1,r8/1:/Table/1{1-2}] added to replica GC queue (peer suggestion)
I180530 08:03:09.939996 138312 storage/store.go:2496 [replicaGC,n1,s1,r8/1:/Table/1{1-2}] removing replica
I180530 08:03:09.971866 138391 storage/replica_raftstorage.go:520 [replicate,n1,s1,r21/1:/Table/5{0-1}] generated preemptive snapshot 9d827d66 at index 24
I180530 08:03:09.990858 138367 storage/replica_raftstorage.go:730 [n5,s5,r8/?:{-}] applying preemptive snapshot at index 33 (id=7df7c51a, encoded size=8789, 1 rocksdb batches, 23 log entries)
I180530 08:03:10.006369 138367 storage/replica_raftstorage.go:736 [n5,s5,r8/?:/Table/1{1-2}] applied preemptive snapshot in 15ms [clear=0ms batch=0ms entries=13ms commit=1ms]
I180530 08:03:10.029771 138312 storage/replica.go:829 [replicaGC,n1,s1,r8/1:/Table/1{1-2}] removed 8 (0+8) keys in 81ms [clear=0ms commit=80ms]
I180530 08:03:10.046214 138391 storage/store_snapshot.go:605 [replicate,n1,s1,r21/1:/Table/5{0-1}] streamed snapshot to (n5,s5):?: kv pairs: 11, log entries: 14, rate-limit: 2.0 MiB/sec, 70ms
I180530 08:03:10.051795 134647 storage/replicate_queue_test.go:94 not balanced: [19 16 15 13 16]
I180530 08:03:10.079905 138370 storage/replica_raftstorage.go:730 [n5,s5,r21/?:{-}] applying preemptive snapshot at index 24 (id=9d827d66, encoded size=4614, 1 rocksdb batches, 14 log entries)
I180530 08:03:10.128625 138364 storage/replica_command.go:1777 [replicate,n2,s2,r8/2:/Table/1{1-2}] change replicas (ADD_REPLICA (n5,s5):5): read existing descriptor r8:/Table/1{1-2} [(n3,s3):4, (n2,s2):2, (n4,s4):3, next=5]
==================
WARNING: DATA RACE
Write at 0x00c42306c8c9 by goroutine 137:
github.com/cockroachdb/cockroach/pkg/roachpb.(*Transaction).Update()
/go/src/github.com/cockroachdb/cockroach/pkg/roachpb/data.go:932 +0xd6f
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).heartbeat()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1126 +0x973
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).heartbeatLoop()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:967 +0x3f8
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).updateState.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1333 +0x55
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:264 +0xf3
Previous read at 0x00c42306c8c9 by goroutine 661:
[failed to restore the stack]
Goroutine 137 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:259 +0x14c
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).updateState()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1331 +0x673
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:494 +0x37f
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).sendUsingSender()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:553 +0x106
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:973 +0x66e
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Send-fm()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:500 +0x96
github.com/cockroachdb/cockroach/pkg/internal/client.sendAndFill()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:468 +0x13d
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Run()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:500 +0xfd
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).changeReplicas.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1790 +0x535
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:516 +0x5e
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Exec()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:748 +0x128
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:515 +0x192
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).changeReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1765 +0xa6f
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).ChangeReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1670 +0xc9
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).removeReplica()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:591 +0xd9
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).processOneChange()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:386 +0x14d1
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).process()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:218 +0x28b
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processReplica()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:734 +0x313
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1.2()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:617 +0xdd
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:264 +0xf3
Goroutine 661 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:259 +0x14c
github.com/cockroachdb/cockroach/pkg/kv.GRPCTransportFactory()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/transport_race.go:61 +0x2f7
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendToReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1263 +0x14e
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendRPC()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:393 +0x2be
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendSingleRange()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:460 +0x265
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendPartialBatch()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1053 +0x3dd
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:738 +0x1a05
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:655 +0x565
github.com/cockroachdb/cockroach/pkg/internal/client.lookupRangeFwdScan()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/range_lookup.go:339 +0x4e7
github.com/cockroachdb/cockroach/pkg/internal/client.RangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/range_lookup.go:202 +0xb0e
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).RangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:299 +0x104
github.com/cockroachdb/cockroach/pkg/kv.(*RangeDescriptorCache).performRangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/range_cache.go:434 +0x42d
github.com/cockroachdb/cockroach/pkg/kv.(*RangeDescriptorCache).lookupRangeDescriptorInternal.func3()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/range_cache.go:310 +0x167
github.com/cockroachdb/cockroach/pkg/util/syncutil/singleflight.(*Group).doCall()
/go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/singleflight/singleflight.go:118 +0x4c
==================
==================
WARNING: DATA RACE
Write at 0x00c42306c8f8 by goroutine 137:
github.com/cockroachdb/cockroach/pkg/roachpb.(*Transaction).Update()
/go/src/github.com/cockroachdb/cockroach/pkg/roachpb/data.go:934 +0xe05
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).heartbeat()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1126 +0x973
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).heartbeatLoop()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:967 +0x3f8
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).updateState.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1333 +0x55
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:264 +0xf3
Previous read at 0x00c42306c8f8 by goroutine 661:
[failed to restore the stack]
Goroutine 137 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:259 +0x14c
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).updateState()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1331 +0x673
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:494 +0x37f
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).sendUsingSender()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:553 +0x106
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:973 +0x66e
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Send-fm()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:500 +0x96
github.com/cockroachdb/cockroach/pkg/internal/client.sendAndFill()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:468 +0x13d
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Run()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:500 +0xfd
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).changeReplicas.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1790 +0x535
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:516 +0x5e
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Exec()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:748 +0x128
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:515 +0x192
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).changeReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1765 +0xa6f
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).ChangeReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1670 +0xc9
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).removeReplica()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:591 +0xd9
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).processOneChange()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:386 +0x14d1
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).process()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:218 +0x28b
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processReplica()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:734 +0x313
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1.2()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:617 +0xdd
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:264 +0xf3
Goroutine 661 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:259 +0x14c
github.com/cockroachdb/cockroach/pkg/kv.GRPCTransportFactory()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/transport_race.go:61 +0x2f7
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendToReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1263 +0x14e
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendRPC()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:393 +0x2be
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendSingleRange()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:460 +0x265
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendPartialBatch()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1053 +0x3dd
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:738 +0x1a05
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:655 +0x565
github.com/cockroachdb/cockroach/pkg/internal/client.lookupRangeFwdScan()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/range_lookup.go:339 +0x4e7
github.com/cockroachdb/cockroach/pkg/internal/client.RangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/range_lookup.go:202 +0xb0e
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).RangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:299 +0x104
github.com/cockroachdb/cockroach/pkg/kv.(*RangeDescriptorCache).performRangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/range_cache.go:434 +0x42d
github.com/cockroachdb/cockroach/pkg/kv.(*RangeDescriptorCache).lookupRangeDescriptorInternal.func3()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/range_cache.go:310 +0x167
github.com/cockroachdb/cockroach/pkg/util/syncutil/singleflight.(*Group).doCall()
/go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/singleflight/singleflight.go:118 +0x4c
==================
==================
WARNING: DATA RACE
Write at 0x00c42306c870 by goroutine 137:
github.com/cockroachdb/cockroach/pkg/roachpb.(*Transaction).Update()
/go/src/github.com/cockroachdb/cockroach/pkg/util/hlc/timestamp.go:106 +0xbfc
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).heartbeat()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1126 +0x973
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).heartbeatLoop()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:967 +0x3f8
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).updateState.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1333 +0x55
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:264 +0xf3
Previous read at 0x00c42306c870 by goroutine 661:
[failed to restore the stack]
Goroutine 137 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:259 +0x14c
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).updateState()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1331 +0x673
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:494 +0x37f
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).sendUsingSender()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:553 +0x106
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:973 +0x66e
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Send-fm()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:500 +0x96
github.com/cockroachdb/cockroach/pkg/internal/client.sendAndFill()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:468 +0x13d
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Run()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:500 +0xfd
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).changeReplicas.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1790 +0x535
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:516 +0x5e
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Exec()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:748 +0x128
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:515 +0x192
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).changeReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1765 +0xa6f
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).ChangeReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1670 +0xc9
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).removeReplica()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:591 +0xd9
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).processOneChange()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:386 +0x14d1
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).process()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:218 +0x28b
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processReplica()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:734 +0x313
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1.2()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:617 +0xdd
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:264 +0xf3
Goroutine 661 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:259 +0x14c
github.com/cockroachdb/cockroach/pkg/kv.GRPCTransportFactory()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/transport_race.go:61 +0x2f7
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendToReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1263 +0x14e
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendRPC()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:393 +0x2be
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendSingleRange()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:460 +0x265
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendPartialBatch()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1053 +0x3dd
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:738 +0x1a05
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:655 +0x565
github.com/cockroachdb/cockroach/pkg/internal/client.lookupRangeFwdScan()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/range_lookup.go:339 +0x4e7
github.com/cockroachdb/cockroach/pkg/internal/client.RangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/range_lookup.go:202 +0xb0e
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).RangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:299 +0x104
github.com/cockroachdb/cockroach/pkg/kv.(*RangeDescriptorCache).performRangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/range_cache.go:434 +0x42d
github.com/cockroachdb/cockroach/pkg/kv.(*RangeDescriptorCache).lookupRangeDescriptorInternal.func3()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/range_cache.go:310 +0x167
github.com/cockroachdb/cockroach/pkg/util/syncutil/singleflight.(*Group).doCall()
/go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/singleflight/singleflight.go:118 +0x4c
==================
==================
WARNING: DATA RACE
Write at 0x00c42306c8b0 by goroutine 137:
github.com/cockroachdb/cockroach/pkg/roachpb.(*Transaction).UpdateObservedTimestamp()
/go/src/github.com/cockroachdb/cockroach/pkg/roachpb/data.go:1027 +0x100
github.com/cockroachdb/cockroach/pkg/roachpb.(*Transaction).Update()
/go/src/github.com/cockroachdb/cockroach/pkg/roachpb/data.go:949 +0x695
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).heartbeat()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1126 +0x973
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).heartbeatLoop()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:967 +0x3f8
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).updateState.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1333 +0x55
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:264 +0xf3
Previous read at 0x00c42306c8b0 by goroutine 661:
[failed to restore the stack]
Goroutine 137 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:259 +0x14c
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).updateState()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1331 +0x673
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:494 +0x37f
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).sendUsingSender()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:553 +0x106
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:973 +0x66e
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Send-fm()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:500 +0x96
github.com/cockroachdb/cockroach/pkg/internal/client.sendAndFill()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:468 +0x13d
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Run()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:500 +0xfd
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).changeReplicas.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1790 +0x535
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:516 +0x5e
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Exec()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:748 +0x128
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:515 +0x192
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).changeReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1765 +0xa6f
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).ChangeReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1670 +0xc9
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).removeReplica()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:591 +0xd9
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).processOneChange()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:386 +0x14d1
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).process()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:218 +0x28b
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processReplica()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:734 +0x313
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1.2()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:617 +0xdd
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:264 +0xf3
Goroutine 661 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:259 +0x14c
github.com/cockroachdb/cockroach/pkg/kv.GRPCTransportFactory()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/transport_race.go:61 +0x2f7
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendToReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1263 +0x14e
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendRPC()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:393 +0x2be
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendSingleRange()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:460 +0x265
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendPartialBatch()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1053 +0x3dd
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:738 +0x1a05
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:655 +0x565
github.com/cockroachdb/cockroach/pkg/internal/client.lookupRangeFwdScan()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/range_lookup.go:339 +0x4e7
github.com/cockroachdb/cockroach/pkg/internal/client.RangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/range_lookup.go:202 +0xb0e
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).RangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:299 +0x104
github.com/cockroachdb/cockroach/pkg/kv.(*RangeDescriptorCache).performRangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/range_cache.go:434 +0x42d
github.com/cockroachdb/cockroach/pkg/kv.(*RangeDescriptorCache).lookupRangeDescriptorInternal.func3()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/range_cache.go:310 +0x167
github.com/cockroachdb/cockroach/pkg/util/syncutil/singleflight.(*Group).doCall()
/go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/singleflight/singleflight.go:118 +0x4c
==================
==================
WARNING: DATA RACE
Write at 0x00c42306c8c8 by goroutine 137:
github.com/cockroachdb/cockroach/pkg/roachpb.(*Transaction).Update()
/go/src/github.com/cockroachdb/cockroach/pkg/roachpb/data.go:955 +0x724
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).heartbeat()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1126 +0x973
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).heartbeatLoop()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:967 +0x3f8
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).updateState.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1333 +0x55
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:264 +0xf3
Previous read at 0x00c42306c8c8 by goroutine 661:
[failed to restore the stack]
Goroutine 137 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:259 +0x14c
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).updateState()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:1331 +0x673
github.com/cockroachdb/cockroach/pkg/kv.(*TxnCoordSender).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/txn_coord_sender.go:494 +0x37f
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).sendUsingSender()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:553 +0x106
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:973 +0x66e
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Send-fm()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:500 +0x96
github.com/cockroachdb/cockroach/pkg/internal/client.sendAndFill()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:468 +0x13d
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Run()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:500 +0xfd
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).changeReplicas.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1790 +0x535
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:516 +0x5e
github.com/cockroachdb/cockroach/pkg/internal/client.(*Txn).Exec()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/txn.go:748 +0x128
github.com/cockroachdb/cockroach/pkg/internal/client.(*DB).Txn()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/db.go:515 +0x192
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).changeReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1765 +0xa6f
github.com/cockroachdb/cockroach/pkg/storage.(*Replica).ChangeReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replica_command.go:1670 +0xc9
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).removeReplica()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:591 +0xd9
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).processOneChange()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:386 +0x14d1
github.com/cockroachdb/cockroach/pkg/storage.(*replicateQueue).process()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/replicate_queue.go:218 +0x28b
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processReplica()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:734 +0x313
github.com/cockroachdb/cockroach/pkg/storage.(*baseQueue).processLoop.func1.2()
/go/src/github.com/cockroachdb/cockroach/pkg/storage/queue.go:617 +0xdd
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:264 +0xf3
Goroutine 661 (running) created at:
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask()
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:259 +0x14c
github.com/cockroachdb/cockroach/pkg/kv.GRPCTransportFactory()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/transport_race.go:61 +0x2f7
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendToReplicas()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1263 +0x14e
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendRPC()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:393 +0x2be
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendSingleRange()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:460 +0x265
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).sendPartialBatch()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:1053 +0x3dd
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).divideAndSendBatchToRanges()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:738 +0x1a05
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).Send()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:655 +0x565
github.com/cockroachdb/cockroach/pkg/internal/client.lookupRangeFwdScan()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/range_lookup.go:339 +0x4e7
github.com/cockroachdb/cockroach/pkg/internal/client.RangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/internal/client/range_lookup.go:202 +0xb0e
github.com/cockroachdb/cockroach/pkg/kv.(*DistSender).RangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/dist_sender.go:299 +0x104
github.com/cockroachdb/cockroach/pkg/kv.(*RangeDescriptorCache).performRangeLookup()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/range_cache.go:434 +0x42d
github.com/cockroachdb/cockroach/pkg/kv.(*RangeDescriptorCache).lookupRangeDescriptorInternal.func3()
/go/src/github.com/cockroachdb/cockroach/pkg/kv/range_cache.go:310 +0x167
github.com/cockroachdb/cockroach/pkg/util/syncutil/singleflight.(*Group).doCall()
/go/src/github.com/cockroachdb/cockroach/pkg/util/syncutil/singleflight/singleflight.go:118 +0x4c
==================
I180530 08:03:10.582713 138487 storage/replica.go:3320 [n5,s5,r10/3:/Table/1{3-4}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n4,s4):4 (n3,s3):2 (n5,s5):3] next=5
I180530 08:03:10.606967 138370 storage/replica_raftstorage.go:736 [n5,s5,r21/?:/Table/5{0-1}] applied preemptive snapshot in 527ms [clear=0ms batch=0ms entries=431ms commit=20ms]
I180530 08:03:10.616792 136293 server/status/runtime.go:219 [n5] runtime stats: 1.5 GiB RSS, 1153 goroutines, 69 MiB/43 MiB/144 MiB GO alloc/idle/total, 157 MiB/204 MiB CGO alloc/total, 1301.03cgo/sec, 1.39/0.11 %(u/s)time, 0.02 %gc (6x)
I180530 08:03:10.732813 138491 storage/replica.go:3320 [n3,s3,r20/2:/Table/{23-50}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n4,s4):4 (n3,s3):2 (n5,s5):3] next=5
I180530 08:03:10.779589 138391 storage/replica_command.go:1777 [replicate,n1,s1,r21/1:/Table/5{0-1}] change replicas (ADD_REPLICA (n5,s5):4): read existing descriptor r21:/Table/5{0-1} [(n1,s1):1, (n2,s2):2, (n4,s4):3, next=4]
W180530 08:03:10.858350 138514 storage/replica.go:3131 [n5,s5,r10/3:/Table/1{3-4}] context canceled before proposing: 1 HeartbeatTxn
I180530 08:03:10.963074 138477 storage/replica_raftstorage.go:520 [replicate,n5,s5,r13/3:/Table/1{6-7}] generated preemptive snapshot d39327b1 at index 33
I180530 08:03:10.980790 136462 storage/store.go:3508 [n1,s1,r10/1:/Table/1{3-4}] added to replica GC queue (peer suggestion)
I180530 08:03:10.981529 136462 storage/store.go:3508 [n1,s1,r10/1:/Table/1{3-4}] added to replica GC queue (peer suggestion)
I180530 08:03:11.000981 136451 storage/store.go:3508 [n1,s1,r20/1:/Table/{23-50}] added to replica GC queue (peer suggestion)
I180530 08:03:11.037791 136451 storage/store.go:3508 [n1,s1,r20/1:/Table/{23-50}] added to replica GC queue (peer suggestion)
I180530 08:03:11.089333 134647 storage/replicate_queue_test.go:94 not balanced: [19 16 15 13 17]
I180530 08:03:11.089912 136451 storage/store.go:3508 [n1,s1,r20/1:/Table/{23-50}] added to replica GC queue (peer suggestion)
I180530 08:03:11.099719 138477 storage/store_snapshot.go:605 [replicate,n5,s5,r13/3:/Table/1{6-7}] streamed snapshot to (n3,s3):?: kv pairs: 13, log entries: 23, rate-limit: 2.0 MiB/sec, 123ms
I180530 08:03:11.108045 138434 storage/replica_raftstorage.go:730 [n3,s3,r13/?:{-}] applying preemptive snapshot at index 33 (id=d39327b1, encoded size=7268, 1 rocksdb batches, 23 log entries)
I180530 08:03:11.159912 138534 storage/store.go:2496 [replicaGC,n1,s1,r10/1:/Table/1{3-4}] removing replica
I180530 08:03:11.196769 138534 storage/replica.go:829 [replicaGC,n1,s1,r10/1:/Table/1{3-4}] removed 457 (449+8) keys in 36ms [clear=0ms commit=35ms]
I180530 08:03:11.213849 138567 storage/replica_command.go:1777 [replicate,n3,s3,r5/2:/System/ts{d-e}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r5:/System/ts{d-e} [(n1,s1):1, (n3,s3):2, (n2,s2):3, (n4,s4):4, next=5]
I180530 08:03:11.250905 138570 storage/store.go:2496 [replicaGC,n1,s1,r20/1:/Table/{23-50}] removing replica
I180530 08:03:11.259926 138434 storage/replica_raftstorage.go:736 [n3,s3,r13/?:/Table/1{6-7}] applied preemptive snapshot in 152ms [clear=0ms batch=0ms entries=139ms commit=11ms]
I180530 08:03:11.260340 138570 storage/replica.go:829 [replicaGC,n1,s1,r20/1:/Table/{23-50}] removed 12 (4+8) keys in 2ms [clear=0ms commit=2ms]
I180530 08:03:11.301258 138477 storage/replica_command.go:1777 [replicate,n5,s5,r13/3:/Table/1{6-7}] change replicas (ADD_REPLICA (n3,s3):5): read existing descriptor r13:/Table/1{6-7} [(n4,s4):4, (n2,s2):2, (n5,s5):3, next=5]
I180530 08:03:11.773875 138576 storage/replica.go:3320 [n2,s2,r8/2:/Table/1{1-2}] proposing ADD_REPLICA((n5,s5):5): updated=[(n3,s3):4 (n2,s2):2 (n4,s4):3 (n5,s5):5] next=6
I180530 08:03:11.808726 138578 storage/replica.go:3320 [n1,s1,r21/1:/Table/5{0-1}] proposing ADD_REPLICA((n5,s5):4): updated=[(n1,s1):1 (n2,s2):2 (n4,s4):3 (n5,s5):4] next=5
I180530 08:03:11.832736 138619 storage/replica_command.go:1777 [replicate,n2,s2,r8/2:/Table/1{1-2}] change replicas (REMOVE_REPLICA (n5,s5):5): read existing descriptor r8:/Table/1{1-2} [(n3,s3):4, (n2,s2):2, (n4,s4):3, (n5,s5):5, next=6]
I180530 08:03:11.861635 138671 storage/replica_command.go:1777 [replicate,n1,s1,r21/1:/Table/5{0-1}] change replicas (REMOVE_REPLICA (n5,s5):4): read existing descriptor r21:/Table/5{0-1} [(n1,s1):1, (n2,s2):2, (n4,s4):3, (n5,s5):4, next=5]
I180530 08:03:12.111461 134647 storage/replicate_queue_test.go:94 not balanced: [17 16 16 13 17]
I180530 08:03:12.236652 138655 storage/replica.go:3320 [n5,s5,r13/3:/Table/1{6-7}] proposing ADD_REPLICA((n3,s3):5): updated=[(n4,s4):4 (n2,s2):2 (n5,s5):3 (n3,s3):5] next=6
I180530 08:03:12.517898 138771 storage/replica.go:3320 [n3,s3,r5/2:/System/ts{d-e}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n4,s4):4 (n3,s3):2 (n2,s2):3] next=5
I180530 08:03:12.656662 138719 storage/replica.go:3320 [n1,s1,r21/1:/Table/5{0-1}] proposing REMOVE_REPLICA((n5,s5):4): updated=[(n1,s1):1 (n2,s2):2 (n4,s4):3] next=5
I180530 08:03:12.728088 136601 storage/store.go:3508 [n5,s5,r21/4:/Table/5{0-1}] added to replica GC queue (peer suggestion)
I180530 08:03:12.744551 136601 storage/store.go:3508 [n5,s5,r21/4:/Table/5{0-1}] added to replica GC queue (peer suggestion)
I180530 08:03:12.922358 138686 storage/store.go:2496 [replicaGC,n5,s5,r21/4:/Table/5{0-1}] removing replica
I180530 08:03:12.943024 136042 storage/replica_proposal.go:202 [n4,s4,r13/4:/Table/1{6-7}] new range lease repl=(n4,s4):4 seq=5 start=1527667392.464660635,0 epo=1 pro=1527667392.464693705,0 following repl=(n5,s5):3 seq=4 start=1527667380.380046322,0 epo=1 pro=1527667380.380059659,0
I180530 08:03:12.966633 138686 storage/replica.go:829 [replicaGC,n5,s5,r21/4:/Table/5{0-1}] removed 8 (0+8) keys in 43ms [clear=0ms commit=43ms]
I180530 08:03:13.000377 136451 storage/store.go:3508 [n1,s1,r5/1:/System/ts{d-e}] added to replica GC queue (peer suggestion)
I180530 08:03:13.000762 136451 storage/store.go:3508 [n1,s1,r5/1:/System/ts{d-e}] added to replica GC queue (peer suggestion)
I180530 08:03:13.023378 138795 storage/store.go:2496 [replicaGC,n1,s1,r5/1:/System/ts{d-e}] removing replica
I180530 08:03:13.025688 138795 storage/replica.go:829 [replicaGC,n1,s1,r5/1:/System/ts{d-e}] removed 440 (431+9) keys in 1ms [clear=0ms commit=1ms]
I180530 08:03:13.129934 134647 storage/replicate_queue_test.go:94 not balanced: [16 16 16 13 16]
I180530 08:03:13.400499 138636 storage/replica.go:3320 [n2,s2,r8/2:/Table/1{1-2}] proposing REMOVE_REPLICA((n5,s5):5): updated=[(n3,s3):4 (n2,s2):2 (n4,s4):3] next=6
I180530 08:03:13.406799 138810 storage/replica_raftstorage.go:520 [replicate,n1,s1,r16/1:/Table/{19-20}] generated preemptive snapshot fc9b5125 at index 24
I180530 08:03:13.459965 138810 storage/store_snapshot.go:605 [replicate,n1,s1,r16/1:/Table/{19-20}] streamed snapshot to (n4,s4):?: kv pairs: 11, log entries: 14, rate-limit: 2.0 MiB/sec, 50ms
I180530 08:03:13.463457 138832 storage/replica_raftstorage.go:730 [n4,s4,r16/?:{-}] applying preemptive snapshot at index 24 (id=fc9b5125, encoded size=4802, 1 rocksdb batches, 14 log entries)
I180530 08:03:13.558236 137255 storage/store.go:3508 [n5,s5,r8/5:/Table/1{1-2}] added to replica GC queue (peer suggestion)
I180530 08:03:13.690396 138832 storage/replica_raftstorage.go:736 [n4,s4,r16/?:/Table/{19-20}] applied preemptive snapshot in 227ms [clear=0ms batch=0ms entries=129ms commit=1ms]
I180530 08:03:13.730213 138810 storage/replica_command.go:1777 [replicate,n1,s1,r16/1:/Table/{19-20}] change replicas (ADD_REPLICA (n4,s4):4): read existing descriptor r16:/Table/{19-20} [(n1,s1):1, (n5,s5):2, (n2,s2):3, next=4]
I180530 08:03:13.835079 138833 storage/store.go:2496 [replicaGC,n5,s5,r8/5:/Table/1{1-2}] removing replica
I180530 08:03:13.867510 138833 storage/replica.go:829 [replicaGC,n5,s5,r8/5:/Table/1{1-2}] removed 8 (0+8) keys in 31ms [clear=0ms commit=31ms]
I180530 08:03:14.131744 138949 util/stop/stopper.go:471 quiescing; tasks left:
1 [async] transport racer
1 [async] storage.replicate: processing replica
1 [async] kv.TxnCoordSender: heartbeat loop
W180530 08:03:14.143274 136367 storage/raft_transport.go:465 [n1] raft transport stream to node 5 failed: rpc error: code = Unavailable desc = transport is closing
I180530 08:03:14.145859 138949 util/stop/stopper.go:471 quiescing; tasks left:
1 [async] transport racer
1 [async] storage.replicate: processing replica
W180530 08:03:14.149388 134895 storage/node_liveness.go:465 [n1,hb] failed node liveness heartbeat: node unavailable; try another peer
W180530 08:03:14.149715 138810 internal/client/txn.go:526 [replicate,n1,s1,r16/1:/Table/{19-20}] failure aborting transaction: node unavailable; try another peer; abort caused by: log-range-event: failed to send RPC: sending to all 4 replicas failed; last error: {<nil> context canceled}
I180530 08:03:14.152797 135895 gossip/gossip.go:1307 [n4] node has connected to cluster via gossip
I180530 08:03:14.154434 135895 storage/stores.go:241 [n4] wrote 4 node addresses to persistent storage
I180530 08:03:14.155346 138953 util/stop/stopper.go:471 quiescing; tasks left:
1 node.Node: batch
W180530 08:03:14.164827 136561 storage/raft_transport.go:465 [n4] raft transport stream to node 1 failed: EOF
W180530 08:03:14.174801 136421 gossip/infostore.go:303 [n5] node unavailable; try another peer
W180530 08:03:14.183405 137487 storage/raft_transport.go:465 [n5] raft transport stream to node 3 failed: EOF
W180530 08:03:14.193702 137389 storage/raft_transport.go:465 [n3] raft transport stream to node 2 failed: EOF
W180530 08:03:14.193813 137219 storage/raft_transport.go:465 [n5] raft transport stream to node 4 failed: EOF
W180530 08:03:14.207193 135833 storage/raft_transport.go:465 [n1] raft transport stream to node 2 failed: rpc error: code = Canceled desc = grpc: the client connection is closing
E180530 08:03:14.211888 138810 storage/queue.go:778 [replicate,n1,s1,r16/1:/Table/{19-20}] change replicas of r16 failed: log-range-event: failed to send RPC: sending to all 4 replicas failed; last error: {<nil> context canceled}
W180530 08:03:14.195876 137801 storage/raft_transport.go:465 [n3] raft transport stream to node 4 failed: EOF
W180530 08:03:14.230633 137309 storage/raft_transport.go:465 [n2] raft transport stream to node 4 failed: rpc error: code = Canceled desc = context canceled
I180530 08:03:14.230886 138949 util/stop/stopper.go:471 quiescing; tasks left:
1 [async] transport racer
W180530 08:03:14.194806 136218 gossip/gossip.go:1293 [n5] no incoming or outgoing connections
W180530 08:03:14.216441 136269 storage/raft_transport.go:465 [n3] raft transport stream to node 1 failed: rpc error: code = Canceled desc = context canceled
W180530 08:03:14.217362 137558 storage/raft_transport.go:465 [n3] raft transport stream to node 5 failed: rpc error: code = Canceled desc = context canceled
W180530 08:03:14.221091 135836 storage/raft_transport.go:465 [n2] raft transport stream to node 1 failed: rpc error: code = Canceled desc = context canceled
W180530 08:03:14.226614 137105 storage/raft_transport.go:465 [n5] raft transport stream to node 2 failed: EOF
W180530 08:03:14.227960 136599 storage/raft_transport.go:465 [n5] raft transport stream to node 1 failed: EOF
W180530 08:03:14.228232 136556 storage/raft_transport.go:465 [n1] raft transport stream to node 4 failed: rpc error: code = Canceled desc = grpc: the client connection is closing
W180530 08:03:14.230332 137299 storage/raft_transport.go:465 [n2] raft transport stream to node 5 failed: rpc error: code = Canceled desc = context canceled
W180530 08:03:14.203037 136433 storage/raft_transport.go:465 [n1] raft transport stream to node 3 failed: rpc error: code = Canceled desc = grpc: the client connection is closing
W180530 08:03:14.231432 137310 storage/raft_transport.go:465 [n2] raft transport stream to node 3 failed: rpc error: code = Canceled desc = grpc: the client connection is closing
W180530 08:03:14.238059 137241 storage/raft_transport.go:465 [n4] raft transport stream to node 5 failed: rpc error: code = Canceled desc = grpc: the client connection is closing
W180530 08:03:14.240091 137459 storage/raft_transport.go:465 [n4] raft transport stream to node 2 failed: rpc error: code = Canceled desc = context canceled
W180530 08:03:14.240671 137941 storage/raft_transport.go:465 [n4] raft transport stream to node 3 failed: rpc error: code = Canceled desc = grpc: the client connection is closing
I180530 08:03:16.398033 134873 kv/transport_race.go:67 transport race promotion: ran 101 iterations on up to 2394 requests
testing.go:730: race detected during execution of test
```
|
non_process
|
github com cockroachdb cockroach pkg storage testreplicatequeuerebalance failed under stress sha parameters tags goflags race failed test storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica go proposing remove replica updated next storage replica raftstorage go applied preemptive snapshot in storage replica go context canceled before proposing heartbeattxn storage replicate queue test go not balanced storage replicate queue test go not balanced storage replicate queue test go not balanced storage replicate queue test go not balanced storage replicate queue test go not balanced storage replica command go change replicas add replica read existing descriptor table storage replicate queue test go not balanced storage store go added to replica gc queue peer suggestion storage replicate queue test go not balanced storage replicate queue test go not balanced storage replicate queue test go not balanced storage replicate queue test go not balanced storage replicate queue test go not balanced storage raft transport go raft transport stream to node established storage replicate queue test go not balanced storage replicate queue test go not balanced storage replicate queue test go not balanced storage replicate queue test go not balanced storage replicate queue test go not balanced storage replicate queue test go not balanced storage replicate queue test go not balanced storage replicate queue test go not balanced storage replicate queue test go not balanced storage replicate queue test go not balanced storage replicate queue test go not balanced storage replicate queue test go not balanced storage replica raftstorage go generated preemptive snapshot at index storage replicate queue test go not balanced storage store go removing replica storage replica go removed keys in storage replicate queue test go not balanced storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replicate queue test go not balanced storage replicate queue test go not balanced storage replicate queue test go not balanced storage replica raftstorage go applied preemptive snapshot in storage replica go proposing add replica updated next storage replica command go change replicas add replica read existing descriptor system nodeliveness storage replicate queue test go not balanced storage replica raftstorage go generated preemptive snapshot at index storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replicate queue test go not balanced storage replica command go change replicas add replica read existing descriptor min system storage replica go proposing add replica updated next storage replicate queue test go not balanced storage replica go proposing add replica updated next storage replica command go change replicas remove replica read existing descriptor table storage replica raftstorage go generated preemptive snapshot at index storage queue go no removable replicas from range that needs a removal storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next storage replica command go change replicas remove replica read existing descriptor table storage raft transport go raft transport stream to node established storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica raftstorage go generated preemptive snapshot at index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replicate queue test go not balanced storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing remove replica updated next storage replica command go change replicas remove replica read existing descriptor system nodeliveness storage store go added to replica gc queue peer suggestion storage store go removing replica storage raft transport go raft transport stream to node established storage replica go removed keys in storage replica go proposing add replica updated next storage intent resolver go failed to push during intent resolution failed to push change replica id key local range table rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replicate queue test go not balanced storage replica go proposing remove replica updated next storage store go added to replica gc queue peer suggestion storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica raftstorage go generated preemptive snapshot at index server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage replica command go change replicas remove replica read existing descriptor min system storage store go removing replica storage replica command go change replicas remove replica read existing descriptor table server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica go removed keys in storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replicate queue test go not balanced storage replica go proposing remove replica updated next storage replica raftstorage go applied preemptive snapshot in storage store go added to replica gc queue peer suggestion storage replica command go change replicas add replica read existing descriptor system ts d e storage replica command go change replicas remove replica read existing descriptor table server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage store go removing replica storage replica go removed keys in storage replica go proposing remove replica updated next storage replicate queue test go not balanced storage store go removing replica storage replica go removed keys in storage replica command go change replicas remove replica read existing descriptor table storage replica go proposing add replica updated next server node go health alerts detected alerts storage replica go proposing remove replica updated next storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica raftstorage go generated preemptive snapshot at index server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage store go added to replica gc queue peer suggestion storage store go removing replica storage replica raftstorage go generated preemptive snapshot at index storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica go removed keys in storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replicate queue test go not balanced storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica command go change replicas add replica read existing descriptor table warning data race write at by goroutine github com cockroachdb cockroach pkg roachpb transaction update go src github com cockroachdb cockroach pkg roachpb data go github com cockroachdb cockroach pkg kv txncoordsender heartbeat go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg kv txncoordsender heartbeatloop go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg kv txncoordsender updatestate go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go previous read at by goroutine goroutine running created at github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg kv txncoordsender updatestate go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg kv txncoordsender send go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg internal client db sendusingsender go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client txn send go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client txn send fm go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client sendandfill go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client txn run go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg storage replica changereplicas go src github com cockroachdb cockroach pkg storage replica command go github com cockroachdb cockroach pkg internal client db txn go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client txn exec go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client db txn go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg storage replica changereplicas go src github com cockroachdb cockroach pkg storage replica command go github com cockroachdb cockroach pkg storage replica changereplicas go src github com cockroachdb cockroach pkg storage replica command go github com cockroachdb cockroach pkg storage replicatequeue removereplica go src github com cockroachdb cockroach pkg storage replicate queue go github com cockroachdb cockroach pkg storage replicatequeue processonechange go src github com cockroachdb cockroach pkg storage replicate queue go github com cockroachdb cockroach pkg storage replicatequeue process go src github com cockroachdb cockroach pkg storage replicate queue go github com cockroachdb cockroach pkg storage basequeue processreplica go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg storage basequeue processloop go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go goroutine running created at github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg kv grpctransportfactory go src github com cockroachdb cockroach pkg kv transport race go github com cockroachdb cockroach pkg kv distsender sendtoreplicas go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendrpc go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendsinglerange go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendpartialbatch go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender divideandsendbatchtoranges go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender send go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg internal client lookuprangefwdscan go src github com cockroachdb cockroach pkg internal client range lookup go github com cockroachdb cockroach pkg internal client rangelookup go src github com cockroachdb cockroach pkg internal client range lookup go github com cockroachdb cockroach pkg kv distsender rangelookup go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv rangedescriptorcache performrangelookup go src github com cockroachdb cockroach pkg kv range cache go github com cockroachdb cockroach pkg kv rangedescriptorcache lookuprangedescriptorinternal go src github com cockroachdb cockroach pkg kv range cache go github com cockroachdb cockroach pkg util syncutil singleflight group docall go src github com cockroachdb cockroach pkg util syncutil singleflight singleflight go warning data race write at by goroutine github com cockroachdb cockroach pkg roachpb transaction update go src github com cockroachdb cockroach pkg roachpb data go github com cockroachdb cockroach pkg kv txncoordsender heartbeat go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg kv txncoordsender heartbeatloop go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg kv txncoordsender updatestate go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go previous read at by goroutine goroutine running created at github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg kv txncoordsender updatestate go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg kv txncoordsender send go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg internal client db sendusingsender go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client txn send go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client txn send fm go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client sendandfill go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client txn run go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg storage replica changereplicas go src github com cockroachdb cockroach pkg storage replica command go github com cockroachdb cockroach pkg internal client db txn go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client txn exec go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client db txn go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg storage replica changereplicas go src github com cockroachdb cockroach pkg storage replica command go github com cockroachdb cockroach pkg storage replica changereplicas go src github com cockroachdb cockroach pkg storage replica command go github com cockroachdb cockroach pkg storage replicatequeue removereplica go src github com cockroachdb cockroach pkg storage replicate queue go github com cockroachdb cockroach pkg storage replicatequeue processonechange go src github com cockroachdb cockroach pkg storage replicate queue go github com cockroachdb cockroach pkg storage replicatequeue process go src github com cockroachdb cockroach pkg storage replicate queue go github com cockroachdb cockroach pkg storage basequeue processreplica go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg storage basequeue processloop go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go goroutine running created at github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg kv grpctransportfactory go src github com cockroachdb cockroach pkg kv transport race go github com cockroachdb cockroach pkg kv distsender sendtoreplicas go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendrpc go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendsinglerange go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendpartialbatch go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender divideandsendbatchtoranges go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender send go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg internal client lookuprangefwdscan go src github com cockroachdb cockroach pkg internal client range lookup go github com cockroachdb cockroach pkg internal client rangelookup go src github com cockroachdb cockroach pkg internal client range lookup go github com cockroachdb cockroach pkg kv distsender rangelookup go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv rangedescriptorcache performrangelookup go src github com cockroachdb cockroach pkg kv range cache go github com cockroachdb cockroach pkg kv rangedescriptorcache lookuprangedescriptorinternal go src github com cockroachdb cockroach pkg kv range cache go github com cockroachdb cockroach pkg util syncutil singleflight group docall go src github com cockroachdb cockroach pkg util syncutil singleflight singleflight go warning data race write at by goroutine github com cockroachdb cockroach pkg roachpb transaction update go src github com cockroachdb cockroach pkg util hlc timestamp go github com cockroachdb cockroach pkg kv txncoordsender heartbeat go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg kv txncoordsender heartbeatloop go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg kv txncoordsender updatestate go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go previous read at by goroutine goroutine running created at github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg kv txncoordsender updatestate go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg kv txncoordsender send go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg internal client db sendusingsender go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client txn send go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client txn send fm go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client sendandfill go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client txn run go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg storage replica changereplicas go src github com cockroachdb cockroach pkg storage replica command go github com cockroachdb cockroach pkg internal client db txn go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client txn exec go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client db txn go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg storage replica changereplicas go src github com cockroachdb cockroach pkg storage replica command go github com cockroachdb cockroach pkg storage replica changereplicas go src github com cockroachdb cockroach pkg storage replica command go github com cockroachdb cockroach pkg storage replicatequeue removereplica go src github com cockroachdb cockroach pkg storage replicate queue go github com cockroachdb cockroach pkg storage replicatequeue processonechange go src github com cockroachdb cockroach pkg storage replicate queue go github com cockroachdb cockroach pkg storage replicatequeue process go src github com cockroachdb cockroach pkg storage replicate queue go github com cockroachdb cockroach pkg storage basequeue processreplica go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg storage basequeue processloop go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go goroutine running created at github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg kv grpctransportfactory go src github com cockroachdb cockroach pkg kv transport race go github com cockroachdb cockroach pkg kv distsender sendtoreplicas go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendrpc go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendsinglerange go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendpartialbatch go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender divideandsendbatchtoranges go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender send go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg internal client lookuprangefwdscan go src github com cockroachdb cockroach pkg internal client range lookup go github com cockroachdb cockroach pkg internal client rangelookup go src github com cockroachdb cockroach pkg internal client range lookup go github com cockroachdb cockroach pkg kv distsender rangelookup go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv rangedescriptorcache performrangelookup go src github com cockroachdb cockroach pkg kv range cache go github com cockroachdb cockroach pkg kv rangedescriptorcache lookuprangedescriptorinternal go src github com cockroachdb cockroach pkg kv range cache go github com cockroachdb cockroach pkg util syncutil singleflight group docall go src github com cockroachdb cockroach pkg util syncutil singleflight singleflight go warning data race write at by goroutine github com cockroachdb cockroach pkg roachpb transaction updateobservedtimestamp go src github com cockroachdb cockroach pkg roachpb data go github com cockroachdb cockroach pkg roachpb transaction update go src github com cockroachdb cockroach pkg roachpb data go github com cockroachdb cockroach pkg kv txncoordsender heartbeat go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg kv txncoordsender heartbeatloop go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg kv txncoordsender updatestate go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go previous read at by goroutine goroutine running created at github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg kv txncoordsender updatestate go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg kv txncoordsender send go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg internal client db sendusingsender go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client txn send go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client txn send fm go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client sendandfill go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client txn run go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg storage replica changereplicas go src github com cockroachdb cockroach pkg storage replica command go github com cockroachdb cockroach pkg internal client db txn go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client txn exec go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client db txn go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg storage replica changereplicas go src github com cockroachdb cockroach pkg storage replica command go github com cockroachdb cockroach pkg storage replica changereplicas go src github com cockroachdb cockroach pkg storage replica command go github com cockroachdb cockroach pkg storage replicatequeue removereplica go src github com cockroachdb cockroach pkg storage replicate queue go github com cockroachdb cockroach pkg storage replicatequeue processonechange go src github com cockroachdb cockroach pkg storage replicate queue go github com cockroachdb cockroach pkg storage replicatequeue process go src github com cockroachdb cockroach pkg storage replicate queue go github com cockroachdb cockroach pkg storage basequeue processreplica go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg storage basequeue processloop go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go goroutine running created at github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg kv grpctransportfactory go src github com cockroachdb cockroach pkg kv transport race go github com cockroachdb cockroach pkg kv distsender sendtoreplicas go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendrpc go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendsinglerange go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendpartialbatch go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender divideandsendbatchtoranges go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender send go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg internal client lookuprangefwdscan go src github com cockroachdb cockroach pkg internal client range lookup go github com cockroachdb cockroach pkg internal client rangelookup go src github com cockroachdb cockroach pkg internal client range lookup go github com cockroachdb cockroach pkg kv distsender rangelookup go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv rangedescriptorcache performrangelookup go src github com cockroachdb cockroach pkg kv range cache go github com cockroachdb cockroach pkg kv rangedescriptorcache lookuprangedescriptorinternal go src github com cockroachdb cockroach pkg kv range cache go github com cockroachdb cockroach pkg util syncutil singleflight group docall go src github com cockroachdb cockroach pkg util syncutil singleflight singleflight go warning data race write at by goroutine github com cockroachdb cockroach pkg roachpb transaction update go src github com cockroachdb cockroach pkg roachpb data go github com cockroachdb cockroach pkg kv txncoordsender heartbeat go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg kv txncoordsender heartbeatloop go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg kv txncoordsender updatestate go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go previous read at by goroutine goroutine running created at github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg kv txncoordsender updatestate go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg kv txncoordsender send go src github com cockroachdb cockroach pkg kv txn coord sender go github com cockroachdb cockroach pkg internal client db sendusingsender go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client txn send go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client txn send fm go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client sendandfill go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client txn run go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg storage replica changereplicas go src github com cockroachdb cockroach pkg storage replica command go github com cockroachdb cockroach pkg internal client db txn go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg internal client txn exec go src github com cockroachdb cockroach pkg internal client txn go github com cockroachdb cockroach pkg internal client db txn go src github com cockroachdb cockroach pkg internal client db go github com cockroachdb cockroach pkg storage replica changereplicas go src github com cockroachdb cockroach pkg storage replica command go github com cockroachdb cockroach pkg storage replica changereplicas go src github com cockroachdb cockroach pkg storage replica command go github com cockroachdb cockroach pkg storage replicatequeue removereplica go src github com cockroachdb cockroach pkg storage replicate queue go github com cockroachdb cockroach pkg storage replicatequeue processonechange go src github com cockroachdb cockroach pkg storage replicate queue go github com cockroachdb cockroach pkg storage replicatequeue process go src github com cockroachdb cockroach pkg storage replicate queue go github com cockroachdb cockroach pkg storage basequeue processreplica go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg storage basequeue processloop go src github com cockroachdb cockroach pkg storage queue go github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go goroutine running created at github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg kv grpctransportfactory go src github com cockroachdb cockroach pkg kv transport race go github com cockroachdb cockroach pkg kv distsender sendtoreplicas go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendrpc go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendsinglerange go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender sendpartialbatch go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender divideandsendbatchtoranges go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv distsender send go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg internal client lookuprangefwdscan go src github com cockroachdb cockroach pkg internal client range lookup go github com cockroachdb cockroach pkg internal client rangelookup go src github com cockroachdb cockroach pkg internal client range lookup go github com cockroachdb cockroach pkg kv distsender rangelookup go src github com cockroachdb cockroach pkg kv dist sender go github com cockroachdb cockroach pkg kv rangedescriptorcache performrangelookup go src github com cockroachdb cockroach pkg kv range cache go github com cockroachdb cockroach pkg kv rangedescriptorcache lookuprangedescriptorinternal go src github com cockroachdb cockroach pkg kv range cache go github com cockroachdb cockroach pkg util syncutil singleflight group docall go src github com cockroachdb cockroach pkg util syncutil singleflight singleflight go storage replica go proposing remove replica updated next storage replica raftstorage go applied preemptive snapshot in server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage replica go proposing remove replica updated next storage replica command go change replicas add replica read existing descriptor table storage replica go context canceled before proposing heartbeattxn storage replica raftstorage go generated preemptive snapshot at index storage store go added to replica gc queue peer suggestion storage store go added to replica gc queue peer suggestion storage store go added to replica gc queue peer suggestion storage store go added to replica gc queue peer suggestion storage replicate queue test go not balanced storage store go added to replica gc queue peer suggestion storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage store go removing replica storage replica go removed keys in storage replica command go change replicas remove replica read existing descriptor system ts d e storage store go removing replica storage replica raftstorage go applied preemptive snapshot in storage replica go removed keys in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next storage replica go proposing add replica updated next storage replica command go change replicas remove replica read existing descriptor table storage replica command go change replicas remove replica read existing descriptor table storage replicate queue test go not balanced storage replica go proposing add replica updated next storage replica go proposing remove replica updated next storage replica go proposing remove replica updated next storage store go added to replica gc queue peer suggestion storage store go added to replica gc queue peer suggestion storage store go removing replica storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica go removed keys in storage store go added to replica gc queue peer suggestion storage store go added to replica gc queue peer suggestion storage store go removing replica storage replica go removed keys in storage replicate queue test go not balanced storage replica go proposing remove replica updated next storage replica raftstorage go generated preemptive snapshot at index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage store go added to replica gc queue peer suggestion storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage store go removing replica storage replica go removed keys in util stop stopper go quiescing tasks left transport racer storage replicate processing replica kv txncoordsender heartbeat loop storage raft transport go raft transport stream to node failed rpc error code unavailable desc transport is closing util stop stopper go quiescing tasks left transport racer storage replicate processing replica storage node liveness go failed node liveness heartbeat node unavailable try another peer internal client txn go failure aborting transaction node unavailable try another peer abort caused by log range event failed to send rpc sending to all replicas failed last error context canceled gossip gossip go node has connected to cluster via gossip storage stores go wrote node addresses to persistent storage util stop stopper go quiescing tasks left node node batch storage raft transport go raft transport stream to node failed eof gossip infostore go node unavailable try another peer storage raft transport go raft transport stream to node failed eof storage raft transport go raft transport stream to node failed eof storage raft transport go raft transport stream to node failed eof storage raft transport go raft transport stream to node failed rpc error code canceled desc grpc the client connection is closing storage queue go change replicas of failed log range event failed to send rpc sending to all replicas failed last error context canceled storage raft transport go raft transport stream to node failed eof storage raft transport go raft transport stream to node failed rpc error code canceled desc context canceled util stop stopper go quiescing tasks left transport racer gossip gossip go no incoming or outgoing connections storage raft transport go raft transport stream to node failed rpc error code canceled desc context canceled storage raft transport go raft transport stream to node failed rpc error code canceled desc context canceled storage raft transport go raft transport stream to node failed rpc error code canceled desc context canceled storage raft transport go raft transport stream to node failed eof storage raft transport go raft transport stream to node failed eof storage raft transport go raft transport stream to node failed rpc error code canceled desc grpc the client connection is closing storage raft transport go raft transport stream to node failed rpc error code canceled desc context canceled storage raft transport go raft transport stream to node failed rpc error code canceled desc grpc the client connection is closing storage raft transport go raft transport stream to node failed rpc error code canceled desc grpc the client connection is closing storage raft transport go raft transport stream to node failed rpc error code canceled desc grpc the client connection is closing storage raft transport go raft transport stream to node failed rpc error code canceled desc context canceled storage raft transport go raft transport stream to node failed rpc error code canceled desc grpc the client connection is closing kv transport race go transport race promotion ran iterations on up to requests testing go race detected during execution of test
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.