Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 4
112
| repo_url
stringlengths 33
141
| action
stringclasses 3
values | title
stringlengths 1
1.02k
| labels
stringlengths 4
1.54k
| body
stringlengths 1
262k
| index
stringclasses 17
values | text_combine
stringlengths 95
262k
| label
stringclasses 2
values | text
stringlengths 96
252k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
102,192
| 16,548,284,026
|
IssuesEvent
|
2021-05-28 04:35:54
|
samq-ghdemo/Java-Demo
|
https://api.github.com/repos/samq-ghdemo/Java-Demo
|
opened
|
CVE-2017-3589 (Low) detected in mysql-connector-java-5.1.25.jar
|
security vulnerability
|
## CVE-2017-3589 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mysql-connector-java-5.1.25.jar</b></p></summary>
<p>MySQL JDBC Type 4 driver</p>
<p>Library home page: <a href="http://dev.mysql.com/doc/connector-j/en/">http://dev.mysql.com/doc/connector-j/en/</a></p>
<p>Path to dependency file: Java-Demo/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/mysql/mysql-connector-java/5.1.25/mysql-connector-java-5.1.25.jar,Java-Demo/target/easybuggy-1-SNAPSHOT/WEB-INF/lib/mysql-connector-java-5.1.25.jar,Java-Demo/.extract/webapps/ROOT/WEB-INF/lib/mysql-connector-java-5.1.25.jar</p>
<p>
Dependency Hierarchy:
- :x: **mysql-connector-java-5.1.25.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/Java-Demo/commit/ebe5ad9510b919e1dd226c70afc5e1115136f8cf">ebe5ad9510b919e1dd226c70afc5e1115136f8cf</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Vulnerability in the MySQL Connectors component of Oracle MySQL (subcomponent: Connector/J). Supported versions that are affected are 5.1.41 and earlier. Easily "exploitable" vulnerability allows low privileged attacker with logon to the infrastructure where MySQL Connectors executes to compromise MySQL Connectors. Successful attacks of this vulnerability can result in unauthorized update, insert or delete access to some of MySQL Connectors accessible data. CVSS 3.0 Base Score 3.3 (Integrity impacts). CVSS Vector: (CVSS:3.0/AV:L/AC:L/PR:L/UI:N/S:U/C:N/I:L/A:N).
<p>Publish Date: 2017-04-24
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-3589>CVE-2017-3589</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-3589">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-3589</a></p>
<p>Release Date: 2017-04-24</p>
<p>Fix Resolution: 5.1.42</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"mysql","packageName":"mysql-connector-java","packageVersion":"5.1.25","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"mysql:mysql-connector-java:5.1.25","isMinimumFixVersionAvailable":true,"minimumFixVersion":"5.1.42"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-3589","vulnerabilityDetails":"Vulnerability in the MySQL Connectors component of Oracle MySQL (subcomponent: Connector/J). Supported versions that are affected are 5.1.41 and earlier. Easily \"exploitable\" vulnerability allows low privileged attacker with logon to the infrastructure where MySQL Connectors executes to compromise MySQL Connectors. Successful attacks of this vulnerability can result in unauthorized update, insert or delete access to some of MySQL Connectors accessible data. CVSS 3.0 Base Score 3.3 (Integrity impacts). CVSS Vector: (CVSS:3.0/AV:L/AC:L/PR:L/UI:N/S:U/C:N/I:L/A:N).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-3589","cvss3Severity":"low","cvss3Score":"3.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Local","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2017-3589 (Low) detected in mysql-connector-java-5.1.25.jar - ## CVE-2017-3589 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mysql-connector-java-5.1.25.jar</b></p></summary>
<p>MySQL JDBC Type 4 driver</p>
<p>Library home page: <a href="http://dev.mysql.com/doc/connector-j/en/">http://dev.mysql.com/doc/connector-j/en/</a></p>
<p>Path to dependency file: Java-Demo/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/mysql/mysql-connector-java/5.1.25/mysql-connector-java-5.1.25.jar,Java-Demo/target/easybuggy-1-SNAPSHOT/WEB-INF/lib/mysql-connector-java-5.1.25.jar,Java-Demo/.extract/webapps/ROOT/WEB-INF/lib/mysql-connector-java-5.1.25.jar</p>
<p>
Dependency Hierarchy:
- :x: **mysql-connector-java-5.1.25.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/Java-Demo/commit/ebe5ad9510b919e1dd226c70afc5e1115136f8cf">ebe5ad9510b919e1dd226c70afc5e1115136f8cf</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Vulnerability in the MySQL Connectors component of Oracle MySQL (subcomponent: Connector/J). Supported versions that are affected are 5.1.41 and earlier. Easily "exploitable" vulnerability allows low privileged attacker with logon to the infrastructure where MySQL Connectors executes to compromise MySQL Connectors. Successful attacks of this vulnerability can result in unauthorized update, insert or delete access to some of MySQL Connectors accessible data. CVSS 3.0 Base Score 3.3 (Integrity impacts). CVSS Vector: (CVSS:3.0/AV:L/AC:L/PR:L/UI:N/S:U/C:N/I:L/A:N).
<p>Publish Date: 2017-04-24
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-3589>CVE-2017-3589</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-3589">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-3589</a></p>
<p>Release Date: 2017-04-24</p>
<p>Fix Resolution: 5.1.42</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"mysql","packageName":"mysql-connector-java","packageVersion":"5.1.25","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"mysql:mysql-connector-java:5.1.25","isMinimumFixVersionAvailable":true,"minimumFixVersion":"5.1.42"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-3589","vulnerabilityDetails":"Vulnerability in the MySQL Connectors component of Oracle MySQL (subcomponent: Connector/J). Supported versions that are affected are 5.1.41 and earlier. Easily \"exploitable\" vulnerability allows low privileged attacker with logon to the infrastructure where MySQL Connectors executes to compromise MySQL Connectors. Successful attacks of this vulnerability can result in unauthorized update, insert or delete access to some of MySQL Connectors accessible data. CVSS 3.0 Base Score 3.3 (Integrity impacts). CVSS Vector: (CVSS:3.0/AV:L/AC:L/PR:L/UI:N/S:U/C:N/I:L/A:N).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-3589","cvss3Severity":"low","cvss3Score":"3.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Local","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
non_test
|
cve low detected in mysql connector java jar cve low severity vulnerability vulnerable library mysql connector java jar mysql jdbc type driver library home page a href path to dependency file java demo pom xml path to vulnerable library canner repository mysql mysql connector java mysql connector java jar java demo target easybuggy snapshot web inf lib mysql connector java jar java demo extract webapps root web inf lib mysql connector java jar dependency hierarchy x mysql connector java jar vulnerable library found in head commit a href found in base branch main vulnerability details vulnerability in the mysql connectors component of oracle mysql subcomponent connector j supported versions that are affected are and earlier easily exploitable vulnerability allows low privileged attacker with logon to the infrastructure where mysql connectors executes to compromise mysql connectors successful attacks of this vulnerability can result in unauthorized update insert or delete access to some of mysql connectors accessible data cvss base score integrity impacts cvss vector cvss av l ac l pr l ui n s u c n i l a n publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree mysql mysql connector java isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails vulnerability in the mysql connectors component of oracle mysql subcomponent connector j supported versions that are affected are and earlier easily exploitable vulnerability allows low privileged attacker with logon to the infrastructure where mysql connectors executes to compromise mysql connectors successful attacks of this vulnerability can result in unauthorized update insert or delete access to some of mysql connectors accessible data cvss base score integrity impacts cvss vector cvss av l ac l pr l ui n s u c n i l a n vulnerabilityurl
| 0
|
33,685
| 4,848,973,578
|
IssuesEvent
|
2016-11-10 19:04:34
|
radare/radare2
|
https://api.github.com/repos/radare/radare2
|
closed
|
Commands are sensitive to double-whitespace
|
has-test
|
Some commands will not work correctly with multiple spaces instead of one, but will not report an error/warning either. For example:
```
// this works
f foobar=0x1000
// this does nothing
f foobar=0x1000
```
version 0.10.6
|
1.0
|
Commands are sensitive to double-whitespace - Some commands will not work correctly with multiple spaces instead of one, but will not report an error/warning either. For example:
```
// this works
f foobar=0x1000
// this does nothing
f foobar=0x1000
```
version 0.10.6
|
test
|
commands are sensitive to double whitespace some commands will not work correctly with multiple spaces instead of one but will not report an error warning either for example this works f foobar this does nothing f foobar version
| 1
|
454,401
| 13,100,218,753
|
IssuesEvent
|
2020-08-03 23:48:11
|
GoogleCloudPlatform/stackdriver-sandbox
|
https://api.github.com/repos/GoogleCloudPlatform/stackdriver-sandbox
|
opened
|
Credential check fail when accessing the storage bucket
|
priority: p2 type: bug
|
Some users may encounter a credential check failure when accessing the storage bucket in Terraform initialization. They need to do "gcloud auth application-default login". We need to put a try-catch block around the initialization.
|
1.0
|
Credential check fail when accessing the storage bucket - Some users may encounter a credential check failure when accessing the storage bucket in Terraform initialization. They need to do "gcloud auth application-default login". We need to put a try-catch block around the initialization.
|
non_test
|
credential check fail when accessing the storage bucket some users may encounter a credential check failure when accessing the storage bucket in terraform initialization they need to do gcloud auth application default login we need to put a try catch block around the initialization
| 0
|
5,715
| 2,790,522,000
|
IssuesEvent
|
2015-05-09 09:27:17
|
Dalmirog/OctoPosh
|
https://api.github.com/repos/Dalmirog/OctoPosh
|
reopened
|
Improve tests for Get-* cmdlets
|
Testing
|
They should be more Unit-test-like and consider scenarios like:
- Get-* should not get a resource that doesnt exist
- Get-* [specific names] should only get resources with those names, and nothing extra
- Get-* with date filters should return results with between the correct date ranges
- Get-* with version filters should return results with between the correct version ranges
Progress tracking
- [ ] Get-OctopusConnectionInfo
- [ ] Get-OctopusDeployment
- [ ] Get-OctopusEnvironment
- [ ] Get-OctopusLifeCycle
- [ ] Get-OctopusMaintenanceMode
- [ ] Get-OctopusProject
- [ ] Get-OctopusProjectGroup
- [ ] Get-OctopusProjectVariable
- [ ] Get-OctopusRelease
- [ ] Get-OctopusResourceModel
- [ ] Get-OctopusSMTPConfig
|
1.0
|
Improve tests for Get-* cmdlets - They should be more Unit-test-like and consider scenarios like:
- Get-* should not get a resource that doesnt exist
- Get-* [specific names] should only get resources with those names, and nothing extra
- Get-* with date filters should return results with between the correct date ranges
- Get-* with version filters should return results with between the correct version ranges
Progress tracking
- [ ] Get-OctopusConnectionInfo
- [ ] Get-OctopusDeployment
- [ ] Get-OctopusEnvironment
- [ ] Get-OctopusLifeCycle
- [ ] Get-OctopusMaintenanceMode
- [ ] Get-OctopusProject
- [ ] Get-OctopusProjectGroup
- [ ] Get-OctopusProjectVariable
- [ ] Get-OctopusRelease
- [ ] Get-OctopusResourceModel
- [ ] Get-OctopusSMTPConfig
|
test
|
improve tests for get cmdlets they should be more unit test like and consider scenarios like get should not get a resource that doesnt exist get should only get resources with those names and nothing extra get with date filters should return results with between the correct date ranges get with version filters should return results with between the correct version ranges progress tracking get octopusconnectioninfo get octopusdeployment get octopusenvironment get octopuslifecycle get octopusmaintenancemode get octopusproject get octopusprojectgroup get octopusprojectvariable get octopusrelease get octopusresourcemodel get octopussmtpconfig
| 1
|
325,821
| 27,964,388,929
|
IssuesEvent
|
2023-03-24 18:08:47
|
Satellite-im/testing-uplink
|
https://api.github.com/repos/Satellite-im/testing-uplink
|
opened
|
UI Tests - Settings Developer - Save Logs In a File
|
test Settings
|
Logs should save in a file when User toggles on Save Logs In A File
|
1.0
|
UI Tests - Settings Developer - Save Logs In a File - Logs should save in a file when User toggles on Save Logs In A File
|
test
|
ui tests settings developer save logs in a file logs should save in a file when user toggles on save logs in a file
| 1
|
92,837
| 11,714,795,081
|
IssuesEvent
|
2020-03-09 13:03:47
|
EightShapes/esds-library
|
https://api.github.com/repos/EightShapes/esds-library
|
closed
|
Required Field Indicator [Design Asset]
|
Component [Design Asset]
|
### Design Starting Point
This is the key for how forms denote required fields, which will usually be positioned below the title of the form.

### Must Have
* Settings: white, light, dark, black
* Leverages existing typography
* Matches indicator on `Label`
|
1.0
|
Required Field Indicator [Design Asset] - ### Design Starting Point
This is the key for how forms denote required fields, which will usually be positioned below the title of the form.

### Must Have
* Settings: white, light, dark, black
* Leverages existing typography
* Matches indicator on `Label`
|
non_test
|
required field indicator design starting point this is the key for how forms denote required fields which will usually be positioned below the title of the form must have settings white light dark black leverages existing typography matches indicator on label
| 0
|
32,133
| 4,751,450,468
|
IssuesEvent
|
2016-10-22 22:02:52
|
flowplayer/flowplayer
|
https://api.github.com/repos/flowplayer/flowplayer
|
closed
|
[Chrome] playlist: error on resume when last clip is in finished state
|
ready to test
|
https://flowplayer.org/standalone/playlist/javascript.html (no plugins)
- finish last clip
- resume by clicking or typing space or `flowplayer(0).resume()` in the console
yields:
```
flowplayer.min.js:6 Uncaught (in promise) DOMException: The play() request was interrupted by a new load request
```
Which describes the problem exactly: play() is requested on the last clip, but interrupted by loading the first.
Proof: entering `next()` in the console does not trigger the error. Or resuming when the playlist has length 1.
Obviously does not happen either when advance is false.
A fix most likely has to be cross-checked in the Flash engine.
|
1.0
|
[Chrome] playlist: error on resume when last clip is in finished state - https://flowplayer.org/standalone/playlist/javascript.html (no plugins)
- finish last clip
- resume by clicking or typing space or `flowplayer(0).resume()` in the console
yields:
```
flowplayer.min.js:6 Uncaught (in promise) DOMException: The play() request was interrupted by a new load request
```
Which describes the problem exactly: play() is requested on the last clip, but interrupted by loading the first.
Proof: entering `next()` in the console does not trigger the error. Or resuming when the playlist has length 1.
Obviously does not happen either when advance is false.
A fix most likely has to be cross-checked in the Flash engine.
|
test
|
playlist error on resume when last clip is in finished state no plugins finish last clip resume by clicking or typing space or flowplayer resume in the console yields flowplayer min js uncaught in promise domexception the play request was interrupted by a new load request which describes the problem exactly play is requested on the last clip but interrupted by loading the first proof entering next in the console does not trigger the error or resuming when the playlist has length obviously does not happen either when advance is false a fix most likely has to be cross checked in the flash engine
| 1
|
141,798
| 11,437,609,410
|
IssuesEvent
|
2020-02-05 00:30:18
|
rancher/rancher
|
https://api.github.com/repos/rancher/rancher
|
closed
|
Go panic in NodeController every 1 second
|
[zube]: To Test internal
|
<!--
Please search for existing issues first, then read https://rancher.com/docs/rancher/v2.x/en/contributing/#bugs-issues-or-questions to see what we expect in an issue
For security issues, please email security@rancher.com instead of posting a public issue in GitHub. You may (but are not required to) use the GPG key located on Keybase.
-->
**What kind of request is this (question/bug/enhancement/feature request):**
bug
**Steps to reproduce (least amount of steps as possible):**
unknown
**Result:**
Seeing this about every 1 second in Rancher logs:
```
2020-01-27T04:01:08.803268734Z 2020/01/27 04:01:08 [ERROR] NodeController xxxxxxxxxxxxxxxxx recovered from panic "runtime error: invalid memory address or nil pointer dereference". (err=<nil>) Call stack:
2020-01-27T04:01:08.803306432Z goroutine 443379 [running]:
2020-01-27T04:01:08.803312366Z github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/runtime.RecoverFromPanic(0xc0045a3d30)
2020-01-27T04:01:08.803316668Z /go/src/github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:158 +0xb5
2020-01-27T04:01:08.803320855Z panic(0x3208c60, 0x62c3810)
2020-01-27T04:01:08.803324801Z /usr/local/go/src/runtime/panic.go:679 +0x1b2
2020-01-27T04:01:08.803350733Z github.com/rancher/rancher/pkg/controllers/user/monitoring.(*clusterMonitoringEnabledHandler).syncWindowsNode(0xc002f03c00, 0xc006b49150, 0xf, 0x0, 0xc000ddd880, 0x7fb72dd1e6d0, 0x0, 0x0)
2020-01-27T04:01:08.803356241Z /go/src/github.com/rancher/rancher/pkg/controllers/user/monitoring/cluster_monitoring_enabled_handler.go:58 +0xe6
2020-01-27T04:01:08.803360044Z github.com/rancher/rancher/vendor/github.com/rancher/types/apis/core/v1.(*nodeController).AddHandler.func1(0xc006b49150, 0xf, 0x0, 0x0, 0xc0045a3cd8, 0x3, 0x3, 0x0)
2020-01-27T04:01:08.803363969Z /go/src/github.com/rancher/rancher/vendor/github.com/rancher/types/apis/core/v1/zz_generated_node_controller.go:149 +0xe0
2020-01-27T04:01:08.803387512Z github.com/rancher/rancher/vendor/github.com/rancher/norman/controller.(*genericController).syncHandler(0xc00207a700, 0x2f92320, 0xc008190e00, 0x0, 0x0)
2020-01-27T04:01:08.803390640Z /go/src/github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:367 +0x371
2020-01-27T04:01:08.803393545Z github.com/rancher/rancher/vendor/github.com/rancher/norman/controller.(*genericController).processNextWorkItem(0xc00207a700, 0x0)
2020-01-27T04:01:08.803396147Z /go/src/github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:276 +0xef
2020-01-27T04:01:08.803398815Z github.com/rancher/rancher/vendor/github.com/rancher/norman/controller.(*genericController).runWorker(0xc00207a700)
2020-01-27T04:01:08.803401547Z /go/src/github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:264 +0x2b
2020-01-27T04:01:08.803404977Z github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1(0xc00378f670)
2020-01-27T04:01:08.803408494Z /go/src/github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:152 +0x5e
2020-01-27T04:01:08.803419953Z github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil(0xc00378f670, 0x3b9aca00, 0x0, 0x1, 0xc003568ae0)
2020-01-27T04:01:08.803424598Z /go/src/github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:153 +0xf8
2020-01-27T04:01:08.803428243Z github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait.Until(0xc00378f670, 0x3b9aca00, 0xc003568ae0)
2020-01-27T04:01:08.803431733Z /go/src/github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88 +0x4d
2020-01-27T04:01:08.803435732Z created by github.com/rancher/rancher/vendor/github.com/rancher/norman/controller.(*genericController).run
2020-01-27T04:01:08.803439922Z /go/src/github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:256 +0xe8
```
**Other details that may be helpful:**
Could be related to https://github.com/rancher/rancher/issues/25164
**Environment information**
- Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI):
v2.3.4
- Installation option (single install/HA):
HA
<!--
If the reported issue is regarding a created cluster, please provide requested info below
-->
**Cluster information**
- Cluster type (Hosted/Infrastructure Provider/Custom/Imported):
Custom
- Machine type (cloud/VM/metal) and specifications (CPU/memory):
- Kubernetes version (use `kubectl version`):
```
1.15
```
- Docker version (use `docker version`):
```
?
```
|
1.0
|
Go panic in NodeController every 1 second - <!--
Please search for existing issues first, then read https://rancher.com/docs/rancher/v2.x/en/contributing/#bugs-issues-or-questions to see what we expect in an issue
For security issues, please email security@rancher.com instead of posting a public issue in GitHub. You may (but are not required to) use the GPG key located on Keybase.
-->
**What kind of request is this (question/bug/enhancement/feature request):**
bug
**Steps to reproduce (least amount of steps as possible):**
unknown
**Result:**
Seeing this about every 1 second in Rancher logs:
```
2020-01-27T04:01:08.803268734Z 2020/01/27 04:01:08 [ERROR] NodeController xxxxxxxxxxxxxxxxx recovered from panic "runtime error: invalid memory address or nil pointer dereference". (err=<nil>) Call stack:
2020-01-27T04:01:08.803306432Z goroutine 443379 [running]:
2020-01-27T04:01:08.803312366Z github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/runtime.RecoverFromPanic(0xc0045a3d30)
2020-01-27T04:01:08.803316668Z /go/src/github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:158 +0xb5
2020-01-27T04:01:08.803320855Z panic(0x3208c60, 0x62c3810)
2020-01-27T04:01:08.803324801Z /usr/local/go/src/runtime/panic.go:679 +0x1b2
2020-01-27T04:01:08.803350733Z github.com/rancher/rancher/pkg/controllers/user/monitoring.(*clusterMonitoringEnabledHandler).syncWindowsNode(0xc002f03c00, 0xc006b49150, 0xf, 0x0, 0xc000ddd880, 0x7fb72dd1e6d0, 0x0, 0x0)
2020-01-27T04:01:08.803356241Z /go/src/github.com/rancher/rancher/pkg/controllers/user/monitoring/cluster_monitoring_enabled_handler.go:58 +0xe6
2020-01-27T04:01:08.803360044Z github.com/rancher/rancher/vendor/github.com/rancher/types/apis/core/v1.(*nodeController).AddHandler.func1(0xc006b49150, 0xf, 0x0, 0x0, 0xc0045a3cd8, 0x3, 0x3, 0x0)
2020-01-27T04:01:08.803363969Z /go/src/github.com/rancher/rancher/vendor/github.com/rancher/types/apis/core/v1/zz_generated_node_controller.go:149 +0xe0
2020-01-27T04:01:08.803387512Z github.com/rancher/rancher/vendor/github.com/rancher/norman/controller.(*genericController).syncHandler(0xc00207a700, 0x2f92320, 0xc008190e00, 0x0, 0x0)
2020-01-27T04:01:08.803390640Z /go/src/github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:367 +0x371
2020-01-27T04:01:08.803393545Z github.com/rancher/rancher/vendor/github.com/rancher/norman/controller.(*genericController).processNextWorkItem(0xc00207a700, 0x0)
2020-01-27T04:01:08.803396147Z /go/src/github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:276 +0xef
2020-01-27T04:01:08.803398815Z github.com/rancher/rancher/vendor/github.com/rancher/norman/controller.(*genericController).runWorker(0xc00207a700)
2020-01-27T04:01:08.803401547Z /go/src/github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:264 +0x2b
2020-01-27T04:01:08.803404977Z github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1(0xc00378f670)
2020-01-27T04:01:08.803408494Z /go/src/github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:152 +0x5e
2020-01-27T04:01:08.803419953Z github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil(0xc00378f670, 0x3b9aca00, 0x0, 0x1, 0xc003568ae0)
2020-01-27T04:01:08.803424598Z /go/src/github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:153 +0xf8
2020-01-27T04:01:08.803428243Z github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait.Until(0xc00378f670, 0x3b9aca00, 0xc003568ae0)
2020-01-27T04:01:08.803431733Z /go/src/github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88 +0x4d
2020-01-27T04:01:08.803435732Z created by github.com/rancher/rancher/vendor/github.com/rancher/norman/controller.(*genericController).run
2020-01-27T04:01:08.803439922Z /go/src/github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:256 +0xe8
```
**Other details that may be helpful:**
Could be related to https://github.com/rancher/rancher/issues/25164
**Environment information**
- Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI):
v2.3.4
- Installation option (single install/HA):
HA
<!--
If the reported issue is regarding a created cluster, please provide requested info below
-->
**Cluster information**
- Cluster type (Hosted/Infrastructure Provider/Custom/Imported):
Custom
- Machine type (cloud/VM/metal) and specifications (CPU/memory):
- Kubernetes version (use `kubectl version`):
```
1.15
```
- Docker version (use `docker version`):
```
?
```
|
test
|
go panic in nodecontroller every second please search for existing issues first then read to see what we expect in an issue for security issues please email security rancher com instead of posting a public issue in github you may but are not required to use the gpg key located on keybase what kind of request is this question bug enhancement feature request bug steps to reproduce least amount of steps as possible unknown result seeing this about every second in rancher logs nodecontroller xxxxxxxxxxxxxxxxx recovered from panic runtime error invalid memory address or nil pointer dereference err call stack goroutine github com rancher rancher vendor io apimachinery pkg util runtime recoverfrompanic go src github com rancher rancher vendor io apimachinery pkg util runtime runtime go panic usr local go src runtime panic go github com rancher rancher pkg controllers user monitoring clustermonitoringenabledhandler syncwindowsnode go src github com rancher rancher pkg controllers user monitoring cluster monitoring enabled handler go github com rancher rancher vendor github com rancher types apis core nodecontroller addhandler go src github com rancher rancher vendor github com rancher types apis core zz generated node controller go github com rancher rancher vendor github com rancher norman controller genericcontroller synchandler go src github com rancher rancher vendor github com rancher norman controller generic controller go github com rancher rancher vendor github com rancher norman controller genericcontroller processnextworkitem go src github com rancher rancher vendor github com rancher norman controller generic controller go github com rancher rancher vendor github com rancher norman controller genericcontroller runworker go src github com rancher rancher vendor github com rancher norman controller generic controller go github com rancher rancher vendor io apimachinery pkg util wait jitteruntil go src github com rancher rancher vendor io apimachinery pkg util wait wait go github com rancher rancher vendor io apimachinery pkg util wait jitteruntil go src github com rancher rancher vendor io apimachinery pkg util wait wait go github com rancher rancher vendor io apimachinery pkg util wait until go src github com rancher rancher vendor io apimachinery pkg util wait wait go created by github com rancher rancher vendor github com rancher norman controller genericcontroller run go src github com rancher rancher vendor github com rancher norman controller generic controller go other details that may be helpful could be related to environment information rancher version rancher rancher rancher server image tag or shown bottom left in the ui installation option single install ha ha if the reported issue is regarding a created cluster please provide requested info below cluster information cluster type hosted infrastructure provider custom imported custom machine type cloud vm metal and specifications cpu memory kubernetes version use kubectl version docker version use docker version
| 1
|
49,887
| 6,044,810,449
|
IssuesEvent
|
2017-06-12 07:24:52
|
pixelhumain/co2
|
https://api.github.com/repos/pixelhumain/co2
|
closed
|
Création URL
|
to test
|
Création d'une URL depuis une organisation :
1\ Erreur JS au chargement de la popup

2\ Erreur à la validation du form au saveurl

|
1.0
|
Création URL - Création d'une URL depuis une organisation :
1\ Erreur JS au chargement de la popup

2\ Erreur à la validation du form au saveurl

|
test
|
création url création d une url depuis une organisation erreur js au chargement de la popup erreur à la validation du form au saveurl
| 1
|
104,700
| 8,998,158,023
|
IssuesEvent
|
2019-02-02 19:08:35
|
rancher/rancher
|
https://api.github.com/repos/rancher/rancher
|
closed
|
When GlobalDNS entry is edited and set to different multicluster App , dns entries relating to the older multicluster App instances continue to be programmed in Route53.
|
area/multi-cluster-app kind/bug-qa release/alpha2 status/reopened status/resolved status/to-test team/ca version/2.0
|
Rancher server version - Build from master - jan 21
Setup - HA setup with 4 clusters.
Steps to reproduce the the problem:
1. Create Multicluster App - mcapp1 for cluster1 and cluster2 with Global DNS enabled Ingress for fqdn - "test1.test.com"
2. Create globalDnsProvider with root domain - "test.com"
3. Create a globalDns entry with fqdn - "test1.test.com" pointing to Multicluster App - mcapp1.
4. Create Multicluster App - mcapp2 for cluster3 and cluster4 with Global DNS enabled Ingress for fqdn - "test1.test.com"
5. Edit GlobalDNS entry and set Multicluster App to point to mcapp2
Actual Result:
Accessing test1.test.com should directs traffic to all worker nodes from cluster1, cluster2, cluster3 and cluster4.
Route53 entry gets updated to point to all worker node ips from cluster1, cluster2, cluster3 and cluster4.
Expected Results:
Accessing test1.test.com should direct traffic to all worker nodes from cluster3 and cluster4.
Route53 entry should get updated to point to all worker node ips from cluster3 and cluster4.
|
1.0
|
When GlobalDNS entry is edited and set to different multicluster App , dns entries relating to the older multicluster App instances continue to be programmed in Route53. - Rancher server version - Build from master - jan 21
Setup - HA setup with 4 clusters.
Steps to reproduce the the problem:
1. Create Multicluster App - mcapp1 for cluster1 and cluster2 with Global DNS enabled Ingress for fqdn - "test1.test.com"
2. Create globalDnsProvider with root domain - "test.com"
3. Create a globalDns entry with fqdn - "test1.test.com" pointing to Multicluster App - mcapp1.
4. Create Multicluster App - mcapp2 for cluster3 and cluster4 with Global DNS enabled Ingress for fqdn - "test1.test.com"
5. Edit GlobalDNS entry and set Multicluster App to point to mcapp2
Actual Result:
Accessing test1.test.com should directs traffic to all worker nodes from cluster1, cluster2, cluster3 and cluster4.
Route53 entry gets updated to point to all worker node ips from cluster1, cluster2, cluster3 and cluster4.
Expected Results:
Accessing test1.test.com should direct traffic to all worker nodes from cluster3 and cluster4.
Route53 entry should get updated to point to all worker node ips from cluster3 and cluster4.
|
test
|
when globaldns entry is edited and set to different multicluster app dns entries relating to the older multicluster app instances continue to be programmed in rancher server version build from master jan setup ha setup with clusters steps to reproduce the the problem create multicluster app for and with global dns enabled ingress for fqdn test com create globaldnsprovider with root domain test com create a globaldns entry with fqdn test com pointing to multicluster app create multicluster app for and with global dns enabled ingress for fqdn test com edit globaldns entry and set multicluster app to point to actual result accessing test com should directs traffic to all worker nodes from and entry gets updated to point to all worker node ips from and expected results accessing test com should direct traffic to all worker nodes from and entry should get updated to point to all worker node ips from and
| 1
|
240,587
| 20,049,706,353
|
IssuesEvent
|
2022-02-03 03:53:31
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
opened
|
Failing test: Chrome X-Pack UI Functional Tests Basic License.x-pack/test/functional/apps/transform/creation_saved_search·ts - apps transform basic license transform creation_saved_search batch transform with unique term and sort by time with saved search filter navigates through the wizard and sets all needed fields
|
failed-test
|
A test failed on a tracked branch
```
Error: Transform id input text should be 'fq_2_1643859986892' (got '')
at Assertion.assert (/opt/local-ssd/buildkite/builds/kb-n2-4-8a4b63cf8afa95a0/elastic/kibana-hourly/kibana/node_modules/@kbn/expect/expect.js:100:11)
at Assertion.eql (/opt/local-ssd/buildkite/builds/kb-n2-4-8a4b63cf8afa95a0/elastic/kibana-hourly/kibana/node_modules/@kbn/expect/expect.js:244:8)
at Object.assertTransformIdValue (test/functional/services/transform/wizard.ts:676:36)
at runMicrotasks (<anonymous>)
at processTicksAndRejections (node:internal/process/task_queues:96:5)
at Object.setTransformId (test/functional/services/transform/wizard.ts:686:7)
at Context.<anonymous> (test/functional/apps/transform/creation_saved_search.ts:222:11)
at Object.apply (/opt/local-ssd/buildkite/builds/kb-n2-4-8a4b63cf8afa95a0/elastic/kibana-hourly/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) {
actual: '',
expected: 'fq_2_1643859986892',
showDiff: true
}
```
First failure: [CI Build - main](https://buildkite.com/elastic/kibana-hourly/builds/9122#823e459c-69ef-4792-9de3-49ba3f4a0b5a)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests Basic License.x-pack/test/functional/apps/transform/creation_saved_search·ts","test.name":"apps transform basic license transform creation_saved_search batch transform with unique term and sort by time with saved search filter navigates through the wizard and sets all needed fields","test.failCount":1}} -->
|
1.0
|
Failing test: Chrome X-Pack UI Functional Tests Basic License.x-pack/test/functional/apps/transform/creation_saved_search·ts - apps transform basic license transform creation_saved_search batch transform with unique term and sort by time with saved search filter navigates through the wizard and sets all needed fields - A test failed on a tracked branch
```
Error: Transform id input text should be 'fq_2_1643859986892' (got '')
at Assertion.assert (/opt/local-ssd/buildkite/builds/kb-n2-4-8a4b63cf8afa95a0/elastic/kibana-hourly/kibana/node_modules/@kbn/expect/expect.js:100:11)
at Assertion.eql (/opt/local-ssd/buildkite/builds/kb-n2-4-8a4b63cf8afa95a0/elastic/kibana-hourly/kibana/node_modules/@kbn/expect/expect.js:244:8)
at Object.assertTransformIdValue (test/functional/services/transform/wizard.ts:676:36)
at runMicrotasks (<anonymous>)
at processTicksAndRejections (node:internal/process/task_queues:96:5)
at Object.setTransformId (test/functional/services/transform/wizard.ts:686:7)
at Context.<anonymous> (test/functional/apps/transform/creation_saved_search.ts:222:11)
at Object.apply (/opt/local-ssd/buildkite/builds/kb-n2-4-8a4b63cf8afa95a0/elastic/kibana-hourly/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) {
actual: '',
expected: 'fq_2_1643859986892',
showDiff: true
}
```
First failure: [CI Build - main](https://buildkite.com/elastic/kibana-hourly/builds/9122#823e459c-69ef-4792-9de3-49ba3f4a0b5a)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests Basic License.x-pack/test/functional/apps/transform/creation_saved_search·ts","test.name":"apps transform basic license transform creation_saved_search batch transform with unique term and sort by time with saved search filter navigates through the wizard and sets all needed fields","test.failCount":1}} -->
|
test
|
failing test chrome x pack ui functional tests basic license x pack test functional apps transform creation saved search·ts apps transform basic license transform creation saved search batch transform with unique term and sort by time with saved search filter navigates through the wizard and sets all needed fields a test failed on a tracked branch error transform id input text should be fq got at assertion assert opt local ssd buildkite builds kb elastic kibana hourly kibana node modules kbn expect expect js at assertion eql opt local ssd buildkite builds kb elastic kibana hourly kibana node modules kbn expect expect js at object asserttransformidvalue test functional services transform wizard ts at runmicrotasks at processticksandrejections node internal process task queues at object settransformid test functional services transform wizard ts at context test functional apps transform creation saved search ts at object apply opt local ssd buildkite builds kb elastic kibana hourly kibana node modules kbn test target node functional test runner lib mocha wrap function js actual expected fq showdiff true first failure
| 1
|
10,724
| 27,329,271,376
|
IssuesEvent
|
2023-02-25 12:12:32
|
Arsenic-ATG/8085
|
https://api.github.com/repos/Arsenic-ATG/8085
|
opened
|
Make instruction functions private
|
good first issue Architecture Instruction set
|
currently all the functions of class `emu::cpu` are publicly visible, ideally these functions should be private and only be used by `emu::cpu::execute ()` to execute the proper instruction.
Doing so would make the program much less error prone and avoid any accidental function calls by users.
|
1.0
|
Make instruction functions private - currently all the functions of class `emu::cpu` are publicly visible, ideally these functions should be private and only be used by `emu::cpu::execute ()` to execute the proper instruction.
Doing so would make the program much less error prone and avoid any accidental function calls by users.
|
non_test
|
make instruction functions private currently all the functions of class emu cpu are publicly visible ideally these functions should be private and only be used by emu cpu execute to execute the proper instruction doing so would make the program much less error prone and avoid any accidental function calls by users
| 0
|
105,246
| 9,048,460,383
|
IssuesEvent
|
2019-02-12 00:10:11
|
knative/serving
|
https://api.github.com/repos/knative/serving
|
closed
|
Create PR check to verify API changes reflected in conformance tests
|
area/API area/test-and-release kind/doc kind/feature
|
/area API
/area test-and-release
/kind dev
/kind doc
## Expected Behavior
Whenever changes are made to the API, in either:
* [The spec doc](https://github.com/elafros/elafros/blob/master/docs/spec/spec.md)
* [The API definition](https://github.com/elafros/elafros/tree/master/pkg/apis/ela)
The [conformance tests](https://github.com/elafros/elafros/tree/master/test/conformance) should (probably) be updated as well.
When a PR is submitted that changes either of those areas, but does not change the conformance tests, the submitter should be required to either update the conformance tests or indicate that they conformance tests do not need to be updated.
## Actual Behavior
Changes to the API that do not break the conformance tests can slip by without being reflected in the conformance tests.
## Steps to Reproduce the Problem
n/a
## Additional Info
n/a
|
1.0
|
Create PR check to verify API changes reflected in conformance tests - /area API
/area test-and-release
/kind dev
/kind doc
## Expected Behavior
Whenever changes are made to the API, in either:
* [The spec doc](https://github.com/elafros/elafros/blob/master/docs/spec/spec.md)
* [The API definition](https://github.com/elafros/elafros/tree/master/pkg/apis/ela)
The [conformance tests](https://github.com/elafros/elafros/tree/master/test/conformance) should (probably) be updated as well.
When a PR is submitted that changes either of those areas, but does not change the conformance tests, the submitter should be required to either update the conformance tests or indicate that they conformance tests do not need to be updated.
## Actual Behavior
Changes to the API that do not break the conformance tests can slip by without being reflected in the conformance tests.
## Steps to Reproduce the Problem
n/a
## Additional Info
n/a
|
test
|
create pr check to verify api changes reflected in conformance tests area api area test and release kind dev kind doc expected behavior whenever changes are made to the api in either the should probably be updated as well when a pr is submitted that changes either of those areas but does not change the conformance tests the submitter should be required to either update the conformance tests or indicate that they conformance tests do not need to be updated actual behavior changes to the api that do not break the conformance tests can slip by without being reflected in the conformance tests steps to reproduce the problem n a additional info n a
| 1
|
107,919
| 9,248,944,446
|
IssuesEvent
|
2019-03-15 08:03:44
|
redhat-developer/odo
|
https://api.github.com/repos/redhat-developer/odo
|
closed
|
Get odo tests as part of OpenShift testgrid
|
kind/testing priority/High state/Ready
|
[kind/Feature]
<!--
Welcome! - We kindly ask you to:
1. Fill out the issue template below
2. Use the chat and talk to us if you have a question rather than a bug or feature request.
The chat room is at: https://chat.openshift.io/developers/channels/odo
Thanks for understanding, and for contributing to the project!
-->
## Which functionality do you think we should add?
As odo is an OpenShift tool, it would be nice to get odo tested, as part of OpenShift's own tests. This means that whenever tests run for OpenShift, odo should get validated for it as well.
### Refer Links
Kubernetes Prow: https://github.com/kubernetes/test-infra
Kubernetes Test Grid: https://k8s-testgrid.appspot.com/
Kubernetes Gubernator: https://gubernator.k8s.io/
Kubernetes Test Triage: https://storage.googleapis.com/k8s-gubernator/triage/index.html
OpenShift Gubernator: https://openshift-gce-devel.appspot.com/
Adding Jobs to prow: https://github.com/kubernetes/test-infra/blob/master/prow/jobs.md
## Why is this needed?
This will enable us to:
- Constantly be up-to date with latest OpenShift releases. When odo is part of OpenShift own testing, we will become aware of new versions very quickly, and also of any breakages, early on in the cycle
- We will be able to communicate more directly with the `OpenShift` team about odo.
- It will help us with continuous delivery of odo with OpenShift
- It will get us better visibility with little cost. We can leverage amazing test infrastructure developed for testing Kubernetes and OpenShift, without having to put the effort
## Tasks
- [x] ~~Create a pre-test script that can bring fresh cluster upto speed with basic expectations of test, such as a user currently being logged in, and authorisation being setup for the users being used for the tests~~ Let's start with just building odo and running unit tests. Those won't have any dependency on OpenShift cluster.
- [x] Get a token for accessing OpenShift CI from OpenShift CI Administrators https://mojo.redhat.com/docs/DOC-1177573
- [x] Setup webhook, to hit https://hook-ci.svc.ci.openshift.org/hook with following parameters :
- Content Type : `application/json`
- Secret : `token retrieved from above`
- SSL verification : enabled
- Event types : all
- Active : yes
- [x] Grant the Team OpenShift Robots team write access to the repository. Since we are currently not in OpenShift Organization
- [x] Invite `openshift-ci-robot` and `openshift-merge-robot` to org or add them as collaborators
- [x] Contact one of CI administrators to accept invite, as needed
- [x] Grant robots write access to repository
- [x] Configure prow as per our requirements
- [x] Decide which plugins we want to use from list available https://deck-ci.svc.ci.openshift.org/plugins
- [x] Update [openshift/release](https://github.com/openshift/release) making following changes
- [x] [plugins.yaml](https://github.com/openshift/release/blob/master/cluster/ci/config/prow/plugins.yaml) to enable CI interacts against repository for plugins under `redhat-developer/odo` ie `$ord/$repo`
- [x] [config.yaml](https://github.com/openshift/release/blob/master/cluster/ci/config/prow/config.yaml) to
- [x] Enable tide merge automation on repository
- [x] Setup CI Operator by creating appropriate files in [ci-operator directory](https://github.com/openshift/release/tree/master/ci-operator)
- [x] `ci-operator/config/redhat-developer/odo/redhat-developer-odo-master.yaml` (ci-operator/config/$org/$repo/$org-$repo-$branch.yaml) to
- [x] Define release artifacts
- [x] Configure ci-operator test definitions for repository
- [x] `ci-operator/jobs/redhat-developer/odo/redhat-developer-odo-master-$jobtype.yaml` to to add test entries for Prow for repository [jobtype = presubmit, postsubmit, periodic]
- [x] Add ourselves to `ci-operator/config/redhat-developer/odo/OWNERS` and `ci-operator/jobs/redhat-developer/odo/OWNERS`
|
1.0
|
Get odo tests as part of OpenShift testgrid - [kind/Feature]
<!--
Welcome! - We kindly ask you to:
1. Fill out the issue template below
2. Use the chat and talk to us if you have a question rather than a bug or feature request.
The chat room is at: https://chat.openshift.io/developers/channels/odo
Thanks for understanding, and for contributing to the project!
-->
## Which functionality do you think we should add?
As odo is an OpenShift tool, it would be nice to get odo tested, as part of OpenShift's own tests. This means that whenever tests run for OpenShift, odo should get validated for it as well.
### Refer Links
Kubernetes Prow: https://github.com/kubernetes/test-infra
Kubernetes Test Grid: https://k8s-testgrid.appspot.com/
Kubernetes Gubernator: https://gubernator.k8s.io/
Kubernetes Test Triage: https://storage.googleapis.com/k8s-gubernator/triage/index.html
OpenShift Gubernator: https://openshift-gce-devel.appspot.com/
Adding Jobs to prow: https://github.com/kubernetes/test-infra/blob/master/prow/jobs.md
## Why is this needed?
This will enable us to:
- Constantly be up-to date with latest OpenShift releases. When odo is part of OpenShift own testing, we will become aware of new versions very quickly, and also of any breakages, early on in the cycle
- We will be able to communicate more directly with the `OpenShift` team about odo.
- It will help us with continuous delivery of odo with OpenShift
- It will get us better visibility with little cost. We can leverage amazing test infrastructure developed for testing Kubernetes and OpenShift, without having to put the effort
## Tasks
- [x] ~~Create a pre-test script that can bring fresh cluster upto speed with basic expectations of test, such as a user currently being logged in, and authorisation being setup for the users being used for the tests~~ Let's start with just building odo and running unit tests. Those won't have any dependency on OpenShift cluster.
- [x] Get a token for accessing OpenShift CI from OpenShift CI Administrators https://mojo.redhat.com/docs/DOC-1177573
- [x] Setup webhook, to hit https://hook-ci.svc.ci.openshift.org/hook with following parameters :
- Content Type : `application/json`
- Secret : `token retrieved from above`
- SSL verification : enabled
- Event types : all
- Active : yes
- [x] Grant the Team OpenShift Robots team write access to the repository. Since we are currently not in OpenShift Organization
- [x] Invite `openshift-ci-robot` and `openshift-merge-robot` to org or add them as collaborators
- [x] Contact one of CI administrators to accept invite, as needed
- [x] Grant robots write access to repository
- [x] Configure prow as per our requirements
- [x] Decide which plugins we want to use from list available https://deck-ci.svc.ci.openshift.org/plugins
- [x] Update [openshift/release](https://github.com/openshift/release) making following changes
- [x] [plugins.yaml](https://github.com/openshift/release/blob/master/cluster/ci/config/prow/plugins.yaml) to enable CI interacts against repository for plugins under `redhat-developer/odo` ie `$ord/$repo`
- [x] [config.yaml](https://github.com/openshift/release/blob/master/cluster/ci/config/prow/config.yaml) to
- [x] Enable tide merge automation on repository
- [x] Setup CI Operator by creating appropriate files in [ci-operator directory](https://github.com/openshift/release/tree/master/ci-operator)
- [x] `ci-operator/config/redhat-developer/odo/redhat-developer-odo-master.yaml` (ci-operator/config/$org/$repo/$org-$repo-$branch.yaml) to
- [x] Define release artifacts
- [x] Configure ci-operator test definitions for repository
- [x] `ci-operator/jobs/redhat-developer/odo/redhat-developer-odo-master-$jobtype.yaml` to to add test entries for Prow for repository [jobtype = presubmit, postsubmit, periodic]
- [x] Add ourselves to `ci-operator/config/redhat-developer/odo/OWNERS` and `ci-operator/jobs/redhat-developer/odo/OWNERS`
|
test
|
get odo tests as part of openshift testgrid welcome we kindly ask you to fill out the issue template below use the chat and talk to us if you have a question rather than a bug or feature request the chat room is at thanks for understanding and for contributing to the project which functionality do you think we should add as odo is an openshift tool it would be nice to get odo tested as part of openshift s own tests this means that whenever tests run for openshift odo should get validated for it as well refer links kubernetes prow kubernetes test grid kubernetes gubernator kubernetes test triage openshift gubernator adding jobs to prow why is this needed this will enable us to constantly be up to date with latest openshift releases when odo is part of openshift own testing we will become aware of new versions very quickly and also of any breakages early on in the cycle we will be able to communicate more directly with the openshift team about odo it will help us with continuous delivery of odo with openshift it will get us better visibility with little cost we can leverage amazing test infrastructure developed for testing kubernetes and openshift without having to put the effort tasks create a pre test script that can bring fresh cluster upto speed with basic expectations of test such as a user currently being logged in and authorisation being setup for the users being used for the tests let s start with just building odo and running unit tests those won t have any dependency on openshift cluster get a token for accessing openshift ci from openshift ci administrators setup webhook to hit with following parameters content type application json secret token retrieved from above ssl verification enabled event types all active yes grant the team openshift robots team write access to the repository since we are currently not in openshift organization invite openshift ci robot and openshift merge robot to org or add them as collaborators contact one of ci administrators to accept invite as needed grant robots write access to repository configure prow as per our requirements decide which plugins we want to use from list available update making following changes to enable ci interacts against repository for plugins under redhat developer odo ie ord repo to enable tide merge automation on repository setup ci operator by creating appropriate files in ci operator config redhat developer odo redhat developer odo master yaml ci operator config org repo org repo branch yaml to define release artifacts configure ci operator test definitions for repository ci operator jobs redhat developer odo redhat developer odo master jobtype yaml to to add test entries for prow for repository add ourselves to ci operator config redhat developer odo owners and ci operator jobs redhat developer odo owners
| 1
|
61,558
| 3,147,475,762
|
IssuesEvent
|
2015-09-15 08:22:12
|
handsontable/handsontable
|
https://api.github.com/repos/handsontable/handsontable
|
closed
|
Broken scrolling on the grouping example.
|
Bug Plugin: drag to scroll Priority: normal
|
Scrolling with shift + mouse wheel can cause sheet to go off the screen.
http://docs.handsontable.com/0.16.0/demo-grouping-and-ungrouping.html?_ga=1.181662767.981730759.1437017583
It also works here
http://docs.handsontable.com/0.16.0/demo-scrollbars.html
http://imgur.com/v4GWthl

To Reproduce:
* Click in a cell eg A2
* Hold Shift + Mouse wheel down to scroll right.
* Let go of shift whilst scrolling down still.
* Hold shift and scroll the mouse wheel up (to scroll to the left) it will go out of bounds.
|
1.0
|
Broken scrolling on the grouping example. - Scrolling with shift + mouse wheel can cause sheet to go off the screen.
http://docs.handsontable.com/0.16.0/demo-grouping-and-ungrouping.html?_ga=1.181662767.981730759.1437017583
It also works here
http://docs.handsontable.com/0.16.0/demo-scrollbars.html
http://imgur.com/v4GWthl

To Reproduce:
* Click in a cell eg A2
* Hold Shift + Mouse wheel down to scroll right.
* Let go of shift whilst scrolling down still.
* Hold shift and scroll the mouse wheel up (to scroll to the left) it will go out of bounds.
|
non_test
|
broken scrolling on the grouping example scrolling with shift mouse wheel can cause sheet to go off the screen it also works here to reproduce click in a cell eg hold shift mouse wheel down to scroll right let go of shift whilst scrolling down still hold shift and scroll the mouse wheel up to scroll to the left it will go out of bounds
| 0
|
87,382
| 25,107,004,353
|
IssuesEvent
|
2022-11-08 17:19:58
|
spack/spack
|
https://api.github.com/repos/spack/spack
|
closed
|
Installation issue: Binary libffi fails checksum verification
|
build-error
|
### Steps to reproduce the issue
```
# On develop branch
spack mirror add binary_mirror https://binaries.spack.io/releases/v0.18
spack buildcache keys --install --trust
spack -vvv install /d6d3lh3
```
### Error message
==> Installing libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd
==> Fetching https://binaries.spack.io/releases/v0.18/build_cache/linux-amzn2-x86_64_v4-gcc-7.3.1-libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd.spec.json.sig
gpg: Signature made Tue 31 May 2022 07:27:40 AM UTC using RSA key ID 3DB0C723
gpg: Good signature from "Spack Project Official Binaries <maintainers@spack.io>"
==> Fetching https://binaries.spack.io/releases/v0.18/build_cache/linux-amzn2-x86_64_v4/gcc-7.3.1/libffi-3.4.2/linux-amzn2-x86_64_v4-gcc-7.3.1-libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd.spack
==> Extracting libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd from binary cache
==> Error: Failed to install libffi due to NoChecksumException: Package tarball failed checksum verification.
It cannot be installed.
### Information on your system
[ec2-user@ip-10-7-228-156 ~]$ spack debug report
* **Spack:** 0.19.0.dev0 (b065d69136023693ace8e4d35d9fdad817d63d93)
* **Python:** 3.7.10
* **Platform:** linux-amzn2-skylake_avx512
* **Concretizer:** clingo
### Additional information
_No response_
### General information
- [X] I have run `spack debug report` and reported the version of Spack/Python/Platform
- [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers
- [X] I have uploaded the build log and environment files
- [X] I have searched the issues of this repo and believe this is not a duplicate
|
1.0
|
Installation issue: Binary libffi fails checksum verification - ### Steps to reproduce the issue
```
# On develop branch
spack mirror add binary_mirror https://binaries.spack.io/releases/v0.18
spack buildcache keys --install --trust
spack -vvv install /d6d3lh3
```
### Error message
==> Installing libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd
==> Fetching https://binaries.spack.io/releases/v0.18/build_cache/linux-amzn2-x86_64_v4-gcc-7.3.1-libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd.spec.json.sig
gpg: Signature made Tue 31 May 2022 07:27:40 AM UTC using RSA key ID 3DB0C723
gpg: Good signature from "Spack Project Official Binaries <maintainers@spack.io>"
==> Fetching https://binaries.spack.io/releases/v0.18/build_cache/linux-amzn2-x86_64_v4/gcc-7.3.1/libffi-3.4.2/linux-amzn2-x86_64_v4-gcc-7.3.1-libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd.spack
==> Extracting libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd from binary cache
==> Error: Failed to install libffi due to NoChecksumException: Package tarball failed checksum verification.
It cannot be installed.
### Information on your system
[ec2-user@ip-10-7-228-156 ~]$ spack debug report
* **Spack:** 0.19.0.dev0 (b065d69136023693ace8e4d35d9fdad817d63d93)
* **Python:** 3.7.10
* **Platform:** linux-amzn2-skylake_avx512
* **Concretizer:** clingo
### Additional information
_No response_
### General information
- [X] I have run `spack debug report` and reported the version of Spack/Python/Platform
- [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers
- [X] I have uploaded the build log and environment files
- [X] I have searched the issues of this repo and believe this is not a duplicate
|
non_test
|
installation issue binary libffi fails checksum verification steps to reproduce the issue on develop branch spack mirror add binary mirror spack buildcache keys install trust spack vvv install error message installing libffi fetching gpg signature made tue may am utc using rsa key id gpg good signature from spack project official binaries fetching extracting libffi from binary cache error failed to install libffi due to nochecksumexception package tarball failed checksum verification it cannot be installed information on your system spack debug report spack python platform linux skylake concretizer clingo additional information no response general information i have run spack debug report and reported the version of spack python platform i have run spack maintainers and mentioned any maintainers i have uploaded the build log and environment files i have searched the issues of this repo and believe this is not a duplicate
| 0
|
363,324
| 25,444,883,865
|
IssuesEvent
|
2022-11-24 04:29:14
|
cdisc-org/conformance-rules-editor
|
https://api.github.com/repos/cdisc-org/conformance-rules-editor
|
closed
|
As a workshop participant, I want sample rules
|
documentation
|
- Rules have been gathered that highlight different scenarios that would be helpful for training.
|
1.0
|
As a workshop participant, I want sample rules - - Rules have been gathered that highlight different scenarios that would be helpful for training.
|
non_test
|
as a workshop participant i want sample rules rules have been gathered that highlight different scenarios that would be helpful for training
| 0
|
677,288
| 23,157,346,824
|
IssuesEvent
|
2022-07-29 14:12:37
|
openghg/openghg
|
https://api.github.com/repos/openghg/openghg
|
closed
|
New xarray and mypy issues
|
high-priority
|
The latest version of xarray (2022.6.0) has some new type hints included. This results in quite a few mypy issues that will need looking at. I'm setting this as high priority as I don't think we want to start pinning versions of commonly used libraries like xarray.
|
1.0
|
New xarray and mypy issues - The latest version of xarray (2022.6.0) has some new type hints included. This results in quite a few mypy issues that will need looking at. I'm setting this as high priority as I don't think we want to start pinning versions of commonly used libraries like xarray.
|
non_test
|
new xarray and mypy issues the latest version of xarray has some new type hints included this results in quite a few mypy issues that will need looking at i m setting this as high priority as i don t think we want to start pinning versions of commonly used libraries like xarray
| 0
|
103,291
| 4,166,283,881
|
IssuesEvent
|
2016-06-20 01:40:07
|
nvs/gem
|
https://api.github.com/repos/nvs/gem
|
opened
|
Introduce a pause before 'starting'
|
Area: JASS Priority: Later Status: Not Started Type: Enhancement
|
This would help some slower computers 'settle', as well as giving players an indication of the game actually starting. This issue is most noticeable when hosted via HCL, and would help with #79.
|
1.0
|
Introduce a pause before 'starting' - This would help some slower computers 'settle', as well as giving players an indication of the game actually starting. This issue is most noticeable when hosted via HCL, and would help with #79.
|
non_test
|
introduce a pause before starting this would help some slower computers settle as well as giving players an indication of the game actually starting this issue is most noticeable when hosted via hcl and would help with
| 0
|
19,170
| 5,814,941,387
|
IssuesEvent
|
2017-05-05 06:42:59
|
joomla/joomla-cms
|
https://api.github.com/repos/joomla/joomla-cms
|
closed
|
Massive loading times after update to 3.7?
|
No Code Attached Yet
|
Hey there,
i updated two pages now to joomla 3.7 and notices that both of them got a high loading time up to 8 seconds to get the page loaded. Does anyone else have that problem?
|
1.0
|
Massive loading times after update to 3.7? - Hey there,
i updated two pages now to joomla 3.7 and notices that both of them got a high loading time up to 8 seconds to get the page loaded. Does anyone else have that problem?
|
non_test
|
massive loading times after update to hey there i updated two pages now to joomla and notices that both of them got a high loading time up to seconds to get the page loaded does anyone else have that problem
| 0
|
308,389
| 26,603,901,914
|
IssuesEvent
|
2023-01-23 17:43:39
|
kedacore/keda
|
https://api.github.com/repos/kedacore/keda
|
opened
|
Add e2e test for Openstack Metrics Scaler
|
help wanted good first issue testing
|
### Proposal
https://github.com/kedacore/keda/tree/main/tests
### Use-Case
_No response_
### Is this a feature you are interested in implementing yourself?
No
### Anything else?
_No response_
|
1.0
|
Add e2e test for Openstack Metrics Scaler - ### Proposal
https://github.com/kedacore/keda/tree/main/tests
### Use-Case
_No response_
### Is this a feature you are interested in implementing yourself?
No
### Anything else?
_No response_
|
test
|
add test for openstack metrics scaler proposal use case no response is this a feature you are interested in implementing yourself no anything else no response
| 1
|
233,117
| 18,949,430,268
|
IssuesEvent
|
2021-11-18 13:48:00
|
BlueCodeSystems/who-anc
|
https://api.github.com/repos/BlueCodeSystems/who-anc
|
closed
|
Use lower case for first letters for the words "Test" and "Screening" for consistency.
|
Tests v1.0.0-beta.4
|

|
1.0
|
Use lower case for first letters for the words "Test" and "Screening" for consistency. -

|
test
|
use lower case for first letters for the words test and screening for consistency
| 1
|
209,490
| 16,024,515,182
|
IssuesEvent
|
2021-04-21 07:21:07
|
CarlosRayon/symfony
|
https://api.github.com/repos/CarlosRayon/symfony
|
opened
|
Metodologias
|
Testing General
|
- Seguir misma estructura directorios que el proyecto bajo un directorio test.
- Hacer una clase de test por cada clase del proyecto principal
- Nombre de la clase sera el mismo que la clase a testear añadiendo la palabra _Test_ en camelCase
(ejemploController.php -> ejemploControllerTest.php).
- La funciones de test empezaran por _test_ (testEjemplo()) y deben ser publicas
- Es buena practica que los datos de test que se recarguen en cada llamada global a los test.
- Podemos ejecutar todos los test de directorio test con `bin/phpunit` o especificar un directorio o un test en concreto `bin/phpunit <ruta-directorio | ruta-fichero>`
- Disponemos de multiples opciones al ejecutar lo test que podemos ver en la [doc de phpunit](https://phpunit.readthedocs.io/en/9.5/textui.html#command-line-options)
|
1.0
|
Metodologias - - Seguir misma estructura directorios que el proyecto bajo un directorio test.
- Hacer una clase de test por cada clase del proyecto principal
- Nombre de la clase sera el mismo que la clase a testear añadiendo la palabra _Test_ en camelCase
(ejemploController.php -> ejemploControllerTest.php).
- La funciones de test empezaran por _test_ (testEjemplo()) y deben ser publicas
- Es buena practica que los datos de test que se recarguen en cada llamada global a los test.
- Podemos ejecutar todos los test de directorio test con `bin/phpunit` o especificar un directorio o un test en concreto `bin/phpunit <ruta-directorio | ruta-fichero>`
- Disponemos de multiples opciones al ejecutar lo test que podemos ver en la [doc de phpunit](https://phpunit.readthedocs.io/en/9.5/textui.html#command-line-options)
|
test
|
metodologias seguir misma estructura directorios que el proyecto bajo un directorio test hacer una clase de test por cada clase del proyecto principal nombre de la clase sera el mismo que la clase a testear añadiendo la palabra test en camelcase ejemplocontroller php ejemplocontrollertest php la funciones de test empezaran por test testejemplo y deben ser publicas es buena practica que los datos de test que se recarguen en cada llamada global a los test podemos ejecutar todos los test de directorio test con bin phpunit o especificar un directorio o un test en concreto bin phpunit disponemos de multiples opciones al ejecutar lo test que podemos ver en la
| 1
|
110,100
| 9,430,499,397
|
IssuesEvent
|
2019-04-12 09:10:45
|
goharbor/harbor
|
https://api.github.com/repos/goharbor/harbor
|
closed
|
Should add retry in keyword <Add Labels To Tag>
|
area/ci area/test automation
|
In keyword <Add Labels To Tag>, one of steps is to click repository name, there should be retry when attempting to go into repository.
https://jenkins11.svc.eng.vmware.com/job/harbor_nightly_result_publisher/2095/robot/report/log.html

|
1.0
|
Should add retry in keyword <Add Labels To Tag> - In keyword <Add Labels To Tag>, one of steps is to click repository name, there should be retry when attempting to go into repository.
https://jenkins11.svc.eng.vmware.com/job/harbor_nightly_result_publisher/2095/robot/report/log.html

|
test
|
should add retry in keyword in keyword one of steps is to click repository name there should be retry when attempting to go into repository
| 1
|
581,505
| 17,295,185,936
|
IssuesEvent
|
2021-07-25 15:26:04
|
DFO-Ocean-Navigator/Ocean-Data-Map-Project
|
https://api.github.com/repos/DFO-Ocean-Navigator/Ocean-Data-Map-Project
|
closed
|
Issue with building the React JS modules for mainline
|
Bug Javascript Priority: High dependencies
|
```
(navigator) vagrant@ubuntu-focal:~/Ocean-Data-Map-Project/oceannavigator/frontend$ yarn install
yarn install v1.22.10
[1/4] Resolving packages...
[2/4] Fetching packages...
info fsevents@2.3.2: The platform "linux" is incompatible with this module.
info "fsevents@2.3.2" is an optional dependency and failed compatibility check. Excluding it from installation.
info fsevents@1.2.13: The platform "linux" is incompatible with this module.
info "fsevents@1.2.13" is an optional dependency and failed compatibility check. Excluding it from installation.
[3/4] Linking dependencies...
[4/4] Building fresh packages...
success Saved lockfile.
Done in 28.12s.
(navigator) vagrant@ubuntu-focal:~/Ocean-Data-Map-Project/oceannavigator/frontend$ yarn build
yarn run v1.22.10
$ NODE_ENV=production node --stack_size=1000000 node_modules/.bin/webpack --progress -p
Hash: 83eb17fbe8f47c58a8d4
Version: webpack 3.6.0
Time: 21890ms
Asset Size Chunks Chunk Names
/glyphicons-halflings-regular.eot 20.1 kB [emitted]
/glyphicons-halflings-regular.woff2 18 kB [emitted]
/glyphicons-halflings-regular.woff 23.4 kB [emitted]
/glyphicons-halflings-regular.ttf 45.4 kB [emitted]
/glyphicons-halflings-regular.svg 109 kB [emitted]
/ui-icons_444444_256x240.png 3.76 kB [emitted]
/ui-icons_555555_256x240.png 3.76 kB [emitted]
/ui-icons_ffffff_256x240.png 3.76 kB [emitted]
/ui-icons_777620_256x240.png 3.76 kB [emitted]
/ui-icons_cc0000_256x240.png 3.76 kB [emitted]
/ui-icons_777777_256x240.png 3.76 kB [emitted]
/fontawesome-webfont.eot 166 kB [emitted]
/fontawesome-webfont.woff2 77.2 kB [emitted]
/fontawesome-webfont.woff 98 kB [emitted]
/fontawesome-webfont.ttf 166 kB [emitted]
/fontawesome-webfont.svg 444 kB [emitted] [big]
/spinner.gif 41.3 kB [emitted]
/x.png 153 bytes [emitted]
/I0.svg 131 bytes [emitted]
/I1.svg 747 bytes [emitted]
/I2.svg 747 bytes [emitted]
/I3.svg 747 bytes [emitted]
/I4.svg 745 bytes [emitted]
/I5.svg 747 bytes [emitted]
/I6.svg 747 bytes [emitted]
/I7.svg 747 bytes [emitted]
/I8.svg 747 bytes [emitted]
/I9.svg 745 bytes [emitted]
/sad-computer.png 2.1 kB [emitted]
/bar_loader.gif 10.8 kB [emitted]
oceannavigator.js 7.57 MB 0 [emitted] [big] main
oceannavigator.css 343 kB 0 [emitted] [big] main
index.html 1.87 kB [emitted]
[20] ./src/i18n.js 828 bytes {0} [built]
[75] (webpack)/buildin/global.js 509 bytes {0} [built]
[626] ./src/currentLanguage.js 93 bytes {0} [built]
[640] multi babel-polyfill ./src/index.jsx 40 bytes {0} [built]
[843] ./src/index.jsx 2.37 kB {0} [built]
[1052] ../translations/fr.json 15.6 kB {0} [built]
[1053] ../translations/en.json 3.29 kB {0} [built]
[1055] ./src/images/x.png 52 bytes {0} [built]
[1056] ./src/images/s111/I0.svg 53 bytes {0} [built]
[1438] ./src/remote/OceanNavigator.js 633 bytes {0} [built]
[1457] ./src/images/bar_loader.gif 61 bytes {0} [built]
[1463] ./src/stylesheets/utils/bootstrap.css 41 bytes {0} [built]
[1464] ./src/stylesheets/main.scss 41 bytes {0} [built]
[1512] ./node_modules/css-loader!./src/stylesheets/utils/bootstrap.css 120 kB [built]
[1513] ./node_modules/css-loader!./node_modules/sass-loader/lib/loader.js!./src/stylesheets/main.scss 13.6 kB [built]
+ 1499 hidden modules
ERROR in oceannavigator.js from UglifyJs
Unexpected token: keyword (const) [oceannavigator.js:60314,9]
Child html-webpack-plugin for "index.html":
1 asset
[0] ./node_modules/html-webpack-plugin/lib/loader.js!./src/index.ejs 2.14 kB {0} [built]
[2] (webpack)/buildin/global.js 509 bytes {0} [built]
[3] (webpack)/buildin/module.js 517 bytes {0} [built]
+ 1 hidden module
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/ol/ol.css:
2 modules
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/jquery-ui-month-picker/src/MonthPicker.css:
2 modules
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/rc-slider/assets/index.css:
2 modules
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/react-datepicker/dist/react-datepicker.css:
2 modules
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/react-bootstrap-toggle/dist/bootstrap2-toggle.css:
2 modules
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/jquery-ui/themes/base/datepicker.css:
2 modules
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!src/stylesheets/utils/bootstrap.css:
[0] ./node_modules/css-loader!./src/stylesheets/utils/bootstrap.css 120 kB {0} [built]
+ 1 hidden module
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/bootstrap/dist/css/bootstrap.css:
5 assets
7 modules
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/jquery-ui/themes/base/theme.css:
6 assets
8 modules
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/jquery-ui/themes/base/base.css:
21 modules
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/sass-loader/lib/loader.js!src/stylesheets/main.scss:
[0] ./node_modules/css-loader!./node_modules/sass-loader/lib/loader.js!./src/stylesheets/main.scss 13.6 kB {0} [built]
+ 1 hidden module
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/sass-loader/lib/loader.js!node_modules/font-awesome/scss/font-awesome.scss:
5 assets
8 modules
error Command failed with exit code 2.
info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command.
```
|
1.0
|
Issue with building the React JS modules for mainline - ```
(navigator) vagrant@ubuntu-focal:~/Ocean-Data-Map-Project/oceannavigator/frontend$ yarn install
yarn install v1.22.10
[1/4] Resolving packages...
[2/4] Fetching packages...
info fsevents@2.3.2: The platform "linux" is incompatible with this module.
info "fsevents@2.3.2" is an optional dependency and failed compatibility check. Excluding it from installation.
info fsevents@1.2.13: The platform "linux" is incompatible with this module.
info "fsevents@1.2.13" is an optional dependency and failed compatibility check. Excluding it from installation.
[3/4] Linking dependencies...
[4/4] Building fresh packages...
success Saved lockfile.
Done in 28.12s.
(navigator) vagrant@ubuntu-focal:~/Ocean-Data-Map-Project/oceannavigator/frontend$ yarn build
yarn run v1.22.10
$ NODE_ENV=production node --stack_size=1000000 node_modules/.bin/webpack --progress -p
Hash: 83eb17fbe8f47c58a8d4
Version: webpack 3.6.0
Time: 21890ms
Asset Size Chunks Chunk Names
/glyphicons-halflings-regular.eot 20.1 kB [emitted]
/glyphicons-halflings-regular.woff2 18 kB [emitted]
/glyphicons-halflings-regular.woff 23.4 kB [emitted]
/glyphicons-halflings-regular.ttf 45.4 kB [emitted]
/glyphicons-halflings-regular.svg 109 kB [emitted]
/ui-icons_444444_256x240.png 3.76 kB [emitted]
/ui-icons_555555_256x240.png 3.76 kB [emitted]
/ui-icons_ffffff_256x240.png 3.76 kB [emitted]
/ui-icons_777620_256x240.png 3.76 kB [emitted]
/ui-icons_cc0000_256x240.png 3.76 kB [emitted]
/ui-icons_777777_256x240.png 3.76 kB [emitted]
/fontawesome-webfont.eot 166 kB [emitted]
/fontawesome-webfont.woff2 77.2 kB [emitted]
/fontawesome-webfont.woff 98 kB [emitted]
/fontawesome-webfont.ttf 166 kB [emitted]
/fontawesome-webfont.svg 444 kB [emitted] [big]
/spinner.gif 41.3 kB [emitted]
/x.png 153 bytes [emitted]
/I0.svg 131 bytes [emitted]
/I1.svg 747 bytes [emitted]
/I2.svg 747 bytes [emitted]
/I3.svg 747 bytes [emitted]
/I4.svg 745 bytes [emitted]
/I5.svg 747 bytes [emitted]
/I6.svg 747 bytes [emitted]
/I7.svg 747 bytes [emitted]
/I8.svg 747 bytes [emitted]
/I9.svg 745 bytes [emitted]
/sad-computer.png 2.1 kB [emitted]
/bar_loader.gif 10.8 kB [emitted]
oceannavigator.js 7.57 MB 0 [emitted] [big] main
oceannavigator.css 343 kB 0 [emitted] [big] main
index.html 1.87 kB [emitted]
[20] ./src/i18n.js 828 bytes {0} [built]
[75] (webpack)/buildin/global.js 509 bytes {0} [built]
[626] ./src/currentLanguage.js 93 bytes {0} [built]
[640] multi babel-polyfill ./src/index.jsx 40 bytes {0} [built]
[843] ./src/index.jsx 2.37 kB {0} [built]
[1052] ../translations/fr.json 15.6 kB {0} [built]
[1053] ../translations/en.json 3.29 kB {0} [built]
[1055] ./src/images/x.png 52 bytes {0} [built]
[1056] ./src/images/s111/I0.svg 53 bytes {0} [built]
[1438] ./src/remote/OceanNavigator.js 633 bytes {0} [built]
[1457] ./src/images/bar_loader.gif 61 bytes {0} [built]
[1463] ./src/stylesheets/utils/bootstrap.css 41 bytes {0} [built]
[1464] ./src/stylesheets/main.scss 41 bytes {0} [built]
[1512] ./node_modules/css-loader!./src/stylesheets/utils/bootstrap.css 120 kB [built]
[1513] ./node_modules/css-loader!./node_modules/sass-loader/lib/loader.js!./src/stylesheets/main.scss 13.6 kB [built]
+ 1499 hidden modules
ERROR in oceannavigator.js from UglifyJs
Unexpected token: keyword (const) [oceannavigator.js:60314,9]
Child html-webpack-plugin for "index.html":
1 asset
[0] ./node_modules/html-webpack-plugin/lib/loader.js!./src/index.ejs 2.14 kB {0} [built]
[2] (webpack)/buildin/global.js 509 bytes {0} [built]
[3] (webpack)/buildin/module.js 517 bytes {0} [built]
+ 1 hidden module
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/ol/ol.css:
2 modules
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/jquery-ui-month-picker/src/MonthPicker.css:
2 modules
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/rc-slider/assets/index.css:
2 modules
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/react-datepicker/dist/react-datepicker.css:
2 modules
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/react-bootstrap-toggle/dist/bootstrap2-toggle.css:
2 modules
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/jquery-ui/themes/base/datepicker.css:
2 modules
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!src/stylesheets/utils/bootstrap.css:
[0] ./node_modules/css-loader!./src/stylesheets/utils/bootstrap.css 120 kB {0} [built]
+ 1 hidden module
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/bootstrap/dist/css/bootstrap.css:
5 assets
7 modules
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/jquery-ui/themes/base/theme.css:
6 assets
8 modules
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/jquery-ui/themes/base/base.css:
21 modules
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/sass-loader/lib/loader.js!src/stylesheets/main.scss:
[0] ./node_modules/css-loader!./node_modules/sass-loader/lib/loader.js!./src/stylesheets/main.scss 13.6 kB {0} [built]
+ 1 hidden module
Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/sass-loader/lib/loader.js!node_modules/font-awesome/scss/font-awesome.scss:
5 assets
8 modules
error Command failed with exit code 2.
info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command.
```
|
non_test
|
issue with building the react js modules for mainline navigator vagrant ubuntu focal ocean data map project oceannavigator frontend yarn install yarn install resolving packages fetching packages info fsevents the platform linux is incompatible with this module info fsevents is an optional dependency and failed compatibility check excluding it from installation info fsevents the platform linux is incompatible with this module info fsevents is an optional dependency and failed compatibility check excluding it from installation linking dependencies building fresh packages success saved lockfile done in navigator vagrant ubuntu focal ocean data map project oceannavigator frontend yarn build yarn run node env production node stack size node modules bin webpack progress p hash version webpack time asset size chunks chunk names glyphicons halflings regular eot kb glyphicons halflings regular kb glyphicons halflings regular woff kb glyphicons halflings regular ttf kb glyphicons halflings regular svg kb ui icons png kb ui icons png kb ui icons ffffff png kb ui icons png kb ui icons png kb ui icons png kb fontawesome webfont eot kb fontawesome webfont kb fontawesome webfont woff kb fontawesome webfont ttf kb fontawesome webfont svg kb spinner gif kb x png bytes svg bytes svg bytes svg bytes svg bytes svg bytes svg bytes svg bytes svg bytes svg bytes svg bytes sad computer png kb bar loader gif kb oceannavigator js mb main oceannavigator css kb main index html kb src js bytes webpack buildin global js bytes src currentlanguage js bytes multi babel polyfill src index jsx bytes src index jsx kb translations fr json kb translations en json kb src images x png bytes src images svg bytes src remote oceannavigator js bytes src images bar loader gif bytes src stylesheets utils bootstrap css bytes src stylesheets main scss bytes node modules css loader src stylesheets utils bootstrap css kb node modules css loader node modules sass loader lib loader js src stylesheets main scss kb hidden modules error in oceannavigator js from uglifyjs unexpected token keyword const child html webpack plugin for index html asset node modules html webpack plugin lib loader js src index ejs kb webpack buildin global js bytes webpack buildin module js bytes hidden module child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules ol ol css modules child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules jquery ui month picker src monthpicker css modules child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules rc slider assets index css modules child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules react datepicker dist react datepicker css modules child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules react bootstrap toggle dist toggle css modules child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules jquery ui themes base datepicker css modules child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js src stylesheets utils bootstrap css node modules css loader src stylesheets utils bootstrap css kb hidden module child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules bootstrap dist css bootstrap css assets modules child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules jquery ui themes base theme css assets modules child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules jquery ui themes base base css modules child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules sass loader lib loader js src stylesheets main scss node modules css loader node modules sass loader lib loader js src stylesheets main scss kb hidden module child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules sass loader lib loader js node modules font awesome scss font awesome scss assets modules error command failed with exit code info visit for documentation about this command
| 0
|
53,962
| 6,353,950,271
|
IssuesEvent
|
2017-07-29 04:05:46
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Investigate flaky sequential/test-fs-readfile-tostring-fail.js on smartos16-64
|
fs smartos test
|
* **Version**: master
* **Platform**: smartos
* **Subsystem**: fs
Example: https://ci.nodejs.org/job/node-test-commit-smartos/7481/nodes=smartos16-64/console
```
not ok 1385 sequential/test-fs-readfile-tostring-fail
---
duration_ms: 0.729
severity: fail
stack: |-
/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos16-64/test/sequential/test-fs-readfile-tostring-fail.js:59
throw err;
^
RangeError: Array buffer allocation failed
at new ArrayBuffer (<anonymous>)
at createUnsafeArrayBuffer (buffer.js:66:12)
at createUnsafeBuffer (buffer.js:60:25)
at Function.Buffer.allocUnsafeSlow (buffer.js:188:10)
at FSReqWrap.readFileAfterStat [as oncomplete] (fs.js:447:27)
```
|
1.0
|
Investigate flaky sequential/test-fs-readfile-tostring-fail.js on smartos16-64 - * **Version**: master
* **Platform**: smartos
* **Subsystem**: fs
Example: https://ci.nodejs.org/job/node-test-commit-smartos/7481/nodes=smartos16-64/console
```
not ok 1385 sequential/test-fs-readfile-tostring-fail
---
duration_ms: 0.729
severity: fail
stack: |-
/home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos16-64/test/sequential/test-fs-readfile-tostring-fail.js:59
throw err;
^
RangeError: Array buffer allocation failed
at new ArrayBuffer (<anonymous>)
at createUnsafeArrayBuffer (buffer.js:66:12)
at createUnsafeBuffer (buffer.js:60:25)
at Function.Buffer.allocUnsafeSlow (buffer.js:188:10)
at FSReqWrap.readFileAfterStat [as oncomplete] (fs.js:447:27)
```
|
test
|
investigate flaky sequential test fs readfile tostring fail js on version master platform smartos subsystem fs example not ok sequential test fs readfile tostring fail duration ms severity fail stack home iojs build workspace node test commit smartos nodes test sequential test fs readfile tostring fail js throw err rangeerror array buffer allocation failed at new arraybuffer at createunsafearraybuffer buffer js at createunsafebuffer buffer js at function buffer allocunsafeslow buffer js at fsreqwrap readfileafterstat fs js
| 1
|
7,560
| 18,245,610,188
|
IssuesEvent
|
2021-10-01 17:58:13
|
arduino/arduino-cli
|
https://api.github.com/repos/arduino/arduino-cli
|
closed
|
Crash after execute "fatal error: unexpected signal during runtime execution" (macOS 11.6)
|
os: macos architecture: arm64 conclusion: resolved type: imperfection
|
## Bug Report
### Current behavior
`arduino-cli` (tag 0.19.1) built with go 1.17.1. Crashes immediately after executing a binary file:
Crash after execution binary (builded with go 1.17.1):
```console
$ arduino-cli
fatal error: unexpected signal during runtime execution
[signal SIGSEGV: segmentation violation code=0x1 addr=0xb01dfacedebac1e pc=0x7fff204ecc9e]
...
```
[arduino-cli-0.19.1-fatal error.txt](https://github.com/arduino/arduino-cli/files/7259691/arduino-cli-0.19.1-fatal.error.txt)
### Environment
- CLI version (~~output of arduino-cli version~~): 0.19.1
- OS and platform: macOS 11.6 x86_64
|
1.0
|
Crash after execute "fatal error: unexpected signal during runtime execution" (macOS 11.6) - ## Bug Report
### Current behavior
`arduino-cli` (tag 0.19.1) built with go 1.17.1. Crashes immediately after executing a binary file:
Crash after execution binary (builded with go 1.17.1):
```console
$ arduino-cli
fatal error: unexpected signal during runtime execution
[signal SIGSEGV: segmentation violation code=0x1 addr=0xb01dfacedebac1e pc=0x7fff204ecc9e]
...
```
[arduino-cli-0.19.1-fatal error.txt](https://github.com/arduino/arduino-cli/files/7259691/arduino-cli-0.19.1-fatal.error.txt)
### Environment
- CLI version (~~output of arduino-cli version~~): 0.19.1
- OS and platform: macOS 11.6 x86_64
|
non_test
|
crash after execute fatal error unexpected signal during runtime execution macos bug report current behavior arduino cli tag built with go crashes immediately after executing a binary file crash after execution binary builded with go console arduino cli fatal error unexpected signal during runtime execution environment cli version output of arduino cli version os and platform macos
| 0
|
268,822
| 23,396,737,046
|
IssuesEvent
|
2022-08-12 01:03:12
|
nim-lang/Nim
|
https://api.github.com/repos/nim-lang/Nim
|
closed
|
Bug with effect system and forward declarations
|
Effect system works_but_needs_test_case
|
```nim
type
SafeFn = proc (): void {. raises: [] }
proc ok() {. raises: [] .} = discard
proc fail() {. raises: [] .}
let f1 : SafeFn = ok
let f2 : SafeFn = fail # Error: type mismatch: got (proc ()) but expected 'SafeFn = proc (){.closure.}'
# .raise effect is 'can raise any'
proc fail() = discard
```
If `fail` is implemented before it's used, it works.
|
1.0
|
Bug with effect system and forward declarations - ```nim
type
SafeFn = proc (): void {. raises: [] }
proc ok() {. raises: [] .} = discard
proc fail() {. raises: [] .}
let f1 : SafeFn = ok
let f2 : SafeFn = fail # Error: type mismatch: got (proc ()) but expected 'SafeFn = proc (){.closure.}'
# .raise effect is 'can raise any'
proc fail() = discard
```
If `fail` is implemented before it's used, it works.
|
test
|
bug with effect system and forward declarations nim type safefn proc void raises proc ok raises discard proc fail raises let safefn ok let safefn fail error type mismatch got proc but expected safefn proc closure raise effect is can raise any proc fail discard if fail is implemented before it s used it works
| 1
|
183,649
| 21,775,132,577
|
IssuesEvent
|
2022-05-13 13:07:46
|
ssobue/redis-demo
|
https://api.github.com/repos/ssobue/redis-demo
|
closed
|
CVE-2021-42550 (Medium) detected in logback-classic-1.2.3.jar
|
security vulnerability
|
## CVE-2021-42550 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>logback-classic-1.2.3.jar</b></p></summary>
<p>logback-classic module</p>
<p>Library home page: <a href="http://logback.qos.ch">http://logback.qos.ch</a></p>
<p>Path to dependency file: redis-demo/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/ch/qos/logback/logback-classic/1.2.3/logback-classic-1.2.3.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.5.0.jar (Root Library)
- spring-boot-starter-2.5.0.jar
- spring-boot-starter-logging-2.5.0.jar
- :x: **logback-classic-1.2.3.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In logback version 1.2.7 and prior versions, an attacker with the required privileges to edit configurations files could craft a malicious configuration allowing to execute arbitrary code loaded from LDAP servers.
<p>Publish Date: 2021-12-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-42550>CVE-2021-42550</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://logback.qos.ch/news.html">http://logback.qos.ch/news.html</a></p>
<p>Release Date: 2021-12-16</p>
<p>Fix Resolution: ch.qos.logback:logback-classic:1.2.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-42550 (Medium) detected in logback-classic-1.2.3.jar - ## CVE-2021-42550 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>logback-classic-1.2.3.jar</b></p></summary>
<p>logback-classic module</p>
<p>Library home page: <a href="http://logback.qos.ch">http://logback.qos.ch</a></p>
<p>Path to dependency file: redis-demo/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/ch/qos/logback/logback-classic/1.2.3/logback-classic-1.2.3.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.5.0.jar (Root Library)
- spring-boot-starter-2.5.0.jar
- spring-boot-starter-logging-2.5.0.jar
- :x: **logback-classic-1.2.3.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In logback version 1.2.7 and prior versions, an attacker with the required privileges to edit configurations files could craft a malicious configuration allowing to execute arbitrary code loaded from LDAP servers.
<p>Publish Date: 2021-12-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-42550>CVE-2021-42550</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://logback.qos.ch/news.html">http://logback.qos.ch/news.html</a></p>
<p>Release Date: 2021-12-16</p>
<p>Fix Resolution: ch.qos.logback:logback-classic:1.2.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in logback classic jar cve medium severity vulnerability vulnerable library logback classic jar logback classic module library home page a href path to dependency file redis demo pom xml path to vulnerable library home wss scanner repository ch qos logback logback classic logback classic jar dependency hierarchy spring boot starter web jar root library spring boot starter jar spring boot starter logging jar x logback classic jar vulnerable library found in base branch master vulnerability details in logback version and prior versions an attacker with the required privileges to edit configurations files could craft a malicious configuration allowing to execute arbitrary code loaded from ldap servers publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ch qos logback logback classic step up your open source security game with whitesource
| 0
|
120,214
| 10,109,722,207
|
IssuesEvent
|
2019-07-30 08:41:03
|
khartec/waltz
|
https://api.github.com/repos/khartec/waltz
|
closed
|
Auth Source fails on Person page
|
bug fixed (test & close)
|
```
ServiceBroker::loadData - AuthSourcesStore.findAuthSources: Internal Server Error - Exception: Cannot make generic selector for kind: PERSON /
```
|
1.0
|
Auth Source fails on Person page - ```
ServiceBroker::loadData - AuthSourcesStore.findAuthSources: Internal Server Error - Exception: Cannot make generic selector for kind: PERSON /
```
|
test
|
auth source fails on person page servicebroker loaddata authsourcesstore findauthsources internal server error exception cannot make generic selector for kind person
| 1
|
344,589
| 30,751,786,665
|
IssuesEvent
|
2023-07-28 20:01:06
|
saltstack/salt
|
https://api.github.com/repos/saltstack/salt
|
opened
|
[Increase Test Coverage] Batch 16
|
Tests
|
Increase the code coverage percent on the following files to at least 80%.
Please be aware that currently the percentage might be inaccurate if the module uses salt due to #64696
File | Percent
salt/modules/junos.py 67
salt/modules/postgres.py 76
salt/_compat.py 64
salt/cloud/clouds/gce.py 18
salt/_logging/impl.py 70
|
1.0
|
[Increase Test Coverage] Batch 16 - Increase the code coverage percent on the following files to at least 80%.
Please be aware that currently the percentage might be inaccurate if the module uses salt due to #64696
File | Percent
salt/modules/junos.py 67
salt/modules/postgres.py 76
salt/_compat.py 64
salt/cloud/clouds/gce.py 18
salt/_logging/impl.py 70
|
test
|
batch increase the code coverage percent on the following files to at least please be aware that currently the percentage might be inaccurate if the module uses salt due to file percent salt modules junos py salt modules postgres py salt compat py salt cloud clouds gce py salt logging impl py
| 1
|
334,069
| 29,820,345,671
|
IssuesEvent
|
2023-06-17 01:31:37
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
reopened
|
Fix manipulation.test_flipud
|
Sub Task Ivy API Experimental Failing Test
|
| | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4561754220/jobs/8048085832" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4561754220/jobs/8048085832" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4561754220/jobs/8048085832" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4561754220/jobs/8048085832" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|
1.0
|
Fix manipulation.test_flipud - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4561754220/jobs/8048085832" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4561754220/jobs/8048085832" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4561754220/jobs/8048085832" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4561754220/jobs/8048085832" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|
test
|
fix manipulation test flipud tensorflow img src torch img src numpy img src jax img src
| 1
|
98,106
| 20,611,770,876
|
IssuesEvent
|
2022-03-07 09:21:14
|
joomla/joomla-cms
|
https://api.github.com/repos/joomla/joomla-cms
|
closed
|
[4.1] Schedule tasks error reporting
|
No Code Attached Yet
|
### Steps to reproduce the issue
Create a new image size action
Do NOT select a path
Save and run a test
### Expected result
Error message that the image path does not exist

### Actual result
No error message on screen.
Not much point in doing a test if you dont get to see any errors
### Additional comments
The error is present in the logs

|
1.0
|
[4.1] Schedule tasks error reporting - ### Steps to reproduce the issue
Create a new image size action
Do NOT select a path
Save and run a test
### Expected result
Error message that the image path does not exist

### Actual result
No error message on screen.
Not much point in doing a test if you dont get to see any errors
### Additional comments
The error is present in the logs

|
non_test
|
schedule tasks error reporting steps to reproduce the issue create a new image size action do not select a path save and run a test expected result error message that the image path does not exist actual result no error message on screen not much point in doing a test if you dont get to see any errors additional comments the error is present in the logs
| 0
|
75,440
| 3,462,464,779
|
IssuesEvent
|
2015-12-20 23:23:34
|
bounswe/bounswe2015group8
|
https://api.github.com/repos/bounswe/bounswe2015group8
|
opened
|
Following Tags Back-end Logic
|
Priority-High Type-Task
|
Users should be able to follow tags based on their interested areas.
Implement the back-end functionality of following a tag for a user. Write the dao, service and controller class functions. It would be better if users can follow multiple tags at once. (similar to adding tags to heritage objects/posts.)
|
1.0
|
Following Tags Back-end Logic - Users should be able to follow tags based on their interested areas.
Implement the back-end functionality of following a tag for a user. Write the dao, service and controller class functions. It would be better if users can follow multiple tags at once. (similar to adding tags to heritage objects/posts.)
|
non_test
|
following tags back end logic users should be able to follow tags based on their interested areas implement the back end functionality of following a tag for a user write the dao service and controller class functions it would be better if users can follow multiple tags at once similar to adding tags to heritage objects posts
| 0
|
285,499
| 24,671,170,002
|
IssuesEvent
|
2022-10-18 13:52:23
|
aldefouw/redcap_cypress
|
https://api.github.com/repos/aldefouw/redcap_cypress
|
reopened
|
Design Forms using Data Dictionary & Online Designer
|
Core Functionality Test Script Feature
|
File Location:
https://github.com/aldefouw/redcap_cypress/blob/v11.1.29/cypress/features/core/pre-requisite/design_forms.feature
Task:
Write the test specs in Gherkin DSL following manual test script as guide
Contact Adam De Fouw ([aldefouw@medicine.wisc.edu](mailto:aldefouw@medicine.wisc.edu)) with any questions.
|
1.0
|
Design Forms using Data Dictionary & Online Designer - File Location:
https://github.com/aldefouw/redcap_cypress/blob/v11.1.29/cypress/features/core/pre-requisite/design_forms.feature
Task:
Write the test specs in Gherkin DSL following manual test script as guide
Contact Adam De Fouw ([aldefouw@medicine.wisc.edu](mailto:aldefouw@medicine.wisc.edu)) with any questions.
|
test
|
design forms using data dictionary online designer file location task write the test specs in gherkin dsl following manual test script as guide contact adam de fouw mailto aldefouw medicine wisc edu with any questions
| 1
|
33,267
| 4,820,388,615
|
IssuesEvent
|
2016-11-04 22:39:48
|
infiniteautomation/ma-core-public
|
https://api.github.com/repos/infiniteautomation/ma-core-public
|
closed
|
Persistent Data Source Throttle Threshold Setting
|
Enhancement Ready for Testing
|
Add system settings and help for the threshold.
|
1.0
|
Persistent Data Source Throttle Threshold Setting - Add system settings and help for the threshold.
|
test
|
persistent data source throttle threshold setting add system settings and help for the threshold
| 1
|
303,482
| 26,212,654,306
|
IssuesEvent
|
2023-01-04 08:16:03
|
WPChill/download-monitor
|
https://api.github.com/repos/WPChill/download-monitor
|
closed
|
prices bigger than 1000 throw an error with PayPal gateway
|
Bug needs testing
|
Error with prices bigger than 1000 on checkout.
https://github.com/WPChill/download-monitor/blob/4.7.70/src/Shop/Checkout/PaymentGateway/PayPal/Api/NumericValidator.php#L22
The if is TRUE because 1,234.00 is not numeric.
Formatted price is set here -> https://github.com/WPChill/download-monitor/blob/4.7.70/src/Shop/Checkout/PaymentGateway/PayPal/PayPalGateway.php#L298
|
1.0
|
prices bigger than 1000 throw an error with PayPal gateway - Error with prices bigger than 1000 on checkout.
https://github.com/WPChill/download-monitor/blob/4.7.70/src/Shop/Checkout/PaymentGateway/PayPal/Api/NumericValidator.php#L22
The if is TRUE because 1,234.00 is not numeric.
Formatted price is set here -> https://github.com/WPChill/download-monitor/blob/4.7.70/src/Shop/Checkout/PaymentGateway/PayPal/PayPalGateway.php#L298
|
test
|
prices bigger than throw an error with paypal gateway error with prices bigger than on checkout the if is true because is not numeric formatted price is set here
| 1
|
135,530
| 11,010,062,856
|
IssuesEvent
|
2019-12-04 13:55:43
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
Failing test: X-Pack API Integration Tests.x-pack/test/api_integration/apis/apm/feature_controls·ts - apis APM apm feature controls APIs can be accessed by global_all user
|
Team:apm failed-test
|
A test failed on a tracked branch
```
Error: Endpoint: POST /api/apm/settings/agent-configuration/search
Status code: 404
Response: Not found
expected 404 to equal 200
at executeRequests (test/api_integration/apis/apm/feature_controls.ts:216:15)
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/1312/)
Test failure details
```json
{
"test.class": "X-Pack API Integration Tests.x-pack/test/api_integration/apis/apm/feature_controls·ts",
"test.name": "apis APM apm feature controls APIs can be accessed by global_all user",
"test.failCount": 2
}
```
Test:
https://github.com/elastic/kibana/blob/873529a044e56b2600d279c50c155d47036d97df/x-pack/test/api_integration/apis/apm/feature_controls.ts#L272-L293
Code under test:
https://github.com/elastic/kibana/blob/4e5bfc7efaa4411b18d97247addba5f7982bc14f/x-pack/legacy/plugins/apm/server/routes/settings/agent_configuration.ts#L151-L183
<!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack API Integration Tests.x-pack/test/api_integration/apis/apm/feature_controls·ts","test.name":"apis APM apm feature controls APIs can be accessed by global_all user","test.failCount":2}} -->
|
1.0
|
Failing test: X-Pack API Integration Tests.x-pack/test/api_integration/apis/apm/feature_controls·ts - apis APM apm feature controls APIs can be accessed by global_all user - A test failed on a tracked branch
```
Error: Endpoint: POST /api/apm/settings/agent-configuration/search
Status code: 404
Response: Not found
expected 404 to equal 200
at executeRequests (test/api_integration/apis/apm/feature_controls.ts:216:15)
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/1312/)
Test failure details
```json
{
"test.class": "X-Pack API Integration Tests.x-pack/test/api_integration/apis/apm/feature_controls·ts",
"test.name": "apis APM apm feature controls APIs can be accessed by global_all user",
"test.failCount": 2
}
```
Test:
https://github.com/elastic/kibana/blob/873529a044e56b2600d279c50c155d47036d97df/x-pack/test/api_integration/apis/apm/feature_controls.ts#L272-L293
Code under test:
https://github.com/elastic/kibana/blob/4e5bfc7efaa4411b18d97247addba5f7982bc14f/x-pack/legacy/plugins/apm/server/routes/settings/agent_configuration.ts#L151-L183
<!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack API Integration Tests.x-pack/test/api_integration/apis/apm/feature_controls·ts","test.name":"apis APM apm feature controls APIs can be accessed by global_all user","test.failCount":2}} -->
|
test
|
failing test x pack api integration tests x pack test api integration apis apm feature controls·ts apis apm apm feature controls apis can be accessed by global all user a test failed on a tracked branch error endpoint post api apm settings agent configuration search status code response not found expected to equal at executerequests test api integration apis apm feature controls ts first failure test failure details json test class x pack api integration tests x pack test api integration apis apm feature controls·ts test name apis apm apm feature controls apis can be accessed by global all user test failcount test code under test
| 1
|
240,629
| 18,363,131,946
|
IssuesEvent
|
2021-10-09 15:28:17
|
girlscript/winter-of-contributing
|
https://api.github.com/repos/girlscript/winter-of-contributing
|
closed
|
C-CPP DSA: Bubble Sorting but printing all the passes
|
documentation GWOC21 DSA Assigned C/CPP
|
### Description
Well, this code will be representing the implementation of bubble sort and printing output of all the passes that occured between the elements.
### Domain
C/CPP
### Type of Contribution
Documentation
### Code of Conduct
- [X] I follow [Contributing Guidelines](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CONTRIBUTING.md) & [Code of conduct](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CODE_OF_CONDUCT.md) of this project.
|
1.0
|
C-CPP DSA: Bubble Sorting but printing all the passes - ### Description
Well, this code will be representing the implementation of bubble sort and printing output of all the passes that occured between the elements.
### Domain
C/CPP
### Type of Contribution
Documentation
### Code of Conduct
- [X] I follow [Contributing Guidelines](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CONTRIBUTING.md) & [Code of conduct](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CODE_OF_CONDUCT.md) of this project.
|
non_test
|
c cpp dsa bubble sorting but printing all the passes description well this code will be representing the implementation of bubble sort and printing output of all the passes that occured between the elements domain c cpp type of contribution documentation code of conduct i follow of this project
| 0
|
136,505
| 12,717,075,730
|
IssuesEvent
|
2020-06-24 03:58:24
|
gardener-attic/issues-foo
|
https://api.github.com/repos/gardener-attic/issues-foo
|
closed
|
A
|
component/dashboard component/documentation component/gardener kind/bug kind/post-mortem kind/regression os/garden-linux os/suse-chost platform/alicloud platform/aws platform/azure platform/converged-cloud platform/gcp priority/normal topology/shoot
|
## Which cluster is affected?
https://dashboard.garden.dev.k8s.ondemand.com/namespace/garden/shoots/aws/
## What happened?
## What you expected to happen?
## When did it happen or started to happen?
<!-- Please provide start time in UTC OR relative time in hours from now, so that we can pull the proper logs. -->
`Absolute`: <!-- e.g. 2029-01-01T15:00:00Z -->
`Relative`: <!-- e.g. 3h -->
## How would we reproduce it (concisely and precisely)?
## Anything else we need to know?
## Help us categorise this issue for faster resolution:
<!-- First word is category, all consecutive words narrow it down. -->
<!-- Please delete every word that doesn't fit here/from your PoV. -->
/area audit-logging auto-scaling backup certification control-plane cost delivery disaster-recovery documentation high-availability logging metering monitoring networking os performance quality security storage usability user-management
/component gardener dashboard documentation
/kind bug regression post-mortem
/os garden-linux suse-chost
/platform alicloud aws azure gcp converged-cloud
<!-- Keep the next line if this issue has no high urgency. Delete the line, if you go for a higher priority. -->
/priority normal
<!-- Uncomment the following line, if you believe this is a critical issue OR... -->
<!-- /priority critical -->
<!-- ...uncomment the following line, if this issue has direct customer impact and requires our SRE staff to be paged (at night/on weekends). Use only if business continuity is at risk! -->
<!-- /priority blocker -->
|
1.0
|
A - ## Which cluster is affected?
https://dashboard.garden.dev.k8s.ondemand.com/namespace/garden/shoots/aws/
## What happened?
## What you expected to happen?
## When did it happen or started to happen?
<!-- Please provide start time in UTC OR relative time in hours from now, so that we can pull the proper logs. -->
`Absolute`: <!-- e.g. 2029-01-01T15:00:00Z -->
`Relative`: <!-- e.g. 3h -->
## How would we reproduce it (concisely and precisely)?
## Anything else we need to know?
## Help us categorise this issue for faster resolution:
<!-- First word is category, all consecutive words narrow it down. -->
<!-- Please delete every word that doesn't fit here/from your PoV. -->
/area audit-logging auto-scaling backup certification control-plane cost delivery disaster-recovery documentation high-availability logging metering monitoring networking os performance quality security storage usability user-management
/component gardener dashboard documentation
/kind bug regression post-mortem
/os garden-linux suse-chost
/platform alicloud aws azure gcp converged-cloud
<!-- Keep the next line if this issue has no high urgency. Delete the line, if you go for a higher priority. -->
/priority normal
<!-- Uncomment the following line, if you believe this is a critical issue OR... -->
<!-- /priority critical -->
<!-- ...uncomment the following line, if this issue has direct customer impact and requires our SRE staff to be paged (at night/on weekends). Use only if business continuity is at risk! -->
<!-- /priority blocker -->
|
non_test
|
a which cluster is affected what happened what you expected to happen when did it happen or started to happen absolute relative how would we reproduce it concisely and precisely anything else we need to know help us categorise this issue for faster resolution area audit logging auto scaling backup certification control plane cost delivery disaster recovery documentation high availability logging metering monitoring networking os performance quality security storage usability user management component gardener dashboard documentation kind bug regression post mortem os garden linux suse chost platform alicloud aws azure gcp converged cloud priority normal
| 0
|
135,997
| 11,032,485,525
|
IssuesEvent
|
2019-12-06 20:18:02
|
brave/brave-browser
|
https://api.github.com/repos/brave/brave-browser
|
opened
|
P3A - Open TOR window response value isn't displayed correctly in Upgraded profile
|
QA/Test-Plan-Specified QA/Yes bug
|
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
Found while testing #6285
P3A - Open TOR window response value isn't displayed correctly in Upgraded profile
Q10: Have you ever used a Tor private window?
(Brave.Core.LastTimeIncognitoUsed)
1. Yes
2. No
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Clean profile 1.0.1
2. Open a TOR window
3. Upgrade profile to 1.1.18
4. Reload local state
4. Open TOR window response value is displayed as `1` instead of `0` in local state file
## Actual result:
<!--Please add screenshots if needed-->
Open TOR window response value is displayed as `1` instead of `0`
## Expected result:
Open TOR window response value should be displayed correctly
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Always
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.1.18 Chromium: 79.0.3945.56 (Official Build) (64-bit)
-- | --
Revision | 73cc6bf591f792b99f8fc7cdfb8addedbd084bf8-refs/branch-heads/3945@{#788}
OS | Windows 10 OS Version 1803 (Build 17134.1006)
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? NA
- Can you reproduce this issue with the beta channel? Yes
- Can you reproduce this issue with the dev channel? Yes
- Can you reproduce this issue with the nightly channel? Yes
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields? NA
- Does the issue resolve itself when disabling Brave Rewards? NA
- Is the issue reproducible on the latest version of Chrome? NA
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
cc: @brave/legacy_qa @iefremov @bsclifton @rebron
|
1.0
|
P3A - Open TOR window response value isn't displayed correctly in Upgraded profile - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
Found while testing #6285
P3A - Open TOR window response value isn't displayed correctly in Upgraded profile
Q10: Have you ever used a Tor private window?
(Brave.Core.LastTimeIncognitoUsed)
1. Yes
2. No
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Clean profile 1.0.1
2. Open a TOR window
3. Upgrade profile to 1.1.18
4. Reload local state
4. Open TOR window response value is displayed as `1` instead of `0` in local state file
## Actual result:
<!--Please add screenshots if needed-->
Open TOR window response value is displayed as `1` instead of `0`
## Expected result:
Open TOR window response value should be displayed correctly
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Always
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.1.18 Chromium: 79.0.3945.56 (Official Build) (64-bit)
-- | --
Revision | 73cc6bf591f792b99f8fc7cdfb8addedbd084bf8-refs/branch-heads/3945@{#788}
OS | Windows 10 OS Version 1803 (Build 17134.1006)
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? NA
- Can you reproduce this issue with the beta channel? Yes
- Can you reproduce this issue with the dev channel? Yes
- Can you reproduce this issue with the nightly channel? Yes
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields? NA
- Does the issue resolve itself when disabling Brave Rewards? NA
- Is the issue reproducible on the latest version of Chrome? NA
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
cc: @brave/legacy_qa @iefremov @bsclifton @rebron
|
test
|
open tor window response value isn t displayed correctly in upgraded profile have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description found while testing open tor window response value isn t displayed correctly in upgraded profile have you ever used a tor private window brave core lasttimeincognitoused yes no steps to reproduce clean profile open a tor window upgrade profile to reload local state open tor window response value is displayed as instead of in local state file actual result open tor window response value is displayed as instead of expected result open tor window response value should be displayed correctly reproduces how often always brave version brave version info brave chromium official build bit revision refs branch heads os windows os version build version channel information can you reproduce this issue with the current release na can you reproduce this issue with the beta channel yes can you reproduce this issue with the dev channel yes can you reproduce this issue with the nightly channel yes other additional information does the issue resolve itself when disabling brave shields na does the issue resolve itself when disabling brave rewards na is the issue reproducible on the latest version of chrome na miscellaneous information cc brave legacy qa iefremov bsclifton rebron
| 1
|
429,744
| 12,427,009,404
|
IssuesEvent
|
2020-05-25 00:28:20
|
eclipse-ee4j/glassfish
|
https://api.github.com/repos/eclipse-ee4j/glassfish
|
closed
|
New annotation @TimedProbe to probe the beginning and end of a method and calculate total time
|
Component: monitoring ERR: Assignee Priority: Major Stale Type: New Feature
|
Commonly needed is how much time is spent in a method. Create a new @TimedProbe which does this automatically.
|
1.0
|
New annotation @TimedProbe to probe the beginning and end of a method and calculate total time - Commonly needed is how much time is spent in a method. Create a new @TimedProbe which does this automatically.
|
non_test
|
new annotation timedprobe to probe the beginning and end of a method and calculate total time commonly needed is how much time is spent in a method create a new timedprobe which does this automatically
| 0
|
63,081
| 14,656,666,922
|
IssuesEvent
|
2020-12-28 13:56:13
|
fu1771695yongxie/learnGitBranching
|
https://api.github.com/repos/fu1771695yongxie/learnGitBranching
|
opened
|
CVE-2015-9251 (Medium) detected in jquery-1.12.4.js
|
security vulnerability
|
## CVE-2015-9251 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.12.4.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.12.4/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.12.4/jquery.js</a></p>
<p>Path to dependency file: learnGitBranching/node_modules/jquery-ui/demos/effect/removeClass.html</p>
<p>Path to vulnerable library: learnGitBranching/node_modules/jquery-ui/demos/effect/../../external/jquery/jquery.js,learnGitBranching/node_modules/jquery-ui/demos/slider/../../external/jquery/jquery.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.12.4.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/learnGitBranching/commit/33cba5147b9149e15d524f7a0f485cf33acd1c2b">33cba5147b9149e15d524f7a0f485cf33acd1c2b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2015-9251 (Medium) detected in jquery-1.12.4.js - ## CVE-2015-9251 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.12.4.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.12.4/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.12.4/jquery.js</a></p>
<p>Path to dependency file: learnGitBranching/node_modules/jquery-ui/demos/effect/removeClass.html</p>
<p>Path to vulnerable library: learnGitBranching/node_modules/jquery-ui/demos/effect/../../external/jquery/jquery.js,learnGitBranching/node_modules/jquery-ui/demos/slider/../../external/jquery/jquery.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.12.4.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/learnGitBranching/commit/33cba5147b9149e15d524f7a0f485cf33acd1c2b">33cba5147b9149e15d524f7a0f485cf33acd1c2b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in jquery js cve medium severity vulnerability vulnerable library jquery js javascript library for dom operations library home page a href path to dependency file learngitbranching node modules jquery ui demos effect removeclass html path to vulnerable library learngitbranching node modules jquery ui demos effect external jquery jquery js learngitbranching node modules jquery ui demos slider external jquery jquery js dependency hierarchy x jquery js vulnerable library found in head commit a href found in base branch master vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
| 0
|
752,129
| 26,274,311,237
|
IssuesEvent
|
2023-01-06 20:15:34
|
craftercms/craftercms
|
https://api.github.com/repos/craftercms/craftercms
|
closed
|
[studio-ui] Tree Nav loses state when background refresh
|
bug priority: high validate
|
### Duplicates
- [X] I have searched the existing issues
### Latest version
- [X] The issue is in the latest released 4.0.x
- [ ] The issue is in the latest released 3.1.x
### Describe the issue
Receiving socket events while you tree nav is expanded past its default results in losing the expanded state
### Steps to reproduce
Steps:
1. Open two different browsers and login with two different users
2. Open preview on both and with one user expand the tree nav past defaults (use show more) and with the other user edit an asset that pertains to the same nav you expanded
3. See the nav tree losing state
### Relevant log output
_No response_
### Screenshots and/or videos
_No response_
|
1.0
|
[studio-ui] Tree Nav loses state when background refresh - ### Duplicates
- [X] I have searched the existing issues
### Latest version
- [X] The issue is in the latest released 4.0.x
- [ ] The issue is in the latest released 3.1.x
### Describe the issue
Receiving socket events while you tree nav is expanded past its default results in losing the expanded state
### Steps to reproduce
Steps:
1. Open two different browsers and login with two different users
2. Open preview on both and with one user expand the tree nav past defaults (use show more) and with the other user edit an asset that pertains to the same nav you expanded
3. See the nav tree losing state
### Relevant log output
_No response_
### Screenshots and/or videos
_No response_
|
non_test
|
tree nav loses state when background refresh duplicates i have searched the existing issues latest version the issue is in the latest released x the issue is in the latest released x describe the issue receiving socket events while you tree nav is expanded past its default results in losing the expanded state steps to reproduce steps open two different browsers and login with two different users open preview on both and with one user expand the tree nav past defaults use show more and with the other user edit an asset that pertains to the same nav you expanded see the nav tree losing state relevant log output no response screenshots and or videos no response
| 0
|
40,128
| 9,852,416,338
|
IssuesEvent
|
2019-06-19 12:49:40
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
opened
|
Possibly wrong MOD implementation in ACCESS and TERADATA
|
C: DB: Access C: DB: Teradata C: Functionality E: Enterprise Edition E: Professional Edition P: Medium T: Defect
|
When generating modulo expressions in `ACCESS` and `TERADATA`, we're currently not wrapping the expression in parentheses:
```sql
a MOD b
```
This could potentially lead to operator precedence issues when combined with operators of higher precedence
|
1.0
|
Possibly wrong MOD implementation in ACCESS and TERADATA - When generating modulo expressions in `ACCESS` and `TERADATA`, we're currently not wrapping the expression in parentheses:
```sql
a MOD b
```
This could potentially lead to operator precedence issues when combined with operators of higher precedence
|
non_test
|
possibly wrong mod implementation in access and teradata when generating modulo expressions in access and teradata we re currently not wrapping the expression in parentheses sql a mod b this could potentially lead to operator precedence issues when combined with operators of higher precedence
| 0
|
148,924
| 11,872,135,752
|
IssuesEvent
|
2020-03-26 15:22:53
|
infinispan/infinispan-operator
|
https://api.github.com/repos/infinispan/infinispan-operator
|
closed
|
Unable to run some of the e2e tests against OpenShift
|
bug test
|
Some of the e2e tests do no work with OpenShift (tested with 4.2 and 4.3). The issue started appearing after migration to Operator SDK 0.15.2 PR was merged (#293).
So far this applies for all the `*Update` tests. Request to change the parameter won't hit the OpenShift.
This issue is probably caused by Kubernetes client update.
|
1.0
|
Unable to run some of the e2e tests against OpenShift - Some of the e2e tests do no work with OpenShift (tested with 4.2 and 4.3). The issue started appearing after migration to Operator SDK 0.15.2 PR was merged (#293).
So far this applies for all the `*Update` tests. Request to change the parameter won't hit the OpenShift.
This issue is probably caused by Kubernetes client update.
|
test
|
unable to run some of the tests against openshift some of the tests do no work with openshift tested with and the issue started appearing after migration to operator sdk pr was merged so far this applies for all the update tests request to change the parameter won t hit the openshift this issue is probably caused by kubernetes client update
| 1
|
666,260
| 22,348,140,570
|
IssuesEvent
|
2022-06-15 09:35:59
|
PCSX2/pcsx2
|
https://api.github.com/repos/PCSX2/pcsx2
|
closed
|
[BUG]: Gran Turismo 4 - Bad Edges on Split Time Vehicle Icons
|
Bug GS: Hardware Regression GS: Texture Cache High Priority
|
### Describe the Bug
On any race, with any vehicle, the vehicle icon which appears while showing split times will have a line on the bottom and right edges. The icon is not supposed to have any boundaries, and is supposed to be transparent except for the vehicle picture on it.
Bug was introduced by 1.7.2126 and present up to latest master as of writing.
### Reproduction Steps
Boot the game, no memory card required, no adjustments to rounding/clamping required. Game must be launched with hardware rendering and upscaling, issue will not occur if the split time first appears on software rendering or native resolution hardware rendering. Go to arcade mode, pick single race. Pick any track, pick any car. Start the race and progress until a split time appears.
### Expected Behavior
Split time icon should appear without any lines on its edges.
### PCSX2 Revision
1.7.2126 to 1.7.2480
### Operating System
Windows 10 (64bit)
### If Linux - Specify Distro
_No response_
### CPU
R7 5800X
### GPU
RX 6900 XT
### GS Settings
4x Native
### Emulation Settings
Preset 3
### GS Window Screenshots

### Logs & Dumps
[gs_20220318210426.zip](https://github.com/PCSX2/pcsx2/files/8307167/gs_20220318210426.zip)
|
1.0
|
[BUG]: Gran Turismo 4 - Bad Edges on Split Time Vehicle Icons - ### Describe the Bug
On any race, with any vehicle, the vehicle icon which appears while showing split times will have a line on the bottom and right edges. The icon is not supposed to have any boundaries, and is supposed to be transparent except for the vehicle picture on it.
Bug was introduced by 1.7.2126 and present up to latest master as of writing.
### Reproduction Steps
Boot the game, no memory card required, no adjustments to rounding/clamping required. Game must be launched with hardware rendering and upscaling, issue will not occur if the split time first appears on software rendering or native resolution hardware rendering. Go to arcade mode, pick single race. Pick any track, pick any car. Start the race and progress until a split time appears.
### Expected Behavior
Split time icon should appear without any lines on its edges.
### PCSX2 Revision
1.7.2126 to 1.7.2480
### Operating System
Windows 10 (64bit)
### If Linux - Specify Distro
_No response_
### CPU
R7 5800X
### GPU
RX 6900 XT
### GS Settings
4x Native
### Emulation Settings
Preset 3
### GS Window Screenshots

### Logs & Dumps
[gs_20220318210426.zip](https://github.com/PCSX2/pcsx2/files/8307167/gs_20220318210426.zip)
|
non_test
|
gran turismo bad edges on split time vehicle icons describe the bug on any race with any vehicle the vehicle icon which appears while showing split times will have a line on the bottom and right edges the icon is not supposed to have any boundaries and is supposed to be transparent except for the vehicle picture on it bug was introduced by and present up to latest master as of writing reproduction steps boot the game no memory card required no adjustments to rounding clamping required game must be launched with hardware rendering and upscaling issue will not occur if the split time first appears on software rendering or native resolution hardware rendering go to arcade mode pick single race pick any track pick any car start the race and progress until a split time appears expected behavior split time icon should appear without any lines on its edges revision to operating system windows if linux specify distro no response cpu gpu rx xt gs settings native emulation settings preset gs window screenshots logs dumps
| 0
|
45,839
| 13,055,754,994
|
IssuesEvent
|
2020-07-30 02:38:18
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
omkey inherits from i3Frame object (Trac #117)
|
IceTray Incomplete Migration Migrated from Trac defect
|
Migrated from https://code.icecube.wisc.edu/ticket/117
```json
{
"status": "closed",
"changetime": "2014-11-23T03:37:56",
"description": "where did this come from?",
"reporter": "troy",
"cc": "",
"resolution": "wont or cant fix",
"_ts": "1416713876900096",
"component": "IceTray",
"summary": "omkey inherits from i3Frame object",
"priority": "major",
"keywords": "",
"time": "2008-08-30T16:06:14",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
|
1.0
|
omkey inherits from i3Frame object (Trac #117) - Migrated from https://code.icecube.wisc.edu/ticket/117
```json
{
"status": "closed",
"changetime": "2014-11-23T03:37:56",
"description": "where did this come from?",
"reporter": "troy",
"cc": "",
"resolution": "wont or cant fix",
"_ts": "1416713876900096",
"component": "IceTray",
"summary": "omkey inherits from i3Frame object",
"priority": "major",
"keywords": "",
"time": "2008-08-30T16:06:14",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
|
non_test
|
omkey inherits from object trac migrated from json status closed changetime description where did this come from reporter troy cc resolution wont or cant fix ts component icetray summary omkey inherits from object priority major keywords time milestone owner troy type defect
| 0
|
91,210
| 8,300,600,604
|
IssuesEvent
|
2018-09-21 08:39:00
|
swarmcity/SwarmCityDapp
|
https://api.github.com/repos/swarmcity/SwarmCityDapp
|
closed
|
on iphone button enter swarm city doesn't appear
|
blocking bug ready to test
|
# Location
/mykeys
# Expected behavior
when creating new account on iphone, after checking the box to use the private key to be the backup, a button appears "enter swarm.city"
# Actual behavior
when creating new account on iphone, after checking the box to use the private key to be the backup, the button appears "enter swarm.city" doesn't appear.
# Steps to reproduce the problem
ON IPHONE: create new/ choose username and avatar/ choose pasword/make backup/show keys
# Operating system / hardware
safari on iphone
|
1.0
|
on iphone button enter swarm city doesn't appear - # Location
/mykeys
# Expected behavior
when creating new account on iphone, after checking the box to use the private key to be the backup, a button appears "enter swarm.city"
# Actual behavior
when creating new account on iphone, after checking the box to use the private key to be the backup, the button appears "enter swarm.city" doesn't appear.
# Steps to reproduce the problem
ON IPHONE: create new/ choose username and avatar/ choose pasword/make backup/show keys
# Operating system / hardware
safari on iphone
|
test
|
on iphone button enter swarm city doesn t appear location mykeys expected behavior when creating new account on iphone after checking the box to use the private key to be the backup a button appears enter swarm city actual behavior when creating new account on iphone after checking the box to use the private key to be the backup the button appears enter swarm city doesn t appear steps to reproduce the problem on iphone create new choose username and avatar choose pasword make backup show keys operating system hardware safari on iphone
| 1
|
338,309
| 30,291,860,715
|
IssuesEvent
|
2023-07-09 11:36:12
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
reopened
|
Fix blas_and_lapack_ops.test_torch_ger
|
PyTorch Frontend Sub Task Failing Test
|
| | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5499384391/jobs/10021539111"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5499384391/jobs/10021539111"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5499384391/jobs/10021539111"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5499384391/jobs/10021539111"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5499384391/jobs/10021539111"><img src=https://img.shields.io/badge/-success-success></a>
|
1.0
|
Fix blas_and_lapack_ops.test_torch_ger - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5499384391/jobs/10021539111"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5499384391/jobs/10021539111"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5499384391/jobs/10021539111"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5499384391/jobs/10021539111"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5499384391/jobs/10021539111"><img src=https://img.shields.io/badge/-success-success></a>
|
test
|
fix blas and lapack ops test torch ger tensorflow a href src jax a href src numpy a href src torch a href src paddle a href src
| 1
|
222,656
| 17,466,766,700
|
IssuesEvent
|
2021-08-06 18:02:25
|
paritytech/polkadot
|
https://api.github.com/repos/paritytech/polkadot
|
closed
|
Approval Voting unit/integration hybrid tests
|
F4-tests
|
In order to adequately test approval voting beyond simple unit tests that validate the behavior of the subsystem, we should create multiple instances of the approval voting subsystem and play them against each other.
|
1.0
|
Approval Voting unit/integration hybrid tests - In order to adequately test approval voting beyond simple unit tests that validate the behavior of the subsystem, we should create multiple instances of the approval voting subsystem and play them against each other.
|
test
|
approval voting unit integration hybrid tests in order to adequately test approval voting beyond simple unit tests that validate the behavior of the subsystem we should create multiple instances of the approval voting subsystem and play them against each other
| 1
|
157,690
| 12,389,088,909
|
IssuesEvent
|
2020-05-20 08:27:22
|
moment/moment
|
https://api.github.com/repos/moment/moment
|
closed
|
2 tests failed. locale:gu:calendar day (1107.6) locale:x-pseudo:calendar day (2655.6)
|
DST Unit Test Failed
|
### Client info
```
Date String : Mon Mar 12 2018 14:40:57 GMT-0700 (Pacific Daylight Time)
Locale String : 3/12/2018, 2:40:57 PM
Offset : 420
User Agent : Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/64.0.3282.186 Safari/537.36
Moment Version: 2.21.0
```
====
### locale:gu:calendar day (1107.6)
yesterday at the same time
```javascript
// Expected ગઇકાલે રાત ૨:૦૦ વાગ્યે
// Actual ગઇકાલે રાત ૩:૦૦ વાગ્યે
"ગઇકાલે રાત ૩:૦૦ વાગ્યે" === "ગઇકાલે રાત ૨:૦૦ વાગ્યે"
```
====
### locale:x-pseudo:calendar day (2655.6)
yesterday at the same time
```javascript
// Expected Ý~ést~érdá~ý át 02:00
// Actual Ý~ést~érdá~ý át 03:00
"Ý~ést~érdá~ý át 03:00" === "Ý~ést~érdá~ý át 02:00"
```
|
1.0
|
2 tests failed. locale:gu:calendar day (1107.6) locale:x-pseudo:calendar day (2655.6) - ### Client info
```
Date String : Mon Mar 12 2018 14:40:57 GMT-0700 (Pacific Daylight Time)
Locale String : 3/12/2018, 2:40:57 PM
Offset : 420
User Agent : Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/64.0.3282.186 Safari/537.36
Moment Version: 2.21.0
```
====
### locale:gu:calendar day (1107.6)
yesterday at the same time
```javascript
// Expected ગઇકાલે રાત ૨:૦૦ વાગ્યે
// Actual ગઇકાલે રાત ૩:૦૦ વાગ્યે
"ગઇકાલે રાત ૩:૦૦ વાગ્યે" === "ગઇકાલે રાત ૨:૦૦ વાગ્યે"
```
====
### locale:x-pseudo:calendar day (2655.6)
yesterday at the same time
```javascript
// Expected Ý~ést~érdá~ý át 02:00
// Actual Ý~ést~érdá~ý át 03:00
"Ý~ést~érdá~ý át 03:00" === "Ý~ést~érdá~ý át 02:00"
```
|
test
|
tests failed locale gu calendar day locale x pseudo calendar day client info date string mon mar gmt pacific daylight time locale string pm offset user agent mozilla windows nt applewebkit khtml like gecko chrome safari moment version locale gu calendar day yesterday at the same time javascript expected ગઇકાલે રાત વાગ્યે actual ગઇકાલે રાત વાગ્યે ગઇકાલે રાત વાગ્યે ગઇકાલે રાત વાગ્યે locale x pseudo calendar day yesterday at the same time javascript expected ý ést érdá ý át actual ý ést érdá ý át ý ést érdá ý át ý ést érdá ý át
| 1
|
615,698
| 19,273,306,644
|
IssuesEvent
|
2021-12-10 08:55:23
|
googleapis/google-cloud-go
|
https://api.github.com/repos/googleapis/google-cloud-go
|
closed
|
logging: memory leak at high frequency logging
|
type: bug api: logging priority: p2 lang: go
|
We have an issue when using the go logging library. We are experiencing a very high memory consumption, potentially a leak, when writing logs at a high frequency.
We first observed this issue in one of our Go services hosted in Cloud Run. When receiving a high amount of requests (6000rps+) we can see the memory utilisation increasing until the instances runs our of allocated memory.
We write one log entry per request. With the logging disabled the service handles the same load consuming around 10-20% of the total memory.
We can replicate this behaviour locally by running a simple for-loop that logs on each iteration. Eventually the client OnError()-function is invoked with the following error: "bundler reached buffered byte limit".
We've tried changing the values of:
[EntryCountThreshold()](https://pkg.go.dev/cloud.google.com/go/logging#EntryCountThreshold)
[ConcurrentWriteLimit()](https://pkg.go.dev/cloud.google.com/go/logging#ConcurrentWriteLimit)
This doesn't mitigate the problem. However when adding a 10ms sleep on each iteration the problem is gone.
Package version: v1.4.2
Container image: golang:1.17.3-alpine3.14
Local environment/version: go1.17.2 darwin/amd64
Sample code to reproduce the issue:
```
package main
import (
"context"
"log"
"sync"
glogging "cloud.google.com/go/logging"
"github.com/google/uuid"
)
func onInternalStackdriverExporterError(err error) {
log.Println(err)
}
const numOfWorkers = 1
func main() {
wg := sync.WaitGroup{}
ctx := context.Background()
c, _ := glogging.NewClient(ctx, "gcp-project-id")
c.OnError = onInternalStackdriverExporterError
logr := c.Logger("log-load-test")
for i := 0; i < numOfWorkers; i++ {
wg.Add(1)
go doSomething(logr)
}
wg.Wait()
}
func doSomething(logger *glogging.Logger) {
for {
//time.Sleep(10 * time.Millisecond) //running this Sleep mitigates the memory leak.
e := glogging.Entry{
Severity: glogging.Severity(100),
Payload: uuid.New().String(),
}
logger.Log(e)
}
}
```
|
1.0
|
logging: memory leak at high frequency logging - We have an issue when using the go logging library. We are experiencing a very high memory consumption, potentially a leak, when writing logs at a high frequency.
We first observed this issue in one of our Go services hosted in Cloud Run. When receiving a high amount of requests (6000rps+) we can see the memory utilisation increasing until the instances runs our of allocated memory.
We write one log entry per request. With the logging disabled the service handles the same load consuming around 10-20% of the total memory.
We can replicate this behaviour locally by running a simple for-loop that logs on each iteration. Eventually the client OnError()-function is invoked with the following error: "bundler reached buffered byte limit".
We've tried changing the values of:
[EntryCountThreshold()](https://pkg.go.dev/cloud.google.com/go/logging#EntryCountThreshold)
[ConcurrentWriteLimit()](https://pkg.go.dev/cloud.google.com/go/logging#ConcurrentWriteLimit)
This doesn't mitigate the problem. However when adding a 10ms sleep on each iteration the problem is gone.
Package version: v1.4.2
Container image: golang:1.17.3-alpine3.14
Local environment/version: go1.17.2 darwin/amd64
Sample code to reproduce the issue:
```
package main
import (
"context"
"log"
"sync"
glogging "cloud.google.com/go/logging"
"github.com/google/uuid"
)
func onInternalStackdriverExporterError(err error) {
log.Println(err)
}
const numOfWorkers = 1
func main() {
wg := sync.WaitGroup{}
ctx := context.Background()
c, _ := glogging.NewClient(ctx, "gcp-project-id")
c.OnError = onInternalStackdriverExporterError
logr := c.Logger("log-load-test")
for i := 0; i < numOfWorkers; i++ {
wg.Add(1)
go doSomething(logr)
}
wg.Wait()
}
func doSomething(logger *glogging.Logger) {
for {
//time.Sleep(10 * time.Millisecond) //running this Sleep mitigates the memory leak.
e := glogging.Entry{
Severity: glogging.Severity(100),
Payload: uuid.New().String(),
}
logger.Log(e)
}
}
```
|
non_test
|
logging memory leak at high frequency logging we have an issue when using the go logging library we are experiencing a very high memory consumption potentially a leak when writing logs at a high frequency we first observed this issue in one of our go services hosted in cloud run when receiving a high amount of requests we can see the memory utilisation increasing until the instances runs our of allocated memory we write one log entry per request with the logging disabled the service handles the same load consuming around of the total memory we can replicate this behaviour locally by running a simple for loop that logs on each iteration eventually the client onerror function is invoked with the following error bundler reached buffered byte limit we ve tried changing the values of this doesn t mitigate the problem however when adding a sleep on each iteration the problem is gone package version container image golang local environment version darwin sample code to reproduce the issue package main import context log sync glogging cloud google com go logging github com google uuid func oninternalstackdriverexportererror err error log println err const numofworkers func main wg sync waitgroup ctx context background c glogging newclient ctx gcp project id c onerror oninternalstackdriverexportererror logr c logger log load test for i i numofworkers i wg add go dosomething logr wg wait func dosomething logger glogging logger for time sleep time millisecond running this sleep mitigates the memory leak e glogging entry severity glogging severity payload uuid new string logger log e
| 0
|
4,814
| 2,875,502,626
|
IssuesEvent
|
2015-06-09 08:36:24
|
bpmn-io/bpmn-js
|
https://api.github.com/repos/bpmn-io/bpmn-js
|
closed
|
Investigate: Document our APIs in a user friendly way
|
documentation in progress
|
Adding a (self) hosted solution with up to date documentation.
[**ReadMe.io**](https://readme.io/) could be a great option since it's a well built, easy to use system and free for open source.
### Primary Use Case
Users should learn about our public API, i.e. the [Overlays](https://github.com/bpmn-io/diagram-js/blob/master/lib/features/overlays/Overlays.js) service or the main entry points into bpmn-js.
* Only some APIs are public
* We want to be able to provide information about these directly in code or as close as possible to it (distance means harder to maintain)
* We want documentation to be generated to HTML + published somewhere
* Linking to GitHub or other parts of the documentation should work
### Options
* Using a library
* Parsing stuff manually and feeding it into a page generator, i.e. [kartoffeldruck](https://github.com/nikku/kartoffeldruck) or the like
### Tasks
* [ ] generate documentation for `Overlays` + `EventBus` (in `MD` format)
* [ ] investigate linking between `Overlays` -> `EventBus`
* [ ] create HTML from the MD documentation; focus on readability / public
|
1.0
|
Investigate: Document our APIs in a user friendly way - Adding a (self) hosted solution with up to date documentation.
[**ReadMe.io**](https://readme.io/) could be a great option since it's a well built, easy to use system and free for open source.
### Primary Use Case
Users should learn about our public API, i.e. the [Overlays](https://github.com/bpmn-io/diagram-js/blob/master/lib/features/overlays/Overlays.js) service or the main entry points into bpmn-js.
* Only some APIs are public
* We want to be able to provide information about these directly in code or as close as possible to it (distance means harder to maintain)
* We want documentation to be generated to HTML + published somewhere
* Linking to GitHub or other parts of the documentation should work
### Options
* Using a library
* Parsing stuff manually and feeding it into a page generator, i.e. [kartoffeldruck](https://github.com/nikku/kartoffeldruck) or the like
### Tasks
* [ ] generate documentation for `Overlays` + `EventBus` (in `MD` format)
* [ ] investigate linking between `Overlays` -> `EventBus`
* [ ] create HTML from the MD documentation; focus on readability / public
|
non_test
|
investigate document our apis in a user friendly way adding a self hosted solution with up to date documentation could be a great option since it s a well built easy to use system and free for open source primary use case users should learn about our public api i e the service or the main entry points into bpmn js only some apis are public we want to be able to provide information about these directly in code or as close as possible to it distance means harder to maintain we want documentation to be generated to html published somewhere linking to github or other parts of the documentation should work options using a library parsing stuff manually and feeding it into a page generator i e or the like tasks generate documentation for overlays eventbus in md format investigate linking between overlays eventbus create html from the md documentation focus on readability public
| 0
|
410,250
| 11,985,432,924
|
IssuesEvent
|
2020-04-07 17:29:51
|
IpsumCapra/project-3-4
|
https://api.github.com/repos/IpsumCapra/project-3-4
|
closed
|
US-B2 - As a user, I want to be able to quickly choose the amount I want to withdraw.
|
3 ATM High priority! MUST UI User Story
|
<h1>Acceptance Criteria</h1>
<ul>
<li>UI software is present, and running on the Raspberry PI.</li>
<li>The Raspberry PI and monitor can communicate correctly.</li>
<li>ATM is functional.</li>
<li>User can interface with ATM terminal.</li>
<li>ATM keypad is communicating with other hardware correctly.</li>
</ul>
<h1>Technical Requirements</h1>
<ul>
<li><a href="https://github.com/IpsumCapra/project-3-4/issues/12">US-A2</a></li>
<li><a href="https://github.com/IpsumCapra/project-3-4/issues/14">US-A3</a></li>
<li><a href="https://github.com/IpsumCapra/project-3-4/issues/16">US-T2</a></li>
</ul>
|
1.0
|
US-B2 - As a user, I want to be able to quickly choose the amount I want to withdraw. -
<h1>Acceptance Criteria</h1>
<ul>
<li>UI software is present, and running on the Raspberry PI.</li>
<li>The Raspberry PI and monitor can communicate correctly.</li>
<li>ATM is functional.</li>
<li>User can interface with ATM terminal.</li>
<li>ATM keypad is communicating with other hardware correctly.</li>
</ul>
<h1>Technical Requirements</h1>
<ul>
<li><a href="https://github.com/IpsumCapra/project-3-4/issues/12">US-A2</a></li>
<li><a href="https://github.com/IpsumCapra/project-3-4/issues/14">US-A3</a></li>
<li><a href="https://github.com/IpsumCapra/project-3-4/issues/16">US-T2</a></li>
</ul>
|
non_test
|
us as a user i want to be able to quickly choose the amount i want to withdraw acceptance criteria ui software is present and running on the raspberry pi the raspberry pi and monitor can communicate correctly atm is functional user can interface with atm terminal atm keypad is communicating with other hardware correctly technical requirements a href a href a href
| 0
|
87,937
| 8,127,236,613
|
IssuesEvent
|
2018-08-17 07:15:36
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
cli: Example_csv_tsv_quoting failed under stress
|
C-test-failure O-robot X-duplicate
|
SHA: https://github.com/cockroachdb/cockroach/commits/eccb4a127dd519375d87d5ffd9f6394c37c3a427
Parameters:
```
TAGS=
GOFLAGS=
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=842992&tab=buildLog
```
W180817 05:15:21.657719 1 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I180817 05:15:21.663213 1 server/server.go:828 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled
I180817 05:15:21.663369 1 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180817 05:15:21.663382 1 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180817 05:15:21.676785 1 server/config.go:496 [n?] 1 storage engine initialized
I180817 05:15:21.676809 1 server/config.go:499 [n?] RocksDB cache size: 128 MiB
I180817 05:15:21.676818 1 server/config.go:499 [n?] store 0: in-memory, size 0 B
I180817 05:15:21.679636 1 server/node.go:373 [n?] **** cluster 9ba6fa1c-6fd3-403e-8418-6edd4e925e3e has been created
I180817 05:15:21.679668 1 server/server.go:1399 [n?] **** add additional nodes by specifying --join=127.0.0.1:39623
I180817 05:15:21.679796 1 gossip/gossip.go:383 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:39623" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v999" started_at:1534482921679733959
I180817 05:15:21.681670 1 storage/store.go:1506 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available
I180817 05:15:21.681856 1 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00}
I180817 05:15:21.681926 1 storage/stores.go:242 [n1] read 0 node addresses from persistent storage
I180817 05:15:21.682006 1 server/node.go:697 [n1] connecting to gossip network to verify cluster ID...
I180817 05:15:21.682030 1 server/node.go:722 [n1] node connected via gossip and verified as part of cluster "9ba6fa1c-6fd3-403e-8418-6edd4e925e3e"
I180817 05:15:21.682055 1 server/node.go:546 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes []
I180817 05:15:21.682290 1 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180817 05:15:21.682315 1 server/server.go:1805 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined
I180817 05:15:21.682368 1 server/server.go:1536 [n1] starting https server at 127.0.0.1:33795 (use: 127.0.0.1:33795)
I180817 05:15:21.682383 1 server/server.go:1538 [n1] starting grpc/postgres server at 127.0.0.1:39623
I180817 05:15:21.682394 1 server/server.go:1539 [n1] advertising CockroachDB node at 127.0.0.1:39623
I180817 05:15:21.682569 30790 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180817 05:15:21.691029 30823 storage/replica_command.go:282 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2]
I180817 05:15:21.701391 30806 storage/replica_command.go:282 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3]
I180817 05:15:21.707183 30838 storage/replica_command.go:282 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4]
I180817 05:15:21.714051 30841 storage/replica_command.go:282 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5]
I180817 05:15:21.726470 30814 storage/replica_command.go:282 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6]
I180817 05:15:21.734352 30849 storage/replica_command.go:282 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7]
I180817 05:15:21.756036 30779 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root}
I180817 05:15:21.758697 30576 storage/replica_command.go:282 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8]
W180817 05:15:21.766816 30857 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=16a55113 key=/Local/Range/Table/SystemConfigSpan/Start/RangeDescriptor rw=true pri=0.11242717 iso=SERIALIZABLE stat=PENDING epo=0 ts=1534482921.758714102,0 orig=1534482921.758714102,0 max=1534482921.758714102,0 wto=false rop=false seq=1
I180817 05:15:21.769233 30871 storage/replica_command.go:282 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9]
I180817 05:15:21.779022 30900 storage/replica_command.go:282 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10]
I180817 05:15:21.784849 30853 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:$1 User:root}
I180817 05:15:21.793390 30562 storage/replica_command.go:282 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11]
I180817 05:15:21.799906 30919 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root}
I180817 05:15:21.803606 30946 storage/replica_command.go:282 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12]
I180817 05:15:21.817114 30964 storage/replica_command.go:282 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13]
I180817 05:15:21.823350 30972 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:gen_random_uuid()::STRING User:root}
I180817 05:15:21.832952 30960 storage/replica_command.go:282 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14]
I180817 05:15:21.836562 30975 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root}
I180817 05:15:21.848127 30925 storage/replica_command.go:282 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15]
I180817 05:15:21.854929 31004 storage/replica_command.go:282 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16]
I180817 05:15:21.857881 30989 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root}
I180817 05:15:21.869594 1 server/server.go:1592 [n1] done ensuring all necessary migrations have run
I180817 05:15:21.869629 1 server/server.go:1595 [n1] serving sql connections
I180817 05:15:21.885520 30863 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:39623} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v999 StartedAt:1534482921679733959 LocalityAddress:[]} ClusterID:9ba6fa1c-6fd3-403e-8418-6edd4e925e3e StartedAt:1534482921679733959 LastUp:1534482921679733959}
I180817 05:15:21.888091 30861 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version
I180817 05:15:22.697045 31075 rpc/nodedialer/nodedialer.go:92 [consistencyChecker,n1,s1,r1/1:/{Min-System/}] connection to n1 established
I180817 05:15:30.686366 30729 storage/replica_proposal.go:214 [n1,s1,r11/1:/Table/1{4-5}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482930.686040857,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0
I180817 05:15:31.683056 30788 server/status/runtime.go:433 [n1] runtime stats: 188 MiB RSS, 253 goroutines, 15 MiB/38 MiB/71 MiB GO alloc/idle/total, 15 MiB/51 MiB CGO alloc/total, 0.00cgo/sec, 0.00/0.00 %(u/s)time, 0.00 %gc (264x)
I180817 05:15:31.686568 30715 storage/replica_proposal.go:214 [n1,s1,r16/1:/{Table/19-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0
I180817 05:15:31.686696 31123 storage/replica_command.go:282 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17]
I180817 05:15:31.687259 30736 storage/replica_proposal.go:214 [n1,s1,r15/1:/Table/1{8-9}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686880553,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0
I180817 05:15:31.689132 30735 storage/replica_proposal.go:214 [n1,s1,r7/1:/Table/{SystemCon…-11}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.688758132,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0
I180817 05:15:31.692734 30749 storage/replica_proposal.go:214 [n1,s1,r10/1:/Table/1{3-4}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.692338326,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0
I180817 05:15:31.695746 30739 storage/replica_proposal.go:214 [n1,s1,r16/1:/{Table/19-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0
I180817 05:15:31.696105 31014 storage/replica_command.go:282 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18]
I180817 05:15:31.702463 30746 storage/replica_proposal.go:214 [n1,s1,r17/1:/{Table/20-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0
I180817 05:15:31.702951 31157 storage/replica_command.go:282 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19]
I180817 05:15:31.704417 30790 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180817 05:15:31.705092 30752 storage/replica_proposal.go:214 [n1,s1,r5/1:/System/ts{d-e}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.704808789,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0
I180817 05:15:31.711731 30690 storage/replica_proposal.go:214 [n1,s1,r18/1:/{Table/21-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0
I180817 05:15:31.712165 31131 storage/replica_command.go:282 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20]
I180817 05:15:31.721651 30723 storage/replica_proposal.go:214 [n1,s1,r4/1:/System/{NodeLive…-tsd}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.721254860,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0
I180817 05:15:31.723804 30724 storage/replica_proposal.go:214 [n1,s1,r19/1:/{Table/22-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0
I180817 05:15:31.726788 31169 storage/replica_command.go:282 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21]
I180817 05:15:31.748832 30706 storage/replica_proposal.go:214 [n1,s1,r20/1:/{Table/23-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0
I180817 05:15:31.749392 31188 storage/replica_command.go:282 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22]
I180817 05:15:31.756080 30732 storage/replica_proposal.go:214 [n1,s1,r21/1:/{Table/50-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0
--- FAIL: Example_csv_tsv_quoting (10.34s)
panic: had 16 ranges at startup, expected 22 [recovered]
panic: had 16 ranges at startup, expected 22
goroutine 1 [running]:
testing.runExample.func2(0xbed5b45a667f7544, 0x7156cb8f0, 0x3adc380, 0xc4200de830, 0xc4200de008, 0xc421acafc0, 0x24f4825, 0x17, 0x2619948, 0x2594e57, ...)
/usr/local/go/src/testing/example.go:117 +0x707
panic(0x225d480, 0xc42291bce0)
/usr/local/go/src/runtime/panic.go:502 +0x229
github.com/cockroachdb/cockroach/pkg/cli.(*cliTest).fail(0xc421993a80, 0x225d480, 0xc42291bce0)
/go/src/github.com/cockroachdb/cockroach/pkg/cli/cli_test.go:79 +0xed
github.com/cockroachdb/cockroach/pkg/cli.newCLITest(0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...)
/go/src/github.com/cockroachdb/cockroach/pkg/cli/cli_test.go:133 +0x798
github.com/cockroachdb/cockroach/pkg/cli.Example_csv_tsv_quoting()
/go/src/github.com/cockroachdb/cockroach/pkg/cli/cli_test.go:1163 +0x54
testing.runExample(0x24f4825, 0x17, 0x2619948, 0x2594e57, 0x998, 0x0, 0x0)
/usr/local/go/src/testing/example.go:122 +0x1f9
testing.runExamples(0xc422f85eb8, 0x3898320, 0x10, 0x10, 0x12acb93810101)
/usr/local/go/src/testing/example.go:46 +0x15b
testing.(*M).Run(0xc42037d900, 0x0)
/usr/local/go/src/testing/testing.go:979 +0x1eb
github.com/cockroachdb/cockroach/pkg/cli_test.TestMain(0xc42037d900)
/go/src/github.com/cockroachdb/cockroach/pkg/cli/main_test.go:38 +0x88
main.main()
_testmain.go:140 +0x151
goroutine 19 [syscall]:
os/signal.signal_recv(0x0)
/usr/local/go/src/runtime/sigqueue.go:139 +0xa6
os/signal.loop()
/usr/local/go/src/os/signal/signal_unix.go:22 +0x22
created by os/signal.init.0
/usr/local/go/src/os/signal/signal_unix.go:28 +0x41
goroutine 22 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.flushDaemon()
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1171 +0xf1
created by github.com/cockroachdb/cockroach/pkg/util/log.init.0
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:590 +0xf4
goroutine 23 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.signalFlusher()
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:597 +0xab
created by github.com/cockroachdb/cockroach/pkg/util/log.init.0
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:591 +0x10c
goroutine 50 [select, locked to thread]:
runtime.gopark(0x261f2c8, 0x0, 0x24d594c, 0x6, 0x18, 0x1)
/usr/local/go/src/runtime/proc.go:291 +0x11a
runtime.selectgo(0xc420532750, 0xc42052a060)
/usr/local/go/src/runtime/select.go:392 +0xe50
runtime.ensureSigM.func1()
/usr/local/go/src/runtime/signal_unix.go:549 +0x1f4
runtime.goexit()
/usr/local/go/src/runtime/asm_amd64.s:2361 +0x1
goroutine 9847 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421045320)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 771 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420267d40)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 770 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420267c20)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 109 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420266d80)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 110 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4202678c0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 9846 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421045200)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 3504 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4a7e0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 29836 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420bf1d40)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 9003 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421327560)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 15006 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4215a7c20)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 23721 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42356e6c0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 11615 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42102cb40)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 2767 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c23200)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 1345 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42142d200)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 19252 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42010cd80)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 10528 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4209125a0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 18480 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420bf0900)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 1346 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42142d320)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 6096 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420e27320)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 20391 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420e26900)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 29206 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b7a0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 15745 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421326120)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 17110 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b0e0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 20392 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420e26a20)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 27154 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421182d80)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 13104 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b9e0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 22709 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421727560)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 30597 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4aa20)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 23722 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42356e7e0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 24295 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4215a6d80)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 19809 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4214370e0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 22710 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421727680)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 29205 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b680)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 25464 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42113a7e0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 1960 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c22d80)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 1961 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c22ea0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 27792 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421183560)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 17111 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b200)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 14346 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4217267e0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 17791 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4218c6c60)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 14345 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4217266c0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 3503 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4a6c0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 26015 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42102c000)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 9004 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421327680)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 21548 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42325a5a0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 29837 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420bf1e60)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 2768 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c23320)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 4828 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4ac60)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 15007 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4215a7d40)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 13103 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b8c0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 8236 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061a900)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 18481 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420bf0a20)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 28436 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4212c6900)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 20972 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061b320)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 28437 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4212c6a20)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 24885 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42356e120)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 30596 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4a900)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 27793 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421183680)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 16442 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4212f5d40)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 27153 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421182c60)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 10527 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420912480)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 19253 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42010cea0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 17792 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4218c6d80)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 5161 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b440)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 5162 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b560)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 15744 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421326000)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 26573 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42090e480)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 3775 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42174d200)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 3776 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42174d320)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 4567 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c23560)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 20971 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061b200)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 6744 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42090f200)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 24884 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42356e000)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 19810 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421437200)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 6745 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42090f320)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 4827 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4ab40)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 7556 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42113a120)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 26016 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42102c120)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 4566 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c23440)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 8235 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061a7e0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 24296 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4215a6ea0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 22116 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061a120)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 21547 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42325a480)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 16441 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4212f5c20)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 11616 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42102cc60)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 26574 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42090e5a0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 22115 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061a000)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 7555 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42113a000)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 25465 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42113a900)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 6097 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420e27440)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
ERROR: exit status 2
24 runs completed, 1 failures, over 1m53s
context canceled
```
|
1.0
|
cli: Example_csv_tsv_quoting failed under stress - SHA: https://github.com/cockroachdb/cockroach/commits/eccb4a127dd519375d87d5ffd9f6394c37c3a427
Parameters:
```
TAGS=
GOFLAGS=
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=842992&tab=buildLog
```
W180817 05:15:21.657719 1 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006"
I180817 05:15:21.663213 1 server/server.go:828 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled
I180817 05:15:21.663369 1 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180817 05:15:21.663382 1 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node
I180817 05:15:21.676785 1 server/config.go:496 [n?] 1 storage engine initialized
I180817 05:15:21.676809 1 server/config.go:499 [n?] RocksDB cache size: 128 MiB
I180817 05:15:21.676818 1 server/config.go:499 [n?] store 0: in-memory, size 0 B
I180817 05:15:21.679636 1 server/node.go:373 [n?] **** cluster 9ba6fa1c-6fd3-403e-8418-6edd4e925e3e has been created
I180817 05:15:21.679668 1 server/server.go:1399 [n?] **** add additional nodes by specifying --join=127.0.0.1:39623
I180817 05:15:21.679796 1 gossip/gossip.go:383 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:39623" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v999" started_at:1534482921679733959
I180817 05:15:21.681670 1 storage/store.go:1506 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available
I180817 05:15:21.681856 1 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00}
I180817 05:15:21.681926 1 storage/stores.go:242 [n1] read 0 node addresses from persistent storage
I180817 05:15:21.682006 1 server/node.go:697 [n1] connecting to gossip network to verify cluster ID...
I180817 05:15:21.682030 1 server/node.go:722 [n1] node connected via gossip and verified as part of cluster "9ba6fa1c-6fd3-403e-8418-6edd4e925e3e"
I180817 05:15:21.682055 1 server/node.go:546 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes []
I180817 05:15:21.682290 1 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180817 05:15:21.682315 1 server/server.go:1805 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined
I180817 05:15:21.682368 1 server/server.go:1536 [n1] starting https server at 127.0.0.1:33795 (use: 127.0.0.1:33795)
I180817 05:15:21.682383 1 server/server.go:1538 [n1] starting grpc/postgres server at 127.0.0.1:39623
I180817 05:15:21.682394 1 server/server.go:1539 [n1] advertising CockroachDB node at 127.0.0.1:39623
I180817 05:15:21.682569 30790 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180817 05:15:21.691029 30823 storage/replica_command.go:282 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2]
I180817 05:15:21.701391 30806 storage/replica_command.go:282 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3]
I180817 05:15:21.707183 30838 storage/replica_command.go:282 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4]
I180817 05:15:21.714051 30841 storage/replica_command.go:282 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5]
I180817 05:15:21.726470 30814 storage/replica_command.go:282 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6]
I180817 05:15:21.734352 30849 storage/replica_command.go:282 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7]
I180817 05:15:21.756036 30779 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root}
I180817 05:15:21.758697 30576 storage/replica_command.go:282 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8]
W180817 05:15:21.766816 30857 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=16a55113 key=/Local/Range/Table/SystemConfigSpan/Start/RangeDescriptor rw=true pri=0.11242717 iso=SERIALIZABLE stat=PENDING epo=0 ts=1534482921.758714102,0 orig=1534482921.758714102,0 max=1534482921.758714102,0 wto=false rop=false seq=1
I180817 05:15:21.769233 30871 storage/replica_command.go:282 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9]
I180817 05:15:21.779022 30900 storage/replica_command.go:282 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10]
I180817 05:15:21.784849 30853 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:$1 User:root}
I180817 05:15:21.793390 30562 storage/replica_command.go:282 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11]
I180817 05:15:21.799906 30919 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root}
I180817 05:15:21.803606 30946 storage/replica_command.go:282 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12]
I180817 05:15:21.817114 30964 storage/replica_command.go:282 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13]
I180817 05:15:21.823350 30972 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:gen_random_uuid()::STRING User:root}
I180817 05:15:21.832952 30960 storage/replica_command.go:282 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14]
I180817 05:15:21.836562 30975 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root}
I180817 05:15:21.848127 30925 storage/replica_command.go:282 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15]
I180817 05:15:21.854929 31004 storage/replica_command.go:282 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16]
I180817 05:15:21.857881 30989 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root}
I180817 05:15:21.869594 1 server/server.go:1592 [n1] done ensuring all necessary migrations have run
I180817 05:15:21.869629 1 server/server.go:1595 [n1] serving sql connections
I180817 05:15:21.885520 30863 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:39623} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v999 StartedAt:1534482921679733959 LocalityAddress:[]} ClusterID:9ba6fa1c-6fd3-403e-8418-6edd4e925e3e StartedAt:1534482921679733959 LastUp:1534482921679733959}
I180817 05:15:21.888091 30861 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version
I180817 05:15:22.697045 31075 rpc/nodedialer/nodedialer.go:92 [consistencyChecker,n1,s1,r1/1:/{Min-System/}] connection to n1 established
I180817 05:15:30.686366 30729 storage/replica_proposal.go:214 [n1,s1,r11/1:/Table/1{4-5}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482930.686040857,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0
I180817 05:15:31.683056 30788 server/status/runtime.go:433 [n1] runtime stats: 188 MiB RSS, 253 goroutines, 15 MiB/38 MiB/71 MiB GO alloc/idle/total, 15 MiB/51 MiB CGO alloc/total, 0.00cgo/sec, 0.00/0.00 %(u/s)time, 0.00 %gc (264x)
I180817 05:15:31.686568 30715 storage/replica_proposal.go:214 [n1,s1,r16/1:/{Table/19-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0
I180817 05:15:31.686696 31123 storage/replica_command.go:282 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17]
I180817 05:15:31.687259 30736 storage/replica_proposal.go:214 [n1,s1,r15/1:/Table/1{8-9}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686880553,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0
I180817 05:15:31.689132 30735 storage/replica_proposal.go:214 [n1,s1,r7/1:/Table/{SystemCon…-11}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.688758132,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0
I180817 05:15:31.692734 30749 storage/replica_proposal.go:214 [n1,s1,r10/1:/Table/1{3-4}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.692338326,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0
I180817 05:15:31.695746 30739 storage/replica_proposal.go:214 [n1,s1,r16/1:/{Table/19-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0
I180817 05:15:31.696105 31014 storage/replica_command.go:282 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18]
I180817 05:15:31.702463 30746 storage/replica_proposal.go:214 [n1,s1,r17/1:/{Table/20-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0
I180817 05:15:31.702951 31157 storage/replica_command.go:282 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19]
I180817 05:15:31.704417 30790 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory
I180817 05:15:31.705092 30752 storage/replica_proposal.go:214 [n1,s1,r5/1:/System/ts{d-e}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.704808789,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0
I180817 05:15:31.711731 30690 storage/replica_proposal.go:214 [n1,s1,r18/1:/{Table/21-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0
I180817 05:15:31.712165 31131 storage/replica_command.go:282 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20]
I180817 05:15:31.721651 30723 storage/replica_proposal.go:214 [n1,s1,r4/1:/System/{NodeLive…-tsd}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.721254860,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0
I180817 05:15:31.723804 30724 storage/replica_proposal.go:214 [n1,s1,r19/1:/{Table/22-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0
I180817 05:15:31.726788 31169 storage/replica_command.go:282 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21]
I180817 05:15:31.748832 30706 storage/replica_proposal.go:214 [n1,s1,r20/1:/{Table/23-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0
I180817 05:15:31.749392 31188 storage/replica_command.go:282 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22]
I180817 05:15:31.756080 30732 storage/replica_proposal.go:214 [n1,s1,r21/1:/{Table/50-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0
--- FAIL: Example_csv_tsv_quoting (10.34s)
panic: had 16 ranges at startup, expected 22 [recovered]
panic: had 16 ranges at startup, expected 22
goroutine 1 [running]:
testing.runExample.func2(0xbed5b45a667f7544, 0x7156cb8f0, 0x3adc380, 0xc4200de830, 0xc4200de008, 0xc421acafc0, 0x24f4825, 0x17, 0x2619948, 0x2594e57, ...)
/usr/local/go/src/testing/example.go:117 +0x707
panic(0x225d480, 0xc42291bce0)
/usr/local/go/src/runtime/panic.go:502 +0x229
github.com/cockroachdb/cockroach/pkg/cli.(*cliTest).fail(0xc421993a80, 0x225d480, 0xc42291bce0)
/go/src/github.com/cockroachdb/cockroach/pkg/cli/cli_test.go:79 +0xed
github.com/cockroachdb/cockroach/pkg/cli.newCLITest(0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...)
/go/src/github.com/cockroachdb/cockroach/pkg/cli/cli_test.go:133 +0x798
github.com/cockroachdb/cockroach/pkg/cli.Example_csv_tsv_quoting()
/go/src/github.com/cockroachdb/cockroach/pkg/cli/cli_test.go:1163 +0x54
testing.runExample(0x24f4825, 0x17, 0x2619948, 0x2594e57, 0x998, 0x0, 0x0)
/usr/local/go/src/testing/example.go:122 +0x1f9
testing.runExamples(0xc422f85eb8, 0x3898320, 0x10, 0x10, 0x12acb93810101)
/usr/local/go/src/testing/example.go:46 +0x15b
testing.(*M).Run(0xc42037d900, 0x0)
/usr/local/go/src/testing/testing.go:979 +0x1eb
github.com/cockroachdb/cockroach/pkg/cli_test.TestMain(0xc42037d900)
/go/src/github.com/cockroachdb/cockroach/pkg/cli/main_test.go:38 +0x88
main.main()
_testmain.go:140 +0x151
goroutine 19 [syscall]:
os/signal.signal_recv(0x0)
/usr/local/go/src/runtime/sigqueue.go:139 +0xa6
os/signal.loop()
/usr/local/go/src/os/signal/signal_unix.go:22 +0x22
created by os/signal.init.0
/usr/local/go/src/os/signal/signal_unix.go:28 +0x41
goroutine 22 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.flushDaemon()
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1171 +0xf1
created by github.com/cockroachdb/cockroach/pkg/util/log.init.0
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:590 +0xf4
goroutine 23 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.signalFlusher()
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:597 +0xab
created by github.com/cockroachdb/cockroach/pkg/util/log.init.0
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:591 +0x10c
goroutine 50 [select, locked to thread]:
runtime.gopark(0x261f2c8, 0x0, 0x24d594c, 0x6, 0x18, 0x1)
/usr/local/go/src/runtime/proc.go:291 +0x11a
runtime.selectgo(0xc420532750, 0xc42052a060)
/usr/local/go/src/runtime/select.go:392 +0xe50
runtime.ensureSigM.func1()
/usr/local/go/src/runtime/signal_unix.go:549 +0x1f4
runtime.goexit()
/usr/local/go/src/runtime/asm_amd64.s:2361 +0x1
goroutine 9847 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421045320)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 771 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420267d40)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 770 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420267c20)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 109 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420266d80)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 110 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4202678c0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 9846 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421045200)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 3504 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4a7e0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 29836 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420bf1d40)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 9003 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421327560)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 15006 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4215a7c20)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 23721 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42356e6c0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 11615 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42102cb40)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 2767 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c23200)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 1345 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42142d200)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 19252 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42010cd80)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 10528 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4209125a0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 18480 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420bf0900)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 1346 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42142d320)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 6096 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420e27320)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 20391 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420e26900)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 29206 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b7a0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 15745 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421326120)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 17110 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b0e0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 20392 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420e26a20)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 27154 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421182d80)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 13104 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b9e0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 22709 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421727560)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 30597 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4aa20)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 23722 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42356e7e0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 24295 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4215a6d80)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 19809 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4214370e0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 22710 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421727680)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 29205 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b680)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 25464 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42113a7e0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 1960 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c22d80)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 1961 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c22ea0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 27792 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421183560)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 17111 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b200)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 14346 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4217267e0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 17791 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4218c6c60)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 14345 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4217266c0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 3503 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4a6c0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 26015 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42102c000)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 9004 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421327680)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 21548 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42325a5a0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 29837 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420bf1e60)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 2768 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c23320)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 4828 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4ac60)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 15007 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4215a7d40)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 13103 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b8c0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 8236 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061a900)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 18481 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420bf0a20)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 28436 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4212c6900)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 20972 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061b320)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 28437 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4212c6a20)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 24885 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42356e120)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 30596 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4a900)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 27793 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421183680)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 16442 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4212f5d40)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 27153 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421182c60)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 10527 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420912480)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 19253 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42010cea0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 17792 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4218c6d80)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 5161 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b440)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 5162 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b560)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 15744 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421326000)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 26573 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42090e480)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 3775 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42174d200)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 3776 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42174d320)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 4567 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c23560)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 20971 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061b200)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 6744 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42090f200)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 24884 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42356e000)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 19810 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421437200)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 6745 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42090f320)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 4827 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4ab40)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 7556 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42113a120)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 26016 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42102c120)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 4566 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c23440)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 8235 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061a7e0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 24296 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4215a6ea0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 22116 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061a120)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 21547 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42325a480)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 16441 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4212f5c20)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 11616 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42102cc60)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 26574 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42090e5a0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 22115 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061a000)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 7555 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42113a000)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 25465 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42113a900)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
goroutine 6097 [chan receive]:
github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420e27440)
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68
created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger
/go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288
ERROR: exit status 2
24 runs completed, 1 failures, over 1m53s
context canceled
```
|
test
|
cli example csv tsv quoting failed under stress sha parameters tags goflags failed test server status runtime go could not parse build timestamp parsing time as cannot parse as server server go monitoring forward clock jumps based on server clock forward jump check enabled base addr validation go server certificate addresses ip dns localhost local cn node base addr validation go web ui certificate addresses ip dns localhost local cn node server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b server node go cluster has been created server server go add additional nodes by specifying join gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag started at storage store go failed initial metrics computation system config not yet available server node go initialized store disk capacity mib available mib used b logicalbytes kib ranges leases queries writes bytesperreplica pmax writesperreplica pmax storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster server node go node started with engine s and attributes server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go could not start heap profiler worker due to directory to store profiles could not be determined server server go starting https server at use server server go starting grpc postgres server at server server go advertising cockroachdb node at server status recorder go available memory from cgroups eib exceeds system memory gib using system memory storage replica command go initiating a split of this range at key system storage replica command go initiating a split of this range at key system nodeliveness storage replica command go initiating a split of this range at key system nodelivenessmax storage replica command go initiating a split of this range at key system tsd storage replica command go initiating a split of this range at key system tse storage replica command go initiating a split of this range at key table systemconfigspan start sql event log go event set cluster setting target info settingname diagnostics reporting enabled value true user root storage replica command go initiating a split of this range at key table storage intent resolver go failed to push during intent resolution failed to push split id key local range table systemconfigspan start rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname version value user root storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname trace debug enable value false user root storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname cluster secret value gen random uuid string user root storage replica command go initiating a split of this range at key table sql event log go event create database target info databasename defaultdb statement create database if not exists defaultdb user root storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event create database target info databasename postgres statement create database if not exists postgres user root server server go done ensuring all necessary migrations have run server server go serving sql connections sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion buildtag startedat localityaddress clusterid startedat lastup server server update go no need to upgrade cluster already at the newest version rpc nodedialer nodedialer go connection to established storage replica proposal go new range lease repl seq start epo pro following repl seq start exp pro server status runtime go runtime stats mib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage replica proposal go new range lease repl seq start epo pro following repl seq start exp pro storage replica command go initiating a split of this range at key table storage replica proposal go new range lease repl seq start epo pro following repl seq start exp pro storage replica proposal go new range lease repl seq start epo pro following repl seq start exp pro storage replica proposal go new range lease repl seq start epo pro following repl seq start exp pro storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica command go initiating a split of this range at key table storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica command go initiating a split of this range at key table server status recorder go available memory from cgroups eib exceeds system memory gib using system memory storage replica proposal go new range lease repl seq start epo pro following repl seq start exp pro storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica command go initiating a split of this range at key table storage replica proposal go new range lease repl seq start epo pro following repl seq start exp pro storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica command go initiating a split of this range at key table storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica command go initiating a split of this range at key table storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro fail example csv tsv quoting panic had ranges at startup expected panic had ranges at startup expected goroutine testing runexample usr local go src testing example go panic usr local go src runtime panic go github com cockroachdb cockroach pkg cli clitest fail go src github com cockroachdb cockroach pkg cli cli test go github com cockroachdb cockroach pkg cli newclitest go src github com cockroachdb cockroach pkg cli cli test go github com cockroachdb cockroach pkg cli example csv tsv quoting go src github com cockroachdb cockroach pkg cli cli test go testing runexample usr local go src testing example go testing runexamples usr local go src testing example go testing m run usr local go src testing testing go github com cockroachdb cockroach pkg cli test testmain go src github com cockroachdb cockroach pkg cli main test go main main testmain go goroutine os signal signal recv usr local go src runtime sigqueue go os signal loop usr local go src os signal signal unix go created by os signal init usr local go src os signal signal unix go goroutine github com cockroachdb cockroach pkg util log flushdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log init go src github com cockroachdb cockroach pkg util log clog go goroutine github com cockroachdb cockroach pkg util log signalflusher go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log init go src github com cockroachdb cockroach pkg util log clog go goroutine runtime gopark usr local go src runtime proc go runtime selectgo usr local go src runtime select go runtime ensuresigm usr local go src runtime signal unix go runtime goexit usr local go src runtime asm s goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go error exit status runs completed failures over context canceled
| 1
|
103,855
| 16,610,450,370
|
IssuesEvent
|
2021-06-02 10:48:58
|
Thanraj/OpenSSL_
|
https://api.github.com/repos/Thanraj/OpenSSL_
|
opened
|
CVE-2015-1787 (Low) detected in opensslOpenSSL_1_0_2
|
security vulnerability
|
## CVE-2015-1787 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opensslOpenSSL_1_0_2</b></p></summary>
<p>
<p>TLS/SSL and crypto library</p>
<p>Library home page: <a href=https://github.com/openssl/openssl.git>https://github.com/openssl/openssl.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Thanraj/OpenSSL_/commit/fa3551a60a0e344c1d79971fbfc4e45646f13720">fa3551a60a0e344c1d79971fbfc4e45646f13720</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>OpenSSL_/ssl/s3_srvr.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The ssl3_get_client_key_exchange function in s3_srvr.c in OpenSSL 1.0.2 before 1.0.2a, when client authentication and an ephemeral Diffie-Hellman ciphersuite are enabled, allows remote attackers to cause a denial of service (daemon crash) via a ClientKeyExchange message with a length of zero.
<p>Publish Date: 2015-03-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-1787>CVE-2015-1787</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>2.6</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-1787">https://nvd.nist.gov/vuln/detail/CVE-2015-1787</a></p>
<p>Release Date: 2015-03-19</p>
<p>Fix Resolution: 1.0.2a</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2015-1787 (Low) detected in opensslOpenSSL_1_0_2 - ## CVE-2015-1787 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opensslOpenSSL_1_0_2</b></p></summary>
<p>
<p>TLS/SSL and crypto library</p>
<p>Library home page: <a href=https://github.com/openssl/openssl.git>https://github.com/openssl/openssl.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Thanraj/OpenSSL_/commit/fa3551a60a0e344c1d79971fbfc4e45646f13720">fa3551a60a0e344c1d79971fbfc4e45646f13720</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>OpenSSL_/ssl/s3_srvr.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The ssl3_get_client_key_exchange function in s3_srvr.c in OpenSSL 1.0.2 before 1.0.2a, when client authentication and an ephemeral Diffie-Hellman ciphersuite are enabled, allows remote attackers to cause a denial of service (daemon crash) via a ClientKeyExchange message with a length of zero.
<p>Publish Date: 2015-03-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-1787>CVE-2015-1787</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>2.6</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-1787">https://nvd.nist.gov/vuln/detail/CVE-2015-1787</a></p>
<p>Release Date: 2015-03-19</p>
<p>Fix Resolution: 1.0.2a</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve low detected in opensslopenssl cve low severity vulnerability vulnerable library opensslopenssl tls ssl and crypto library library home page a href found in head commit a href found in base branch master vulnerable source files openssl ssl srvr c vulnerability details the get client key exchange function in srvr c in openssl before when client authentication and an ephemeral diffie hellman ciphersuite are enabled allows remote attackers to cause a denial of service daemon crash via a clientkeyexchange message with a length of zero publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
69,517
| 14,992,608,921
|
IssuesEvent
|
2021-01-29 10:06:27
|
KorAP/Kalamar
|
https://api.github.com/repos/KorAP/Kalamar
|
closed
|
Establish Content Security Policy
|
security
|
Before integrating the widget plugin mechanism in Kalamar, the server should establish strict rules to limit security risks. Currently there are some violations of basic JS and CSS inline rules, that could easily be fixed.
|
True
|
Establish Content Security Policy - Before integrating the widget plugin mechanism in Kalamar, the server should establish strict rules to limit security risks. Currently there are some violations of basic JS and CSS inline rules, that could easily be fixed.
|
non_test
|
establish content security policy before integrating the widget plugin mechanism in kalamar the server should establish strict rules to limit security risks currently there are some violations of basic js and css inline rules that could easily be fixed
| 0
|
241,367
| 20,118,266,167
|
IssuesEvent
|
2022-02-07 22:06:02
|
eclipse-openj9/openj9
|
https://api.github.com/repos/eclipse-openj9/openj9
|
closed
|
JDK11 MacOS jdk_imageio_0_FAILED - AWTError: WindowServer is not available & others
|
comp:vm test failure os:macos
|
Failure link
------------
From an internal build `Test_openjdk11_j9_extended.openjdk_x86-64_mac/1`
```
05:23:03 openjdk version "11.0.11" 2021-04-20
05:23:03 OpenJDK Runtime Environment AdoptOpenJDK (build 11.0.11+4)
05:23:03 Eclipse OpenJ9 VM AdoptOpenJDK (build master-f021812fb, JRE 11 Mac OS X amd64-64-Bit Compressed References 20210227_55 (JIT enabled, AOT enabled)
05:23:03 OpenJ9 - f021812fb
05:23:03 OMR - eb4f5a875
05:23:03 JCL - 7222dc7018 based on jdk-11.0.11+4)
```
rebuild the failed tests in one link:
07:09:55 https://hyc-runtimes-jenkins.swg-devops.com/job/Grinder/parambuild/?JDK_VERSION=11&JDK_IMPL=openj9&JDK_VENDOR=adoptopenjdk&BUILD_LIST=openjdk&PLATFORM=x86-64_mac_mixed&TARGET=testList%20TESTLIST=jdk_imageio_0,jdk_imageio_1
Optional info
-------------
Failure output (captured from console output)
---------------------------------------------
```
07:03:29 TEST: javax/imageio/plugins/wbmp/WbmpBigDestinationTest.java
07:03:29 TEST JDK: /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdkbinary/j2sdk-image/Contents/Home/bin/..
07:03:29
07:03:29 ACTION: build -- Passed. Build successful
07:03:29 REASON: Named class compiled on demand
07:03:29 TIME: 1.184 seconds
07:03:29 messages:
07:03:29 command: build WbmpBigDestinationTest
07:03:29 reason: Named class compiled on demand
07:03:29 Test directory:
07:03:29 compile: WbmpBigDestinationTest
07:03:29 elapsed time (seconds): 1.184
07:03:29
07:03:29 ACTION: compile -- Passed. Compilation successful
07:03:29 REASON: .class file out of date or does not exist
07:03:29 TIME: 1.184 seconds
07:03:29 messages:
07:03:29 command: compile /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.java
07:03:29 reason: .class file out of date or does not exist
07:03:29 Additional options from @modules: --add-modules java.desktop
07:03:29 Mode: othervm
07:03:29 elapsed time (seconds): 1.184
07:03:29 configuration:
07:03:29 javac compilation environment
07:03:29 add modules: java.desktop
07:03:29 source path: /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp
07:03:29 class path: /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp
07:03:29 /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/classes/1/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.d
07:03:29
07:03:29 rerun:
07:03:29 cd /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/javax/imageio/plugins/wbmp/WbmpBigDestinationTest && \
07:03:29 HOME=/Users/jenkins \
07:03:29 PATH=/bin:/usr/bin:/usr/sbin \
07:03:29 /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdkbinary/j2sdk-image/Contents/Home/bin/../bin/javac \
07:03:29 -J-ea \
07:03:29 -J-esa \
07:03:29 -J-Xmx512m \
07:03:29 -J-XX:+UseCompressedOops \
07:03:29 -J-Dtest.vm.opts='-ea -esa -Xmx512m -XX:+UseCompressedOops' \
07:03:29 -J-Dtest.tool.vm.opts='-J-ea -J-esa -J-Xmx512m -J-XX:+UseCompressedOops' \
07:03:29 -J-Dtest.compiler.opts= \
07:03:29 -J-Dtest.java.opts= \
07:03:29 -J-Dtest.jdk=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdkbinary/j2sdk-image/Contents/Home/bin/.. \
07:03:29 -J-Dcompile.jdk=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdkbinary/j2sdk-image/Contents/Home/bin/.. \
07:03:29 -J-Dtest.timeout.factor=8.0 \
07:03:29 -J-Dtest.root=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk \
07:03:29 -J-Dtest.name=javax/imageio/plugins/wbmp/WbmpBigDestinationTest.java \
07:03:29 -J-Dtest.file=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.java \
07:03:29 -J-Dtest.src=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp \
07:03:29 -J-Dtest.src.path=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp \
07:03:29 -J-Dtest.classes=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/classes/1/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.d \
07:03:29 -J-Dtest.class.path=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/classes/1/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.d \
07:03:29 -J-Dtest.modules=java.desktop \
07:03:29 --add-modules java.desktop \
07:03:29 -d /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/classes/1/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.d \
07:03:29 -sourcepath /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp \
07:03:29 -classpath /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp:/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/classes/1/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.d /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.java
07:03:29 STDOUT:
07:03:29 STDERR:
07:03:29
07:03:29 ACTION: main -- Failed. Execution failed: `main' threw exception: java.awt.AWTError: WindowServer is not available
07:03:29 REASON: Assumed action based on file name: run main WbmpBigDestinationTest
07:03:29 TIME: 0.349 seconds
07:03:29 messages:
07:03:29 command: main WbmpBigDestinationTest
07:03:29 reason: Assumed action based on file name: run main WbmpBigDestinationTest
07:03:29 Mode: othervm
07:03:29 Additional options from @modules: --add-modules java.desktop
07:03:29 elapsed time (seconds): 0.349
07:03:29 configuration:
07:03:29 Boot Layer
07:03:29 add modules: java.desktop
07:03:29
07:03:29 STDOUT:
07:03:29 STDERR:
07:03:29 java.awt.AWTError: WindowServer is not available
07:03:29 at java.desktop/sun.lwawt.macosx.LWCToolkit.<clinit>(LWCToolkit.java:167)
07:03:29 at java.base/java.lang.Class.forNameImpl(Native Method)
07:03:29 at java.base/java.lang.Class.forName(Class.java:337)
07:03:29 at java.desktop/java.awt.Toolkit$2.run(Toolkit.java:588)
07:03:29 at java.desktop/java.awt.Toolkit$2.run(Toolkit.java:583)
07:03:29 at java.base/java.security.AccessController.doPrivileged(AccessController.java:682)
07:03:29 at java.desktop/java.awt.Toolkit.getDefaultToolkit(Toolkit.java:582)
07:03:29 at java.desktop/sun.awt.CGraphicsEnvironment.<clinit>(CGraphicsEnvironment.java:73)
07:03:29 at java.base/java.lang.Class.forNameImpl(Native Method)
07:03:29 at java.base/java.lang.Class.forName(Class.java:337)
07:03:29 at java.desktop/java.awt.GraphicsEnvironment$LocalGE.createGE(GraphicsEnvironment.java:101)
07:03:29 at java.desktop/java.awt.GraphicsEnvironment$LocalGE.<clinit>(GraphicsEnvironment.java:83)
07:03:29 at java.desktop/java.awt.GraphicsEnvironment.getLocalGraphicsEnvironment(GraphicsEnvironment.java:129)
07:03:29 at java.desktop/java.awt.image.BufferedImage.createGraphics(BufferedImage.java:1181)
07:03:29 at WbmpBigDestinationTest.main(WbmpBigDestinationTest.java:51)
07:03:29 at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
07:03:29 at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
07:03:29 at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
07:03:29 at java.base/java.lang.reflect.Method.invoke(Method.java:566)
07:03:29 at com.sun.javatest.regtest.agent.MainWrapper$MainThread.run(MainWrapper.java:127)
07:03:29 at java.base/java.lang.Thread.run(Thread.java:836)
07:03:29
07:03:29 JavaTest Message: Test threw exception: java.awt.AWTError: WindowServer is not available
07:03:29 JavaTest Message: shutting down test
07:03:29
07:03:29 STATUS:Failed.`main' threw exception: java.awt.AWTError: WindowServer is not available
07:03:58 --------------------------------------------------
07:04:27 Test results: passed: 122; failed: 68
07:04:37 Report written to /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/jvmtest/openjdk/report/html/report.html
07:04:37 Results written to /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work
07:04:37 Error: Some tests failed or other problems occurred.
07:04:37
07:04:37 jdk_imageio_0_FAILED
```
It appears related to the test setup.
|
1.0
|
JDK11 MacOS jdk_imageio_0_FAILED - AWTError: WindowServer is not available & others - Failure link
------------
From an internal build `Test_openjdk11_j9_extended.openjdk_x86-64_mac/1`
```
05:23:03 openjdk version "11.0.11" 2021-04-20
05:23:03 OpenJDK Runtime Environment AdoptOpenJDK (build 11.0.11+4)
05:23:03 Eclipse OpenJ9 VM AdoptOpenJDK (build master-f021812fb, JRE 11 Mac OS X amd64-64-Bit Compressed References 20210227_55 (JIT enabled, AOT enabled)
05:23:03 OpenJ9 - f021812fb
05:23:03 OMR - eb4f5a875
05:23:03 JCL - 7222dc7018 based on jdk-11.0.11+4)
```
rebuild the failed tests in one link:
07:09:55 https://hyc-runtimes-jenkins.swg-devops.com/job/Grinder/parambuild/?JDK_VERSION=11&JDK_IMPL=openj9&JDK_VENDOR=adoptopenjdk&BUILD_LIST=openjdk&PLATFORM=x86-64_mac_mixed&TARGET=testList%20TESTLIST=jdk_imageio_0,jdk_imageio_1
Optional info
-------------
Failure output (captured from console output)
---------------------------------------------
```
07:03:29 TEST: javax/imageio/plugins/wbmp/WbmpBigDestinationTest.java
07:03:29 TEST JDK: /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdkbinary/j2sdk-image/Contents/Home/bin/..
07:03:29
07:03:29 ACTION: build -- Passed. Build successful
07:03:29 REASON: Named class compiled on demand
07:03:29 TIME: 1.184 seconds
07:03:29 messages:
07:03:29 command: build WbmpBigDestinationTest
07:03:29 reason: Named class compiled on demand
07:03:29 Test directory:
07:03:29 compile: WbmpBigDestinationTest
07:03:29 elapsed time (seconds): 1.184
07:03:29
07:03:29 ACTION: compile -- Passed. Compilation successful
07:03:29 REASON: .class file out of date or does not exist
07:03:29 TIME: 1.184 seconds
07:03:29 messages:
07:03:29 command: compile /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.java
07:03:29 reason: .class file out of date or does not exist
07:03:29 Additional options from @modules: --add-modules java.desktop
07:03:29 Mode: othervm
07:03:29 elapsed time (seconds): 1.184
07:03:29 configuration:
07:03:29 javac compilation environment
07:03:29 add modules: java.desktop
07:03:29 source path: /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp
07:03:29 class path: /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp
07:03:29 /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/classes/1/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.d
07:03:29
07:03:29 rerun:
07:03:29 cd /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/javax/imageio/plugins/wbmp/WbmpBigDestinationTest && \
07:03:29 HOME=/Users/jenkins \
07:03:29 PATH=/bin:/usr/bin:/usr/sbin \
07:03:29 /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdkbinary/j2sdk-image/Contents/Home/bin/../bin/javac \
07:03:29 -J-ea \
07:03:29 -J-esa \
07:03:29 -J-Xmx512m \
07:03:29 -J-XX:+UseCompressedOops \
07:03:29 -J-Dtest.vm.opts='-ea -esa -Xmx512m -XX:+UseCompressedOops' \
07:03:29 -J-Dtest.tool.vm.opts='-J-ea -J-esa -J-Xmx512m -J-XX:+UseCompressedOops' \
07:03:29 -J-Dtest.compiler.opts= \
07:03:29 -J-Dtest.java.opts= \
07:03:29 -J-Dtest.jdk=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdkbinary/j2sdk-image/Contents/Home/bin/.. \
07:03:29 -J-Dcompile.jdk=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdkbinary/j2sdk-image/Contents/Home/bin/.. \
07:03:29 -J-Dtest.timeout.factor=8.0 \
07:03:29 -J-Dtest.root=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk \
07:03:29 -J-Dtest.name=javax/imageio/plugins/wbmp/WbmpBigDestinationTest.java \
07:03:29 -J-Dtest.file=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.java \
07:03:29 -J-Dtest.src=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp \
07:03:29 -J-Dtest.src.path=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp \
07:03:29 -J-Dtest.classes=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/classes/1/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.d \
07:03:29 -J-Dtest.class.path=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/classes/1/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.d \
07:03:29 -J-Dtest.modules=java.desktop \
07:03:29 --add-modules java.desktop \
07:03:29 -d /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/classes/1/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.d \
07:03:29 -sourcepath /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp \
07:03:29 -classpath /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp:/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/classes/1/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.d /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.java
07:03:29 STDOUT:
07:03:29 STDERR:
07:03:29
07:03:29 ACTION: main -- Failed. Execution failed: `main' threw exception: java.awt.AWTError: WindowServer is not available
07:03:29 REASON: Assumed action based on file name: run main WbmpBigDestinationTest
07:03:29 TIME: 0.349 seconds
07:03:29 messages:
07:03:29 command: main WbmpBigDestinationTest
07:03:29 reason: Assumed action based on file name: run main WbmpBigDestinationTest
07:03:29 Mode: othervm
07:03:29 Additional options from @modules: --add-modules java.desktop
07:03:29 elapsed time (seconds): 0.349
07:03:29 configuration:
07:03:29 Boot Layer
07:03:29 add modules: java.desktop
07:03:29
07:03:29 STDOUT:
07:03:29 STDERR:
07:03:29 java.awt.AWTError: WindowServer is not available
07:03:29 at java.desktop/sun.lwawt.macosx.LWCToolkit.<clinit>(LWCToolkit.java:167)
07:03:29 at java.base/java.lang.Class.forNameImpl(Native Method)
07:03:29 at java.base/java.lang.Class.forName(Class.java:337)
07:03:29 at java.desktop/java.awt.Toolkit$2.run(Toolkit.java:588)
07:03:29 at java.desktop/java.awt.Toolkit$2.run(Toolkit.java:583)
07:03:29 at java.base/java.security.AccessController.doPrivileged(AccessController.java:682)
07:03:29 at java.desktop/java.awt.Toolkit.getDefaultToolkit(Toolkit.java:582)
07:03:29 at java.desktop/sun.awt.CGraphicsEnvironment.<clinit>(CGraphicsEnvironment.java:73)
07:03:29 at java.base/java.lang.Class.forNameImpl(Native Method)
07:03:29 at java.base/java.lang.Class.forName(Class.java:337)
07:03:29 at java.desktop/java.awt.GraphicsEnvironment$LocalGE.createGE(GraphicsEnvironment.java:101)
07:03:29 at java.desktop/java.awt.GraphicsEnvironment$LocalGE.<clinit>(GraphicsEnvironment.java:83)
07:03:29 at java.desktop/java.awt.GraphicsEnvironment.getLocalGraphicsEnvironment(GraphicsEnvironment.java:129)
07:03:29 at java.desktop/java.awt.image.BufferedImage.createGraphics(BufferedImage.java:1181)
07:03:29 at WbmpBigDestinationTest.main(WbmpBigDestinationTest.java:51)
07:03:29 at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
07:03:29 at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
07:03:29 at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
07:03:29 at java.base/java.lang.reflect.Method.invoke(Method.java:566)
07:03:29 at com.sun.javatest.regtest.agent.MainWrapper$MainThread.run(MainWrapper.java:127)
07:03:29 at java.base/java.lang.Thread.run(Thread.java:836)
07:03:29
07:03:29 JavaTest Message: Test threw exception: java.awt.AWTError: WindowServer is not available
07:03:29 JavaTest Message: shutting down test
07:03:29
07:03:29 STATUS:Failed.`main' threw exception: java.awt.AWTError: WindowServer is not available
07:03:58 --------------------------------------------------
07:04:27 Test results: passed: 122; failed: 68
07:04:37 Report written to /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/jvmtest/openjdk/report/html/report.html
07:04:37 Results written to /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work
07:04:37 Error: Some tests failed or other problems occurred.
07:04:37
07:04:37 jdk_imageio_0_FAILED
```
It appears related to the test setup.
|
test
|
macos jdk imageio failed awterror windowserver is not available others failure link from an internal build test extended openjdk mac openjdk version openjdk runtime environment adoptopenjdk build eclipse vm adoptopenjdk build master jre mac os x bit compressed references jit enabled aot enabled omr jcl based on jdk rebuild the failed tests in one link optional info failure output captured from console output test javax imageio plugins wbmp wbmpbigdestinationtest java test jdk users jenkins workspace test extended openjdk mac openjdkbinary image contents home bin action build passed build successful reason named class compiled on demand time seconds messages command build wbmpbigdestinationtest reason named class compiled on demand test directory compile wbmpbigdestinationtest elapsed time seconds action compile passed compilation successful reason class file out of date or does not exist time seconds messages command compile users jenkins workspace test extended openjdk mac openjdk tests openjdk openjdk jdk test jdk javax imageio plugins wbmp wbmpbigdestinationtest java reason class file out of date or does not exist additional options from modules add modules java desktop mode othervm elapsed time seconds configuration javac compilation environment add modules java desktop source path users jenkins workspace test extended openjdk mac openjdk tests openjdk openjdk jdk test jdk javax imageio plugins wbmp class path users jenkins workspace test extended openjdk mac openjdk tests openjdk openjdk jdk test jdk javax imageio plugins wbmp users jenkins workspace test extended openjdk mac openjdk tests tkg output jdk imageio work classes javax imageio plugins wbmp wbmpbigdestinationtest d rerun cd users jenkins workspace test extended openjdk mac openjdk tests tkg output jdk imageio work javax imageio plugins wbmp wbmpbigdestinationtest home users jenkins path bin usr bin usr sbin users jenkins workspace test extended openjdk mac openjdkbinary image contents home bin bin javac j ea j esa j j xx usecompressedoops j dtest vm opts ea esa xx usecompressedoops j dtest tool vm opts j ea j esa j j xx usecompressedoops j dtest compiler opts j dtest java opts j dtest jdk users jenkins workspace test extended openjdk mac openjdkbinary image contents home bin j dcompile jdk users jenkins workspace test extended openjdk mac openjdkbinary image contents home bin j dtest timeout factor j dtest root users jenkins workspace test extended openjdk mac openjdk tests openjdk openjdk jdk test jdk j dtest name javax imageio plugins wbmp wbmpbigdestinationtest java j dtest file users jenkins workspace test extended openjdk mac openjdk tests openjdk openjdk jdk test jdk javax imageio plugins wbmp wbmpbigdestinationtest java j dtest src users jenkins workspace test extended openjdk mac openjdk tests openjdk openjdk jdk test jdk javax imageio plugins wbmp j dtest src path users jenkins workspace test extended openjdk mac openjdk tests openjdk openjdk jdk test jdk javax imageio plugins wbmp j dtest classes users jenkins workspace test extended openjdk mac openjdk tests tkg output jdk imageio work classes javax imageio plugins wbmp wbmpbigdestinationtest d j dtest class path users jenkins workspace test extended openjdk mac openjdk tests tkg output jdk imageio work classes javax imageio plugins wbmp wbmpbigdestinationtest d j dtest modules java desktop add modules java desktop d users jenkins workspace test extended openjdk mac openjdk tests tkg output jdk imageio work classes javax imageio plugins wbmp wbmpbigdestinationtest d sourcepath users jenkins workspace test extended openjdk mac openjdk tests openjdk openjdk jdk test jdk javax imageio plugins wbmp classpath users jenkins workspace test extended openjdk mac openjdk tests openjdk openjdk jdk test jdk javax imageio plugins wbmp users jenkins workspace test extended openjdk mac openjdk tests tkg output jdk imageio work classes javax imageio plugins wbmp wbmpbigdestinationtest d users jenkins workspace test extended openjdk mac openjdk tests openjdk openjdk jdk test jdk javax imageio plugins wbmp wbmpbigdestinationtest java stdout stderr action main failed execution failed main threw exception java awt awterror windowserver is not available reason assumed action based on file name run main wbmpbigdestinationtest time seconds messages command main wbmpbigdestinationtest reason assumed action based on file name run main wbmpbigdestinationtest mode othervm additional options from modules add modules java desktop elapsed time seconds configuration boot layer add modules java desktop stdout stderr java awt awterror windowserver is not available at java desktop sun lwawt macosx lwctoolkit lwctoolkit java at java base java lang class fornameimpl native method at java base java lang class forname class java at java desktop java awt toolkit run toolkit java at java desktop java awt toolkit run toolkit java at java base java security accesscontroller doprivileged accesscontroller java at java desktop java awt toolkit getdefaulttoolkit toolkit java at java desktop sun awt cgraphicsenvironment cgraphicsenvironment java at java base java lang class fornameimpl native method at java base java lang class forname class java at java desktop java awt graphicsenvironment localge createge graphicsenvironment java at java desktop java awt graphicsenvironment localge graphicsenvironment java at java desktop java awt graphicsenvironment getlocalgraphicsenvironment graphicsenvironment java at java desktop java awt image bufferedimage creategraphics bufferedimage java at wbmpbigdestinationtest main wbmpbigdestinationtest java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at com sun javatest regtest agent mainwrapper mainthread run mainwrapper java at java base java lang thread run thread java javatest message test threw exception java awt awterror windowserver is not available javatest message shutting down test status failed main threw exception java awt awterror windowserver is not available test results passed failed report written to users jenkins workspace test extended openjdk mac jvmtest openjdk report html report html results written to users jenkins workspace test extended openjdk mac openjdk tests tkg output jdk imageio work error some tests failed or other problems occurred jdk imageio failed it appears related to the test setup
| 1
|
145,288
| 11,683,459,254
|
IssuesEvent
|
2020-03-05 03:27:35
|
creativecommons/cc-chooser
|
https://api.github.com/repos/creativecommons/cc-chooser
|
opened
|
Add unit and e2e tests for the LicenseCopy component
|
good first issue help wanted test-coverage
|
Unit and e2e tests need to be written for the LicenseCopy component. Unit tests are done with [Jest](https://jestjs.io/), and e2e tests are done with [nightwatch](https://nightwatchjs.org/).
Please remember to test the following things:
- That individual parts of the component are present when appropriate. (unit and e2e)
- That any computed props and methods work properly, if there are any. (unit)
- Any common interactions between the user and component, if there are any. (e2e)
- Any other functionality unique to the component being tested!
### Additional Context
- [./src/components/LicenseCopy.vue](https://github.com/creativecommons/cc-chooser/blob/master/src/components/LicenseCopy.vue)
- [This repo's testing README](https://github.com/creativecommons/cc-chooser/blob/master/tests/README.md)
- [Vue's guide on unit testing](https://vuejs.org/v2/guide/unit-testing.html)
- [Vue's guide on unit testing with VueX](https://vue-test-utils.vuejs.org/guides/using-with-vuex.html)
|
1.0
|
Add unit and e2e tests for the LicenseCopy component - Unit and e2e tests need to be written for the LicenseCopy component. Unit tests are done with [Jest](https://jestjs.io/), and e2e tests are done with [nightwatch](https://nightwatchjs.org/).
Please remember to test the following things:
- That individual parts of the component are present when appropriate. (unit and e2e)
- That any computed props and methods work properly, if there are any. (unit)
- Any common interactions between the user and component, if there are any. (e2e)
- Any other functionality unique to the component being tested!
### Additional Context
- [./src/components/LicenseCopy.vue](https://github.com/creativecommons/cc-chooser/blob/master/src/components/LicenseCopy.vue)
- [This repo's testing README](https://github.com/creativecommons/cc-chooser/blob/master/tests/README.md)
- [Vue's guide on unit testing](https://vuejs.org/v2/guide/unit-testing.html)
- [Vue's guide on unit testing with VueX](https://vue-test-utils.vuejs.org/guides/using-with-vuex.html)
|
test
|
add unit and tests for the licensecopy component unit and tests need to be written for the licensecopy component unit tests are done with and tests are done with please remember to test the following things that individual parts of the component are present when appropriate unit and that any computed props and methods work properly if there are any unit any common interactions between the user and component if there are any any other functionality unique to the component being tested additional context
| 1
|
75,471
| 9,855,940,053
|
IssuesEvent
|
2019-06-19 20:42:24
|
ofpinewood/http-exceptions
|
https://api.github.com/repos/ofpinewood/http-exceptions
|
closed
|
Update documentation and sample project
|
documentation
|
Update the documentation for the projects and add more sample code to the sample project.
|
1.0
|
Update documentation and sample project - Update the documentation for the projects and add more sample code to the sample project.
|
non_test
|
update documentation and sample project update the documentation for the projects and add more sample code to the sample project
| 0
|
349,467
| 31,806,832,428
|
IssuesEvent
|
2023-09-13 14:24:43
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: cdc/pubsub-sink failed
|
C-test-failure O-robot O-roachtest branch-master release-blocker T-cdc
|
roachtest.cdc/pubsub-sink [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/11734739?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/11734739?buildTab=artifacts#/cdc/pubsub-sink) on master @ [0310d119569b59b9efe1b7bbc439c2c604217a4c](https://github.com/cockroachdb/cockroach/commits/0310d119569b59b9efe1b7bbc439c2c604217a4c):
```
(latency_verifier.go:192).assertValid: latency never dropped to acceptable steady level: 1m0s
(monitor.go:153).Wait: monitor failure: monitor user task failed: t.Fatal() was called
test artifacts and logs in: /artifacts/cdc/pubsub-sink/run_1
```
<p>Parameters: <code>ROACHTEST_arch=amd64</code>
, <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=16</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_fs=ext4</code>
, <code>ROACHTEST_localSSD=true</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
See: [Grafana](https://go.crdb.dev/roachtest-grafana/teamcity-11734739/cdc-pubsub-sink/1694515846448/1694517902058)
</p>
</details>
/cc @cockroachdb/cdc
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*cdc/pubsub-sink.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-31422
Epic CRDB-11732
|
2.0
|
roachtest: cdc/pubsub-sink failed - roachtest.cdc/pubsub-sink [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/11734739?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/11734739?buildTab=artifacts#/cdc/pubsub-sink) on master @ [0310d119569b59b9efe1b7bbc439c2c604217a4c](https://github.com/cockroachdb/cockroach/commits/0310d119569b59b9efe1b7bbc439c2c604217a4c):
```
(latency_verifier.go:192).assertValid: latency never dropped to acceptable steady level: 1m0s
(monitor.go:153).Wait: monitor failure: monitor user task failed: t.Fatal() was called
test artifacts and logs in: /artifacts/cdc/pubsub-sink/run_1
```
<p>Parameters: <code>ROACHTEST_arch=amd64</code>
, <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=16</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_fs=ext4</code>
, <code>ROACHTEST_localSSD=true</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
See: [Grafana](https://go.crdb.dev/roachtest-grafana/teamcity-11734739/cdc-pubsub-sink/1694515846448/1694517902058)
</p>
</details>
/cc @cockroachdb/cdc
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*cdc/pubsub-sink.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-31422
Epic CRDB-11732
|
test
|
roachtest cdc pubsub sink failed roachtest cdc pubsub sink with on master latency verifier go assertvalid latency never dropped to acceptable steady level monitor go wait monitor failure monitor user task failed t fatal was called test artifacts and logs in artifacts cdc pubsub sink run parameters roachtest arch roachtest cloud gce roachtest cpu roachtest encrypted false roachtest fs roachtest localssd true roachtest ssd help see see see cc cockroachdb cdc jira issue crdb epic crdb
| 1
|
292,469
| 25,216,141,519
|
IssuesEvent
|
2022-11-14 09:18:07
|
Joystream/pioneer
|
https://api.github.com/repos/Joystream/pioneer
|
closed
|
Set Working Group Lead Reward
|
enhancement scope:proposals qa-task qa-tests-failed
|
[ ] Set Working Group Lead Reward
- Same as when updating reward of a worker in group with given inputs, except signer check
|
1.0
|
Set Working Group Lead Reward - [ ] Set Working Group Lead Reward
- Same as when updating reward of a worker in group with given inputs, except signer check
|
test
|
set working group lead reward set working group lead reward same as when updating reward of a worker in group with given inputs except signer check
| 1
|
568,767
| 16,988,328,393
|
IssuesEvent
|
2021-06-30 16:55:03
|
npm/cli
|
https://api.github.com/repos/npm/cli
|
closed
|
[BUG] progress=false is ignored on npm 7
|
Bug Priority 2 Release 7.x
|
<!--
Note: Please search to see if an issue already exists for your problem: https://github.com/npm/cli/issues
-->
### Current Behavior:
Have an .npmrc with progress=false for aesthetic reasons.
On npm6 that disables the progress bar, on npm7 it does not although the docs say it still should.
This is my full .npmrc if that helps
```
loglevel=http
progress=false
package-lock=false
save=false
```
### Expected Behavior:
Expected the above to disable the progress bar like it does on 6
### Steps To Reproduce:
Save the .npmrc in home and install something. Progress bar is still shown.
### Environment:
- OS: macOS big sur
- Node: 15.7.0
- npm: 7.4.3
|
1.0
|
[BUG] progress=false is ignored on npm 7 - <!--
Note: Please search to see if an issue already exists for your problem: https://github.com/npm/cli/issues
-->
### Current Behavior:
Have an .npmrc with progress=false for aesthetic reasons.
On npm6 that disables the progress bar, on npm7 it does not although the docs say it still should.
This is my full .npmrc if that helps
```
loglevel=http
progress=false
package-lock=false
save=false
```
### Expected Behavior:
Expected the above to disable the progress bar like it does on 6
### Steps To Reproduce:
Save the .npmrc in home and install something. Progress bar is still shown.
### Environment:
- OS: macOS big sur
- Node: 15.7.0
- npm: 7.4.3
|
non_test
|
progress false is ignored on npm note please search to see if an issue already exists for your problem current behavior have an npmrc with progress false for aesthetic reasons on that disables the progress bar on it does not although the docs say it still should this is my full npmrc if that helps loglevel http progress false package lock false save false expected behavior expected the above to disable the progress bar like it does on steps to reproduce save the npmrc in home and install something progress bar is still shown environment os macos big sur node npm
| 0
|
321,146
| 27,508,935,704
|
IssuesEvent
|
2023-03-06 07:07:08
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
ccl/serverccl: TestServerControllerHTTP failed
|
C-test-failure O-robot branch-master
|
ccl/serverccl.TestServerControllerHTTP [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/8933299?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/8933299?buildTab=artifacts#/) on master @ [14b43be03c1c246765be17aa15245493a42c401d](https://github.com/cockroachdb/cockroach/commits/14b43be03c1c246765be17aa15245493a42c401d):
```
=== RUN TestServerControllerHTTP
test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/c101b7a464a1afc1f5af0cd85792187e/logTestServerControllerHTTP452427513
test_log_scope.go:79: use -show-logs to present logs inline
server_controller_test.go:67:
Error Trace: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3439/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/ccl/serverccl/serverccl_test_/serverccl_test.runfiles/com_github_cockroachdb_cockroach/pkg/ccl/serverccl/server_controller_test.go:67
Error: Received unexpected error:
server stop before start
(1) attached stack trace
-- stack trace:
| github.com/cockroachdb/cockroach/pkg/server.(*serverController).startControlledServer.func3.1
| github.com/cockroachdb/cockroach/pkg/server/server_controller_orchestration.go:257
| github.com/cockroachdb/cockroach/pkg/server.(*serverController).startControlledServer.func3
| github.com/cockroachdb/cockroach/pkg/server/server_controller_orchestration.go:306
| github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTaskEx.func2
| github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:470
| runtime.goexit
| GOROOT/src/runtime/asm_amd64.s:1594
Wraps: (2) server stop before start
Error types: (1) *withstack.withStack (2) *errutil.leafError
Test: TestServerControllerHTTP
panic.go:522: -- test log scope end --
test logs left over in: /artifacts/tmp/_tmp/c101b7a464a1afc1f5af0cd85792187e/logTestServerControllerHTTP452427513
--- FAIL: TestServerControllerHTTP (3.77s)
```
<p>Parameters: <code>TAGS=bazel,gss,deadlock</code>
</p>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/server
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestServerControllerHTTP.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
1.0
|
ccl/serverccl: TestServerControllerHTTP failed - ccl/serverccl.TestServerControllerHTTP [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/8933299?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/8933299?buildTab=artifacts#/) on master @ [14b43be03c1c246765be17aa15245493a42c401d](https://github.com/cockroachdb/cockroach/commits/14b43be03c1c246765be17aa15245493a42c401d):
```
=== RUN TestServerControllerHTTP
test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/c101b7a464a1afc1f5af0cd85792187e/logTestServerControllerHTTP452427513
test_log_scope.go:79: use -show-logs to present logs inline
server_controller_test.go:67:
Error Trace: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3439/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/ccl/serverccl/serverccl_test_/serverccl_test.runfiles/com_github_cockroachdb_cockroach/pkg/ccl/serverccl/server_controller_test.go:67
Error: Received unexpected error:
server stop before start
(1) attached stack trace
-- stack trace:
| github.com/cockroachdb/cockroach/pkg/server.(*serverController).startControlledServer.func3.1
| github.com/cockroachdb/cockroach/pkg/server/server_controller_orchestration.go:257
| github.com/cockroachdb/cockroach/pkg/server.(*serverController).startControlledServer.func3
| github.com/cockroachdb/cockroach/pkg/server/server_controller_orchestration.go:306
| github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTaskEx.func2
| github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:470
| runtime.goexit
| GOROOT/src/runtime/asm_amd64.s:1594
Wraps: (2) server stop before start
Error types: (1) *withstack.withStack (2) *errutil.leafError
Test: TestServerControllerHTTP
panic.go:522: -- test log scope end --
test logs left over in: /artifacts/tmp/_tmp/c101b7a464a1afc1f5af0cd85792187e/logTestServerControllerHTTP452427513
--- FAIL: TestServerControllerHTTP (3.77s)
```
<p>Parameters: <code>TAGS=bazel,gss,deadlock</code>
</p>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/server
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestServerControllerHTTP.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
test
|
ccl serverccl testservercontrollerhttp failed ccl serverccl testservercontrollerhttp with on master run testservercontrollerhttp test log scope go test logs captured to artifacts tmp tmp test log scope go use show logs to present logs inline server controller test go error trace home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg ccl serverccl serverccl test serverccl test runfiles com github cockroachdb cockroach pkg ccl serverccl server controller test go error received unexpected error server stop before start attached stack trace stack trace github com cockroachdb cockroach pkg server servercontroller startcontrolledserver github com cockroachdb cockroach pkg server server controller orchestration go github com cockroachdb cockroach pkg server servercontroller startcontrolledserver github com cockroachdb cockroach pkg server server controller orchestration go github com cockroachdb cockroach pkg util stop stopper runasynctaskex github com cockroachdb cockroach pkg util stop stopper go runtime goexit goroot src runtime asm s wraps server stop before start error types withstack withstack errutil leaferror test testservercontrollerhttp panic go test log scope end test logs left over in artifacts tmp tmp fail testservercontrollerhttp parameters tags bazel gss deadlock help see also cc cockroachdb server
| 1
|
105,759
| 9,100,592,774
|
IssuesEvent
|
2019-02-20 08:58:48
|
intermine/intermine
|
https://api.github.com/repos/intermine/intermine
|
closed
|
OMIM - failed on bad line
|
data in-progress please-test
|
```
Caused by: java.lang.RuntimeException: bad line: '[Beta-glycopyranoside tasting], 617956 (3) {Alcohol dependence, susceptibility to}, 103780 (3) TAS2R16, T2R16, BGLPT 604867 7q31.32'
at org.intermine.bio.dataconversion.OmimConverter.processMorbidMapFile(OmimConverter.java:168)
at org.intermine.bio.dataconversion.OmimConverter.process(OmimConverter.java:101)
at org.intermine.task.DirectoryConverterTask.execute(DirectoryConverterTask.java:104)
... 42 more
```
|
1.0
|
OMIM - failed on bad line - ```
Caused by: java.lang.RuntimeException: bad line: '[Beta-glycopyranoside tasting], 617956 (3) {Alcohol dependence, susceptibility to}, 103780 (3) TAS2R16, T2R16, BGLPT 604867 7q31.32'
at org.intermine.bio.dataconversion.OmimConverter.processMorbidMapFile(OmimConverter.java:168)
at org.intermine.bio.dataconversion.OmimConverter.process(OmimConverter.java:101)
at org.intermine.task.DirectoryConverterTask.execute(DirectoryConverterTask.java:104)
... 42 more
```
|
test
|
omim failed on bad line caused by java lang runtimeexception bad line alcohol dependence susceptibility to bglpt at org intermine bio dataconversion omimconverter processmorbidmapfile omimconverter java at org intermine bio dataconversion omimconverter process omimconverter java at org intermine task directoryconvertertask execute directoryconvertertask java more
| 1
|
93,021
| 8,391,895,924
|
IssuesEvent
|
2018-10-09 16:05:18
|
kabirbaidhya/boss
|
https://api.github.com/repos/kabirbaidhya/boss
|
closed
|
Write tests for boss api and utils
|
good first issue hacktoberfest help wanted test
|
Write tests for:
* Modules and functions under `boss.api`
* Code under `boss.util`
* Any other code that requires tests
|
1.0
|
Write tests for boss api and utils - Write tests for:
* Modules and functions under `boss.api`
* Code under `boss.util`
* Any other code that requires tests
|
test
|
write tests for boss api and utils write tests for modules and functions under boss api code under boss util any other code that requires tests
| 1
|
487,676
| 14,050,375,076
|
IssuesEvent
|
2020-11-02 11:39:48
|
drashland/dmm
|
https://api.github.com/repos/drashland/dmm
|
opened
|
Support https://raw.githubusercontent.com
|
Priority: Low Type: Enhancement
|
## Summary
What:
Alongside supporting deno.land and x.nest.land, add support for https://raw.githubusercontent
Why:
Mainly, so we can use dmm to update dependencies pulled from the services repo, but it does have it's own good use
## Acceptance Criteria
Below is a list of tasks that must be completed before this issue can be closed.
- [ ] Documentation (link)
- [ ] Tests
- [ ] The feature
## Example Pseudo Code (for implementation)
```typescript
// Add example pseudo code for implementation
```
|
1.0
|
Support https://raw.githubusercontent.com - ## Summary
What:
Alongside supporting deno.land and x.nest.land, add support for https://raw.githubusercontent
Why:
Mainly, so we can use dmm to update dependencies pulled from the services repo, but it does have it's own good use
## Acceptance Criteria
Below is a list of tasks that must be completed before this issue can be closed.
- [ ] Documentation (link)
- [ ] Tests
- [ ] The feature
## Example Pseudo Code (for implementation)
```typescript
// Add example pseudo code for implementation
```
|
non_test
|
support summary what alongside supporting deno land and x nest land add support for why mainly so we can use dmm to update dependencies pulled from the services repo but it does have it s own good use acceptance criteria below is a list of tasks that must be completed before this issue can be closed documentation link tests the feature example pseudo code for implementation typescript add example pseudo code for implementation
| 0
|
147,712
| 23,260,569,060
|
IssuesEvent
|
2022-08-04 13:11:33
|
excalidraw/excalidraw
|
https://api.github.com/repos/excalidraw/excalidraw
|
closed
|
Flip horizontal does not work on linear elements since the redesign (#5501). Flip vertical works.
|
bug arrow-redesign
| ERROR: type should be string, got "\r\nhttps://user-images.githubusercontent.com/14358394/182805730-1ad74bca-da46-4383-b03b-6d89f2a68ae8.mp4\r\n\r\n"
|
1.0
|
Flip horizontal does not work on linear elements since the redesign (#5501). Flip vertical works. -
https://user-images.githubusercontent.com/14358394/182805730-1ad74bca-da46-4383-b03b-6d89f2a68ae8.mp4
|
non_test
|
flip horizontal does not work on linear elements since the redesign flip vertical works
| 0
|
293,471
| 22,059,411,895
|
IssuesEvent
|
2022-05-30 15:49:52
|
ms-club-sliit/ms-meeting-manager
|
https://api.github.com/repos/ms-club-sliit/ms-meeting-manager
|
opened
|
README for the repository
|
documentation help wanted
|
Need to create a Readme file for this project. The readme file should contains the following information.
* Technologies that used in the project
* How to run the project
* How to contribute the project
* Recent contributors
* CI/ CD pipeline status
|
1.0
|
README for the repository - Need to create a Readme file for this project. The readme file should contains the following information.
* Technologies that used in the project
* How to run the project
* How to contribute the project
* Recent contributors
* CI/ CD pipeline status
|
non_test
|
readme for the repository need to create a readme file for this project the readme file should contains the following information technologies that used in the project how to run the project how to contribute the project recent contributors ci cd pipeline status
| 0
|
120,167
| 12,060,612,333
|
IssuesEvent
|
2020-04-15 21:34:48
|
deathlyrage/pot-demo-bugs
|
https://api.github.com/repos/deathlyrage/pot-demo-bugs
|
closed
|
Camera spawns in Trees (stuck) and Stuck Daspleto cannot move
|
documentation needs testing
|
Spawned in to try AI mode and find my camera unable to move after having been spawned inside the crown of a tree

as well as daspletosaurus being stuck between a rock and a tree the moment it spawned in

|
1.0
|
Camera spawns in Trees (stuck) and Stuck Daspleto cannot move - Spawned in to try AI mode and find my camera unable to move after having been spawned inside the crown of a tree

as well as daspletosaurus being stuck between a rock and a tree the moment it spawned in

|
non_test
|
camera spawns in trees stuck and stuck daspleto cannot move spawned in to try ai mode and find my camera unable to move after having been spawned inside the crown of a tree as well as daspletosaurus being stuck between a rock and a tree the moment it spawned in
| 0
|
811,365
| 30,285,275,602
|
IssuesEvent
|
2023-07-08 15:43:33
|
SuffolkLITLab/ALKiln
|
https://api.github.com/repos/SuffolkLITLab/ALKiln
|
closed
|
Restore multi-language tests set by env vars
|
priority
|
As per user request, restore running language tests and write internal tests for such (though I'm not sure how we check that the right tests have been run). These should primarily be triggered manually through the workflow dispatch, though allowing an env var for them might still be useful to carry over from previous behavior. Include:
- [x] Access language page url as per #360
- [x] Add language to Scenario tags
- [x] Internal tests of some kind, maybe using tags in Scenario name section. Check that certain text appears multiple times won't be sufficient as errors can also cause that behavior. [Maybe create our own translation file with different text that we can then look for. Use [MADE](https://github.com/GBLS/docassemble-MAEvictionDefense)'s methods as a model for that as that is our primary user need at the moment?]
- [x] ~[If we're creating the language files each time] Ensure InThePlayground deletes the language files during post-test cleanup.~
Should close #360.
Edit:
- [ ] Add [new issue to document this in the documentation repo](https://github.com/SuffolkLITLab/docassemble-AssemblyLine-documentation/issues/new)
|
1.0
|
Restore multi-language tests set by env vars - As per user request, restore running language tests and write internal tests for such (though I'm not sure how we check that the right tests have been run). These should primarily be triggered manually through the workflow dispatch, though allowing an env var for them might still be useful to carry over from previous behavior. Include:
- [x] Access language page url as per #360
- [x] Add language to Scenario tags
- [x] Internal tests of some kind, maybe using tags in Scenario name section. Check that certain text appears multiple times won't be sufficient as errors can also cause that behavior. [Maybe create our own translation file with different text that we can then look for. Use [MADE](https://github.com/GBLS/docassemble-MAEvictionDefense)'s methods as a model for that as that is our primary user need at the moment?]
- [x] ~[If we're creating the language files each time] Ensure InThePlayground deletes the language files during post-test cleanup.~
Should close #360.
Edit:
- [ ] Add [new issue to document this in the documentation repo](https://github.com/SuffolkLITLab/docassemble-AssemblyLine-documentation/issues/new)
|
non_test
|
restore multi language tests set by env vars as per user request restore running language tests and write internal tests for such though i m not sure how we check that the right tests have been run these should primarily be triggered manually through the workflow dispatch though allowing an env var for them might still be useful to carry over from previous behavior include access language page url as per add language to scenario tags internal tests of some kind maybe using tags in scenario name section check that certain text appears multiple times won t be sufficient as errors can also cause that behavior methods as a model for that as that is our primary user need at the moment ensure intheplayground deletes the language files during post test cleanup should close edit add
| 0
|
821,730
| 30,833,468,475
|
IssuesEvent
|
2023-08-02 05:02:22
|
GSM-MSG/SMS-Android
|
https://api.github.com/repos/GSM-MSG/SMS-Android
|
opened
|
Show the snack bar when a screen capture is detected.
|
0️⃣ Priority: Critical ✨ Feature
|
### Describe
화면 캡쳐가 감지되었을 때 스낵바를 표시합니다.
### Additional
_No response_
|
1.0
|
Show the snack bar when a screen capture is detected. - ### Describe
화면 캡쳐가 감지되었을 때 스낵바를 표시합니다.
### Additional
_No response_
|
non_test
|
show the snack bar when a screen capture is detected describe 화면 캡쳐가 감지되었을 때 스낵바를 표시합니다 additional no response
| 0
|
610,596
| 18,911,904,600
|
IssuesEvent
|
2021-11-16 14:53:38
|
googleapis/google-api-dotnet-client
|
https://api.github.com/repos/googleapis/google-api-dotnet-client
|
closed
|
How to Create Custom Dimension and Custom metrics in google analytics
|
type: question priority: p2 api: analytics
|
We have tried below steps to create Custom Dimension and Custom metrics from code side.
1) We have installed Nuget package "Install-Package Google.Apis.Analytics.v3 -Version 1.55.0.1679" into solution and Enabled Google Analytics API's in "Google Analytics Account".
2) We have created service account for google analytics project and Generated ".p12" file for google analytic account authentication and programmatically we are validating the .p12 file with EmailId. Please find the below code.
===================================================================
var certificate = new X509Certificate2(keyPath, "notasecret", X509KeyStorageFlags.Exportable);
var credentials = new ServiceAccountCredential(
new ServiceAccountCredential.Initializer(emailID)
{
Scopes = new[] { AnalyticsService.Scope.AnalyticsEdit}
}.FromCertificate(certificate));
var service = new AnalyticsService(new BaseClientService.Initializer()
{
HttpClientInitializer = credentials,
//ApplicationName = "Analytics API Sample",
});
Note:
We have tried with Analytics, AnalyticsEdit, and AnalyticsManageUsers Scope.
===================================================================
3) We are creating custom dimension using below code.
public static string CreateCustomDimension(AnalyticsService ga, string accountId, string webPropertyId, string dimensionName, string scope, bool active)
{
List customDimensions = ga.Management.CustomDimensions.List(accountId, webPropertyId).Execute().Items as List;
if (!customDimensions.Select(c => c.Name).Contains(dimensionName))
{
CustomDimension body = new CustomDimension
{
AccountId = accountId,
WebPropertyId = webPropertyId,
Name = dimensionName,
Scope = scope,
Active = active
};
ManagementResource.CustomDimensionsResource.InsertRequest request = ga.Management.CustomDimensions.Insert(body, accountId, webPropertyId);
CustomDimension newDimension = request.Execute();
List<CustomDimension> newDimensions = ga.Management.CustomDimensions.List(accountId, webPropertyId).Execute().Items as List<CustomDimension>;
if (newDimensions.Select(d => d.Id).Contains(newDimension.Id)) return "Success: Custom Dimension Added";
else return "Failure: Unknown";
}
else
{
return "Custom Dimension Already Exists";
}
}
we are getting exception at [ga.Management.CustomDimensions.List(accountId, webPropertyId).Execute().Items as List;] in above method.
Exception Details:
Request had insufficient authentication scopes. [403]
Errors [Message[Insufficient Permission] Location[ - ] Reason[insufficientPermissions] Domain[global]]
|
1.0
|
How to Create Custom Dimension and Custom metrics in google analytics - We have tried below steps to create Custom Dimension and Custom metrics from code side.
1) We have installed Nuget package "Install-Package Google.Apis.Analytics.v3 -Version 1.55.0.1679" into solution and Enabled Google Analytics API's in "Google Analytics Account".
2) We have created service account for google analytics project and Generated ".p12" file for google analytic account authentication and programmatically we are validating the .p12 file with EmailId. Please find the below code.
===================================================================
var certificate = new X509Certificate2(keyPath, "notasecret", X509KeyStorageFlags.Exportable);
var credentials = new ServiceAccountCredential(
new ServiceAccountCredential.Initializer(emailID)
{
Scopes = new[] { AnalyticsService.Scope.AnalyticsEdit}
}.FromCertificate(certificate));
var service = new AnalyticsService(new BaseClientService.Initializer()
{
HttpClientInitializer = credentials,
//ApplicationName = "Analytics API Sample",
});
Note:
We have tried with Analytics, AnalyticsEdit, and AnalyticsManageUsers Scope.
===================================================================
3) We are creating custom dimension using below code.
public static string CreateCustomDimension(AnalyticsService ga, string accountId, string webPropertyId, string dimensionName, string scope, bool active)
{
List customDimensions = ga.Management.CustomDimensions.List(accountId, webPropertyId).Execute().Items as List;
if (!customDimensions.Select(c => c.Name).Contains(dimensionName))
{
CustomDimension body = new CustomDimension
{
AccountId = accountId,
WebPropertyId = webPropertyId,
Name = dimensionName,
Scope = scope,
Active = active
};
ManagementResource.CustomDimensionsResource.InsertRequest request = ga.Management.CustomDimensions.Insert(body, accountId, webPropertyId);
CustomDimension newDimension = request.Execute();
List<CustomDimension> newDimensions = ga.Management.CustomDimensions.List(accountId, webPropertyId).Execute().Items as List<CustomDimension>;
if (newDimensions.Select(d => d.Id).Contains(newDimension.Id)) return "Success: Custom Dimension Added";
else return "Failure: Unknown";
}
else
{
return "Custom Dimension Already Exists";
}
}
we are getting exception at [ga.Management.CustomDimensions.List(accountId, webPropertyId).Execute().Items as List;] in above method.
Exception Details:
Request had insufficient authentication scopes. [403]
Errors [Message[Insufficient Permission] Location[ - ] Reason[insufficientPermissions] Domain[global]]
|
non_test
|
how to create custom dimension and custom metrics in google analytics we have tried below steps to create custom dimension and custom metrics from code side we have installed nuget package install package google apis analytics version into solution and enabled google analytics api s in google analytics account we have created service account for google analytics project and generated file for google analytic account authentication and programmatically we are validating the file with emailid please find the below code var certificate new keypath notasecret exportable var credentials new serviceaccountcredential new serviceaccountcredential initializer emailid scopes new analyticsservice scope analyticsedit fromcertificate certificate var service new analyticsservice new baseclientservice initializer httpclientinitializer credentials applicationname analytics api sample note we have tried with analytics analyticsedit and analyticsmanageusers scope we are creating custom dimension using below code public static string createcustomdimension analyticsservice ga string accountid string webpropertyid string dimensionname string scope bool active list customdimensions ga management customdimensions list accountid webpropertyid execute items as list if customdimensions select c c name contains dimensionname customdimension body new customdimension accountid accountid webpropertyid webpropertyid name dimensionname scope scope active active managementresource customdimensionsresource insertrequest request ga management customdimensions insert body accountid webpropertyid customdimension newdimension request execute list newdimensions ga management customdimensions list accountid webpropertyid execute items as list if newdimensions select d d id contains newdimension id return success custom dimension added else return failure unknown else return custom dimension already exists we are getting exception at in above method exception details request had insufficient authentication scopes errors location reason domain
| 0
|
65,592
| 12,625,364,989
|
IssuesEvent
|
2020-06-14 11:32:14
|
intellij-rust/intellij-rust
|
https://api.github.com/repos/intellij-rust/intellij-rust
|
opened
|
No E0368/E0369 for binary operations
|
feature subsystem::code insight
|
<!--
Hello and thank you for the issue!
If you would like to report a bug, we have added some points below that you can fill out.
Feel free to remove all the irrelevant text to request a new feature.
-->
## Environment
* **IntelliJ Rust plugin version:** 0.2.125.3158-202-nightly
* **Rust toolchain version:** 1.46.0-nightly (0262de554 2020-06-07) x86_64-apple-darwin
* **IDE name and version:** IntelliJ IDEA 2020.2 EAP Ultimate Edition (IU-202.5428.22)
* **Operating system:** macOS 10.13.6
## Problem description
If the compiler can't find a proper implementation of trait responsible for the particular binary operator, it produces an error. Currently, the plugin says nothing in such cases.
## Steps to reproduce
```rust
struct Foo(u8);
fn add(lhs: Foo, rhs: Foo) {
lhs + rhs; // error[E0369]: cannot add `Foo` to `Foo`
}
fn munis_assign(mut lhs: Foo, rhs: Foo) {
lhs -= rhs; // error[E0368]: binary assignment operation `-=` cannot be applied to type `Foo`
}
```
<!--
Please include as much of your codebase as needed to reproduce the error.
If the relevant files are large, please provide a link to a public repository or a [Gist](https://gist.github.com/).
-->
|
1.0
|
No E0368/E0369 for binary operations - <!--
Hello and thank you for the issue!
If you would like to report a bug, we have added some points below that you can fill out.
Feel free to remove all the irrelevant text to request a new feature.
-->
## Environment
* **IntelliJ Rust plugin version:** 0.2.125.3158-202-nightly
* **Rust toolchain version:** 1.46.0-nightly (0262de554 2020-06-07) x86_64-apple-darwin
* **IDE name and version:** IntelliJ IDEA 2020.2 EAP Ultimate Edition (IU-202.5428.22)
* **Operating system:** macOS 10.13.6
## Problem description
If the compiler can't find a proper implementation of trait responsible for the particular binary operator, it produces an error. Currently, the plugin says nothing in such cases.
## Steps to reproduce
```rust
struct Foo(u8);
fn add(lhs: Foo, rhs: Foo) {
lhs + rhs; // error[E0369]: cannot add `Foo` to `Foo`
}
fn munis_assign(mut lhs: Foo, rhs: Foo) {
lhs -= rhs; // error[E0368]: binary assignment operation `-=` cannot be applied to type `Foo`
}
```
<!--
Please include as much of your codebase as needed to reproduce the error.
If the relevant files are large, please provide a link to a public repository or a [Gist](https://gist.github.com/).
-->
|
non_test
|
no for binary operations hello and thank you for the issue if you would like to report a bug we have added some points below that you can fill out feel free to remove all the irrelevant text to request a new feature environment intellij rust plugin version nightly rust toolchain version nightly apple darwin ide name and version intellij idea eap ultimate edition iu operating system macos problem description if the compiler can t find a proper implementation of trait responsible for the particular binary operator it produces an error currently the plugin says nothing in such cases steps to reproduce rust struct foo fn add lhs foo rhs foo lhs rhs error cannot add foo to foo fn munis assign mut lhs foo rhs foo lhs rhs error binary assignment operation cannot be applied to type foo please include as much of your codebase as needed to reproduce the error if the relevant files are large please provide a link to a public repository or a
| 0
|
20,861
| 6,114,254,682
|
IssuesEvent
|
2017-06-22 00:24:09
|
ganeti/ganeti
|
https://api.github.com/repos/ganeti/ganeti
|
closed
|
burnin instructions & possible bug
|
imported_from_google_code Status:Invalid
|
Originally reported of Google Code with ID 106.
```
What software version are you running? Please provide the output of "gnt-
cluster --version" and "gnt-cluster version".
<b>What distribution are you using?</b>
gnt-cluster (ganeti) 2.1.1
software version 2.1.1
internode protocol: 30
configuration format: 201000
os api version: 15
export interface: 0
comments relate to latest debootstrap download (v 0.9)
note: doing install from source, download from code.google.com onto Debian
Lenny
immediately post installation, following the instructions to run burnin
(http://ganeti-doc.googlecode.com/svn/ganeti-2.1/html/admin.html#burnin-label)
results in an error "can't get the OS list"
turns out that the instructions in
http://ganeti-doc.googlecode.com/svn/ganeti-2.1/html/install.html#installing-the-operating-system-support-packages
say to simply do "make" followed by "make install" - which does not put the
operating system in /srv/ganeti/os
the instructions in the README file are more specific re. configure options
to put files in the proper places
following a proper config, make, make install:
gnt-os diagnose shows:
OS: debootstrap [global status: valid]
but,
burnin -o debootstrap test
now responds with a new error:
OS 'debootstrap' not found
```
Originally added on 2010-05-02 11:21:02 +0000 UTC.
|
1.0
|
burnin instructions & possible bug - Originally reported of Google Code with ID 106.
```
What software version are you running? Please provide the output of "gnt-
cluster --version" and "gnt-cluster version".
<b>What distribution are you using?</b>
gnt-cluster (ganeti) 2.1.1
software version 2.1.1
internode protocol: 30
configuration format: 201000
os api version: 15
export interface: 0
comments relate to latest debootstrap download (v 0.9)
note: doing install from source, download from code.google.com onto Debian
Lenny
immediately post installation, following the instructions to run burnin
(http://ganeti-doc.googlecode.com/svn/ganeti-2.1/html/admin.html#burnin-label)
results in an error "can't get the OS list"
turns out that the instructions in
http://ganeti-doc.googlecode.com/svn/ganeti-2.1/html/install.html#installing-the-operating-system-support-packages
say to simply do "make" followed by "make install" - which does not put the
operating system in /srv/ganeti/os
the instructions in the README file are more specific re. configure options
to put files in the proper places
following a proper config, make, make install:
gnt-os diagnose shows:
OS: debootstrap [global status: valid]
but,
burnin -o debootstrap test
now responds with a new error:
OS 'debootstrap' not found
```
Originally added on 2010-05-02 11:21:02 +0000 UTC.
|
non_test
|
burnin instructions possible bug originally reported of google code with id what software version are you running please provide the output of gnt cluster version and gnt cluster version what distribution are you using gnt cluster ganeti software version internode protocol configuration format os api version export interface comments relate to latest debootstrap download v note doing install from source download from code google com onto debian lenny immediately post installation following the instructions to run burnin results in an error can t get the os list turns out that the instructions in say to simply do make followed by make install which does not put the operating system in srv ganeti os the instructions in the readme file are more specific re configure options to put files in the proper places following a proper config make make install gnt os diagnose shows os debootstrap but burnin o debootstrap test now responds with a new error os debootstrap not found originally added on utc
| 0
|
87,624
| 25,165,008,263
|
IssuesEvent
|
2022-11-10 20:00:17
|
libjxl/libjxl
|
https://api.github.com/repos/libjxl/libjxl
|
closed
|
StoreInterleaved: 2 3 4
|
building/portability unrelated to 1.0 highway
|
**Describe the bug**
in order to compile `main` branch I had to comment out all lines with `StoreInterleaved2` `StoreInterleaved3` `StoreInterleaved4`
(in `dec_group_jpeg.cc` and `stage_write.cc`)
**To Reproduce**
try to compile `main` branch
**Expected behavior**
`main` branch compiles successfully
**Environment**
- OS: Gentoo Linux
- Compiler version: gcc-12.2.1
- CPU type: x86_64
- cjxl/djxl version string: JPEG XL encoder v0.8.0 [AVX2]
**Additional context**
`emerge =media-libs/libjxl-9999` with `-DJXL_HWY_DISABLED_TARGETS_FORCED:BOOL=ON`
|
1.0
|
StoreInterleaved: 2 3 4 - **Describe the bug**
in order to compile `main` branch I had to comment out all lines with `StoreInterleaved2` `StoreInterleaved3` `StoreInterleaved4`
(in `dec_group_jpeg.cc` and `stage_write.cc`)
**To Reproduce**
try to compile `main` branch
**Expected behavior**
`main` branch compiles successfully
**Environment**
- OS: Gentoo Linux
- Compiler version: gcc-12.2.1
- CPU type: x86_64
- cjxl/djxl version string: JPEG XL encoder v0.8.0 [AVX2]
**Additional context**
`emerge =media-libs/libjxl-9999` with `-DJXL_HWY_DISABLED_TARGETS_FORCED:BOOL=ON`
|
non_test
|
storeinterleaved describe the bug in order to compile main branch i had to comment out all lines with in dec group jpeg cc and stage write cc to reproduce try to compile main branch expected behavior main branch compiles successfully environment os gentoo linux compiler version gcc cpu type cjxl djxl version string jpeg xl encoder additional context emerge media libs libjxl with djxl hwy disabled targets forced bool on
| 0
|
41,907
| 2,869,088,009
|
IssuesEvent
|
2015-06-05 23:14:12
|
dart-lang/polymer-dart
|
https://api.github.com/repos/dart-lang/polymer-dart
|
opened
|
Provide a way to easily trace async stack traces
|
bug Priority-Medium
|
<a href="https://github.com/sigmundch"><img src="https://avatars.githubusercontent.com/u/2049220?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [sigmundch](https://github.com/sigmundch)**
_Originally opened as dart-lang/sdk#20322_
----
maybe a toggle UI like the one we have for logger?
|
1.0
|
Provide a way to easily trace async stack traces - <a href="https://github.com/sigmundch"><img src="https://avatars.githubusercontent.com/u/2049220?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [sigmundch](https://github.com/sigmundch)**
_Originally opened as dart-lang/sdk#20322_
----
maybe a toggle UI like the one we have for logger?
|
non_test
|
provide a way to easily trace async stack traces issue by originally opened as dart lang sdk maybe a toggle ui like the one we have for logger
| 0
|
323,802
| 27,753,378,396
|
IssuesEvent
|
2023-03-15 23:04:15
|
sanktjodel/cctest1
|
https://api.github.com/repos/sanktjodel/cctest1
|
opened
|
Fix "method_complexity" issue in Signing.java
|
test1 test2'"><h1>tt
|
Method `signRequest` has a Cognitive Complexity of 34 (exceeds 5 allowed). Consider refactoring.
https://codeclimate.com/github/sanktjodel/cctest1/Signing.java#issue_64124ec1a9c49c0001000012
|
2.0
|
Fix "method_complexity" issue in Signing.java - Method `signRequest` has a Cognitive Complexity of 34 (exceeds 5 allowed). Consider refactoring.
https://codeclimate.com/github/sanktjodel/cctest1/Signing.java#issue_64124ec1a9c49c0001000012
|
test
|
fix method complexity issue in signing java method signrequest has a cognitive complexity of exceeds allowed consider refactoring
| 1
|
266,972
| 8,377,573,801
|
IssuesEvent
|
2018-10-06 02:53:54
|
medic/medic-webapp
|
https://api.github.com/repos/medic/medic-webapp
|
closed
|
Horti crashes on subsequent upgrade
|
Priority: 1 - High Status: 5 - Ready Type: Bug horticulturalist
|
Each time an app (api or sentinel) is updated, horti creates a symlink to the previous running version (called `old`).
When doing multiple upgrades, this `old` symlink is not removed prior to attempting to write it again resulting in the following fatal error (which crashes horti):
```
horti:debug Primary ddoc written +3s
Updating symlinks for changed apps… [ { name: 'medic-api',
attachmentName: 'medic-api-0.1.0.tgz',
digest: 'md5-wmKqGbAc5bAKugPtyXZdKQ==',
deployPath: [Function: deployPath] } ]
********FATAL********
{ Error: EEXIST: file already exists, symlink '/home/user/.horticulturalist/deployments/medic-api/md5-vU2CqBOkCe+HMwOM3oCb+w==' -> '/home/user/.horticulturalist/deployments/medic-api/old'
at Object.fs.symlinkSync (fs.js:1031:18)
at Promise.all.changedApps.map.app (/usr/lib/node_modules/horticulturalist/src/install/deploySteps.js:112:14)
at Array.map (<anonymous>)
at updateSymlink (/usr/lib/node_modules/horticulturalist/src/install/deploySteps.js:105:36)
at Promise.resolve.then.then (/usr/lib/node_modules/horticulturalist/src/install/deploySteps.js:153:25)
at <anonymous>
at process._tickCallback (internal/process/next_tick.js:182:7)
errno: -17,
code: 'EEXIST',
syscall: 'symlink',
path: '/home/user/.horticulturalist/deployments/medic-api/md5-vU2CqBOkCe+HMwOM3oCb+w==',
dest: '/home/user/.horticulturalist/deployments/medic-api/old' }
```
|
1.0
|
Horti crashes on subsequent upgrade - Each time an app (api or sentinel) is updated, horti creates a symlink to the previous running version (called `old`).
When doing multiple upgrades, this `old` symlink is not removed prior to attempting to write it again resulting in the following fatal error (which crashes horti):
```
horti:debug Primary ddoc written +3s
Updating symlinks for changed apps… [ { name: 'medic-api',
attachmentName: 'medic-api-0.1.0.tgz',
digest: 'md5-wmKqGbAc5bAKugPtyXZdKQ==',
deployPath: [Function: deployPath] } ]
********FATAL********
{ Error: EEXIST: file already exists, symlink '/home/user/.horticulturalist/deployments/medic-api/md5-vU2CqBOkCe+HMwOM3oCb+w==' -> '/home/user/.horticulturalist/deployments/medic-api/old'
at Object.fs.symlinkSync (fs.js:1031:18)
at Promise.all.changedApps.map.app (/usr/lib/node_modules/horticulturalist/src/install/deploySteps.js:112:14)
at Array.map (<anonymous>)
at updateSymlink (/usr/lib/node_modules/horticulturalist/src/install/deploySteps.js:105:36)
at Promise.resolve.then.then (/usr/lib/node_modules/horticulturalist/src/install/deploySteps.js:153:25)
at <anonymous>
at process._tickCallback (internal/process/next_tick.js:182:7)
errno: -17,
code: 'EEXIST',
syscall: 'symlink',
path: '/home/user/.horticulturalist/deployments/medic-api/md5-vU2CqBOkCe+HMwOM3oCb+w==',
dest: '/home/user/.horticulturalist/deployments/medic-api/old' }
```
|
non_test
|
horti crashes on subsequent upgrade each time an app api or sentinel is updated horti creates a symlink to the previous running version called old when doing multiple upgrades this old symlink is not removed prior to attempting to write it again resulting in the following fatal error which crashes horti horti debug primary ddoc written updating symlinks for changed apps… name medic api attachmentname medic api tgz digest deploypath fatal error eexist file already exists symlink home user horticulturalist deployments medic api w home user horticulturalist deployments medic api old at object fs symlinksync fs js at promise all changedapps map app usr lib node modules horticulturalist src install deploysteps js at array map at updatesymlink usr lib node modules horticulturalist src install deploysteps js at promise resolve then then usr lib node modules horticulturalist src install deploysteps js at at process tickcallback internal process next tick js errno code eexist syscall symlink path home user horticulturalist deployments medic api w dest home user horticulturalist deployments medic api old
| 0
|
158,570
| 6,031,907,622
|
IssuesEvent
|
2017-06-09 01:09:57
|
chartjs/Chart.js
|
https://api.github.com/repos/chartjs/Chart.js
|
closed
|
First and last bars display problem
|
Category: Bug Help wanted Inactive: duplicate Priority: p1 Time Scale
|
I have the following issue:

The first bar is not fully displayed and the last one is hidden at all.
Any solution to this?
|
1.0
|
First and last bars display problem - I have the following issue:

The first bar is not fully displayed and the last one is hidden at all.
Any solution to this?
|
non_test
|
first and last bars display problem i have the following issue the first bar is not fully displayed and the last one is hidden at all any solution to this
| 0
|
488,037
| 14,073,876,166
|
IssuesEvent
|
2020-11-04 06:05:17
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.rawstory.com - video or audio doesn't play
|
browser-fenix engine-gecko ml-needsdiagnosis-false ml-probability-high priority-normal
|
<!-- @browser: Firefox Mobile 83.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:83.0) Gecko/83.0 Firefox/83.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/61027 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.rawstory.com/2020/11/watch-florida-deputies-caught-on-tape-beating-biden-supporter-at-political-rally/
**Browser / Version**: Firefox Mobile 83.0
**Operating System**: Android
**Tested Another Browser**: No
**Problem type**: Video or audio doesn't play
**Description**: There is no video
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201025174155</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/11/a2757cf2-64eb-4019-80f2-9be21ee7edc7)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.rawstory.com - video or audio doesn't play - <!-- @browser: Firefox Mobile 83.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:83.0) Gecko/83.0 Firefox/83.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/61027 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.rawstory.com/2020/11/watch-florida-deputies-caught-on-tape-beating-biden-supporter-at-political-rally/
**Browser / Version**: Firefox Mobile 83.0
**Operating System**: Android
**Tested Another Browser**: No
**Problem type**: Video or audio doesn't play
**Description**: There is no video
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201025174155</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/11/a2757cf2-64eb-4019-80f2-9be21ee7edc7)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_test
|
video or audio doesn t play url browser version firefox mobile operating system android tested another browser no problem type video or audio doesn t play description there is no video steps to reproduce browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
115,319
| 24,746,390,426
|
IssuesEvent
|
2022-10-21 10:02:44
|
arduino-libraries/Arduino_DebugUtils
|
https://api.github.com/repos/arduino-libraries/Arduino_DebugUtils
|
closed
|
Adding Debug.info() style functions
|
type: enhancement topic: code
|
Hi, has there been any discussion around adding individual functions for each debug logging level, instead of passing in the level explicitly, similar to how the logging module in python works? I believe something like this should work:
```
void Arduino_DebugUtils::info(const char * fmt, ...)
{
va_list args;
va_start(args, fmt);
print(DBG_INFO, fmt, args);
va_end(args);
}
```
I would be more than happy to open a PR if this feels like something that should be added
|
1.0
|
Adding Debug.info() style functions - Hi, has there been any discussion around adding individual functions for each debug logging level, instead of passing in the level explicitly, similar to how the logging module in python works? I believe something like this should work:
```
void Arduino_DebugUtils::info(const char * fmt, ...)
{
va_list args;
va_start(args, fmt);
print(DBG_INFO, fmt, args);
va_end(args);
}
```
I would be more than happy to open a PR if this feels like something that should be added
|
non_test
|
adding debug info style functions hi has there been any discussion around adding individual functions for each debug logging level instead of passing in the level explicitly similar to how the logging module in python works i believe something like this should work void arduino debugutils info const char fmt va list args va start args fmt print dbg info fmt args va end args i would be more than happy to open a pr if this feels like something that should be added
| 0
|
328,134
| 28,102,012,833
|
IssuesEvent
|
2023-03-30 20:18:25
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: kv/splits/nodes=3/quiesce=false failed
|
C-test-failure O-robot X-duplicate O-roachtest branch-master release-blocker T-kv
|
roachtest.kv/splits/nodes=3/quiesce=false [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/9329887?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/9329887?buildTab=artifacts#/kv/splits/nodes=3/quiesce=false) on master @ [1f8024bf14433ca169e5a8c3768c5d223dc5018c](https://github.com/cockroachdb/cockroach/commits/1f8024bf14433ca169e5a8c3768c5d223dc5018c):
```
test artifacts and logs in: /artifacts/kv/splits/nodes=3/quiesce=false/run_1
(cluster.go:1977).Run: output in run_053210.523548575_n4_workload-run-kv-init: ./workload run kv --init --max-ops=1 --concurrency=192 --splits=30000 {pgurl:1-3} returned: COMMAND_PROBLEM: ssh verbose log retained in ssh_053211.295244109_n4_workload-run-kv-init.log: exit status 1
(monitor.go:127).Wait: monitor failure: monitor task failed: t.Fatal() was called
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*kv/splits/nodes=3/quiesce=false.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-26198
|
2.0
|
roachtest: kv/splits/nodes=3/quiesce=false failed - roachtest.kv/splits/nodes=3/quiesce=false [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/9329887?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/9329887?buildTab=artifacts#/kv/splits/nodes=3/quiesce=false) on master @ [1f8024bf14433ca169e5a8c3768c5d223dc5018c](https://github.com/cockroachdb/cockroach/commits/1f8024bf14433ca169e5a8c3768c5d223dc5018c):
```
test artifacts and logs in: /artifacts/kv/splits/nodes=3/quiesce=false/run_1
(cluster.go:1977).Run: output in run_053210.523548575_n4_workload-run-kv-init: ./workload run kv --init --max-ops=1 --concurrency=192 --splits=30000 {pgurl:1-3} returned: COMMAND_PROBLEM: ssh verbose log retained in ssh_053211.295244109_n4_workload-run-kv-init.log: exit status 1
(monitor.go:127).Wait: monitor failure: monitor task failed: t.Fatal() was called
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*kv/splits/nodes=3/quiesce=false.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-26198
|
test
|
roachtest kv splits nodes quiesce false failed roachtest kv splits nodes quiesce false with on master test artifacts and logs in artifacts kv splits nodes quiesce false run cluster go run output in run workload run kv init workload run kv init max ops concurrency splits pgurl returned command problem ssh verbose log retained in ssh workload run kv init log exit status monitor go wait monitor failure monitor task failed t fatal was called parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see cc cockroachdb kv triage jira issue crdb
| 1
|
81,035
| 7,763,681,590
|
IssuesEvent
|
2018-06-01 17:27:32
|
mozilla-mobile/focus-android
|
https://api.github.com/repos/mozilla-mobile/focus-android
|
closed
|
Create UI Test Script for performance comparision on website loading
|
size M testing
|
This is a subtask of #2201
The script will initially load about 10~20 websites in order, where the loading time will be measured and collected.
Since the bookends will be hit multiple times, they need to be defined with an iterator inside the client code.
|
1.0
|
Create UI Test Script for performance comparision on website loading - This is a subtask of #2201
The script will initially load about 10~20 websites in order, where the loading time will be measured and collected.
Since the bookends will be hit multiple times, they need to be defined with an iterator inside the client code.
|
test
|
create ui test script for performance comparision on website loading this is a subtask of the script will initially load about websites in order where the loading time will be measured and collected since the bookends will be hit multiple times they need to be defined with an iterator inside the client code
| 1
|
106,760
| 9,181,639,191
|
IssuesEvent
|
2019-03-05 10:42:33
|
elastic/elasticsearch
|
https://api.github.com/repos/elastic/elasticsearch
|
closed
|
DeleteExpiredDataIT.testDeleteExpiredData failure
|
:ml >test-failure
|
Failed build at https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+master+periodic/813/
Does not reproduce locally.
Reproduce block:
```
./gradlew :x-pack:plugin:ml:qa:native-multi-node-tests:integTestRunner \
-Dtests.seed=26207D14EDCE60BC \
-Dtests.class=org.elasticsearch.xpack.ml.integration.DeleteExpiredDataIT \
-Dtests.method="testDeleteExpiredData" \
-Dtests.security.manager=true \
-Dtests.locale=pl-PL \
-Dtests.timezone=Etc/GMT \
-Dcompiler.java=11 \
-Druntime.java=8
```
There are a few interesting stack traces here, but this one jumped out at me, and happens a *lot*:
```
[2019-03-01T07:39:15,383][DEBUG][o.e.a.b.TransportShardBulkAction] [node-1] [.ml-state][3] failed to execute bulk item (index) index {[.ml-state-write][doc][non_existing_job_categorizer_state#859], source[{}]}
--
java.lang.IllegalArgumentException: Rejecting mapping update to [.ml-state] as the final mapping would have more than 1 type: [_doc, doc]
at org.elasticsearch.index.mapper.MapperService.internalMerge(MapperService.java:449) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.index.mapper.MapperService.internalMerge(MapperService.java:398) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.index.mapper.MapperService.merge(MapperService.java:331) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.cluster.metadata.MetaDataMappingService$PutMappingExecutor.applyRequest(MetaDataMappingService.java:315) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.cluster.metadata.MetaDataMappingService$PutMappingExecutor.execute(MetaDataMappingService.java:238) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.cluster.service.MasterService.executeTasks(MasterService.java:687) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.cluster.service.MasterService.calculateTaskOutputs(MasterService.java:310) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.cluster.service.MasterService.runTasks(MasterService.java:210) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.cluster.service.MasterService$Batcher.run(MasterService.java:142) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.cluster.service.TaskBatcher.runIfNotProcessed(TaskBatcher.java:150) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.cluster.service.TaskBatcher$BatchedTask.run(TaskBatcher.java:188) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:681) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.common.util.concurrent.PrioritizedEsThreadPoolExecutor$TieBreakingPrioritizedRunnable.runAndClean(PrioritizedEsThreadPoolExecutor.java:252) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.common.util.concurrent.PrioritizedEsThreadPoolExecutor$TieBreakingPrioritizedRunnable.run(PrioritizedEsThreadPoolExecutor.java:215) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?]
at java.lang.Thread.run(Thread.java:834) [?:?]
```
But this is also present:
```
[2019-03-01T07:46:45,108][WARN ][o.e.p.PersistentTasksNodeService] [node-2] task datafeed-realtime-job-given-process-is-killed-datafeed failed with an exception
--
org.elasticsearch.xpack.ml.datafeed.DatafeedJob$AnalysisProblemException: ElasticsearchStatusException[Cannot perform requested action because job [realtime-job-given-process-is-killed] is not open]
at org.elasticsearch.xpack.ml.datafeed.DatafeedJob.flushJob(DatafeedJob.java:457) ~[x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.ml.datafeed.DatafeedJob.run(DatafeedJob.java:396) ~[x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.ml.datafeed.DatafeedJob.runRealtime(DatafeedJob.java:177) ~[x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.ml.datafeed.DatafeedManager$Holder.executeRealTime(DatafeedManager.java:397) ~[x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.ml.datafeed.DatafeedManager$Holder.access$600(DatafeedManager.java:290) [x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.ml.datafeed.DatafeedManager$3.doRun(DatafeedManager.java:227) [x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingAbstractRunnable.doRun(ThreadContext.java:751) [elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:37) [elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?]
at java.lang.Thread.run(Thread.java:834) [?:?]
Caused by: org.elasticsearch.ElasticsearchStatusException: Cannot perform requested action because job [realtime-job-given-process-is-killed] is not open
at org.elasticsearch.xpack.core.ml.utils.ExceptionsHelper.conflictStatusException(ExceptionsHelper.java:50) ~[?:?]
at org.elasticsearch.xpack.ml.action.TransportJobTaskAction.doExecute(TransportJobTaskAction.java:55) ~[?:?]
at org.elasticsearch.xpack.ml.action.TransportJobTaskAction.doExecute(TransportJobTaskAction.java:31) ~[?:?]
at org.elasticsearch.action.support.TransportAction$RequestFilterChain.proceed(TransportAction.java:145) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$apply$0(SecurityActionFilter.java:86) ~[?:?]
at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:61) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$authorizeRequest$4(SecurityActionFilter.java:171) ~[?:?]
at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:61) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.security.authz.AuthorizationService.lambda$authorizeAction$4(AuthorizationService.java:237) ~[?:?]
at org.elasticsearch.xpack.security.authz.AuthorizationService$AuthorizationResultListener.onResponse(AuthorizationService.java:599) ~[?:?]
at org.elasticsearch.xpack.security.authz.AuthorizationService$AuthorizationResultListener.onResponse(AuthorizationService.java:574) ~[?:?]
at org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.security.authz.RBACEngine.authorizeClusterAction(RBACEngine.java:140) ~[?:?]
at org.elasticsearch.xpack.security.authz.AuthorizationService.authorizeAction(AuthorizationService.java:239) ~[?:?]
at org.elasticsearch.xpack.security.authz.AuthorizationService.maybeAuthorizeRunAs(AuthorizationService.java:223) ~[?:?]
at org.elasticsearch.xpack.security.authz.AuthorizationService.lambda$authorize$1(AuthorizationService.java:189) ~[?:?]
at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:61) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.security.authz.RBACEngine.lambda$resolveAuthorizationInfo$1(RBACEngine.java:113) ~[?:?]
at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:61) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.security.authz.store.CompositeRolesStore.getRoles(CompositeRolesStore.java:193) ~[?:?]
at org.elasticsearch.xpack.security.authz.RBACEngine.getRoles(RBACEngine.java:119) ~[?:?]
at org.elasticsearch.xpack.security.authz.RBACEngine.resolveAuthorizationInfo(RBACEngine.java:107) ~[?:?]
at org.elasticsearch.xpack.security.authz.AuthorizationService.authorize(AuthorizationService.java:191) ~[?:?]
at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.authorizeRequest(SecurityActionFilter.java:171) ~[?:?]
at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$applyInternal$3(SecurityActionFilter.java:157) ~[?:?]
at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:61) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$authenticateAsync$2(AuthenticationService.java:245) ~[?:?]
at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$lookForExistingAuthentication$6(AuthenticationService.java:305) ~[?:?]
at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lookForExistingAuthentication(AuthenticationService.java:316) ~[?:?]
at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.authenticateAsync(AuthenticationService.java:243) ~[?:?]
at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.access$000(AuthenticationService.java:195) ~[?:?]
at org.elasticsearch.xpack.security.authc.AuthenticationService.authenticate(AuthenticationService.java:138) ~[?:?]
at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.applyInternal(SecurityActionFilter.java:154) ~[?:?]
at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$apply$2(SecurityActionFilter.java:100) ~[?:?]
at org.elasticsearch.xpack.core.security.SecurityContext.executeAsUser(SecurityContext.java:97) ~[?:?]
at org.elasticsearch.xpack.security.authz.AuthorizationUtils.switchUserBasedOnActionOriginAndExecute(AuthorizationUtils.java:115) ~[?:?]
at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.apply(SecurityActionFilter.java:98) ~[?:?]
at org.elasticsearch.action.support.TransportAction$RequestFilterChain.proceed(TransportAction.java:143) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.action.support.TransportAction.execute(TransportAction.java:121) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.action.support.TransportAction.execute(TransportAction.java:64) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.client.node.NodeClient.executeLocally(NodeClient.java:83) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.client.node.NodeClient.doExecute(NodeClient.java:72) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.client.support.AbstractClient.execute(AbstractClient.java:394) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.client.support.AbstractClient.execute(AbstractClient.java:383) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.ml.datafeed.DatafeedJob.flushJob(DatafeedJob.java:444) ~[x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
... 10 more
[2019-03-01T07:46:45,112][INFO ][o.e.x.m.d.DatafeedManager] [node-2] [realtime_analysis_error] datafeed [realtime-job-given-process-is-killed-datafeed] for job [realtime-job-given-process-is-killed] has been stopped
```
|
1.0
|
DeleteExpiredDataIT.testDeleteExpiredData failure - Failed build at https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+master+periodic/813/
Does not reproduce locally.
Reproduce block:
```
./gradlew :x-pack:plugin:ml:qa:native-multi-node-tests:integTestRunner \
-Dtests.seed=26207D14EDCE60BC \
-Dtests.class=org.elasticsearch.xpack.ml.integration.DeleteExpiredDataIT \
-Dtests.method="testDeleteExpiredData" \
-Dtests.security.manager=true \
-Dtests.locale=pl-PL \
-Dtests.timezone=Etc/GMT \
-Dcompiler.java=11 \
-Druntime.java=8
```
There are a few interesting stack traces here, but this one jumped out at me, and happens a *lot*:
```
[2019-03-01T07:39:15,383][DEBUG][o.e.a.b.TransportShardBulkAction] [node-1] [.ml-state][3] failed to execute bulk item (index) index {[.ml-state-write][doc][non_existing_job_categorizer_state#859], source[{}]}
--
java.lang.IllegalArgumentException: Rejecting mapping update to [.ml-state] as the final mapping would have more than 1 type: [_doc, doc]
at org.elasticsearch.index.mapper.MapperService.internalMerge(MapperService.java:449) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.index.mapper.MapperService.internalMerge(MapperService.java:398) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.index.mapper.MapperService.merge(MapperService.java:331) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.cluster.metadata.MetaDataMappingService$PutMappingExecutor.applyRequest(MetaDataMappingService.java:315) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.cluster.metadata.MetaDataMappingService$PutMappingExecutor.execute(MetaDataMappingService.java:238) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.cluster.service.MasterService.executeTasks(MasterService.java:687) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.cluster.service.MasterService.calculateTaskOutputs(MasterService.java:310) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.cluster.service.MasterService.runTasks(MasterService.java:210) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.cluster.service.MasterService$Batcher.run(MasterService.java:142) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.cluster.service.TaskBatcher.runIfNotProcessed(TaskBatcher.java:150) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.cluster.service.TaskBatcher$BatchedTask.run(TaskBatcher.java:188) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:681) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.common.util.concurrent.PrioritizedEsThreadPoolExecutor$TieBreakingPrioritizedRunnable.runAndClean(PrioritizedEsThreadPoolExecutor.java:252) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.common.util.concurrent.PrioritizedEsThreadPoolExecutor$TieBreakingPrioritizedRunnable.run(PrioritizedEsThreadPoolExecutor.java:215) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?]
at java.lang.Thread.run(Thread.java:834) [?:?]
```
But this is also present:
```
[2019-03-01T07:46:45,108][WARN ][o.e.p.PersistentTasksNodeService] [node-2] task datafeed-realtime-job-given-process-is-killed-datafeed failed with an exception
--
org.elasticsearch.xpack.ml.datafeed.DatafeedJob$AnalysisProblemException: ElasticsearchStatusException[Cannot perform requested action because job [realtime-job-given-process-is-killed] is not open]
at org.elasticsearch.xpack.ml.datafeed.DatafeedJob.flushJob(DatafeedJob.java:457) ~[x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.ml.datafeed.DatafeedJob.run(DatafeedJob.java:396) ~[x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.ml.datafeed.DatafeedJob.runRealtime(DatafeedJob.java:177) ~[x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.ml.datafeed.DatafeedManager$Holder.executeRealTime(DatafeedManager.java:397) ~[x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.ml.datafeed.DatafeedManager$Holder.access$600(DatafeedManager.java:290) [x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.ml.datafeed.DatafeedManager$3.doRun(DatafeedManager.java:227) [x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingAbstractRunnable.doRun(ThreadContext.java:751) [elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:37) [elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?]
at java.lang.Thread.run(Thread.java:834) [?:?]
Caused by: org.elasticsearch.ElasticsearchStatusException: Cannot perform requested action because job [realtime-job-given-process-is-killed] is not open
at org.elasticsearch.xpack.core.ml.utils.ExceptionsHelper.conflictStatusException(ExceptionsHelper.java:50) ~[?:?]
at org.elasticsearch.xpack.ml.action.TransportJobTaskAction.doExecute(TransportJobTaskAction.java:55) ~[?:?]
at org.elasticsearch.xpack.ml.action.TransportJobTaskAction.doExecute(TransportJobTaskAction.java:31) ~[?:?]
at org.elasticsearch.action.support.TransportAction$RequestFilterChain.proceed(TransportAction.java:145) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$apply$0(SecurityActionFilter.java:86) ~[?:?]
at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:61) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$authorizeRequest$4(SecurityActionFilter.java:171) ~[?:?]
at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:61) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.security.authz.AuthorizationService.lambda$authorizeAction$4(AuthorizationService.java:237) ~[?:?]
at org.elasticsearch.xpack.security.authz.AuthorizationService$AuthorizationResultListener.onResponse(AuthorizationService.java:599) ~[?:?]
at org.elasticsearch.xpack.security.authz.AuthorizationService$AuthorizationResultListener.onResponse(AuthorizationService.java:574) ~[?:?]
at org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.security.authz.RBACEngine.authorizeClusterAction(RBACEngine.java:140) ~[?:?]
at org.elasticsearch.xpack.security.authz.AuthorizationService.authorizeAction(AuthorizationService.java:239) ~[?:?]
at org.elasticsearch.xpack.security.authz.AuthorizationService.maybeAuthorizeRunAs(AuthorizationService.java:223) ~[?:?]
at org.elasticsearch.xpack.security.authz.AuthorizationService.lambda$authorize$1(AuthorizationService.java:189) ~[?:?]
at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:61) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.security.authz.RBACEngine.lambda$resolveAuthorizationInfo$1(RBACEngine.java:113) ~[?:?]
at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:61) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.security.authz.store.CompositeRolesStore.getRoles(CompositeRolesStore.java:193) ~[?:?]
at org.elasticsearch.xpack.security.authz.RBACEngine.getRoles(RBACEngine.java:119) ~[?:?]
at org.elasticsearch.xpack.security.authz.RBACEngine.resolveAuthorizationInfo(RBACEngine.java:107) ~[?:?]
at org.elasticsearch.xpack.security.authz.AuthorizationService.authorize(AuthorizationService.java:191) ~[?:?]
at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.authorizeRequest(SecurityActionFilter.java:171) ~[?:?]
at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$applyInternal$3(SecurityActionFilter.java:157) ~[?:?]
at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:61) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$authenticateAsync$2(AuthenticationService.java:245) ~[?:?]
at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$lookForExistingAuthentication$6(AuthenticationService.java:305) ~[?:?]
at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lookForExistingAuthentication(AuthenticationService.java:316) ~[?:?]
at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.authenticateAsync(AuthenticationService.java:243) ~[?:?]
at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.access$000(AuthenticationService.java:195) ~[?:?]
at org.elasticsearch.xpack.security.authc.AuthenticationService.authenticate(AuthenticationService.java:138) ~[?:?]
at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.applyInternal(SecurityActionFilter.java:154) ~[?:?]
at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$apply$2(SecurityActionFilter.java:100) ~[?:?]
at org.elasticsearch.xpack.core.security.SecurityContext.executeAsUser(SecurityContext.java:97) ~[?:?]
at org.elasticsearch.xpack.security.authz.AuthorizationUtils.switchUserBasedOnActionOriginAndExecute(AuthorizationUtils.java:115) ~[?:?]
at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.apply(SecurityActionFilter.java:98) ~[?:?]
at org.elasticsearch.action.support.TransportAction$RequestFilterChain.proceed(TransportAction.java:143) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.action.support.TransportAction.execute(TransportAction.java:121) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.action.support.TransportAction.execute(TransportAction.java:64) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.client.node.NodeClient.executeLocally(NodeClient.java:83) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.client.node.NodeClient.doExecute(NodeClient.java:72) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.client.support.AbstractClient.execute(AbstractClient.java:394) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.client.support.AbstractClient.execute(AbstractClient.java:383) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
at org.elasticsearch.xpack.ml.datafeed.DatafeedJob.flushJob(DatafeedJob.java:444) ~[x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT]
... 10 more
[2019-03-01T07:46:45,112][INFO ][o.e.x.m.d.DatafeedManager] [node-2] [realtime_analysis_error] datafeed [realtime-job-given-process-is-killed-datafeed] for job [realtime-job-given-process-is-killed] has been stopped
```
|
test
|
deleteexpireddatait testdeleteexpireddata failure failed build at does not reproduce locally reproduce block gradlew x pack plugin ml qa native multi node tests integtestrunner dtests seed dtests class org elasticsearch xpack ml integration deleteexpireddatait dtests method testdeleteexpireddata dtests security manager true dtests locale pl pl dtests timezone etc gmt dcompiler java druntime java there are a few interesting stack traces here but this one jumped out at me and happens a lot failed to execute bulk item index index source java lang illegalargumentexception rejecting mapping update to as the final mapping would have more than type at org elasticsearch index mapper mapperservice internalmerge mapperservice java at org elasticsearch index mapper mapperservice internalmerge mapperservice java at org elasticsearch index mapper mapperservice merge mapperservice java at org elasticsearch cluster metadata metadatamappingservice putmappingexecutor applyrequest metadatamappingservice java at org elasticsearch cluster metadata metadatamappingservice putmappingexecutor execute metadatamappingservice java at org elasticsearch cluster service masterservice executetasks masterservice java at org elasticsearch cluster service masterservice calculatetaskoutputs masterservice java at org elasticsearch cluster service masterservice runtasks masterservice java at org elasticsearch cluster service masterservice batcher run masterservice java at org elasticsearch cluster service taskbatcher runifnotprocessed taskbatcher java at org elasticsearch cluster service taskbatcher batchedtask run taskbatcher java at org elasticsearch common util concurrent threadcontext contextpreservingrunnable run threadcontext java at org elasticsearch common util concurrent prioritizedesthreadpoolexecutor tiebreakingprioritizedrunnable runandclean prioritizedesthreadpoolexecutor java at org elasticsearch common util concurrent prioritizedesthreadpoolexecutor tiebreakingprioritizedrunnable run prioritizedesthreadpoolexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java but this is also present task datafeed realtime job given process is killed datafeed failed with an exception org elasticsearch xpack ml datafeed datafeedjob analysisproblemexception elasticsearchstatusexception is not open at org elasticsearch xpack ml datafeed datafeedjob flushjob datafeedjob java at org elasticsearch xpack ml datafeed datafeedjob run datafeedjob java at org elasticsearch xpack ml datafeed datafeedjob runrealtime datafeedjob java at org elasticsearch xpack ml datafeed datafeedmanager holder executerealtime datafeedmanager java at org elasticsearch xpack ml datafeed datafeedmanager holder access datafeedmanager java at org elasticsearch xpack ml datafeed datafeedmanager dorun datafeedmanager java at org elasticsearch common util concurrent threadcontext contextpreservingabstractrunnable dorun threadcontext java at org elasticsearch common util concurrent abstractrunnable run abstractrunnable java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java caused by org elasticsearch elasticsearchstatusexception cannot perform requested action because job is not open at org elasticsearch xpack core ml utils exceptionshelper conflictstatusexception exceptionshelper java at org elasticsearch xpack ml action transportjobtaskaction doexecute transportjobtaskaction java at org elasticsearch xpack ml action transportjobtaskaction doexecute transportjobtaskaction java at org elasticsearch action support transportaction requestfilterchain proceed transportaction java at org elasticsearch xpack security action filter securityactionfilter lambda apply securityactionfilter java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch xpack security action filter securityactionfilter lambda authorizerequest securityactionfilter java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch xpack security authz authorizationservice lambda authorizeaction authorizationservice java at org elasticsearch xpack security authz authorizationservice authorizationresultlistener onresponse authorizationservice java at org elasticsearch xpack security authz authorizationservice authorizationresultlistener onresponse authorizationservice java at org elasticsearch action support contextpreservingactionlistener onresponse contextpreservingactionlistener java at org elasticsearch xpack security authz rbacengine authorizeclusteraction rbacengine java at org elasticsearch xpack security authz authorizationservice authorizeaction authorizationservice java at org elasticsearch xpack security authz authorizationservice maybeauthorizerunas authorizationservice java at org elasticsearch xpack security authz authorizationservice lambda authorize authorizationservice java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch action support contextpreservingactionlistener onresponse contextpreservingactionlistener java at org elasticsearch xpack security authz rbacengine lambda resolveauthorizationinfo rbacengine java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch xpack security authz store compositerolesstore getroles compositerolesstore java at org elasticsearch xpack security authz rbacengine getroles rbacengine java at org elasticsearch xpack security authz rbacengine resolveauthorizationinfo rbacengine java at org elasticsearch xpack security authz authorizationservice authorize authorizationservice java at org elasticsearch xpack security action filter securityactionfilter authorizerequest securityactionfilter java at org elasticsearch xpack security action filter securityactionfilter lambda applyinternal securityactionfilter java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch xpack security authc authenticationservice authenticator lambda authenticateasync authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator lambda lookforexistingauthentication authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator lookforexistingauthentication authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator authenticateasync authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator access authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticate authenticationservice java at org elasticsearch xpack security action filter securityactionfilter applyinternal securityactionfilter java at org elasticsearch xpack security action filter securityactionfilter lambda apply securityactionfilter java at org elasticsearch xpack core security securitycontext executeasuser securitycontext java at org elasticsearch xpack security authz authorizationutils switchuserbasedonactionoriginandexecute authorizationutils java at org elasticsearch xpack security action filter securityactionfilter apply securityactionfilter java at org elasticsearch action support transportaction requestfilterchain proceed transportaction java at org elasticsearch action support transportaction execute transportaction java at org elasticsearch action support transportaction execute transportaction java at org elasticsearch client node nodeclient executelocally nodeclient java at org elasticsearch client node nodeclient doexecute nodeclient java at org elasticsearch client support abstractclient execute abstractclient java at org elasticsearch client support abstractclient execute abstractclient java at org elasticsearch xpack ml datafeed datafeedjob flushjob datafeedjob java more datafeed for job has been stopped
| 1
|
188,430
| 22,046,416,088
|
IssuesEvent
|
2022-05-30 02:35:35
|
DavidSpek/kale
|
https://api.github.com/repos/DavidSpek/kale
|
opened
|
CVE-2022-29204 (Medium) detected in tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl, tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl
|
security vulnerability
|
## CVE-2022-29204 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl</b>, <b>tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary>
<p>
<details><summary><b>tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/7b/c5/a97ed48fcc878e36bb05a3ea700c077360853c0994473a8f6b0ab4c2ddd2/tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/7b/c5/a97ed48fcc878e36bb05a3ea700c077360853c0994473a8f6b0ab4c2ddd2/tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /examples/dog-breed-classification/requirements/requirements.txt</p>
<p>Path to vulnerable library: /examples/dog-breed-classification/requirements/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</details>
<details><summary><b>tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/ef/73/205b5e7f8fe086ffe4165d984acb2c49fa3086f330f03099378753982d2e/tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/ef/73/205b5e7f8fe086ffe4165d984acb2c49fa3086f330f03099378753982d2e/tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl</a></p>
<p>Path to dependency file: /examples/taxi-cab-classification/requirements.txt</p>
<p>Path to vulnerable library: /examples/taxi-cab-classification/requirements.txt</p>
<p>
Dependency Hierarchy:
- tfx_bsl-0.21.4-cp27-cp27mu-manylinux2010_x86_64.whl (Root Library)
- :x: **tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an open source platform for machine learning. Prior to versions 2.9.0, 2.8.1, 2.7.2, and 2.6.4, the implementation of `tf.raw_ops.UnsortedSegmentJoin` does not fully validate the input arguments. This results in a `CHECK`-failure which can be used to trigger a denial of service attack. The code assumes `num_segments` is a positive scalar but there is no validation. Since this value is used to allocate the output tensor, a negative value would result in a `CHECK`-failure (assertion failure), as per TFSA-2021-198. Versions 2.9.0, 2.8.1, 2.7.2, and 2.6.4 contain a patch for this issue.
<p>Publish Date: 2022-05-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-29204>CVE-2022-29204</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-29204">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-29204</a></p>
<p>Release Date: 2022-05-20</p>
<p>Fix Resolution: tensorflow - 2.6.4,2.7.2,2.8.1,2.9.0;tensorflow-cpu - 2.6.4,2.7.2,2.8.1,2.9.0;tensorflow-gpu - 2.6.4,2.7.2,2.8.1,2.9.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-29204 (Medium) detected in tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl, tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl - ## CVE-2022-29204 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl</b>, <b>tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary>
<p>
<details><summary><b>tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/7b/c5/a97ed48fcc878e36bb05a3ea700c077360853c0994473a8f6b0ab4c2ddd2/tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/7b/c5/a97ed48fcc878e36bb05a3ea700c077360853c0994473a8f6b0ab4c2ddd2/tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /examples/dog-breed-classification/requirements/requirements.txt</p>
<p>Path to vulnerable library: /examples/dog-breed-classification/requirements/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</details>
<details><summary><b>tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/ef/73/205b5e7f8fe086ffe4165d984acb2c49fa3086f330f03099378753982d2e/tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/ef/73/205b5e7f8fe086ffe4165d984acb2c49fa3086f330f03099378753982d2e/tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl</a></p>
<p>Path to dependency file: /examples/taxi-cab-classification/requirements.txt</p>
<p>Path to vulnerable library: /examples/taxi-cab-classification/requirements.txt</p>
<p>
Dependency Hierarchy:
- tfx_bsl-0.21.4-cp27-cp27mu-manylinux2010_x86_64.whl (Root Library)
- :x: **tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an open source platform for machine learning. Prior to versions 2.9.0, 2.8.1, 2.7.2, and 2.6.4, the implementation of `tf.raw_ops.UnsortedSegmentJoin` does not fully validate the input arguments. This results in a `CHECK`-failure which can be used to trigger a denial of service attack. The code assumes `num_segments` is a positive scalar but there is no validation. Since this value is used to allocate the output tensor, a negative value would result in a `CHECK`-failure (assertion failure), as per TFSA-2021-198. Versions 2.9.0, 2.8.1, 2.7.2, and 2.6.4 contain a patch for this issue.
<p>Publish Date: 2022-05-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-29204>CVE-2022-29204</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-29204">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-29204</a></p>
<p>Release Date: 2022-05-20</p>
<p>Fix Resolution: tensorflow - 2.6.4,2.7.2,2.8.1,2.9.0;tensorflow-cpu - 2.6.4,2.7.2,2.8.1,2.9.0;tensorflow-gpu - 2.6.4,2.7.2,2.8.1,2.9.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in tensorflow whl tensorflow whl cve medium severity vulnerability vulnerable libraries tensorflow whl tensorflow whl tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file examples dog breed classification requirements requirements txt path to vulnerable library examples dog breed classification requirements requirements txt dependency hierarchy x tensorflow whl vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file examples taxi cab classification requirements txt path to vulnerable library examples taxi cab classification requirements txt dependency hierarchy tfx bsl whl root library x tensorflow whl vulnerable library found in base branch master vulnerability details tensorflow is an open source platform for machine learning prior to versions and the implementation of tf raw ops unsortedsegmentjoin does not fully validate the input arguments this results in a check failure which can be used to trigger a denial of service attack the code assumes num segments is a positive scalar but there is no validation since this value is used to allocate the output tensor a negative value would result in a check failure assertion failure as per tfsa versions and contain a patch for this issue publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with mend
| 0
|
252,477
| 21,580,133,159
|
IssuesEvent
|
2022-05-02 17:47:26
|
rancher/dashboard
|
https://api.github.com/repos/rancher/dashboard
|
closed
|
Error messages appear below form buttons
|
kind/bug [zube]: To Test team/area3
|
Create a deployment (e.g. name: 'test' and image : 'nginx').
Clone this deployment (from the row action menu for the deployment) and leave the name blank - click 'Create'.
The Create button changes to error but the message is not visible - scroll down and you will see it beneath the buttons.
The simple fix would be to scroll the page to bottom when there is an error - might be worth just doing that in a PR - that fixes the immediate issue - but I think we need to move the errors, if we want the buttons to stick to the bottom of the page. This will need some UX design.
|
1.0
|
Error messages appear below form buttons - Create a deployment (e.g. name: 'test' and image : 'nginx').
Clone this deployment (from the row action menu for the deployment) and leave the name blank - click 'Create'.
The Create button changes to error but the message is not visible - scroll down and you will see it beneath the buttons.
The simple fix would be to scroll the page to bottom when there is an error - might be worth just doing that in a PR - that fixes the immediate issue - but I think we need to move the errors, if we want the buttons to stick to the bottom of the page. This will need some UX design.
|
test
|
error messages appear below form buttons create a deployment e g name test and image nginx clone this deployment from the row action menu for the deployment and leave the name blank click create the create button changes to error but the message is not visible scroll down and you will see it beneath the buttons the simple fix would be to scroll the page to bottom when there is an error might be worth just doing that in a pr that fixes the immediate issue but i think we need to move the errors if we want the buttons to stick to the bottom of the page this will need some ux design
| 1
|
109,894
| 11,661,568,021
|
IssuesEvent
|
2020-03-03 07:08:04
|
ift-gftc/doc.ift.org
|
https://api.github.com/repos/ift-gftc/doc.ift.org
|
opened
|
DOC | How to launch Gatsby server locally?
|
documentation
|
## 📖 Document
Create documentation on how to run the Gatsby server locally so that future contributors know how to do this when testing branches they are working on.
|
1.0
|
DOC | How to launch Gatsby server locally? - ## 📖 Document
Create documentation on how to run the Gatsby server locally so that future contributors know how to do this when testing branches they are working on.
|
non_test
|
doc how to launch gatsby server locally 📖 document create documentation on how to run the gatsby server locally so that future contributors know how to do this when testing branches they are working on
| 0
|
296,735
| 25,572,508,271
|
IssuesEvent
|
2022-11-30 18:55:30
|
MD-Anderson-Bioinformatics/NG-CHM
|
https://api.github.com/repos/MD-Anderson-Bioinformatics/NG-CHM
|
closed
|
All plugin info not removed when one plugin replaced by another after opening a pane below plot pane.
|
bug passed retest 2.21.3
|
To reproduce,
Load a map that has extra coordinates for use by 2D or 3D scatterplots (like deprez1000).
Replace Summary pane with 3D Scatterplot. (Use default settings).
Open pane below 3D Scatterplot pane.
Replace 3D Scatterplot pane with 2D Scatterplot. (Use default settings).
Save Map.
Dialog telling that some parts of Scatterplot can't be saved reports that info for both 2D and 3D Scatterplots even though current map only contains a 2D Scatterplot.
|
1.0
|
All plugin info not removed when one plugin replaced by another after opening a pane below plot pane. - To reproduce,
Load a map that has extra coordinates for use by 2D or 3D scatterplots (like deprez1000).
Replace Summary pane with 3D Scatterplot. (Use default settings).
Open pane below 3D Scatterplot pane.
Replace 3D Scatterplot pane with 2D Scatterplot. (Use default settings).
Save Map.
Dialog telling that some parts of Scatterplot can't be saved reports that info for both 2D and 3D Scatterplots even though current map only contains a 2D Scatterplot.
|
test
|
all plugin info not removed when one plugin replaced by another after opening a pane below plot pane to reproduce load a map that has extra coordinates for use by or scatterplots like replace summary pane with scatterplot use default settings open pane below scatterplot pane replace scatterplot pane with scatterplot use default settings save map dialog telling that some parts of scatterplot can t be saved reports that info for both and scatterplots even though current map only contains a scatterplot
| 1
|
140,100
| 11,302,418,821
|
IssuesEvent
|
2020-01-17 17:37:15
|
aliasrobotics/RVD
|
https://api.github.com/repos/aliasrobotics/RVD
|
opened
|
CWE-362/CWE-367 (race), This usually indicates a security flaw. If an attacker can change any... @ /canopen_master/objdict.h:525
|
CWE-362 CWE-367 bug flawfinder flawfinder_level_4 static analysis testing triage
|
```yaml
{
"system": "./install/canopen_master/include/canopen_master/objdict.h:525:42",
"id": 1,
"title": "CWE-362/CWE-367 (race), This usually indicates a security flaw. If an attacker can change any... @ /canopen_master/objdict.h:525",
"cwe": [
"CWE-362",
"CWE-367"
],
"exploitation": {
"exploitation-image": "",
"exploitation-vector": "",
"description": ""
},
"keywords": [
"flawfinder",
"flawfinder_level_4",
"static analysis",
"testing",
"triage",
"CWE-362",
"CWE-367",
"bug"
],
"type": "bug",
"cve": "None",
"severity": {
"rvss-vector": "",
"rvss-score": 0,
"cvss-score": 0,
"severity-description": "",
"cvss-vector": ""
},
"vendor": null,
"links": "",
"mitigation": {
"pull-request": "",
"date-mitigation": "",
"description": "Set up the correct permissions (e.g., using setuid()) and try to open the file directly"
},
"flaw": {
"application": "N/A",
"reproducibility": "always",
"detected-by": "Alias Robotics",
"languages": "None",
"package": "N/A",
"reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_test_flawfinder/-/jobs/407215108/artifacts/download",
"date-detected": "2020-01-17 (17:36)",
"issue": "",
"specificity": "subject-specific",
"phase": "testing",
"trace": "(context) template<> String & ObjectStorage::Data::access();",
"architectural-location": "application-specific",
"reproduction": "See artifacts below (if available)",
"reported-by-relationship": "automatic",
"date-reported": "2020-01-17 (17:36)",
"reported-by": "Alias Robotics",
"detected-by-method": "testing static",
"subsystem": "N/A"
},
"description": "This usually indicates a security flaw. If an attacker can change anything along the path between the call to access() and the files actual use (e.g., by moving files), the attacker can exploit the race condition (CWE-362/CWE-367!). Set up the correct permissions (e.g., using setuid()) and try to open the file directly. . Happening @ .../canopen_master/objdict.h:525"
}
```
|
1.0
|
CWE-362/CWE-367 (race), This usually indicates a security flaw. If an attacker can change any... @ /canopen_master/objdict.h:525 - ```yaml
{
"system": "./install/canopen_master/include/canopen_master/objdict.h:525:42",
"id": 1,
"title": "CWE-362/CWE-367 (race), This usually indicates a security flaw. If an attacker can change any... @ /canopen_master/objdict.h:525",
"cwe": [
"CWE-362",
"CWE-367"
],
"exploitation": {
"exploitation-image": "",
"exploitation-vector": "",
"description": ""
},
"keywords": [
"flawfinder",
"flawfinder_level_4",
"static analysis",
"testing",
"triage",
"CWE-362",
"CWE-367",
"bug"
],
"type": "bug",
"cve": "None",
"severity": {
"rvss-vector": "",
"rvss-score": 0,
"cvss-score": 0,
"severity-description": "",
"cvss-vector": ""
},
"vendor": null,
"links": "",
"mitigation": {
"pull-request": "",
"date-mitigation": "",
"description": "Set up the correct permissions (e.g., using setuid()) and try to open the file directly"
},
"flaw": {
"application": "N/A",
"reproducibility": "always",
"detected-by": "Alias Robotics",
"languages": "None",
"package": "N/A",
"reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_test_flawfinder/-/jobs/407215108/artifacts/download",
"date-detected": "2020-01-17 (17:36)",
"issue": "",
"specificity": "subject-specific",
"phase": "testing",
"trace": "(context) template<> String & ObjectStorage::Data::access();",
"architectural-location": "application-specific",
"reproduction": "See artifacts below (if available)",
"reported-by-relationship": "automatic",
"date-reported": "2020-01-17 (17:36)",
"reported-by": "Alias Robotics",
"detected-by-method": "testing static",
"subsystem": "N/A"
},
"description": "This usually indicates a security flaw. If an attacker can change anything along the path between the call to access() and the files actual use (e.g., by moving files), the attacker can exploit the race condition (CWE-362/CWE-367!). Set up the correct permissions (e.g., using setuid()) and try to open the file directly. . Happening @ .../canopen_master/objdict.h:525"
}
```
|
test
|
cwe cwe race this usually indicates a security flaw if an attacker can change any canopen master objdict h yaml system install canopen master include canopen master objdict h id title cwe cwe race this usually indicates a security flaw if an attacker can change any canopen master objdict h cwe cwe cwe exploitation exploitation image exploitation vector description keywords flawfinder flawfinder level static analysis testing triage cwe cwe bug type bug cve none severity rvss vector rvss score cvss score severity description cvss vector vendor null links mitigation pull request date mitigation description set up the correct permissions e g using setuid and try to open the file directly flaw application n a reproducibility always detected by alias robotics languages none package n a reproduction image gitlab com aliasrobotics offensive alurity pipelines active pipeline test flawfinder jobs artifacts download date detected issue specificity subject specific phase testing trace context template string objectstorage data access architectural location application specific reproduction see artifacts below if available reported by relationship automatic date reported reported by alias robotics detected by method testing static subsystem n a description this usually indicates a security flaw if an attacker can change anything along the path between the call to access and the files actual use e g by moving files the attacker can exploit the race condition cwe cwe set up the correct permissions e g using setuid and try to open the file directly happening canopen master objdict h
| 1
|
23,990
| 11,996,347,598
|
IssuesEvent
|
2020-04-08 16:36:37
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Syntax Error on KEY
|
Pri2 cognitive-services/svc face-api/subsvc
|
Hi,
I followed your instruction ... KEY = os.environ['FACE_SUBSCRIPTION_KEY']
I replace it something like KEY = os.environ['123567890'] according to the Key value from my Azure portal
but I got Key error message. Any suggestions? What's the correct way to set up the Key? My email address is mchou@cdm.depaul.edu Thanks.
Regards,
Marco
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 9d930c16-d536-f5d3-fed6-8e08a9d4ae19
* Version Independent ID: 505b497d-5bb1-57c9-c400-8305e0a9300b
* Content: [Quickstart: Face client library for Python - Azure Cognitive Services](https://docs.microsoft.com/en-us/azure/cognitive-services/face/quickstarts/python-sdk#feedback)
* Content Source: [articles/cognitive-services/Face/QuickStarts/python-sdk.md](https://github.com/Microsoft/azure-docs/blob/master/articles/cognitive-services/Face/QuickStarts/python-sdk.md)
* Service: **cognitive-services**
* Sub-service: **face-api**
* GitHub Login: @PatrickFarley
* Microsoft Alias: **pafarley**
|
1.0
|
Syntax Error on KEY - Hi,
I followed your instruction ... KEY = os.environ['FACE_SUBSCRIPTION_KEY']
I replace it something like KEY = os.environ['123567890'] according to the Key value from my Azure portal
but I got Key error message. Any suggestions? What's the correct way to set up the Key? My email address is mchou@cdm.depaul.edu Thanks.
Regards,
Marco
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 9d930c16-d536-f5d3-fed6-8e08a9d4ae19
* Version Independent ID: 505b497d-5bb1-57c9-c400-8305e0a9300b
* Content: [Quickstart: Face client library for Python - Azure Cognitive Services](https://docs.microsoft.com/en-us/azure/cognitive-services/face/quickstarts/python-sdk#feedback)
* Content Source: [articles/cognitive-services/Face/QuickStarts/python-sdk.md](https://github.com/Microsoft/azure-docs/blob/master/articles/cognitive-services/Face/QuickStarts/python-sdk.md)
* Service: **cognitive-services**
* Sub-service: **face-api**
* GitHub Login: @PatrickFarley
* Microsoft Alias: **pafarley**
|
non_test
|
syntax error on key hi i followed your instruction key os environ i replace it something like key os environ according to the key value from my azure portal but i got key error message any suggestions what s the correct way to set up the key my email address is mchou cdm depaul edu thanks regards marco document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service cognitive services sub service face api github login patrickfarley microsoft alias pafarley
| 0
|
65,353
| 19,429,147,209
|
IssuesEvent
|
2021-12-21 09:54:06
|
GoldenSoftwareLtd/gedemin
|
https://api.github.com/repos/GoldenSoftwareLtd/gedemin
|
opened
|
На вкладке с множеством в гриде не отображаются колонки
|
Type-Defect Component-UI Priority-Critical
|
Сейчас в диалоговых окнах, где есть атрибуты типа множество на вкладке с элементами множества грид по умолчанию отображается со всеми невидимыми колонками. Например, окно группа компаний:

Надо посмотреть по коду. По умолчанию должна быть видима колонка с полем наименование для бизнес-класса. Получается методом getListField.
|
1.0
|
На вкладке с множеством в гриде не отображаются колонки - Сейчас в диалоговых окнах, где есть атрибуты типа множество на вкладке с элементами множества грид по умолчанию отображается со всеми невидимыми колонками. Например, окно группа компаний:

Надо посмотреть по коду. По умолчанию должна быть видима колонка с полем наименование для бизнес-класса. Получается методом getListField.
|
non_test
|
на вкладке с множеством в гриде не отображаются колонки сейчас в диалоговых окнах где есть атрибуты типа множество на вкладке с элементами множества грид по умолчанию отображается со всеми невидимыми колонками например окно группа компаний надо посмотреть по коду по умолчанию должна быть видима колонка с полем наименование для бизнес класса получается методом getlistfield
| 0
|
371,352
| 10,965,325,437
|
IssuesEvent
|
2019-11-28 02:24:59
|
connect-foundation/2019-21
|
https://api.github.com/repos/connect-foundation/2019-21
|
opened
|
host-app FE 질문 state 에 따른 출력 방식 변경
|
DB discussion priority:middle react
|
현재 질문에 대한 state 는 dummyData 를 출력하기 위해
- moderation
- newQuestion
- popularQuestion
- completeQuestion
인데, `DB` 상에는
- moderation
- active
- completeQuestion
으로 저장되는 상태.
DB 에서 가져온
`active` 상태의 question 을 생성 시간에 따라 newQuestion,
`like` 수에 따라 popularQuestion 이라는 state 를 붙이고 sorting 해주는 중간 과정이 필요.
|
1.0
|
host-app FE 질문 state 에 따른 출력 방식 변경 - 현재 질문에 대한 state 는 dummyData 를 출력하기 위해
- moderation
- newQuestion
- popularQuestion
- completeQuestion
인데, `DB` 상에는
- moderation
- active
- completeQuestion
으로 저장되는 상태.
DB 에서 가져온
`active` 상태의 question 을 생성 시간에 따라 newQuestion,
`like` 수에 따라 popularQuestion 이라는 state 를 붙이고 sorting 해주는 중간 과정이 필요.
|
non_test
|
host app fe 질문 state 에 따른 출력 방식 변경 현재 질문에 대한 state 는 dummydata 를 출력하기 위해 moderation newquestion popularquestion completequestion 인데 db 상에는 moderation active completequestion 으로 저장되는 상태 db 에서 가져온 active 상태의 question 을 생성 시간에 따라 newquestion like 수에 따라 popularquestion 이라는 state 를 붙이고 sorting 해주는 중간 과정이 필요
| 0
|
3,353
| 2,610,061,139
|
IssuesEvent
|
2015-02-26 18:17:59
|
chrsmith/jsjsj122
|
https://api.github.com/repos/chrsmith/jsjsj122
|
opened
|
路桥看不育去哪里最好
|
auto-migrated Priority-Medium Type-Defect
|
```
路桥看不育去哪里最好【台州五洲生殖医院】24小时健康咨询
热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州市
椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108、1
18、198及椒江一金清公交车直达枫南小区,乘坐107、105、109、
112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:19
|
1.0
|
路桥看不育去哪里最好 - ```
路桥看不育去哪里最好【台州五洲生殖医院】24小时健康咨询
热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州市
椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108、1
18、198及椒江一金清公交车直达枫南小区,乘坐107、105、109、
112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:19
|
non_test
|
路桥看不育去哪里最好 路桥看不育去哪里最好【台州五洲生殖医院】 热线 微信号tzwzszyy 医院地址 台州市 (枫南大转盘旁)乘车线路 、 、 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
| 0
|
237,635
| 19,662,298,332
|
IssuesEvent
|
2022-01-10 18:18:30
|
RocketChat/Rocket.Chat.Electron
|
https://api.github.com/repos/RocketChat/Rocket.Chat.Electron
|
closed
|
Updates on macOS do not install
|
type: bug plat: osx stat: needs testing
|
## My Setup
- Operating System: 10.13.6
- App Version: 2.14.0 (previous version also has this issue)
- Installation type: dmg
<!-- Answer questions by putting x in box, e.g. [x] -->
- [ x] I have tested with the latest version
- [ x] I can simulate the issue easily
### Current Behavior
0. Click on rocket chat icon.
1. New update is available!
2. Click install update ... blah blah. you will be notified when its ready...
3. Click install now.
4. Program closes.
5. Click on rocket chat icon.
6. Goto step #1. (repeat forever).
### Expected Behavior
Not an endless loop as described above.
I did figure out a workaround. Basically if you run chat from the command line, a zip file is listed. You can download the zip file and uncompress/install manually which installs the update!! Yay!
|
1.0
|
Updates on macOS do not install -
## My Setup
- Operating System: 10.13.6
- App Version: 2.14.0 (previous version also has this issue)
- Installation type: dmg
<!-- Answer questions by putting x in box, e.g. [x] -->
- [ x] I have tested with the latest version
- [ x] I can simulate the issue easily
### Current Behavior
0. Click on rocket chat icon.
1. New update is available!
2. Click install update ... blah blah. you will be notified when its ready...
3. Click install now.
4. Program closes.
5. Click on rocket chat icon.
6. Goto step #1. (repeat forever).
### Expected Behavior
Not an endless loop as described above.
I did figure out a workaround. Basically if you run chat from the command line, a zip file is listed. You can download the zip file and uncompress/install manually which installs the update!! Yay!
|
test
|
updates on macos do not install my setup operating system app version previous version also has this issue installation type dmg i have tested with the latest version i can simulate the issue easily current behavior click on rocket chat icon new update is available click install update blah blah you will be notified when its ready click install now program closes click on rocket chat icon goto step repeat forever expected behavior not an endless loop as described above i did figure out a workaround basically if you run chat from the command line a zip file is listed you can download the zip file and uncompress install manually which installs the update yay
| 1
|
795,696
| 28,082,759,214
|
IssuesEvent
|
2023-03-30 07:44:57
|
AY2223S2-CS2103-F11-3/tp
|
https://api.github.com/repos/AY2223S2-CS2103-F11-3/tp
|
closed
|
As a user, I can enter relative time as dates
|
type.Story priority.Low
|
... so that I need not enter the entire date.
Improve date parser method in `ParserUtil` to allow relative times.
* `today` - today at 0000.
* If now is 2023-05-03 0455, `today` will parse to 2023-05-0455 0000
* `hhmm` - earliest time from now to the time hhmm that has not yet passed.
* If now is 2023-05-03 0455, `0400` will parse to 2023-05-04 0400
|
1.0
|
As a user, I can enter relative time as dates - ... so that I need not enter the entire date.
Improve date parser method in `ParserUtil` to allow relative times.
* `today` - today at 0000.
* If now is 2023-05-03 0455, `today` will parse to 2023-05-0455 0000
* `hhmm` - earliest time from now to the time hhmm that has not yet passed.
* If now is 2023-05-03 0455, `0400` will parse to 2023-05-04 0400
|
non_test
|
as a user i can enter relative time as dates so that i need not enter the entire date improve date parser method in parserutil to allow relative times today today at if now is today will parse to hhmm earliest time from now to the time hhmm that has not yet passed if now is will parse to
| 0
|
89,120
| 3,789,897,007
|
IssuesEvent
|
2016-03-21 19:31:54
|
NuGet/Home
|
https://api.github.com/repos/NuGet/Home
|
closed
|
All sources "browse" can hit a null ref, and continually retry
|
NuGet Visual Studio UI Priority:1 ShipRoom Check
|
My configuration is two Klondike servers, one accessed via localhost and the other via remote request to a VM.
Both of them have ~5 different packages, with a few versions in common, etc...
Browse via All sources...and you'll see an error with Null Ref in output for package manager.
@alpaix is investigating.
|
1.0
|
All sources "browse" can hit a null ref, and continually retry - My configuration is two Klondike servers, one accessed via localhost and the other via remote request to a VM.
Both of them have ~5 different packages, with a few versions in common, etc...
Browse via All sources...and you'll see an error with Null Ref in output for package manager.
@alpaix is investigating.
|
non_test
|
all sources browse can hit a null ref and continually retry my configuration is two klondike servers one accessed via localhost and the other via remote request to a vm both of them have different packages with a few versions in common etc browse via all sources and you ll see an error with null ref in output for package manager alpaix is investigating
| 0
|
641,311
| 20,823,790,961
|
IssuesEvent
|
2022-03-18 18:10:44
|
rathena/rathena
|
https://api.github.com/repos/rathena/rathena
|
closed
|
4th Class Windhawk can attack normally when riding the wolf?
|
status:confirmed component:core priority:low mode:renewal type:bug
|
<!-- NOTE: Anything within these brackets will be hidden on the preview of the Issue. -->
* **rAthena Hash**:
latest
<!-- Please specify the rAthena [GitHub hash](https://help.github.com/articles/autolinked-references-and-urls/#commit-shas) on which you encountered this issue.
How to get your GitHub Hash:
1. cd your/rAthena/directory/
2. git rev-parse --short HEAD
3. Copy the resulting hash.
-->
* **Client Date**: 20211103
<!-- Please specify the client date you used. -->
* **Server Mode**: Re
<!-- Which mode does your server use: Pre-Renewal or Renewal? -->
* **Description of Issue**: Windhawk can attack normally when riding the wolf, is this the offical behaviour?
* Result: <!-- Describe the issue that you experienced in detail. -->
* Expected Result: <!-- Describe what you would expect to happen in detail. -->
* How to Reproduce: <!-- If you have not stated in the description of the result already, please give us a short guide how we can reproduce your issue. -->4th class Windhawk, use WUGRIDER skill, then attack target
* Official Information: <!-- If possible, provide information from official servers (kRO or other sources) which prove that the result is wrong. Please take into account that iRO (especially iRO Wiki) is not always the same as kRO. -->
<!-- * _NOTE: Make sure you quote ``` `@atcommands` ``` just like this so that you do not tag uninvolved GitHub users!_ -->
* **Modifications that may affect results**:
<!-- * Please provide any information that could influence the expected result. -->
<!-- * This can be either configurations you changed, database values you changed, or even external source modifications. -->
|
1.0
|
4th Class Windhawk can attack normally when riding the wolf? - <!-- NOTE: Anything within these brackets will be hidden on the preview of the Issue. -->
* **rAthena Hash**:
latest
<!-- Please specify the rAthena [GitHub hash](https://help.github.com/articles/autolinked-references-and-urls/#commit-shas) on which you encountered this issue.
How to get your GitHub Hash:
1. cd your/rAthena/directory/
2. git rev-parse --short HEAD
3. Copy the resulting hash.
-->
* **Client Date**: 20211103
<!-- Please specify the client date you used. -->
* **Server Mode**: Re
<!-- Which mode does your server use: Pre-Renewal or Renewal? -->
* **Description of Issue**: Windhawk can attack normally when riding the wolf, is this the offical behaviour?
* Result: <!-- Describe the issue that you experienced in detail. -->
* Expected Result: <!-- Describe what you would expect to happen in detail. -->
* How to Reproduce: <!-- If you have not stated in the description of the result already, please give us a short guide how we can reproduce your issue. -->4th class Windhawk, use WUGRIDER skill, then attack target
* Official Information: <!-- If possible, provide information from official servers (kRO or other sources) which prove that the result is wrong. Please take into account that iRO (especially iRO Wiki) is not always the same as kRO. -->
<!-- * _NOTE: Make sure you quote ``` `@atcommands` ``` just like this so that you do not tag uninvolved GitHub users!_ -->
* **Modifications that may affect results**:
<!-- * Please provide any information that could influence the expected result. -->
<!-- * This can be either configurations you changed, database values you changed, or even external source modifications. -->
|
non_test
|
class windhawk can attack normally when riding the wolf rathena hash latest please specify the rathena on which you encountered this issue how to get your github hash cd your rathena directory git rev parse short head copy the resulting hash client date server mode re description of issue windhawk can attack normally when riding the wolf is this the offical behaviour result expected result how to reproduce class windhawk use wugrider skill then attack target official information modifications that may affect results
| 0
|
162,086
| 12,619,360,474
|
IssuesEvent
|
2020-06-13 00:14:42
|
avidrucker/autofocus-exp
|
https://api.github.com/repos/avidrucker/autofocus-exp
|
closed
|
User can focus in on current todo
|
awaiting tests clarify no-issue-activity question user story
|
ala the zen design of https://momentumdash.com/
Also called "focus mode", the user sees simply the current most want to do item, and a prompt (or button) to leave the screen / progress forward.
Note: This is not a "screen" per se, but simply a clearing of the console, a display of the current in-progress todo item, and a prompt to hit the enter key to answer the question, "Is there further work left on this todo item? (y/n, or, hit 'q' then 'enter' to quit without saving progress) " where the next "screen" (display) is the return to main menu (prompt).
To sum it up:
- [x] 1. User sees the console cleared
- [x] 2. User sees current in-progress todo item (the cmwtd)
- [ ] 3. User can hit any key to indicate they are "done working for now"
- [ ] 4A. User can answer whether there is work left to do in the future (y/n/quit)
- [ ] 4B. User can quit to indicate that the cmwtd shouldn't be marked as complete
- [ ] 5. User is returned to the main menu
|
1.0
|
User can focus in on current todo - ala the zen design of https://momentumdash.com/
Also called "focus mode", the user sees simply the current most want to do item, and a prompt (or button) to leave the screen / progress forward.
Note: This is not a "screen" per se, but simply a clearing of the console, a display of the current in-progress todo item, and a prompt to hit the enter key to answer the question, "Is there further work left on this todo item? (y/n, or, hit 'q' then 'enter' to quit without saving progress) " where the next "screen" (display) is the return to main menu (prompt).
To sum it up:
- [x] 1. User sees the console cleared
- [x] 2. User sees current in-progress todo item (the cmwtd)
- [ ] 3. User can hit any key to indicate they are "done working for now"
- [ ] 4A. User can answer whether there is work left to do in the future (y/n/quit)
- [ ] 4B. User can quit to indicate that the cmwtd shouldn't be marked as complete
- [ ] 5. User is returned to the main menu
|
test
|
user can focus in on current todo ala the zen design of also called focus mode the user sees simply the current most want to do item and a prompt or button to leave the screen progress forward note this is not a screen per se but simply a clearing of the console a display of the current in progress todo item and a prompt to hit the enter key to answer the question is there further work left on this todo item y n or hit q then enter to quit without saving progress where the next screen display is the return to main menu prompt to sum it up user sees the console cleared user sees current in progress todo item the cmwtd user can hit any key to indicate they are done working for now user can answer whether there is work left to do in the future y n quit user can quit to indicate that the cmwtd shouldn t be marked as complete user is returned to the main menu
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.