Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
102,192
16,548,284,026
IssuesEvent
2021-05-28 04:35:54
samq-ghdemo/Java-Demo
https://api.github.com/repos/samq-ghdemo/Java-Demo
opened
CVE-2017-3589 (Low) detected in mysql-connector-java-5.1.25.jar
security vulnerability
## CVE-2017-3589 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mysql-connector-java-5.1.25.jar</b></p></summary> <p>MySQL JDBC Type 4 driver</p> <p>Library home page: <a href="http://dev.mysql.com/doc/connector-j/en/">http://dev.mysql.com/doc/connector-j/en/</a></p> <p>Path to dependency file: Java-Demo/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/mysql/mysql-connector-java/5.1.25/mysql-connector-java-5.1.25.jar,Java-Demo/target/easybuggy-1-SNAPSHOT/WEB-INF/lib/mysql-connector-java-5.1.25.jar,Java-Demo/.extract/webapps/ROOT/WEB-INF/lib/mysql-connector-java-5.1.25.jar</p> <p> Dependency Hierarchy: - :x: **mysql-connector-java-5.1.25.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/Java-Demo/commit/ebe5ad9510b919e1dd226c70afc5e1115136f8cf">ebe5ad9510b919e1dd226c70afc5e1115136f8cf</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Vulnerability in the MySQL Connectors component of Oracle MySQL (subcomponent: Connector/J). Supported versions that are affected are 5.1.41 and earlier. Easily "exploitable" vulnerability allows low privileged attacker with logon to the infrastructure where MySQL Connectors executes to compromise MySQL Connectors. Successful attacks of this vulnerability can result in unauthorized update, insert or delete access to some of MySQL Connectors accessible data. CVSS 3.0 Base Score 3.3 (Integrity impacts). CVSS Vector: (CVSS:3.0/AV:L/AC:L/PR:L/UI:N/S:U/C:N/I:L/A:N). <p>Publish Date: 2017-04-24 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-3589>CVE-2017-3589</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-3589">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-3589</a></p> <p>Release Date: 2017-04-24</p> <p>Fix Resolution: 5.1.42</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"mysql","packageName":"mysql-connector-java","packageVersion":"5.1.25","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"mysql:mysql-connector-java:5.1.25","isMinimumFixVersionAvailable":true,"minimumFixVersion":"5.1.42"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-3589","vulnerabilityDetails":"Vulnerability in the MySQL Connectors component of Oracle MySQL (subcomponent: Connector/J). Supported versions that are affected are 5.1.41 and earlier. Easily \"exploitable\" vulnerability allows low privileged attacker with logon to the infrastructure where MySQL Connectors executes to compromise MySQL Connectors. Successful attacks of this vulnerability can result in unauthorized update, insert or delete access to some of MySQL Connectors accessible data. CVSS 3.0 Base Score 3.3 (Integrity impacts). CVSS Vector: (CVSS:3.0/AV:L/AC:L/PR:L/UI:N/S:U/C:N/I:L/A:N).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-3589","cvss3Severity":"low","cvss3Score":"3.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Local","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2017-3589 (Low) detected in mysql-connector-java-5.1.25.jar - ## CVE-2017-3589 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mysql-connector-java-5.1.25.jar</b></p></summary> <p>MySQL JDBC Type 4 driver</p> <p>Library home page: <a href="http://dev.mysql.com/doc/connector-j/en/">http://dev.mysql.com/doc/connector-j/en/</a></p> <p>Path to dependency file: Java-Demo/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/mysql/mysql-connector-java/5.1.25/mysql-connector-java-5.1.25.jar,Java-Demo/target/easybuggy-1-SNAPSHOT/WEB-INF/lib/mysql-connector-java-5.1.25.jar,Java-Demo/.extract/webapps/ROOT/WEB-INF/lib/mysql-connector-java-5.1.25.jar</p> <p> Dependency Hierarchy: - :x: **mysql-connector-java-5.1.25.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/Java-Demo/commit/ebe5ad9510b919e1dd226c70afc5e1115136f8cf">ebe5ad9510b919e1dd226c70afc5e1115136f8cf</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Vulnerability in the MySQL Connectors component of Oracle MySQL (subcomponent: Connector/J). Supported versions that are affected are 5.1.41 and earlier. Easily "exploitable" vulnerability allows low privileged attacker with logon to the infrastructure where MySQL Connectors executes to compromise MySQL Connectors. Successful attacks of this vulnerability can result in unauthorized update, insert or delete access to some of MySQL Connectors accessible data. CVSS 3.0 Base Score 3.3 (Integrity impacts). CVSS Vector: (CVSS:3.0/AV:L/AC:L/PR:L/UI:N/S:U/C:N/I:L/A:N). <p>Publish Date: 2017-04-24 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-3589>CVE-2017-3589</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-3589">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-3589</a></p> <p>Release Date: 2017-04-24</p> <p>Fix Resolution: 5.1.42</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"mysql","packageName":"mysql-connector-java","packageVersion":"5.1.25","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"mysql:mysql-connector-java:5.1.25","isMinimumFixVersionAvailable":true,"minimumFixVersion":"5.1.42"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-3589","vulnerabilityDetails":"Vulnerability in the MySQL Connectors component of Oracle MySQL (subcomponent: Connector/J). Supported versions that are affected are 5.1.41 and earlier. Easily \"exploitable\" vulnerability allows low privileged attacker with logon to the infrastructure where MySQL Connectors executes to compromise MySQL Connectors. Successful attacks of this vulnerability can result in unauthorized update, insert or delete access to some of MySQL Connectors accessible data. CVSS 3.0 Base Score 3.3 (Integrity impacts). CVSS Vector: (CVSS:3.0/AV:L/AC:L/PR:L/UI:N/S:U/C:N/I:L/A:N).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-3589","cvss3Severity":"low","cvss3Score":"3.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Local","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_test
cve low detected in mysql connector java jar cve low severity vulnerability vulnerable library mysql connector java jar mysql jdbc type driver library home page a href path to dependency file java demo pom xml path to vulnerable library canner repository mysql mysql connector java mysql connector java jar java demo target easybuggy snapshot web inf lib mysql connector java jar java demo extract webapps root web inf lib mysql connector java jar dependency hierarchy x mysql connector java jar vulnerable library found in head commit a href found in base branch main vulnerability details vulnerability in the mysql connectors component of oracle mysql subcomponent connector j supported versions that are affected are and earlier easily exploitable vulnerability allows low privileged attacker with logon to the infrastructure where mysql connectors executes to compromise mysql connectors successful attacks of this vulnerability can result in unauthorized update insert or delete access to some of mysql connectors accessible data cvss base score integrity impacts cvss vector cvss av l ac l pr l ui n s u c n i l a n publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree mysql mysql connector java isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails vulnerability in the mysql connectors component of oracle mysql subcomponent connector j supported versions that are affected are and earlier easily exploitable vulnerability allows low privileged attacker with logon to the infrastructure where mysql connectors executes to compromise mysql connectors successful attacks of this vulnerability can result in unauthorized update insert or delete access to some of mysql connectors accessible data cvss base score integrity impacts cvss vector cvss av l ac l pr l ui n s u c n i l a n vulnerabilityurl
0
33,685
4,848,973,578
IssuesEvent
2016-11-10 19:04:34
radare/radare2
https://api.github.com/repos/radare/radare2
closed
Commands are sensitive to double-whitespace
has-test
Some commands will not work correctly with multiple spaces instead of one, but will not report an error/warning either. For example: ``` // this works f foobar=0x1000 // this does nothing f foobar=0x1000 ``` version 0.10.6
1.0
Commands are sensitive to double-whitespace - Some commands will not work correctly with multiple spaces instead of one, but will not report an error/warning either. For example: ``` // this works f foobar=0x1000 // this does nothing f foobar=0x1000 ``` version 0.10.6
test
commands are sensitive to double whitespace some commands will not work correctly with multiple spaces instead of one but will not report an error warning either for example this works f foobar this does nothing f foobar version
1
454,401
13,100,218,753
IssuesEvent
2020-08-03 23:48:11
GoogleCloudPlatform/stackdriver-sandbox
https://api.github.com/repos/GoogleCloudPlatform/stackdriver-sandbox
opened
Credential check fail when accessing the storage bucket
priority: p2 type: bug
Some users may encounter a credential check failure when accessing the storage bucket in Terraform initialization. They need to do "gcloud auth application-default login". We need to put a try-catch block around the initialization.
1.0
Credential check fail when accessing the storage bucket - Some users may encounter a credential check failure when accessing the storage bucket in Terraform initialization. They need to do "gcloud auth application-default login". We need to put a try-catch block around the initialization.
non_test
credential check fail when accessing the storage bucket some users may encounter a credential check failure when accessing the storage bucket in terraform initialization they need to do gcloud auth application default login we need to put a try catch block around the initialization
0
5,715
2,790,522,000
IssuesEvent
2015-05-09 09:27:17
Dalmirog/OctoPosh
https://api.github.com/repos/Dalmirog/OctoPosh
reopened
Improve tests for Get-* cmdlets
Testing
They should be more Unit-test-like and consider scenarios like: - Get-* should not get a resource that doesnt exist - Get-* [specific names] should only get resources with those names, and nothing extra - Get-* with date filters should return results with between the correct date ranges - Get-* with version filters should return results with between the correct version ranges Progress tracking - [ ] Get-OctopusConnectionInfo - [ ] Get-OctopusDeployment - [ ] Get-OctopusEnvironment - [ ] Get-OctopusLifeCycle - [ ] Get-OctopusMaintenanceMode - [ ] Get-OctopusProject - [ ] Get-OctopusProjectGroup - [ ] Get-OctopusProjectVariable - [ ] Get-OctopusRelease - [ ] Get-OctopusResourceModel - [ ] Get-OctopusSMTPConfig
1.0
Improve tests for Get-* cmdlets - They should be more Unit-test-like and consider scenarios like: - Get-* should not get a resource that doesnt exist - Get-* [specific names] should only get resources with those names, and nothing extra - Get-* with date filters should return results with between the correct date ranges - Get-* with version filters should return results with between the correct version ranges Progress tracking - [ ] Get-OctopusConnectionInfo - [ ] Get-OctopusDeployment - [ ] Get-OctopusEnvironment - [ ] Get-OctopusLifeCycle - [ ] Get-OctopusMaintenanceMode - [ ] Get-OctopusProject - [ ] Get-OctopusProjectGroup - [ ] Get-OctopusProjectVariable - [ ] Get-OctopusRelease - [ ] Get-OctopusResourceModel - [ ] Get-OctopusSMTPConfig
test
improve tests for get cmdlets they should be more unit test like and consider scenarios like get should not get a resource that doesnt exist get should only get resources with those names and nothing extra get with date filters should return results with between the correct date ranges get with version filters should return results with between the correct version ranges progress tracking get octopusconnectioninfo get octopusdeployment get octopusenvironment get octopuslifecycle get octopusmaintenancemode get octopusproject get octopusprojectgroup get octopusprojectvariable get octopusrelease get octopusresourcemodel get octopussmtpconfig
1
325,821
27,964,388,929
IssuesEvent
2023-03-24 18:08:47
Satellite-im/testing-uplink
https://api.github.com/repos/Satellite-im/testing-uplink
opened
UI Tests - Settings Developer - Save Logs In a File
test Settings
Logs should save in a file when User toggles on Save Logs In A File
1.0
UI Tests - Settings Developer - Save Logs In a File - Logs should save in a file when User toggles on Save Logs In A File
test
ui tests settings developer save logs in a file logs should save in a file when user toggles on save logs in a file
1
92,837
11,714,795,081
IssuesEvent
2020-03-09 13:03:47
EightShapes/esds-library
https://api.github.com/repos/EightShapes/esds-library
closed
Required Field Indicator [Design Asset]
Component [Design Asset]
### Design Starting Point This is the key for how forms denote required fields, which will usually be positioned below the title of the form. ![Screen Shot 2020-02-24 at 10 50 03 AM](https://user-images.githubusercontent.com/1207666/75168394-c7a96d80-56f4-11ea-9d58-47b0c65f8fd0.png) ### Must Have * Settings: white, light, dark, black * Leverages existing typography * Matches indicator on `Label`
1.0
Required Field Indicator [Design Asset] - ### Design Starting Point This is the key for how forms denote required fields, which will usually be positioned below the title of the form. ![Screen Shot 2020-02-24 at 10 50 03 AM](https://user-images.githubusercontent.com/1207666/75168394-c7a96d80-56f4-11ea-9d58-47b0c65f8fd0.png) ### Must Have * Settings: white, light, dark, black * Leverages existing typography * Matches indicator on `Label`
non_test
required field indicator design starting point this is the key for how forms denote required fields which will usually be positioned below the title of the form must have settings white light dark black leverages existing typography matches indicator on label
0
32,133
4,751,450,468
IssuesEvent
2016-10-22 22:02:52
flowplayer/flowplayer
https://api.github.com/repos/flowplayer/flowplayer
closed
[Chrome] playlist: error on resume when last clip is in finished state
ready to test
https://flowplayer.org/standalone/playlist/javascript.html (no plugins) - finish last clip - resume by clicking or typing space or `flowplayer(0).resume()` in the console yields: ``` flowplayer.min.js:6 Uncaught (in promise) DOMException: The play() request was interrupted by a new load request ``` Which describes the problem exactly: play() is requested on the last clip, but interrupted by loading the first. Proof: entering `next()` in the console does not trigger the error. Or resuming when the playlist has length 1. Obviously does not happen either when advance is false. A fix most likely has to be cross-checked in the Flash engine.
1.0
[Chrome] playlist: error on resume when last clip is in finished state - https://flowplayer.org/standalone/playlist/javascript.html (no plugins) - finish last clip - resume by clicking or typing space or `flowplayer(0).resume()` in the console yields: ``` flowplayer.min.js:6 Uncaught (in promise) DOMException: The play() request was interrupted by a new load request ``` Which describes the problem exactly: play() is requested on the last clip, but interrupted by loading the first. Proof: entering `next()` in the console does not trigger the error. Or resuming when the playlist has length 1. Obviously does not happen either when advance is false. A fix most likely has to be cross-checked in the Flash engine.
test
playlist error on resume when last clip is in finished state no plugins finish last clip resume by clicking or typing space or flowplayer resume in the console yields flowplayer min js uncaught in promise domexception the play request was interrupted by a new load request which describes the problem exactly play is requested on the last clip but interrupted by loading the first proof entering next in the console does not trigger the error or resuming when the playlist has length obviously does not happen either when advance is false a fix most likely has to be cross checked in the flash engine
1
141,798
11,437,609,410
IssuesEvent
2020-02-05 00:30:18
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Go panic in NodeController every 1 second
[zube]: To Test internal
<!-- Please search for existing issues first, then read https://rancher.com/docs/rancher/v2.x/en/contributing/#bugs-issues-or-questions to see what we expect in an issue For security issues, please email security@rancher.com instead of posting a public issue in GitHub. You may (but are not required to) use the GPG key located on Keybase. --> **What kind of request is this (question/bug/enhancement/feature request):** bug **Steps to reproduce (least amount of steps as possible):** unknown **Result:** Seeing this about every 1 second in Rancher logs: ``` 2020-01-27T04:01:08.803268734Z 2020/01/27 04:01:08 [ERROR] NodeController xxxxxxxxxxxxxxxxx recovered from panic "runtime error: invalid memory address or nil pointer dereference". (err=<nil>) Call stack: 2020-01-27T04:01:08.803306432Z goroutine 443379 [running]: 2020-01-27T04:01:08.803312366Z github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/runtime.RecoverFromPanic(0xc0045a3d30) 2020-01-27T04:01:08.803316668Z /go/src/github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:158 +0xb5 2020-01-27T04:01:08.803320855Z panic(0x3208c60, 0x62c3810) 2020-01-27T04:01:08.803324801Z /usr/local/go/src/runtime/panic.go:679 +0x1b2 2020-01-27T04:01:08.803350733Z github.com/rancher/rancher/pkg/controllers/user/monitoring.(*clusterMonitoringEnabledHandler).syncWindowsNode(0xc002f03c00, 0xc006b49150, 0xf, 0x0, 0xc000ddd880, 0x7fb72dd1e6d0, 0x0, 0x0) 2020-01-27T04:01:08.803356241Z /go/src/github.com/rancher/rancher/pkg/controllers/user/monitoring/cluster_monitoring_enabled_handler.go:58 +0xe6 2020-01-27T04:01:08.803360044Z github.com/rancher/rancher/vendor/github.com/rancher/types/apis/core/v1.(*nodeController).AddHandler.func1(0xc006b49150, 0xf, 0x0, 0x0, 0xc0045a3cd8, 0x3, 0x3, 0x0) 2020-01-27T04:01:08.803363969Z /go/src/github.com/rancher/rancher/vendor/github.com/rancher/types/apis/core/v1/zz_generated_node_controller.go:149 +0xe0 2020-01-27T04:01:08.803387512Z github.com/rancher/rancher/vendor/github.com/rancher/norman/controller.(*genericController).syncHandler(0xc00207a700, 0x2f92320, 0xc008190e00, 0x0, 0x0) 2020-01-27T04:01:08.803390640Z /go/src/github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:367 +0x371 2020-01-27T04:01:08.803393545Z github.com/rancher/rancher/vendor/github.com/rancher/norman/controller.(*genericController).processNextWorkItem(0xc00207a700, 0x0) 2020-01-27T04:01:08.803396147Z /go/src/github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:276 +0xef 2020-01-27T04:01:08.803398815Z github.com/rancher/rancher/vendor/github.com/rancher/norman/controller.(*genericController).runWorker(0xc00207a700) 2020-01-27T04:01:08.803401547Z /go/src/github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:264 +0x2b 2020-01-27T04:01:08.803404977Z github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1(0xc00378f670) 2020-01-27T04:01:08.803408494Z /go/src/github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:152 +0x5e 2020-01-27T04:01:08.803419953Z github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil(0xc00378f670, 0x3b9aca00, 0x0, 0x1, 0xc003568ae0) 2020-01-27T04:01:08.803424598Z /go/src/github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:153 +0xf8 2020-01-27T04:01:08.803428243Z github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait.Until(0xc00378f670, 0x3b9aca00, 0xc003568ae0) 2020-01-27T04:01:08.803431733Z /go/src/github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88 +0x4d 2020-01-27T04:01:08.803435732Z created by github.com/rancher/rancher/vendor/github.com/rancher/norman/controller.(*genericController).run 2020-01-27T04:01:08.803439922Z /go/src/github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:256 +0xe8 ``` **Other details that may be helpful:** Could be related to https://github.com/rancher/rancher/issues/25164 **Environment information** - Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): v2.3.4 - Installation option (single install/HA): HA <!-- If the reported issue is regarding a created cluster, please provide requested info below --> **Cluster information** - Cluster type (Hosted/Infrastructure Provider/Custom/Imported): Custom - Machine type (cloud/VM/metal) and specifications (CPU/memory): - Kubernetes version (use `kubectl version`): ``` 1.15 ``` - Docker version (use `docker version`): ``` ? ```
1.0
Go panic in NodeController every 1 second - <!-- Please search for existing issues first, then read https://rancher.com/docs/rancher/v2.x/en/contributing/#bugs-issues-or-questions to see what we expect in an issue For security issues, please email security@rancher.com instead of posting a public issue in GitHub. You may (but are not required to) use the GPG key located on Keybase. --> **What kind of request is this (question/bug/enhancement/feature request):** bug **Steps to reproduce (least amount of steps as possible):** unknown **Result:** Seeing this about every 1 second in Rancher logs: ``` 2020-01-27T04:01:08.803268734Z 2020/01/27 04:01:08 [ERROR] NodeController xxxxxxxxxxxxxxxxx recovered from panic "runtime error: invalid memory address or nil pointer dereference". (err=<nil>) Call stack: 2020-01-27T04:01:08.803306432Z goroutine 443379 [running]: 2020-01-27T04:01:08.803312366Z github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/runtime.RecoverFromPanic(0xc0045a3d30) 2020-01-27T04:01:08.803316668Z /go/src/github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:158 +0xb5 2020-01-27T04:01:08.803320855Z panic(0x3208c60, 0x62c3810) 2020-01-27T04:01:08.803324801Z /usr/local/go/src/runtime/panic.go:679 +0x1b2 2020-01-27T04:01:08.803350733Z github.com/rancher/rancher/pkg/controllers/user/monitoring.(*clusterMonitoringEnabledHandler).syncWindowsNode(0xc002f03c00, 0xc006b49150, 0xf, 0x0, 0xc000ddd880, 0x7fb72dd1e6d0, 0x0, 0x0) 2020-01-27T04:01:08.803356241Z /go/src/github.com/rancher/rancher/pkg/controllers/user/monitoring/cluster_monitoring_enabled_handler.go:58 +0xe6 2020-01-27T04:01:08.803360044Z github.com/rancher/rancher/vendor/github.com/rancher/types/apis/core/v1.(*nodeController).AddHandler.func1(0xc006b49150, 0xf, 0x0, 0x0, 0xc0045a3cd8, 0x3, 0x3, 0x0) 2020-01-27T04:01:08.803363969Z /go/src/github.com/rancher/rancher/vendor/github.com/rancher/types/apis/core/v1/zz_generated_node_controller.go:149 +0xe0 2020-01-27T04:01:08.803387512Z github.com/rancher/rancher/vendor/github.com/rancher/norman/controller.(*genericController).syncHandler(0xc00207a700, 0x2f92320, 0xc008190e00, 0x0, 0x0) 2020-01-27T04:01:08.803390640Z /go/src/github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:367 +0x371 2020-01-27T04:01:08.803393545Z github.com/rancher/rancher/vendor/github.com/rancher/norman/controller.(*genericController).processNextWorkItem(0xc00207a700, 0x0) 2020-01-27T04:01:08.803396147Z /go/src/github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:276 +0xef 2020-01-27T04:01:08.803398815Z github.com/rancher/rancher/vendor/github.com/rancher/norman/controller.(*genericController).runWorker(0xc00207a700) 2020-01-27T04:01:08.803401547Z /go/src/github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:264 +0x2b 2020-01-27T04:01:08.803404977Z github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1(0xc00378f670) 2020-01-27T04:01:08.803408494Z /go/src/github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:152 +0x5e 2020-01-27T04:01:08.803419953Z github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil(0xc00378f670, 0x3b9aca00, 0x0, 0x1, 0xc003568ae0) 2020-01-27T04:01:08.803424598Z /go/src/github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:153 +0xf8 2020-01-27T04:01:08.803428243Z github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait.Until(0xc00378f670, 0x3b9aca00, 0xc003568ae0) 2020-01-27T04:01:08.803431733Z /go/src/github.com/rancher/rancher/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88 +0x4d 2020-01-27T04:01:08.803435732Z created by github.com/rancher/rancher/vendor/github.com/rancher/norman/controller.(*genericController).run 2020-01-27T04:01:08.803439922Z /go/src/github.com/rancher/rancher/vendor/github.com/rancher/norman/controller/generic_controller.go:256 +0xe8 ``` **Other details that may be helpful:** Could be related to https://github.com/rancher/rancher/issues/25164 **Environment information** - Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): v2.3.4 - Installation option (single install/HA): HA <!-- If the reported issue is regarding a created cluster, please provide requested info below --> **Cluster information** - Cluster type (Hosted/Infrastructure Provider/Custom/Imported): Custom - Machine type (cloud/VM/metal) and specifications (CPU/memory): - Kubernetes version (use `kubectl version`): ``` 1.15 ``` - Docker version (use `docker version`): ``` ? ```
test
go panic in nodecontroller every second please search for existing issues first then read to see what we expect in an issue for security issues please email security rancher com instead of posting a public issue in github you may but are not required to use the gpg key located on keybase what kind of request is this question bug enhancement feature request bug steps to reproduce least amount of steps as possible unknown result seeing this about every second in rancher logs nodecontroller xxxxxxxxxxxxxxxxx recovered from panic runtime error invalid memory address or nil pointer dereference err call stack goroutine github com rancher rancher vendor io apimachinery pkg util runtime recoverfrompanic go src github com rancher rancher vendor io apimachinery pkg util runtime runtime go panic usr local go src runtime panic go github com rancher rancher pkg controllers user monitoring clustermonitoringenabledhandler syncwindowsnode go src github com rancher rancher pkg controllers user monitoring cluster monitoring enabled handler go github com rancher rancher vendor github com rancher types apis core nodecontroller addhandler go src github com rancher rancher vendor github com rancher types apis core zz generated node controller go github com rancher rancher vendor github com rancher norman controller genericcontroller synchandler go src github com rancher rancher vendor github com rancher norman controller generic controller go github com rancher rancher vendor github com rancher norman controller genericcontroller processnextworkitem go src github com rancher rancher vendor github com rancher norman controller generic controller go github com rancher rancher vendor github com rancher norman controller genericcontroller runworker go src github com rancher rancher vendor github com rancher norman controller generic controller go github com rancher rancher vendor io apimachinery pkg util wait jitteruntil go src github com rancher rancher vendor io apimachinery pkg util wait wait go github com rancher rancher vendor io apimachinery pkg util wait jitteruntil go src github com rancher rancher vendor io apimachinery pkg util wait wait go github com rancher rancher vendor io apimachinery pkg util wait until go src github com rancher rancher vendor io apimachinery pkg util wait wait go created by github com rancher rancher vendor github com rancher norman controller genericcontroller run go src github com rancher rancher vendor github com rancher norman controller generic controller go other details that may be helpful could be related to environment information rancher version rancher rancher rancher server image tag or shown bottom left in the ui installation option single install ha ha if the reported issue is regarding a created cluster please provide requested info below cluster information cluster type hosted infrastructure provider custom imported custom machine type cloud vm metal and specifications cpu memory kubernetes version use kubectl version docker version use docker version
1
49,887
6,044,810,449
IssuesEvent
2017-06-12 07:24:52
pixelhumain/co2
https://api.github.com/repos/pixelhumain/co2
closed
Création URL
to test
Création d'une URL depuis une organisation : 1\ Erreur JS au chargement de la popup ![capture d ecran 2017-05-24 a 16 15 39](https://cloud.githubusercontent.com/assets/2662262/26402675/4de4940a-409c-11e7-8346-ef2a2ba7d5fa.png) 2\ Erreur à la validation du form au saveurl ![capture d ecran 2017-05-24 a 16 17 06](https://cloud.githubusercontent.com/assets/2662262/26402727/7f739e30-409c-11e7-935f-38f2e988c29a.png)
1.0
Création URL - Création d'une URL depuis une organisation : 1\ Erreur JS au chargement de la popup ![capture d ecran 2017-05-24 a 16 15 39](https://cloud.githubusercontent.com/assets/2662262/26402675/4de4940a-409c-11e7-8346-ef2a2ba7d5fa.png) 2\ Erreur à la validation du form au saveurl ![capture d ecran 2017-05-24 a 16 17 06](https://cloud.githubusercontent.com/assets/2662262/26402727/7f739e30-409c-11e7-935f-38f2e988c29a.png)
test
création url création d une url depuis une organisation erreur js au chargement de la popup erreur à la validation du form au saveurl
1
104,700
8,998,158,023
IssuesEvent
2019-02-02 19:08:35
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
When GlobalDNS entry is edited and set to different multicluster App , dns entries relating to the older multicluster App instances continue to be programmed in Route53.
area/multi-cluster-app kind/bug-qa release/alpha2 status/reopened status/resolved status/to-test team/ca version/2.0
Rancher server version - Build from master - jan 21 Setup - HA setup with 4 clusters. Steps to reproduce the the problem: 1. Create Multicluster App - mcapp1 for cluster1 and cluster2 with Global DNS enabled Ingress for fqdn - "test1.test.com" 2. Create globalDnsProvider with root domain - "test.com" 3. Create a globalDns entry with fqdn - "test1.test.com" pointing to Multicluster App - mcapp1. 4. Create Multicluster App - mcapp2 for cluster3 and cluster4 with Global DNS enabled Ingress for fqdn - "test1.test.com" 5. Edit GlobalDNS entry and set Multicluster App to point to mcapp2 Actual Result: Accessing test1.test.com should directs traffic to all worker nodes from cluster1, cluster2, cluster3 and cluster4. Route53 entry gets updated to point to all worker node ips from cluster1, cluster2, cluster3 and cluster4. Expected Results: Accessing test1.test.com should direct traffic to all worker nodes from cluster3 and cluster4. Route53 entry should get updated to point to all worker node ips from cluster3 and cluster4.
1.0
When GlobalDNS entry is edited and set to different multicluster App , dns entries relating to the older multicluster App instances continue to be programmed in Route53. - Rancher server version - Build from master - jan 21 Setup - HA setup with 4 clusters. Steps to reproduce the the problem: 1. Create Multicluster App - mcapp1 for cluster1 and cluster2 with Global DNS enabled Ingress for fqdn - "test1.test.com" 2. Create globalDnsProvider with root domain - "test.com" 3. Create a globalDns entry with fqdn - "test1.test.com" pointing to Multicluster App - mcapp1. 4. Create Multicluster App - mcapp2 for cluster3 and cluster4 with Global DNS enabled Ingress for fqdn - "test1.test.com" 5. Edit GlobalDNS entry and set Multicluster App to point to mcapp2 Actual Result: Accessing test1.test.com should directs traffic to all worker nodes from cluster1, cluster2, cluster3 and cluster4. Route53 entry gets updated to point to all worker node ips from cluster1, cluster2, cluster3 and cluster4. Expected Results: Accessing test1.test.com should direct traffic to all worker nodes from cluster3 and cluster4. Route53 entry should get updated to point to all worker node ips from cluster3 and cluster4.
test
when globaldns entry is edited and set to different multicluster app dns entries relating to the older multicluster app instances continue to be programmed in rancher server version build from master jan setup ha setup with clusters steps to reproduce the the problem create multicluster app for and with global dns enabled ingress for fqdn test com create globaldnsprovider with root domain test com create a globaldns entry with fqdn test com pointing to multicluster app create multicluster app for and with global dns enabled ingress for fqdn test com edit globaldns entry and set multicluster app to point to actual result accessing test com should directs traffic to all worker nodes from and entry gets updated to point to all worker node ips from and expected results accessing test com should direct traffic to all worker nodes from and entry should get updated to point to all worker node ips from and
1
240,587
20,049,706,353
IssuesEvent
2022-02-03 03:53:31
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
Failing test: Chrome X-Pack UI Functional Tests Basic License.x-pack/test/functional/apps/transform/creation_saved_search·ts - apps transform basic license transform creation_saved_search batch transform with unique term and sort by time with saved search filter navigates through the wizard and sets all needed fields
failed-test
A test failed on a tracked branch ``` Error: Transform id input text should be 'fq_2_1643859986892' (got '') at Assertion.assert (/opt/local-ssd/buildkite/builds/kb-n2-4-8a4b63cf8afa95a0/elastic/kibana-hourly/kibana/node_modules/@kbn/expect/expect.js:100:11) at Assertion.eql (/opt/local-ssd/buildkite/builds/kb-n2-4-8a4b63cf8afa95a0/elastic/kibana-hourly/kibana/node_modules/@kbn/expect/expect.js:244:8) at Object.assertTransformIdValue (test/functional/services/transform/wizard.ts:676:36) at runMicrotasks (<anonymous>) at processTicksAndRejections (node:internal/process/task_queues:96:5) at Object.setTransformId (test/functional/services/transform/wizard.ts:686:7) at Context.<anonymous> (test/functional/apps/transform/creation_saved_search.ts:222:11) at Object.apply (/opt/local-ssd/buildkite/builds/kb-n2-4-8a4b63cf8afa95a0/elastic/kibana-hourly/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) { actual: '', expected: 'fq_2_1643859986892', showDiff: true } ``` First failure: [CI Build - main](https://buildkite.com/elastic/kibana-hourly/builds/9122#823e459c-69ef-4792-9de3-49ba3f4a0b5a) <!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests Basic License.x-pack/test/functional/apps/transform/creation_saved_search·ts","test.name":"apps transform basic license transform creation_saved_search batch transform with unique term and sort by time with saved search filter navigates through the wizard and sets all needed fields","test.failCount":1}} -->
1.0
Failing test: Chrome X-Pack UI Functional Tests Basic License.x-pack/test/functional/apps/transform/creation_saved_search·ts - apps transform basic license transform creation_saved_search batch transform with unique term and sort by time with saved search filter navigates through the wizard and sets all needed fields - A test failed on a tracked branch ``` Error: Transform id input text should be 'fq_2_1643859986892' (got '') at Assertion.assert (/opt/local-ssd/buildkite/builds/kb-n2-4-8a4b63cf8afa95a0/elastic/kibana-hourly/kibana/node_modules/@kbn/expect/expect.js:100:11) at Assertion.eql (/opt/local-ssd/buildkite/builds/kb-n2-4-8a4b63cf8afa95a0/elastic/kibana-hourly/kibana/node_modules/@kbn/expect/expect.js:244:8) at Object.assertTransformIdValue (test/functional/services/transform/wizard.ts:676:36) at runMicrotasks (<anonymous>) at processTicksAndRejections (node:internal/process/task_queues:96:5) at Object.setTransformId (test/functional/services/transform/wizard.ts:686:7) at Context.<anonymous> (test/functional/apps/transform/creation_saved_search.ts:222:11) at Object.apply (/opt/local-ssd/buildkite/builds/kb-n2-4-8a4b63cf8afa95a0/elastic/kibana-hourly/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) { actual: '', expected: 'fq_2_1643859986892', showDiff: true } ``` First failure: [CI Build - main](https://buildkite.com/elastic/kibana-hourly/builds/9122#823e459c-69ef-4792-9de3-49ba3f4a0b5a) <!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests Basic License.x-pack/test/functional/apps/transform/creation_saved_search·ts","test.name":"apps transform basic license transform creation_saved_search batch transform with unique term and sort by time with saved search filter navigates through the wizard and sets all needed fields","test.failCount":1}} -->
test
failing test chrome x pack ui functional tests basic license x pack test functional apps transform creation saved search·ts apps transform basic license transform creation saved search batch transform with unique term and sort by time with saved search filter navigates through the wizard and sets all needed fields a test failed on a tracked branch error transform id input text should be fq got at assertion assert opt local ssd buildkite builds kb elastic kibana hourly kibana node modules kbn expect expect js at assertion eql opt local ssd buildkite builds kb elastic kibana hourly kibana node modules kbn expect expect js at object asserttransformidvalue test functional services transform wizard ts at runmicrotasks at processticksandrejections node internal process task queues at object settransformid test functional services transform wizard ts at context test functional apps transform creation saved search ts at object apply opt local ssd buildkite builds kb elastic kibana hourly kibana node modules kbn test target node functional test runner lib mocha wrap function js actual expected fq showdiff true first failure
1
10,724
27,329,271,376
IssuesEvent
2023-02-25 12:12:32
Arsenic-ATG/8085
https://api.github.com/repos/Arsenic-ATG/8085
opened
Make instruction functions private
good first issue Architecture Instruction set
currently all the functions of class `emu::cpu` are publicly visible, ideally these functions should be private and only be used by `emu::cpu::execute ()` to execute the proper instruction. Doing so would make the program much less error prone and avoid any accidental function calls by users.
1.0
Make instruction functions private - currently all the functions of class `emu::cpu` are publicly visible, ideally these functions should be private and only be used by `emu::cpu::execute ()` to execute the proper instruction. Doing so would make the program much less error prone and avoid any accidental function calls by users.
non_test
make instruction functions private currently all the functions of class emu cpu are publicly visible ideally these functions should be private and only be used by emu cpu execute to execute the proper instruction doing so would make the program much less error prone and avoid any accidental function calls by users
0
105,246
9,048,460,383
IssuesEvent
2019-02-12 00:10:11
knative/serving
https://api.github.com/repos/knative/serving
closed
Create PR check to verify API changes reflected in conformance tests
area/API area/test-and-release kind/doc kind/feature
/area API /area test-and-release /kind dev /kind doc ## Expected Behavior Whenever changes are made to the API, in either: * [The spec doc](https://github.com/elafros/elafros/blob/master/docs/spec/spec.md) * [The API definition](https://github.com/elafros/elafros/tree/master/pkg/apis/ela) The [conformance tests](https://github.com/elafros/elafros/tree/master/test/conformance) should (probably) be updated as well. When a PR is submitted that changes either of those areas, but does not change the conformance tests, the submitter should be required to either update the conformance tests or indicate that they conformance tests do not need to be updated. ## Actual Behavior Changes to the API that do not break the conformance tests can slip by without being reflected in the conformance tests. ## Steps to Reproduce the Problem n/a ## Additional Info n/a
1.0
Create PR check to verify API changes reflected in conformance tests - /area API /area test-and-release /kind dev /kind doc ## Expected Behavior Whenever changes are made to the API, in either: * [The spec doc](https://github.com/elafros/elafros/blob/master/docs/spec/spec.md) * [The API definition](https://github.com/elafros/elafros/tree/master/pkg/apis/ela) The [conformance tests](https://github.com/elafros/elafros/tree/master/test/conformance) should (probably) be updated as well. When a PR is submitted that changes either of those areas, but does not change the conformance tests, the submitter should be required to either update the conformance tests or indicate that they conformance tests do not need to be updated. ## Actual Behavior Changes to the API that do not break the conformance tests can slip by without being reflected in the conformance tests. ## Steps to Reproduce the Problem n/a ## Additional Info n/a
test
create pr check to verify api changes reflected in conformance tests area api area test and release kind dev kind doc expected behavior whenever changes are made to the api in either the should probably be updated as well when a pr is submitted that changes either of those areas but does not change the conformance tests the submitter should be required to either update the conformance tests or indicate that they conformance tests do not need to be updated actual behavior changes to the api that do not break the conformance tests can slip by without being reflected in the conformance tests steps to reproduce the problem n a additional info n a
1
107,919
9,248,944,446
IssuesEvent
2019-03-15 08:03:44
redhat-developer/odo
https://api.github.com/repos/redhat-developer/odo
closed
Get odo tests as part of OpenShift testgrid
kind/testing priority/High state/Ready
[kind/Feature] <!-- Welcome! - We kindly ask you to: 1. Fill out the issue template below 2. Use the chat and talk to us if you have a question rather than a bug or feature request. The chat room is at: https://chat.openshift.io/developers/channels/odo Thanks for understanding, and for contributing to the project! --> ## Which functionality do you think we should add? As odo is an OpenShift tool, it would be nice to get odo tested, as part of OpenShift's own tests. This means that whenever tests run for OpenShift, odo should get validated for it as well. ### Refer Links Kubernetes Prow: https://github.com/kubernetes/test-infra Kubernetes Test Grid: https://k8s-testgrid.appspot.com/ Kubernetes Gubernator: https://gubernator.k8s.io/ Kubernetes Test Triage: https://storage.googleapis.com/k8s-gubernator/triage/index.html OpenShift Gubernator: https://openshift-gce-devel.appspot.com/ Adding Jobs to prow: https://github.com/kubernetes/test-infra/blob/master/prow/jobs.md ## Why is this needed? This will enable us to: - Constantly be up-to date with latest OpenShift releases. When odo is part of OpenShift own testing, we will become aware of new versions very quickly, and also of any breakages, early on in the cycle - We will be able to communicate more directly with the `OpenShift` team about odo. - It will help us with continuous delivery of odo with OpenShift - It will get us better visibility with little cost. We can leverage amazing test infrastructure developed for testing Kubernetes and OpenShift, without having to put the effort ## Tasks - [x] ~~Create a pre-test script that can bring fresh cluster upto speed with basic expectations of test, such as a user currently being logged in, and authorisation being setup for the users being used for the tests~~ Let's start with just building odo and running unit tests. Those won't have any dependency on OpenShift cluster. - [x] Get a token for accessing OpenShift CI from OpenShift CI Administrators https://mojo.redhat.com/docs/DOC-1177573 - [x] Setup webhook, to hit https://hook-ci.svc.ci.openshift.org/hook with following parameters : - Content Type : `application/json` - Secret : `token retrieved from above` - SSL verification : enabled - Event types : all - Active : yes - [x] Grant the Team OpenShift Robots team write access to the repository. Since we are currently not in OpenShift Organization - [x] Invite `openshift-ci-robot` and `openshift-merge-robot` to org or add them as collaborators - [x] Contact one of CI administrators to accept invite, as needed - [x] Grant robots write access to repository - [x] Configure prow as per our requirements - [x] Decide which plugins we want to use from list available https://deck-ci.svc.ci.openshift.org/plugins - [x] Update [openshift/release](https://github.com/openshift/release) making following changes - [x] [plugins.yaml](https://github.com/openshift/release/blob/master/cluster/ci/config/prow/plugins.yaml) to enable CI interacts against repository for plugins under `redhat-developer/odo` ie `$ord/$repo` - [x] [config.yaml](https://github.com/openshift/release/blob/master/cluster/ci/config/prow/config.yaml) to - [x] Enable tide merge automation on repository - [x] Setup CI Operator by creating appropriate files in [ci-operator directory](https://github.com/openshift/release/tree/master/ci-operator) - [x] `ci-operator/config/redhat-developer/odo/redhat-developer-odo-master.yaml` (ci-operator/config/$org/$repo/$org-$repo-$branch.yaml) to - [x] Define release artifacts - [x] Configure ci-operator test definitions for repository - [x] `ci-operator/jobs/redhat-developer/odo/redhat-developer-odo-master-$jobtype.yaml` to to add test entries for Prow for repository [jobtype = presubmit, postsubmit, periodic] - [x] Add ourselves to `ci-operator/config/redhat-developer/odo/OWNERS` and `ci-operator/jobs/redhat-developer/odo/OWNERS`
1.0
Get odo tests as part of OpenShift testgrid - [kind/Feature] <!-- Welcome! - We kindly ask you to: 1. Fill out the issue template below 2. Use the chat and talk to us if you have a question rather than a bug or feature request. The chat room is at: https://chat.openshift.io/developers/channels/odo Thanks for understanding, and for contributing to the project! --> ## Which functionality do you think we should add? As odo is an OpenShift tool, it would be nice to get odo tested, as part of OpenShift's own tests. This means that whenever tests run for OpenShift, odo should get validated for it as well. ### Refer Links Kubernetes Prow: https://github.com/kubernetes/test-infra Kubernetes Test Grid: https://k8s-testgrid.appspot.com/ Kubernetes Gubernator: https://gubernator.k8s.io/ Kubernetes Test Triage: https://storage.googleapis.com/k8s-gubernator/triage/index.html OpenShift Gubernator: https://openshift-gce-devel.appspot.com/ Adding Jobs to prow: https://github.com/kubernetes/test-infra/blob/master/prow/jobs.md ## Why is this needed? This will enable us to: - Constantly be up-to date with latest OpenShift releases. When odo is part of OpenShift own testing, we will become aware of new versions very quickly, and also of any breakages, early on in the cycle - We will be able to communicate more directly with the `OpenShift` team about odo. - It will help us with continuous delivery of odo with OpenShift - It will get us better visibility with little cost. We can leverage amazing test infrastructure developed for testing Kubernetes and OpenShift, without having to put the effort ## Tasks - [x] ~~Create a pre-test script that can bring fresh cluster upto speed with basic expectations of test, such as a user currently being logged in, and authorisation being setup for the users being used for the tests~~ Let's start with just building odo and running unit tests. Those won't have any dependency on OpenShift cluster. - [x] Get a token for accessing OpenShift CI from OpenShift CI Administrators https://mojo.redhat.com/docs/DOC-1177573 - [x] Setup webhook, to hit https://hook-ci.svc.ci.openshift.org/hook with following parameters : - Content Type : `application/json` - Secret : `token retrieved from above` - SSL verification : enabled - Event types : all - Active : yes - [x] Grant the Team OpenShift Robots team write access to the repository. Since we are currently not in OpenShift Organization - [x] Invite `openshift-ci-robot` and `openshift-merge-robot` to org or add them as collaborators - [x] Contact one of CI administrators to accept invite, as needed - [x] Grant robots write access to repository - [x] Configure prow as per our requirements - [x] Decide which plugins we want to use from list available https://deck-ci.svc.ci.openshift.org/plugins - [x] Update [openshift/release](https://github.com/openshift/release) making following changes - [x] [plugins.yaml](https://github.com/openshift/release/blob/master/cluster/ci/config/prow/plugins.yaml) to enable CI interacts against repository for plugins under `redhat-developer/odo` ie `$ord/$repo` - [x] [config.yaml](https://github.com/openshift/release/blob/master/cluster/ci/config/prow/config.yaml) to - [x] Enable tide merge automation on repository - [x] Setup CI Operator by creating appropriate files in [ci-operator directory](https://github.com/openshift/release/tree/master/ci-operator) - [x] `ci-operator/config/redhat-developer/odo/redhat-developer-odo-master.yaml` (ci-operator/config/$org/$repo/$org-$repo-$branch.yaml) to - [x] Define release artifacts - [x] Configure ci-operator test definitions for repository - [x] `ci-operator/jobs/redhat-developer/odo/redhat-developer-odo-master-$jobtype.yaml` to to add test entries for Prow for repository [jobtype = presubmit, postsubmit, periodic] - [x] Add ourselves to `ci-operator/config/redhat-developer/odo/OWNERS` and `ci-operator/jobs/redhat-developer/odo/OWNERS`
test
get odo tests as part of openshift testgrid welcome we kindly ask you to fill out the issue template below use the chat and talk to us if you have a question rather than a bug or feature request the chat room is at thanks for understanding and for contributing to the project which functionality do you think we should add as odo is an openshift tool it would be nice to get odo tested as part of openshift s own tests this means that whenever tests run for openshift odo should get validated for it as well refer links kubernetes prow kubernetes test grid kubernetes gubernator kubernetes test triage openshift gubernator adding jobs to prow why is this needed this will enable us to constantly be up to date with latest openshift releases when odo is part of openshift own testing we will become aware of new versions very quickly and also of any breakages early on in the cycle we will be able to communicate more directly with the openshift team about odo it will help us with continuous delivery of odo with openshift it will get us better visibility with little cost we can leverage amazing test infrastructure developed for testing kubernetes and openshift without having to put the effort tasks create a pre test script that can bring fresh cluster upto speed with basic expectations of test such as a user currently being logged in and authorisation being setup for the users being used for the tests let s start with just building odo and running unit tests those won t have any dependency on openshift cluster get a token for accessing openshift ci from openshift ci administrators setup webhook to hit with following parameters content type application json secret token retrieved from above ssl verification enabled event types all active yes grant the team openshift robots team write access to the repository since we are currently not in openshift organization invite openshift ci robot and openshift merge robot to org or add them as collaborators contact one of ci administrators to accept invite as needed grant robots write access to repository configure prow as per our requirements decide which plugins we want to use from list available update making following changes to enable ci interacts against repository for plugins under redhat developer odo ie ord repo to enable tide merge automation on repository setup ci operator by creating appropriate files in ci operator config redhat developer odo redhat developer odo master yaml ci operator config org repo org repo branch yaml to define release artifacts configure ci operator test definitions for repository ci operator jobs redhat developer odo redhat developer odo master jobtype yaml to to add test entries for prow for repository add ourselves to ci operator config redhat developer odo owners and ci operator jobs redhat developer odo owners
1
61,558
3,147,475,762
IssuesEvent
2015-09-15 08:22:12
handsontable/handsontable
https://api.github.com/repos/handsontable/handsontable
closed
Broken scrolling on the grouping example.
Bug Plugin: drag to scroll Priority: normal
Scrolling with shift + mouse wheel can cause sheet to go off the screen. http://docs.handsontable.com/0.16.0/demo-grouping-and-ungrouping.html?_ga=1.181662767.981730759.1437017583 It also works here http://docs.handsontable.com/0.16.0/demo-scrollbars.html http://imgur.com/v4GWthl ![image](https://cloud.githubusercontent.com/assets/1968109/8715284/3ff01296-2bd3-11e5-9184-fc760e699ef6.png) To Reproduce: * Click in a cell eg A2 * Hold Shift + Mouse wheel down to scroll right. * Let go of shift whilst scrolling down still. * Hold shift and scroll the mouse wheel up (to scroll to the left) it will go out of bounds.
1.0
Broken scrolling on the grouping example. - Scrolling with shift + mouse wheel can cause sheet to go off the screen. http://docs.handsontable.com/0.16.0/demo-grouping-and-ungrouping.html?_ga=1.181662767.981730759.1437017583 It also works here http://docs.handsontable.com/0.16.0/demo-scrollbars.html http://imgur.com/v4GWthl ![image](https://cloud.githubusercontent.com/assets/1968109/8715284/3ff01296-2bd3-11e5-9184-fc760e699ef6.png) To Reproduce: * Click in a cell eg A2 * Hold Shift + Mouse wheel down to scroll right. * Let go of shift whilst scrolling down still. * Hold shift and scroll the mouse wheel up (to scroll to the left) it will go out of bounds.
non_test
broken scrolling on the grouping example scrolling with shift mouse wheel can cause sheet to go off the screen it also works here to reproduce click in a cell eg hold shift mouse wheel down to scroll right let go of shift whilst scrolling down still hold shift and scroll the mouse wheel up to scroll to the left it will go out of bounds
0
87,382
25,107,004,353
IssuesEvent
2022-11-08 17:19:58
spack/spack
https://api.github.com/repos/spack/spack
closed
Installation issue: Binary libffi fails checksum verification
build-error
### Steps to reproduce the issue ``` # On develop branch spack mirror add binary_mirror https://binaries.spack.io/releases/v0.18 spack buildcache keys --install --trust spack -vvv install /d6d3lh3 ``` ### Error message ==> Installing libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd ==> Fetching https://binaries.spack.io/releases/v0.18/build_cache/linux-amzn2-x86_64_v4-gcc-7.3.1-libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd.spec.json.sig gpg: Signature made Tue 31 May 2022 07:27:40 AM UTC using RSA key ID 3DB0C723 gpg: Good signature from "Spack Project Official Binaries <maintainers@spack.io>" ==> Fetching https://binaries.spack.io/releases/v0.18/build_cache/linux-amzn2-x86_64_v4/gcc-7.3.1/libffi-3.4.2/linux-amzn2-x86_64_v4-gcc-7.3.1-libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd.spack ==> Extracting libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd from binary cache ==> Error: Failed to install libffi due to NoChecksumException: Package tarball failed checksum verification. It cannot be installed. ### Information on your system [ec2-user@ip-10-7-228-156 ~]$ spack debug report * **Spack:** 0.19.0.dev0 (b065d69136023693ace8e4d35d9fdad817d63d93) * **Python:** 3.7.10 * **Platform:** linux-amzn2-skylake_avx512 * **Concretizer:** clingo ### Additional information _No response_ ### General information - [X] I have run `spack debug report` and reported the version of Spack/Python/Platform - [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers - [X] I have uploaded the build log and environment files - [X] I have searched the issues of this repo and believe this is not a duplicate
1.0
Installation issue: Binary libffi fails checksum verification - ### Steps to reproduce the issue ``` # On develop branch spack mirror add binary_mirror https://binaries.spack.io/releases/v0.18 spack buildcache keys --install --trust spack -vvv install /d6d3lh3 ``` ### Error message ==> Installing libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd ==> Fetching https://binaries.spack.io/releases/v0.18/build_cache/linux-amzn2-x86_64_v4-gcc-7.3.1-libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd.spec.json.sig gpg: Signature made Tue 31 May 2022 07:27:40 AM UTC using RSA key ID 3DB0C723 gpg: Good signature from "Spack Project Official Binaries <maintainers@spack.io>" ==> Fetching https://binaries.spack.io/releases/v0.18/build_cache/linux-amzn2-x86_64_v4/gcc-7.3.1/libffi-3.4.2/linux-amzn2-x86_64_v4-gcc-7.3.1-libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd.spack ==> Extracting libffi-3.4.2-d6d3lh3hgkjwbifbbjuvqs2a3w5xj7pd from binary cache ==> Error: Failed to install libffi due to NoChecksumException: Package tarball failed checksum verification. It cannot be installed. ### Information on your system [ec2-user@ip-10-7-228-156 ~]$ spack debug report * **Spack:** 0.19.0.dev0 (b065d69136023693ace8e4d35d9fdad817d63d93) * **Python:** 3.7.10 * **Platform:** linux-amzn2-skylake_avx512 * **Concretizer:** clingo ### Additional information _No response_ ### General information - [X] I have run `spack debug report` and reported the version of Spack/Python/Platform - [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers - [X] I have uploaded the build log and environment files - [X] I have searched the issues of this repo and believe this is not a duplicate
non_test
installation issue binary libffi fails checksum verification steps to reproduce the issue on develop branch spack mirror add binary mirror spack buildcache keys install trust spack vvv install error message installing libffi fetching gpg signature made tue may am utc using rsa key id gpg good signature from spack project official binaries fetching extracting libffi from binary cache error failed to install libffi due to nochecksumexception package tarball failed checksum verification it cannot be installed information on your system spack debug report spack python platform linux skylake concretizer clingo additional information no response general information i have run spack debug report and reported the version of spack python platform i have run spack maintainers and mentioned any maintainers i have uploaded the build log and environment files i have searched the issues of this repo and believe this is not a duplicate
0
363,324
25,444,883,865
IssuesEvent
2022-11-24 04:29:14
cdisc-org/conformance-rules-editor
https://api.github.com/repos/cdisc-org/conformance-rules-editor
closed
As a workshop participant, I want sample rules
documentation
- Rules have been gathered that highlight different scenarios that would be helpful for training.
1.0
As a workshop participant, I want sample rules - - Rules have been gathered that highlight different scenarios that would be helpful for training.
non_test
as a workshop participant i want sample rules rules have been gathered that highlight different scenarios that would be helpful for training
0
677,288
23,157,346,824
IssuesEvent
2022-07-29 14:12:37
openghg/openghg
https://api.github.com/repos/openghg/openghg
closed
New xarray and mypy issues
high-priority
The latest version of xarray (2022.6.0) has some new type hints included. This results in quite a few mypy issues that will need looking at. I'm setting this as high priority as I don't think we want to start pinning versions of commonly used libraries like xarray.
1.0
New xarray and mypy issues - The latest version of xarray (2022.6.0) has some new type hints included. This results in quite a few mypy issues that will need looking at. I'm setting this as high priority as I don't think we want to start pinning versions of commonly used libraries like xarray.
non_test
new xarray and mypy issues the latest version of xarray has some new type hints included this results in quite a few mypy issues that will need looking at i m setting this as high priority as i don t think we want to start pinning versions of commonly used libraries like xarray
0
103,291
4,166,283,881
IssuesEvent
2016-06-20 01:40:07
nvs/gem
https://api.github.com/repos/nvs/gem
opened
Introduce a pause before 'starting'
Area: JASS Priority: Later Status: Not Started Type: Enhancement
This would help some slower computers 'settle', as well as giving players an indication of the game actually starting. This issue is most noticeable when hosted via HCL, and would help with #79.
1.0
Introduce a pause before 'starting' - This would help some slower computers 'settle', as well as giving players an indication of the game actually starting. This issue is most noticeable when hosted via HCL, and would help with #79.
non_test
introduce a pause before starting this would help some slower computers settle as well as giving players an indication of the game actually starting this issue is most noticeable when hosted via hcl and would help with
0
19,170
5,814,941,387
IssuesEvent
2017-05-05 06:42:59
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
Massive loading times after update to 3.7?
No Code Attached Yet
Hey there, i updated two pages now to joomla 3.7 and notices that both of them got a high loading time up to 8 seconds to get the page loaded. Does anyone else have that problem?
1.0
Massive loading times after update to 3.7? - Hey there, i updated two pages now to joomla 3.7 and notices that both of them got a high loading time up to 8 seconds to get the page loaded. Does anyone else have that problem?
non_test
massive loading times after update to hey there i updated two pages now to joomla and notices that both of them got a high loading time up to seconds to get the page loaded does anyone else have that problem
0
308,389
26,603,901,914
IssuesEvent
2023-01-23 17:43:39
kedacore/keda
https://api.github.com/repos/kedacore/keda
opened
Add e2e test for Openstack Metrics Scaler
help wanted good first issue testing
### Proposal https://github.com/kedacore/keda/tree/main/tests ### Use-Case _No response_ ### Is this a feature you are interested in implementing yourself? No ### Anything else? _No response_
1.0
Add e2e test for Openstack Metrics Scaler - ### Proposal https://github.com/kedacore/keda/tree/main/tests ### Use-Case _No response_ ### Is this a feature you are interested in implementing yourself? No ### Anything else? _No response_
test
add test for openstack metrics scaler proposal use case no response is this a feature you are interested in implementing yourself no anything else no response
1
233,117
18,949,430,268
IssuesEvent
2021-11-18 13:48:00
BlueCodeSystems/who-anc
https://api.github.com/repos/BlueCodeSystems/who-anc
closed
Use lower case for first letters for the words "Test" and "Screening" for consistency.
Tests v1.0.0-beta.4
![Screenshot_20211115-145640_WHO ANC](https://user-images.githubusercontent.com/25362528/141786255-c8dbd457-7ab8-4398-9373-9465a8f29b36.jpg)
1.0
Use lower case for first letters for the words "Test" and "Screening" for consistency. - ![Screenshot_20211115-145640_WHO ANC](https://user-images.githubusercontent.com/25362528/141786255-c8dbd457-7ab8-4398-9373-9465a8f29b36.jpg)
test
use lower case for first letters for the words test and screening for consistency
1
209,490
16,024,515,182
IssuesEvent
2021-04-21 07:21:07
CarlosRayon/symfony
https://api.github.com/repos/CarlosRayon/symfony
opened
Metodologias
Testing General
- Seguir misma estructura directorios que el proyecto bajo un directorio test. - Hacer una clase de test por cada clase del proyecto principal - Nombre de la clase sera el mismo que la clase a testear añadiendo la palabra _Test_ en camelCase (ejemploController.php -> ejemploControllerTest.php). - La funciones de test empezaran por _test_ (testEjemplo()) y deben ser publicas - Es buena practica que los datos de test que se recarguen en cada llamada global a los test. - Podemos ejecutar todos los test de directorio test con `bin/phpunit` o especificar un directorio o un test en concreto `bin/phpunit <ruta-directorio | ruta-fichero>` - Disponemos de multiples opciones al ejecutar lo test que podemos ver en la [doc de phpunit](https://phpunit.readthedocs.io/en/9.5/textui.html#command-line-options)
1.0
Metodologias - - Seguir misma estructura directorios que el proyecto bajo un directorio test. - Hacer una clase de test por cada clase del proyecto principal - Nombre de la clase sera el mismo que la clase a testear añadiendo la palabra _Test_ en camelCase (ejemploController.php -> ejemploControllerTest.php). - La funciones de test empezaran por _test_ (testEjemplo()) y deben ser publicas - Es buena practica que los datos de test que se recarguen en cada llamada global a los test. - Podemos ejecutar todos los test de directorio test con `bin/phpunit` o especificar un directorio o un test en concreto `bin/phpunit <ruta-directorio | ruta-fichero>` - Disponemos de multiples opciones al ejecutar lo test que podemos ver en la [doc de phpunit](https://phpunit.readthedocs.io/en/9.5/textui.html#command-line-options)
test
metodologias seguir misma estructura directorios que el proyecto bajo un directorio test hacer una clase de test por cada clase del proyecto principal nombre de la clase sera el mismo que la clase a testear añadiendo la palabra test en camelcase ejemplocontroller php ejemplocontrollertest php la funciones de test empezaran por test testejemplo y deben ser publicas es buena practica que los datos de test que se recarguen en cada llamada global a los test podemos ejecutar todos los test de directorio test con bin phpunit o especificar un directorio o un test en concreto bin phpunit disponemos de multiples opciones al ejecutar lo test que podemos ver en la
1
110,100
9,430,499,397
IssuesEvent
2019-04-12 09:10:45
goharbor/harbor
https://api.github.com/repos/goharbor/harbor
closed
Should add retry in keyword <Add Labels To Tag>
area/ci area/test automation
In keyword <Add Labels To Tag>, one of steps is to click repository name, there should be retry when attempting to go into repository. https://jenkins11.svc.eng.vmware.com/job/harbor_nightly_result_publisher/2095/robot/report/log.html ![image](https://user-images.githubusercontent.com/44018302/54327943-07dc2100-4647-11e9-8012-5227a02c28ed.png)
1.0
Should add retry in keyword <Add Labels To Tag> - In keyword <Add Labels To Tag>, one of steps is to click repository name, there should be retry when attempting to go into repository. https://jenkins11.svc.eng.vmware.com/job/harbor_nightly_result_publisher/2095/robot/report/log.html ![image](https://user-images.githubusercontent.com/44018302/54327943-07dc2100-4647-11e9-8012-5227a02c28ed.png)
test
should add retry in keyword in keyword one of steps is to click repository name there should be retry when attempting to go into repository
1
581,505
17,295,185,936
IssuesEvent
2021-07-25 15:26:04
DFO-Ocean-Navigator/Ocean-Data-Map-Project
https://api.github.com/repos/DFO-Ocean-Navigator/Ocean-Data-Map-Project
closed
Issue with building the React JS modules for mainline
Bug Javascript Priority: High dependencies
``` (navigator) vagrant@ubuntu-focal:~/Ocean-Data-Map-Project/oceannavigator/frontend$ yarn install yarn install v1.22.10 [1/4] Resolving packages... [2/4] Fetching packages... info fsevents@2.3.2: The platform "linux" is incompatible with this module. info "fsevents@2.3.2" is an optional dependency and failed compatibility check. Excluding it from installation. info fsevents@1.2.13: The platform "linux" is incompatible with this module. info "fsevents@1.2.13" is an optional dependency and failed compatibility check. Excluding it from installation. [3/4] Linking dependencies... [4/4] Building fresh packages... success Saved lockfile. Done in 28.12s. (navigator) vagrant@ubuntu-focal:~/Ocean-Data-Map-Project/oceannavigator/frontend$ yarn build yarn run v1.22.10 $ NODE_ENV=production node --stack_size=1000000 node_modules/.bin/webpack --progress -p Hash: 83eb17fbe8f47c58a8d4 Version: webpack 3.6.0 Time: 21890ms Asset Size Chunks Chunk Names /glyphicons-halflings-regular.eot 20.1 kB [emitted] /glyphicons-halflings-regular.woff2 18 kB [emitted] /glyphicons-halflings-regular.woff 23.4 kB [emitted] /glyphicons-halflings-regular.ttf 45.4 kB [emitted] /glyphicons-halflings-regular.svg 109 kB [emitted] /ui-icons_444444_256x240.png 3.76 kB [emitted] /ui-icons_555555_256x240.png 3.76 kB [emitted] /ui-icons_ffffff_256x240.png 3.76 kB [emitted] /ui-icons_777620_256x240.png 3.76 kB [emitted] /ui-icons_cc0000_256x240.png 3.76 kB [emitted] /ui-icons_777777_256x240.png 3.76 kB [emitted] /fontawesome-webfont.eot 166 kB [emitted] /fontawesome-webfont.woff2 77.2 kB [emitted] /fontawesome-webfont.woff 98 kB [emitted] /fontawesome-webfont.ttf 166 kB [emitted] /fontawesome-webfont.svg 444 kB [emitted] [big] /spinner.gif 41.3 kB [emitted] /x.png 153 bytes [emitted] /I0.svg 131 bytes [emitted] /I1.svg 747 bytes [emitted] /I2.svg 747 bytes [emitted] /I3.svg 747 bytes [emitted] /I4.svg 745 bytes [emitted] /I5.svg 747 bytes [emitted] /I6.svg 747 bytes [emitted] /I7.svg 747 bytes [emitted] /I8.svg 747 bytes [emitted] /I9.svg 745 bytes [emitted] /sad-computer.png 2.1 kB [emitted] /bar_loader.gif 10.8 kB [emitted] oceannavigator.js 7.57 MB 0 [emitted] [big] main oceannavigator.css 343 kB 0 [emitted] [big] main index.html 1.87 kB [emitted] [20] ./src/i18n.js 828 bytes {0} [built] [75] (webpack)/buildin/global.js 509 bytes {0} [built] [626] ./src/currentLanguage.js 93 bytes {0} [built] [640] multi babel-polyfill ./src/index.jsx 40 bytes {0} [built] [843] ./src/index.jsx 2.37 kB {0} [built] [1052] ../translations/fr.json 15.6 kB {0} [built] [1053] ../translations/en.json 3.29 kB {0} [built] [1055] ./src/images/x.png 52 bytes {0} [built] [1056] ./src/images/s111/I0.svg 53 bytes {0} [built] [1438] ./src/remote/OceanNavigator.js 633 bytes {0} [built] [1457] ./src/images/bar_loader.gif 61 bytes {0} [built] [1463] ./src/stylesheets/utils/bootstrap.css 41 bytes {0} [built] [1464] ./src/stylesheets/main.scss 41 bytes {0} [built] [1512] ./node_modules/css-loader!./src/stylesheets/utils/bootstrap.css 120 kB [built] [1513] ./node_modules/css-loader!./node_modules/sass-loader/lib/loader.js!./src/stylesheets/main.scss 13.6 kB [built] + 1499 hidden modules ERROR in oceannavigator.js from UglifyJs Unexpected token: keyword (const) [oceannavigator.js:60314,9] Child html-webpack-plugin for "index.html": 1 asset [0] ./node_modules/html-webpack-plugin/lib/loader.js!./src/index.ejs 2.14 kB {0} [built] [2] (webpack)/buildin/global.js 509 bytes {0} [built] [3] (webpack)/buildin/module.js 517 bytes {0} [built] + 1 hidden module Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/ol/ol.css: 2 modules Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/jquery-ui-month-picker/src/MonthPicker.css: 2 modules Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/rc-slider/assets/index.css: 2 modules Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/react-datepicker/dist/react-datepicker.css: 2 modules Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/react-bootstrap-toggle/dist/bootstrap2-toggle.css: 2 modules Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/jquery-ui/themes/base/datepicker.css: 2 modules Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!src/stylesheets/utils/bootstrap.css: [0] ./node_modules/css-loader!./src/stylesheets/utils/bootstrap.css 120 kB {0} [built] + 1 hidden module Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/bootstrap/dist/css/bootstrap.css: 5 assets 7 modules Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/jquery-ui/themes/base/theme.css: 6 assets 8 modules Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/jquery-ui/themes/base/base.css: 21 modules Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/sass-loader/lib/loader.js!src/stylesheets/main.scss: [0] ./node_modules/css-loader!./node_modules/sass-loader/lib/loader.js!./src/stylesheets/main.scss 13.6 kB {0} [built] + 1 hidden module Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/sass-loader/lib/loader.js!node_modules/font-awesome/scss/font-awesome.scss: 5 assets 8 modules error Command failed with exit code 2. info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command. ```
1.0
Issue with building the React JS modules for mainline - ``` (navigator) vagrant@ubuntu-focal:~/Ocean-Data-Map-Project/oceannavigator/frontend$ yarn install yarn install v1.22.10 [1/4] Resolving packages... [2/4] Fetching packages... info fsevents@2.3.2: The platform "linux" is incompatible with this module. info "fsevents@2.3.2" is an optional dependency and failed compatibility check. Excluding it from installation. info fsevents@1.2.13: The platform "linux" is incompatible with this module. info "fsevents@1.2.13" is an optional dependency and failed compatibility check. Excluding it from installation. [3/4] Linking dependencies... [4/4] Building fresh packages... success Saved lockfile. Done in 28.12s. (navigator) vagrant@ubuntu-focal:~/Ocean-Data-Map-Project/oceannavigator/frontend$ yarn build yarn run v1.22.10 $ NODE_ENV=production node --stack_size=1000000 node_modules/.bin/webpack --progress -p Hash: 83eb17fbe8f47c58a8d4 Version: webpack 3.6.0 Time: 21890ms Asset Size Chunks Chunk Names /glyphicons-halflings-regular.eot 20.1 kB [emitted] /glyphicons-halflings-regular.woff2 18 kB [emitted] /glyphicons-halflings-regular.woff 23.4 kB [emitted] /glyphicons-halflings-regular.ttf 45.4 kB [emitted] /glyphicons-halflings-regular.svg 109 kB [emitted] /ui-icons_444444_256x240.png 3.76 kB [emitted] /ui-icons_555555_256x240.png 3.76 kB [emitted] /ui-icons_ffffff_256x240.png 3.76 kB [emitted] /ui-icons_777620_256x240.png 3.76 kB [emitted] /ui-icons_cc0000_256x240.png 3.76 kB [emitted] /ui-icons_777777_256x240.png 3.76 kB [emitted] /fontawesome-webfont.eot 166 kB [emitted] /fontawesome-webfont.woff2 77.2 kB [emitted] /fontawesome-webfont.woff 98 kB [emitted] /fontawesome-webfont.ttf 166 kB [emitted] /fontawesome-webfont.svg 444 kB [emitted] [big] /spinner.gif 41.3 kB [emitted] /x.png 153 bytes [emitted] /I0.svg 131 bytes [emitted] /I1.svg 747 bytes [emitted] /I2.svg 747 bytes [emitted] /I3.svg 747 bytes [emitted] /I4.svg 745 bytes [emitted] /I5.svg 747 bytes [emitted] /I6.svg 747 bytes [emitted] /I7.svg 747 bytes [emitted] /I8.svg 747 bytes [emitted] /I9.svg 745 bytes [emitted] /sad-computer.png 2.1 kB [emitted] /bar_loader.gif 10.8 kB [emitted] oceannavigator.js 7.57 MB 0 [emitted] [big] main oceannavigator.css 343 kB 0 [emitted] [big] main index.html 1.87 kB [emitted] [20] ./src/i18n.js 828 bytes {0} [built] [75] (webpack)/buildin/global.js 509 bytes {0} [built] [626] ./src/currentLanguage.js 93 bytes {0} [built] [640] multi babel-polyfill ./src/index.jsx 40 bytes {0} [built] [843] ./src/index.jsx 2.37 kB {0} [built] [1052] ../translations/fr.json 15.6 kB {0} [built] [1053] ../translations/en.json 3.29 kB {0} [built] [1055] ./src/images/x.png 52 bytes {0} [built] [1056] ./src/images/s111/I0.svg 53 bytes {0} [built] [1438] ./src/remote/OceanNavigator.js 633 bytes {0} [built] [1457] ./src/images/bar_loader.gif 61 bytes {0} [built] [1463] ./src/stylesheets/utils/bootstrap.css 41 bytes {0} [built] [1464] ./src/stylesheets/main.scss 41 bytes {0} [built] [1512] ./node_modules/css-loader!./src/stylesheets/utils/bootstrap.css 120 kB [built] [1513] ./node_modules/css-loader!./node_modules/sass-loader/lib/loader.js!./src/stylesheets/main.scss 13.6 kB [built] + 1499 hidden modules ERROR in oceannavigator.js from UglifyJs Unexpected token: keyword (const) [oceannavigator.js:60314,9] Child html-webpack-plugin for "index.html": 1 asset [0] ./node_modules/html-webpack-plugin/lib/loader.js!./src/index.ejs 2.14 kB {0} [built] [2] (webpack)/buildin/global.js 509 bytes {0} [built] [3] (webpack)/buildin/module.js 517 bytes {0} [built] + 1 hidden module Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/ol/ol.css: 2 modules Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/jquery-ui-month-picker/src/MonthPicker.css: 2 modules Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/rc-slider/assets/index.css: 2 modules Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/react-datepicker/dist/react-datepicker.css: 2 modules Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/react-bootstrap-toggle/dist/bootstrap2-toggle.css: 2 modules Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/jquery-ui/themes/base/datepicker.css: 2 modules Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!src/stylesheets/utils/bootstrap.css: [0] ./node_modules/css-loader!./src/stylesheets/utils/bootstrap.css 120 kB {0} [built] + 1 hidden module Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/bootstrap/dist/css/bootstrap.css: 5 assets 7 modules Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/jquery-ui/themes/base/theme.css: 6 assets 8 modules Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/jquery-ui/themes/base/base.css: 21 modules Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/sass-loader/lib/loader.js!src/stylesheets/main.scss: [0] ./node_modules/css-loader!./node_modules/sass-loader/lib/loader.js!./src/stylesheets/main.scss 13.6 kB {0} [built] + 1 hidden module Child extract-text-webpack-plugin node_modules/extract-text-webpack-plugin/dist node_modules/css-loader/index.js!node_modules/sass-loader/lib/loader.js!node_modules/font-awesome/scss/font-awesome.scss: 5 assets 8 modules error Command failed with exit code 2. info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command. ```
non_test
issue with building the react js modules for mainline navigator vagrant ubuntu focal ocean data map project oceannavigator frontend yarn install yarn install resolving packages fetching packages info fsevents the platform linux is incompatible with this module info fsevents is an optional dependency and failed compatibility check excluding it from installation info fsevents the platform linux is incompatible with this module info fsevents is an optional dependency and failed compatibility check excluding it from installation linking dependencies building fresh packages success saved lockfile done in navigator vagrant ubuntu focal ocean data map project oceannavigator frontend yarn build yarn run node env production node stack size node modules bin webpack progress p hash version webpack time asset size chunks chunk names glyphicons halflings regular eot kb glyphicons halflings regular kb glyphicons halflings regular woff kb glyphicons halflings regular ttf kb glyphicons halflings regular svg kb ui icons png kb ui icons png kb ui icons ffffff png kb ui icons png kb ui icons png kb ui icons png kb fontawesome webfont eot kb fontawesome webfont kb fontawesome webfont woff kb fontawesome webfont ttf kb fontawesome webfont svg kb spinner gif kb x png bytes svg bytes svg bytes svg bytes svg bytes svg bytes svg bytes svg bytes svg bytes svg bytes svg bytes sad computer png kb bar loader gif kb oceannavigator js mb main oceannavigator css kb main index html kb src js bytes webpack buildin global js bytes src currentlanguage js bytes multi babel polyfill src index jsx bytes src index jsx kb translations fr json kb translations en json kb src images x png bytes src images svg bytes src remote oceannavigator js bytes src images bar loader gif bytes src stylesheets utils bootstrap css bytes src stylesheets main scss bytes node modules css loader src stylesheets utils bootstrap css kb node modules css loader node modules sass loader lib loader js src stylesheets main scss kb hidden modules error in oceannavigator js from uglifyjs unexpected token keyword const child html webpack plugin for index html asset node modules html webpack plugin lib loader js src index ejs kb webpack buildin global js bytes webpack buildin module js bytes hidden module child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules ol ol css modules child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules jquery ui month picker src monthpicker css modules child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules rc slider assets index css modules child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules react datepicker dist react datepicker css modules child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules react bootstrap toggle dist toggle css modules child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules jquery ui themes base datepicker css modules child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js src stylesheets utils bootstrap css node modules css loader src stylesheets utils bootstrap css kb hidden module child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules bootstrap dist css bootstrap css assets modules child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules jquery ui themes base theme css assets modules child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules jquery ui themes base base css modules child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules sass loader lib loader js src stylesheets main scss node modules css loader node modules sass loader lib loader js src stylesheets main scss kb hidden module child extract text webpack plugin node modules extract text webpack plugin dist node modules css loader index js node modules sass loader lib loader js node modules font awesome scss font awesome scss assets modules error command failed with exit code info visit for documentation about this command
0
53,962
6,353,950,271
IssuesEvent
2017-07-29 04:05:46
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Investigate flaky sequential/test-fs-readfile-tostring-fail.js on smartos16-64
fs smartos test
* **Version**: master * **Platform**: smartos * **Subsystem**: fs Example: https://ci.nodejs.org/job/node-test-commit-smartos/7481/nodes=smartos16-64/console ``` not ok 1385 sequential/test-fs-readfile-tostring-fail --- duration_ms: 0.729 severity: fail stack: |- /home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos16-64/test/sequential/test-fs-readfile-tostring-fail.js:59 throw err; ^ RangeError: Array buffer allocation failed at new ArrayBuffer (<anonymous>) at createUnsafeArrayBuffer (buffer.js:66:12) at createUnsafeBuffer (buffer.js:60:25) at Function.Buffer.allocUnsafeSlow (buffer.js:188:10) at FSReqWrap.readFileAfterStat [as oncomplete] (fs.js:447:27) ```
1.0
Investigate flaky sequential/test-fs-readfile-tostring-fail.js on smartos16-64 - * **Version**: master * **Platform**: smartos * **Subsystem**: fs Example: https://ci.nodejs.org/job/node-test-commit-smartos/7481/nodes=smartos16-64/console ``` not ok 1385 sequential/test-fs-readfile-tostring-fail --- duration_ms: 0.729 severity: fail stack: |- /home/iojs/build/workspace/node-test-commit-smartos/nodes/smartos16-64/test/sequential/test-fs-readfile-tostring-fail.js:59 throw err; ^ RangeError: Array buffer allocation failed at new ArrayBuffer (<anonymous>) at createUnsafeArrayBuffer (buffer.js:66:12) at createUnsafeBuffer (buffer.js:60:25) at Function.Buffer.allocUnsafeSlow (buffer.js:188:10) at FSReqWrap.readFileAfterStat [as oncomplete] (fs.js:447:27) ```
test
investigate flaky sequential test fs readfile tostring fail js on version master platform smartos subsystem fs example not ok sequential test fs readfile tostring fail duration ms severity fail stack home iojs build workspace node test commit smartos nodes test sequential test fs readfile tostring fail js throw err rangeerror array buffer allocation failed at new arraybuffer at createunsafearraybuffer buffer js at createunsafebuffer buffer js at function buffer allocunsafeslow buffer js at fsreqwrap readfileafterstat fs js
1
7,560
18,245,610,188
IssuesEvent
2021-10-01 17:58:13
arduino/arduino-cli
https://api.github.com/repos/arduino/arduino-cli
closed
Crash after execute "fatal error: unexpected signal during runtime execution" (macOS 11.6)
os: macos architecture: arm64 conclusion: resolved type: imperfection
## Bug Report ### Current behavior `arduino-cli` (tag 0.19.1) built with go 1.17.1. Crashes immediately after executing a binary file: Crash after execution binary (builded with go 1.17.1): ```console $ arduino-cli fatal error: unexpected signal during runtime execution [signal SIGSEGV: segmentation violation code=0x1 addr=0xb01dfacedebac1e pc=0x7fff204ecc9e] ... ``` [arduino-cli-0.19.1-fatal error.txt](https://github.com/arduino/arduino-cli/files/7259691/arduino-cli-0.19.1-fatal.error.txt) ### Environment - CLI version (~~output of arduino-cli version~~): 0.19.1 - OS and platform: macOS 11.6 x86_64
1.0
Crash after execute "fatal error: unexpected signal during runtime execution" (macOS 11.6) - ## Bug Report ### Current behavior `arduino-cli` (tag 0.19.1) built with go 1.17.1. Crashes immediately after executing a binary file: Crash after execution binary (builded with go 1.17.1): ```console $ arduino-cli fatal error: unexpected signal during runtime execution [signal SIGSEGV: segmentation violation code=0x1 addr=0xb01dfacedebac1e pc=0x7fff204ecc9e] ... ``` [arduino-cli-0.19.1-fatal error.txt](https://github.com/arduino/arduino-cli/files/7259691/arduino-cli-0.19.1-fatal.error.txt) ### Environment - CLI version (~~output of arduino-cli version~~): 0.19.1 - OS and platform: macOS 11.6 x86_64
non_test
crash after execute fatal error unexpected signal during runtime execution macos bug report current behavior arduino cli tag built with go crashes immediately after executing a binary file crash after execution binary builded with go console arduino cli fatal error unexpected signal during runtime execution environment cli version output of arduino cli version os and platform macos
0
268,822
23,396,737,046
IssuesEvent
2022-08-12 01:03:12
nim-lang/Nim
https://api.github.com/repos/nim-lang/Nim
closed
Bug with effect system and forward declarations
Effect system works_but_needs_test_case
```nim type SafeFn = proc (): void {. raises: [] } proc ok() {. raises: [] .} = discard proc fail() {. raises: [] .} let f1 : SafeFn = ok let f2 : SafeFn = fail # Error: type mismatch: got (proc ()) but expected 'SafeFn = proc (){.closure.}' # .raise effect is 'can raise any' proc fail() = discard ``` If `fail` is implemented before it's used, it works.
1.0
Bug with effect system and forward declarations - ```nim type SafeFn = proc (): void {. raises: [] } proc ok() {. raises: [] .} = discard proc fail() {. raises: [] .} let f1 : SafeFn = ok let f2 : SafeFn = fail # Error: type mismatch: got (proc ()) but expected 'SafeFn = proc (){.closure.}' # .raise effect is 'can raise any' proc fail() = discard ``` If `fail` is implemented before it's used, it works.
test
bug with effect system and forward declarations nim type safefn proc void raises proc ok raises discard proc fail raises let safefn ok let safefn fail error type mismatch got proc but expected safefn proc closure raise effect is can raise any proc fail discard if fail is implemented before it s used it works
1
183,649
21,775,132,577
IssuesEvent
2022-05-13 13:07:46
ssobue/redis-demo
https://api.github.com/repos/ssobue/redis-demo
closed
CVE-2021-42550 (Medium) detected in logback-classic-1.2.3.jar
security vulnerability
## CVE-2021-42550 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>logback-classic-1.2.3.jar</b></p></summary> <p>logback-classic module</p> <p>Library home page: <a href="http://logback.qos.ch">http://logback.qos.ch</a></p> <p>Path to dependency file: redis-demo/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/ch/qos/logback/logback-classic/1.2.3/logback-classic-1.2.3.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.5.0.jar (Root Library) - spring-boot-starter-2.5.0.jar - spring-boot-starter-logging-2.5.0.jar - :x: **logback-classic-1.2.3.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In logback version 1.2.7 and prior versions, an attacker with the required privileges to edit configurations files could craft a malicious configuration allowing to execute arbitrary code loaded from LDAP servers. <p>Publish Date: 2021-12-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-42550>CVE-2021-42550</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://logback.qos.ch/news.html">http://logback.qos.ch/news.html</a></p> <p>Release Date: 2021-12-16</p> <p>Fix Resolution: ch.qos.logback:logback-classic:1.2.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-42550 (Medium) detected in logback-classic-1.2.3.jar - ## CVE-2021-42550 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>logback-classic-1.2.3.jar</b></p></summary> <p>logback-classic module</p> <p>Library home page: <a href="http://logback.qos.ch">http://logback.qos.ch</a></p> <p>Path to dependency file: redis-demo/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/ch/qos/logback/logback-classic/1.2.3/logback-classic-1.2.3.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.5.0.jar (Root Library) - spring-boot-starter-2.5.0.jar - spring-boot-starter-logging-2.5.0.jar - :x: **logback-classic-1.2.3.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In logback version 1.2.7 and prior versions, an attacker with the required privileges to edit configurations files could craft a malicious configuration allowing to execute arbitrary code loaded from LDAP servers. <p>Publish Date: 2021-12-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-42550>CVE-2021-42550</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://logback.qos.ch/news.html">http://logback.qos.ch/news.html</a></p> <p>Release Date: 2021-12-16</p> <p>Fix Resolution: ch.qos.logback:logback-classic:1.2.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in logback classic jar cve medium severity vulnerability vulnerable library logback classic jar logback classic module library home page a href path to dependency file redis demo pom xml path to vulnerable library home wss scanner repository ch qos logback logback classic logback classic jar dependency hierarchy spring boot starter web jar root library spring boot starter jar spring boot starter logging jar x logback classic jar vulnerable library found in base branch master vulnerability details in logback version and prior versions an attacker with the required privileges to edit configurations files could craft a malicious configuration allowing to execute arbitrary code loaded from ldap servers publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ch qos logback logback classic step up your open source security game with whitesource
0
120,214
10,109,722,207
IssuesEvent
2019-07-30 08:41:03
khartec/waltz
https://api.github.com/repos/khartec/waltz
closed
Auth Source fails on Person page
bug fixed (test & close)
``` ServiceBroker::loadData - AuthSourcesStore.findAuthSources: Internal Server Error - Exception: Cannot make generic selector for kind: PERSON / ```
1.0
Auth Source fails on Person page - ``` ServiceBroker::loadData - AuthSourcesStore.findAuthSources: Internal Server Error - Exception: Cannot make generic selector for kind: PERSON / ```
test
auth source fails on person page servicebroker loaddata authsourcesstore findauthsources internal server error exception cannot make generic selector for kind person
1
344,589
30,751,786,665
IssuesEvent
2023-07-28 20:01:06
saltstack/salt
https://api.github.com/repos/saltstack/salt
opened
[Increase Test Coverage] Batch 16
Tests
Increase the code coverage percent on the following files to at least 80%. Please be aware that currently the percentage might be inaccurate if the module uses salt due to #64696 File | Percent salt/modules/junos.py 67 salt/modules/postgres.py 76 salt/_compat.py 64 salt/cloud/clouds/gce.py 18 salt/_logging/impl.py 70
1.0
[Increase Test Coverage] Batch 16 - Increase the code coverage percent on the following files to at least 80%. Please be aware that currently the percentage might be inaccurate if the module uses salt due to #64696 File | Percent salt/modules/junos.py 67 salt/modules/postgres.py 76 salt/_compat.py 64 salt/cloud/clouds/gce.py 18 salt/_logging/impl.py 70
test
batch increase the code coverage percent on the following files to at least please be aware that currently the percentage might be inaccurate if the module uses salt due to file percent salt modules junos py salt modules postgres py salt compat py salt cloud clouds gce py salt logging impl py
1
334,069
29,820,345,671
IssuesEvent
2023-06-17 01:31:37
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix manipulation.test_flipud
Sub Task Ivy API Experimental Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4561754220/jobs/8048085832" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4561754220/jobs/8048085832" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4561754220/jobs/8048085832" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4561754220/jobs/8048085832" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix manipulation.test_flipud - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4561754220/jobs/8048085832" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4561754220/jobs/8048085832" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4561754220/jobs/8048085832" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4561754220/jobs/8048085832" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
test
fix manipulation test flipud tensorflow img src torch img src numpy img src jax img src
1
98,106
20,611,770,876
IssuesEvent
2022-03-07 09:21:14
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
[4.1] Schedule tasks error reporting
No Code Attached Yet
### Steps to reproduce the issue Create a new image size action Do NOT select a path Save and run a test ### Expected result Error message that the image path does not exist ![image](https://user-images.githubusercontent.com/1296369/144852162-0635c2d4-357d-4cf3-ba55-e1346fd93aee.png) ### Actual result No error message on screen. Not much point in doing a test if you dont get to see any errors ### Additional comments The error is present in the logs ![image](https://user-images.githubusercontent.com/1296369/144852064-d495acf2-921c-420a-a8c7-856cfc47b070.png)
1.0
[4.1] Schedule tasks error reporting - ### Steps to reproduce the issue Create a new image size action Do NOT select a path Save and run a test ### Expected result Error message that the image path does not exist ![image](https://user-images.githubusercontent.com/1296369/144852162-0635c2d4-357d-4cf3-ba55-e1346fd93aee.png) ### Actual result No error message on screen. Not much point in doing a test if you dont get to see any errors ### Additional comments The error is present in the logs ![image](https://user-images.githubusercontent.com/1296369/144852064-d495acf2-921c-420a-a8c7-856cfc47b070.png)
non_test
schedule tasks error reporting steps to reproduce the issue create a new image size action do not select a path save and run a test expected result error message that the image path does not exist actual result no error message on screen not much point in doing a test if you dont get to see any errors additional comments the error is present in the logs
0
75,440
3,462,464,779
IssuesEvent
2015-12-20 23:23:34
bounswe/bounswe2015group8
https://api.github.com/repos/bounswe/bounswe2015group8
opened
Following Tags Back-end Logic
Priority-High Type-Task
Users should be able to follow tags based on their interested areas. Implement the back-end functionality of following a tag for a user. Write the dao, service and controller class functions. It would be better if users can follow multiple tags at once. (similar to adding tags to heritage objects/posts.)
1.0
Following Tags Back-end Logic - Users should be able to follow tags based on their interested areas. Implement the back-end functionality of following a tag for a user. Write the dao, service and controller class functions. It would be better if users can follow multiple tags at once. (similar to adding tags to heritage objects/posts.)
non_test
following tags back end logic users should be able to follow tags based on their interested areas implement the back end functionality of following a tag for a user write the dao service and controller class functions it would be better if users can follow multiple tags at once similar to adding tags to heritage objects posts
0
285,499
24,671,170,002
IssuesEvent
2022-10-18 13:52:23
aldefouw/redcap_cypress
https://api.github.com/repos/aldefouw/redcap_cypress
reopened
Design Forms using Data Dictionary & Online Designer
Core Functionality Test Script Feature
File Location: https://github.com/aldefouw/redcap_cypress/blob/v11.1.29/cypress/features/core/pre-requisite/design_forms.feature Task: Write the test specs in Gherkin DSL following manual test script as guide Contact Adam De Fouw ([aldefouw@medicine.wisc.edu](mailto:aldefouw@medicine.wisc.edu)) with any questions.
1.0
Design Forms using Data Dictionary & Online Designer - File Location: https://github.com/aldefouw/redcap_cypress/blob/v11.1.29/cypress/features/core/pre-requisite/design_forms.feature Task: Write the test specs in Gherkin DSL following manual test script as guide Contact Adam De Fouw ([aldefouw@medicine.wisc.edu](mailto:aldefouw@medicine.wisc.edu)) with any questions.
test
design forms using data dictionary online designer file location task write the test specs in gherkin dsl following manual test script as guide contact adam de fouw mailto aldefouw medicine wisc edu with any questions
1
33,267
4,820,388,615
IssuesEvent
2016-11-04 22:39:48
infiniteautomation/ma-core-public
https://api.github.com/repos/infiniteautomation/ma-core-public
closed
Persistent Data Source Throttle Threshold Setting
Enhancement Ready for Testing
Add system settings and help for the threshold.
1.0
Persistent Data Source Throttle Threshold Setting - Add system settings and help for the threshold.
test
persistent data source throttle threshold setting add system settings and help for the threshold
1
303,482
26,212,654,306
IssuesEvent
2023-01-04 08:16:03
WPChill/download-monitor
https://api.github.com/repos/WPChill/download-monitor
closed
prices bigger than 1000 throw an error with PayPal gateway
Bug needs testing
Error with prices bigger than 1000 on checkout. https://github.com/WPChill/download-monitor/blob/4.7.70/src/Shop/Checkout/PaymentGateway/PayPal/Api/NumericValidator.php#L22 The if is TRUE because 1,234.00 is not numeric. Formatted price is set here -> https://github.com/WPChill/download-monitor/blob/4.7.70/src/Shop/Checkout/PaymentGateway/PayPal/PayPalGateway.php#L298
1.0
prices bigger than 1000 throw an error with PayPal gateway - Error with prices bigger than 1000 on checkout. https://github.com/WPChill/download-monitor/blob/4.7.70/src/Shop/Checkout/PaymentGateway/PayPal/Api/NumericValidator.php#L22 The if is TRUE because 1,234.00 is not numeric. Formatted price is set here -> https://github.com/WPChill/download-monitor/blob/4.7.70/src/Shop/Checkout/PaymentGateway/PayPal/PayPalGateway.php#L298
test
prices bigger than throw an error with paypal gateway error with prices bigger than on checkout the if is true because is not numeric formatted price is set here
1
135,530
11,010,062,856
IssuesEvent
2019-12-04 13:55:43
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Failing test: X-Pack API Integration Tests.x-pack/test/api_integration/apis/apm/feature_controls·ts - apis APM apm feature controls APIs can be accessed by global_all user
Team:apm failed-test
A test failed on a tracked branch ``` Error: Endpoint: POST /api/apm/settings/agent-configuration/search Status code: 404 Response: Not found expected 404 to equal 200 at executeRequests (test/api_integration/apis/apm/feature_controls.ts:216:15) ``` First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/1312/) Test failure details ```json { "test.class": "X-Pack API Integration Tests.x-pack/test/api_integration/apis/apm/feature_controls·ts", "test.name": "apis APM apm feature controls APIs can be accessed by global_all user", "test.failCount": 2 } ``` Test: https://github.com/elastic/kibana/blob/873529a044e56b2600d279c50c155d47036d97df/x-pack/test/api_integration/apis/apm/feature_controls.ts#L272-L293 Code under test: https://github.com/elastic/kibana/blob/4e5bfc7efaa4411b18d97247addba5f7982bc14f/x-pack/legacy/plugins/apm/server/routes/settings/agent_configuration.ts#L151-L183 <!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack API Integration Tests.x-pack/test/api_integration/apis/apm/feature_controls·ts","test.name":"apis APM apm feature controls APIs can be accessed by global_all user","test.failCount":2}} -->
1.0
Failing test: X-Pack API Integration Tests.x-pack/test/api_integration/apis/apm/feature_controls·ts - apis APM apm feature controls APIs can be accessed by global_all user - A test failed on a tracked branch ``` Error: Endpoint: POST /api/apm/settings/agent-configuration/search Status code: 404 Response: Not found expected 404 to equal 200 at executeRequests (test/api_integration/apis/apm/feature_controls.ts:216:15) ``` First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/1312/) Test failure details ```json { "test.class": "X-Pack API Integration Tests.x-pack/test/api_integration/apis/apm/feature_controls·ts", "test.name": "apis APM apm feature controls APIs can be accessed by global_all user", "test.failCount": 2 } ``` Test: https://github.com/elastic/kibana/blob/873529a044e56b2600d279c50c155d47036d97df/x-pack/test/api_integration/apis/apm/feature_controls.ts#L272-L293 Code under test: https://github.com/elastic/kibana/blob/4e5bfc7efaa4411b18d97247addba5f7982bc14f/x-pack/legacy/plugins/apm/server/routes/settings/agent_configuration.ts#L151-L183 <!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack API Integration Tests.x-pack/test/api_integration/apis/apm/feature_controls·ts","test.name":"apis APM apm feature controls APIs can be accessed by global_all user","test.failCount":2}} -->
test
failing test x pack api integration tests x pack test api integration apis apm feature controls·ts apis apm apm feature controls apis can be accessed by global all user a test failed on a tracked branch error endpoint post api apm settings agent configuration search status code response not found expected to equal at executerequests test api integration apis apm feature controls ts first failure test failure details json test class x pack api integration tests x pack test api integration apis apm feature controls·ts test name apis apm apm feature controls apis can be accessed by global all user test failcount test code under test
1
240,629
18,363,131,946
IssuesEvent
2021-10-09 15:28:17
girlscript/winter-of-contributing
https://api.github.com/repos/girlscript/winter-of-contributing
closed
C-CPP DSA: Bubble Sorting but printing all the passes
documentation GWOC21 DSA Assigned C/CPP
### Description Well, this code will be representing the implementation of bubble sort and printing output of all the passes that occured between the elements. ### Domain C/CPP ### Type of Contribution Documentation ### Code of Conduct - [X] I follow [Contributing Guidelines](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CONTRIBUTING.md) & [Code of conduct](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CODE_OF_CONDUCT.md) of this project.
1.0
C-CPP DSA: Bubble Sorting but printing all the passes - ### Description Well, this code will be representing the implementation of bubble sort and printing output of all the passes that occured between the elements. ### Domain C/CPP ### Type of Contribution Documentation ### Code of Conduct - [X] I follow [Contributing Guidelines](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CONTRIBUTING.md) & [Code of conduct](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CODE_OF_CONDUCT.md) of this project.
non_test
c cpp dsa bubble sorting but printing all the passes description well this code will be representing the implementation of bubble sort and printing output of all the passes that occured between the elements domain c cpp type of contribution documentation code of conduct i follow of this project
0
136,505
12,717,075,730
IssuesEvent
2020-06-24 03:58:24
gardener-attic/issues-foo
https://api.github.com/repos/gardener-attic/issues-foo
closed
A
component/dashboard component/documentation component/gardener kind/bug kind/post-mortem kind/regression os/garden-linux os/suse-chost platform/alicloud platform/aws platform/azure platform/converged-cloud platform/gcp priority/normal topology/shoot
## Which cluster is affected? https://dashboard.garden.dev.k8s.ondemand.com/namespace/garden/shoots/aws/ ## What happened? ## What you expected to happen? ## When did it happen or started to happen? <!-- Please provide start time in UTC OR relative time in hours from now, so that we can pull the proper logs. --> `Absolute`: <!-- e.g. 2029-01-01T15:00:00Z --> `Relative`: <!-- e.g. 3h --> ## How would we reproduce it (concisely and precisely)? ## Anything else we need to know? ## Help us categorise this issue for faster resolution: <!-- First word is category, all consecutive words narrow it down. --> <!-- Please delete every word that doesn't fit here/from your PoV. --> /area audit-logging auto-scaling backup certification control-plane cost delivery disaster-recovery documentation high-availability logging metering monitoring networking os performance quality security storage usability user-management /component gardener dashboard documentation /kind bug regression post-mortem /os garden-linux suse-chost /platform alicloud aws azure gcp converged-cloud <!-- Keep the next line if this issue has no high urgency. Delete the line, if you go for a higher priority. --> /priority normal <!-- Uncomment the following line, if you believe this is a critical issue OR... --> <!-- /priority critical --> <!-- ...uncomment the following line, if this issue has direct customer impact and requires our SRE staff to be paged (at night/on weekends). Use only if business continuity is at risk! --> <!-- /priority blocker -->
1.0
A - ## Which cluster is affected? https://dashboard.garden.dev.k8s.ondemand.com/namespace/garden/shoots/aws/ ## What happened? ## What you expected to happen? ## When did it happen or started to happen? <!-- Please provide start time in UTC OR relative time in hours from now, so that we can pull the proper logs. --> `Absolute`: <!-- e.g. 2029-01-01T15:00:00Z --> `Relative`: <!-- e.g. 3h --> ## How would we reproduce it (concisely and precisely)? ## Anything else we need to know? ## Help us categorise this issue for faster resolution: <!-- First word is category, all consecutive words narrow it down. --> <!-- Please delete every word that doesn't fit here/from your PoV. --> /area audit-logging auto-scaling backup certification control-plane cost delivery disaster-recovery documentation high-availability logging metering monitoring networking os performance quality security storage usability user-management /component gardener dashboard documentation /kind bug regression post-mortem /os garden-linux suse-chost /platform alicloud aws azure gcp converged-cloud <!-- Keep the next line if this issue has no high urgency. Delete the line, if you go for a higher priority. --> /priority normal <!-- Uncomment the following line, if you believe this is a critical issue OR... --> <!-- /priority critical --> <!-- ...uncomment the following line, if this issue has direct customer impact and requires our SRE staff to be paged (at night/on weekends). Use only if business continuity is at risk! --> <!-- /priority blocker -->
non_test
a which cluster is affected what happened what you expected to happen when did it happen or started to happen absolute relative how would we reproduce it concisely and precisely anything else we need to know help us categorise this issue for faster resolution area audit logging auto scaling backup certification control plane cost delivery disaster recovery documentation high availability logging metering monitoring networking os performance quality security storage usability user management component gardener dashboard documentation kind bug regression post mortem os garden linux suse chost platform alicloud aws azure gcp converged cloud priority normal
0
135,997
11,032,485,525
IssuesEvent
2019-12-06 20:18:02
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
opened
P3A - Open TOR window response value isn't displayed correctly in Upgraded profile
QA/Test-Plan-Specified QA/Yes bug
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> Found while testing #6285 P3A - Open TOR window response value isn't displayed correctly in Upgraded profile Q10: Have you ever used a Tor private window? (Brave.Core.LastTimeIncognitoUsed) 1. Yes 2. No ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Clean profile 1.0.1 2. Open a TOR window 3. Upgrade profile to 1.1.18 4. Reload local state 4. Open TOR window response value is displayed as `1` instead of `0` in local state file ## Actual result: <!--Please add screenshots if needed--> Open TOR window response value is displayed as `1` instead of `0` ## Expected result: Open TOR window response value should be displayed correctly ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Always ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 1.1.18 Chromium: 79.0.3945.56 (Official Build) (64-bit) -- | -- Revision | 73cc6bf591f792b99f8fc7cdfb8addedbd084bf8-refs/branch-heads/3945@{#788} OS | Windows 10 OS Version 1803 (Build 17134.1006) ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? NA - Can you reproduce this issue with the beta channel? Yes - Can you reproduce this issue with the dev channel? Yes - Can you reproduce this issue with the nightly channel? Yes ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? NA - Does the issue resolve itself when disabling Brave Rewards? NA - Is the issue reproducible on the latest version of Chrome? NA ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> cc: @brave/legacy_qa @iefremov @bsclifton @rebron
1.0
P3A - Open TOR window response value isn't displayed correctly in Upgraded profile - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> Found while testing #6285 P3A - Open TOR window response value isn't displayed correctly in Upgraded profile Q10: Have you ever used a Tor private window? (Brave.Core.LastTimeIncognitoUsed) 1. Yes 2. No ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Clean profile 1.0.1 2. Open a TOR window 3. Upgrade profile to 1.1.18 4. Reload local state 4. Open TOR window response value is displayed as `1` instead of `0` in local state file ## Actual result: <!--Please add screenshots if needed--> Open TOR window response value is displayed as `1` instead of `0` ## Expected result: Open TOR window response value should be displayed correctly ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Always ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 1.1.18 Chromium: 79.0.3945.56 (Official Build) (64-bit) -- | -- Revision | 73cc6bf591f792b99f8fc7cdfb8addedbd084bf8-refs/branch-heads/3945@{#788} OS | Windows 10 OS Version 1803 (Build 17134.1006) ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? NA - Can you reproduce this issue with the beta channel? Yes - Can you reproduce this issue with the dev channel? Yes - Can you reproduce this issue with the nightly channel? Yes ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? NA - Does the issue resolve itself when disabling Brave Rewards? NA - Is the issue reproducible on the latest version of Chrome? NA ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> cc: @brave/legacy_qa @iefremov @bsclifton @rebron
test
open tor window response value isn t displayed correctly in upgraded profile have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description found while testing open tor window response value isn t displayed correctly in upgraded profile have you ever used a tor private window brave core lasttimeincognitoused yes no steps to reproduce clean profile open a tor window upgrade profile to reload local state open tor window response value is displayed as instead of in local state file actual result open tor window response value is displayed as instead of expected result open tor window response value should be displayed correctly reproduces how often always brave version brave version info brave chromium   official build   bit revision refs branch heads os windows  os version build version channel information can you reproduce this issue with the current release na can you reproduce this issue with the beta channel yes can you reproduce this issue with the dev channel yes can you reproduce this issue with the nightly channel yes other additional information does the issue resolve itself when disabling brave shields na does the issue resolve itself when disabling brave rewards na is the issue reproducible on the latest version of chrome na miscellaneous information cc brave legacy qa iefremov bsclifton rebron
1
429,744
12,427,009,404
IssuesEvent
2020-05-25 00:28:20
eclipse-ee4j/glassfish
https://api.github.com/repos/eclipse-ee4j/glassfish
closed
New annotation @TimedProbe to probe the beginning and end of a method and calculate total time
Component: monitoring ERR: Assignee Priority: Major Stale Type: New Feature
Commonly needed is how much time is spent in a method. Create a new @TimedProbe which does this automatically.
1.0
New annotation @TimedProbe to probe the beginning and end of a method and calculate total time - Commonly needed is how much time is spent in a method. Create a new @TimedProbe which does this automatically.
non_test
new annotation timedprobe to probe the beginning and end of a method and calculate total time commonly needed is how much time is spent in a method create a new timedprobe which does this automatically
0
63,081
14,656,666,922
IssuesEvent
2020-12-28 13:56:13
fu1771695yongxie/learnGitBranching
https://api.github.com/repos/fu1771695yongxie/learnGitBranching
opened
CVE-2015-9251 (Medium) detected in jquery-1.12.4.js
security vulnerability
## CVE-2015-9251 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.12.4.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.12.4/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.12.4/jquery.js</a></p> <p>Path to dependency file: learnGitBranching/node_modules/jquery-ui/demos/effect/removeClass.html</p> <p>Path to vulnerable library: learnGitBranching/node_modules/jquery-ui/demos/effect/../../external/jquery/jquery.js,learnGitBranching/node_modules/jquery-ui/demos/slider/../../external/jquery/jquery.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.12.4.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/learnGitBranching/commit/33cba5147b9149e15d524f7a0f485cf33acd1c2b">33cba5147b9149e15d524f7a0f485cf33acd1c2b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v3.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2015-9251 (Medium) detected in jquery-1.12.4.js - ## CVE-2015-9251 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.12.4.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.12.4/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.12.4/jquery.js</a></p> <p>Path to dependency file: learnGitBranching/node_modules/jquery-ui/demos/effect/removeClass.html</p> <p>Path to vulnerable library: learnGitBranching/node_modules/jquery-ui/demos/effect/../../external/jquery/jquery.js,learnGitBranching/node_modules/jquery-ui/demos/slider/../../external/jquery/jquery.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.12.4.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/learnGitBranching/commit/33cba5147b9149e15d524f7a0f485cf33acd1c2b">33cba5147b9149e15d524f7a0f485cf33acd1c2b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v3.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in jquery js cve medium severity vulnerability vulnerable library jquery js javascript library for dom operations library home page a href path to dependency file learngitbranching node modules jquery ui demos effect removeclass html path to vulnerable library learngitbranching node modules jquery ui demos effect external jquery jquery js learngitbranching node modules jquery ui demos slider external jquery jquery js dependency hierarchy x jquery js vulnerable library found in head commit a href found in base branch master vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
0
752,129
26,274,311,237
IssuesEvent
2023-01-06 20:15:34
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
closed
[studio-ui] Tree Nav loses state when background refresh
bug priority: high validate
### Duplicates - [X] I have searched the existing issues ### Latest version - [X] The issue is in the latest released 4.0.x - [ ] The issue is in the latest released 3.1.x ### Describe the issue Receiving socket events while you tree nav is expanded past its default results in losing the expanded state ### Steps to reproduce Steps: 1. Open two different browsers and login with two different users 2. Open preview on both and with one user expand the tree nav past defaults (use show more) and with the other user edit an asset that pertains to the same nav you expanded 3. See the nav tree losing state ### Relevant log output _No response_ ### Screenshots and/or videos _No response_
1.0
[studio-ui] Tree Nav loses state when background refresh - ### Duplicates - [X] I have searched the existing issues ### Latest version - [X] The issue is in the latest released 4.0.x - [ ] The issue is in the latest released 3.1.x ### Describe the issue Receiving socket events while you tree nav is expanded past its default results in losing the expanded state ### Steps to reproduce Steps: 1. Open two different browsers and login with two different users 2. Open preview on both and with one user expand the tree nav past defaults (use show more) and with the other user edit an asset that pertains to the same nav you expanded 3. See the nav tree losing state ### Relevant log output _No response_ ### Screenshots and/or videos _No response_
non_test
tree nav loses state when background refresh duplicates i have searched the existing issues latest version the issue is in the latest released x the issue is in the latest released x describe the issue receiving socket events while you tree nav is expanded past its default results in losing the expanded state steps to reproduce steps open two different browsers and login with two different users open preview on both and with one user expand the tree nav past defaults use show more and with the other user edit an asset that pertains to the same nav you expanded see the nav tree losing state relevant log output no response screenshots and or videos no response
0
40,128
9,852,416,338
IssuesEvent
2019-06-19 12:49:40
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
opened
Possibly wrong MOD implementation in ACCESS and TERADATA
C: DB: Access C: DB: Teradata C: Functionality E: Enterprise Edition E: Professional Edition P: Medium T: Defect
When generating modulo expressions in `ACCESS` and `TERADATA`, we're currently not wrapping the expression in parentheses: ```sql a MOD b ``` This could potentially lead to operator precedence issues when combined with operators of higher precedence
1.0
Possibly wrong MOD implementation in ACCESS and TERADATA - When generating modulo expressions in `ACCESS` and `TERADATA`, we're currently not wrapping the expression in parentheses: ```sql a MOD b ``` This could potentially lead to operator precedence issues when combined with operators of higher precedence
non_test
possibly wrong mod implementation in access and teradata when generating modulo expressions in access and teradata we re currently not wrapping the expression in parentheses sql a mod b this could potentially lead to operator precedence issues when combined with operators of higher precedence
0
148,924
11,872,135,752
IssuesEvent
2020-03-26 15:22:53
infinispan/infinispan-operator
https://api.github.com/repos/infinispan/infinispan-operator
closed
Unable to run some of the e2e tests against OpenShift
bug test
Some of the e2e tests do no work with OpenShift (tested with 4.2 and 4.3). The issue started appearing after migration to Operator SDK 0.15.2 PR was merged (#293). So far this applies for all the `*Update` tests. Request to change the parameter won't hit the OpenShift. This issue is probably caused by Kubernetes client update.
1.0
Unable to run some of the e2e tests against OpenShift - Some of the e2e tests do no work with OpenShift (tested with 4.2 and 4.3). The issue started appearing after migration to Operator SDK 0.15.2 PR was merged (#293). So far this applies for all the `*Update` tests. Request to change the parameter won't hit the OpenShift. This issue is probably caused by Kubernetes client update.
test
unable to run some of the tests against openshift some of the tests do no work with openshift tested with and the issue started appearing after migration to operator sdk pr was merged so far this applies for all the update tests request to change the parameter won t hit the openshift this issue is probably caused by kubernetes client update
1
666,260
22,348,140,570
IssuesEvent
2022-06-15 09:35:59
PCSX2/pcsx2
https://api.github.com/repos/PCSX2/pcsx2
closed
[BUG]: Gran Turismo 4 - Bad Edges on Split Time Vehicle Icons
Bug GS: Hardware Regression GS: Texture Cache High Priority
### Describe the Bug On any race, with any vehicle, the vehicle icon which appears while showing split times will have a line on the bottom and right edges. The icon is not supposed to have any boundaries, and is supposed to be transparent except for the vehicle picture on it. Bug was introduced by 1.7.2126 and present up to latest master as of writing. ### Reproduction Steps Boot the game, no memory card required, no adjustments to rounding/clamping required. Game must be launched with hardware rendering and upscaling, issue will not occur if the split time first appears on software rendering or native resolution hardware rendering. Go to arcade mode, pick single race. Pick any track, pick any car. Start the race and progress until a split time appears. ### Expected Behavior Split time icon should appear without any lines on its edges. ### PCSX2 Revision 1.7.2126 to 1.7.2480 ### Operating System Windows 10 (64bit) ### If Linux - Specify Distro _No response_ ### CPU R7 5800X ### GPU RX 6900 XT ### GS Settings 4x Native ### Emulation Settings Preset 3 ### GS Window Screenshots ![image](https://user-images.githubusercontent.com/6377490/159072083-d1321791-b54a-4b9b-b625-cc8fde8e3f6d.png) ### Logs & Dumps [gs_20220318210426.zip](https://github.com/PCSX2/pcsx2/files/8307167/gs_20220318210426.zip)
1.0
[BUG]: Gran Turismo 4 - Bad Edges on Split Time Vehicle Icons - ### Describe the Bug On any race, with any vehicle, the vehicle icon which appears while showing split times will have a line on the bottom and right edges. The icon is not supposed to have any boundaries, and is supposed to be transparent except for the vehicle picture on it. Bug was introduced by 1.7.2126 and present up to latest master as of writing. ### Reproduction Steps Boot the game, no memory card required, no adjustments to rounding/clamping required. Game must be launched with hardware rendering and upscaling, issue will not occur if the split time first appears on software rendering or native resolution hardware rendering. Go to arcade mode, pick single race. Pick any track, pick any car. Start the race and progress until a split time appears. ### Expected Behavior Split time icon should appear without any lines on its edges. ### PCSX2 Revision 1.7.2126 to 1.7.2480 ### Operating System Windows 10 (64bit) ### If Linux - Specify Distro _No response_ ### CPU R7 5800X ### GPU RX 6900 XT ### GS Settings 4x Native ### Emulation Settings Preset 3 ### GS Window Screenshots ![image](https://user-images.githubusercontent.com/6377490/159072083-d1321791-b54a-4b9b-b625-cc8fde8e3f6d.png) ### Logs & Dumps [gs_20220318210426.zip](https://github.com/PCSX2/pcsx2/files/8307167/gs_20220318210426.zip)
non_test
gran turismo bad edges on split time vehicle icons describe the bug on any race with any vehicle the vehicle icon which appears while showing split times will have a line on the bottom and right edges the icon is not supposed to have any boundaries and is supposed to be transparent except for the vehicle picture on it bug was introduced by and present up to latest master as of writing reproduction steps boot the game no memory card required no adjustments to rounding clamping required game must be launched with hardware rendering and upscaling issue will not occur if the split time first appears on software rendering or native resolution hardware rendering go to arcade mode pick single race pick any track pick any car start the race and progress until a split time appears expected behavior split time icon should appear without any lines on its edges revision to operating system windows if linux specify distro no response cpu gpu rx xt gs settings native emulation settings preset gs window screenshots logs dumps
0
45,839
13,055,754,994
IssuesEvent
2020-07-30 02:38:18
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
omkey inherits from i3Frame object (Trac #117)
IceTray Incomplete Migration Migrated from Trac defect
Migrated from https://code.icecube.wisc.edu/ticket/117 ```json { "status": "closed", "changetime": "2014-11-23T03:37:56", "description": "where did this come from?", "reporter": "troy", "cc": "", "resolution": "wont or cant fix", "_ts": "1416713876900096", "component": "IceTray", "summary": "omkey inherits from i3Frame object", "priority": "major", "keywords": "", "time": "2008-08-30T16:06:14", "milestone": "", "owner": "troy", "type": "defect" } ```
1.0
omkey inherits from i3Frame object (Trac #117) - Migrated from https://code.icecube.wisc.edu/ticket/117 ```json { "status": "closed", "changetime": "2014-11-23T03:37:56", "description": "where did this come from?", "reporter": "troy", "cc": "", "resolution": "wont or cant fix", "_ts": "1416713876900096", "component": "IceTray", "summary": "omkey inherits from i3Frame object", "priority": "major", "keywords": "", "time": "2008-08-30T16:06:14", "milestone": "", "owner": "troy", "type": "defect" } ```
non_test
omkey inherits from object trac migrated from json status closed changetime description where did this come from reporter troy cc resolution wont or cant fix ts component icetray summary omkey inherits from object priority major keywords time milestone owner troy type defect
0
91,210
8,300,600,604
IssuesEvent
2018-09-21 08:39:00
swarmcity/SwarmCityDapp
https://api.github.com/repos/swarmcity/SwarmCityDapp
closed
on iphone button enter swarm city doesn't appear
blocking bug ready to test
# Location /mykeys # Expected behavior when creating new account on iphone, after checking the box to use the private key to be the backup, a button appears "enter swarm.city" # Actual behavior when creating new account on iphone, after checking the box to use the private key to be the backup, the button appears "enter swarm.city" doesn't appear. # Steps to reproduce the problem ON IPHONE: create new/ choose username and avatar/ choose pasword/make backup/show keys # Operating system / hardware safari on iphone
1.0
on iphone button enter swarm city doesn't appear - # Location /mykeys # Expected behavior when creating new account on iphone, after checking the box to use the private key to be the backup, a button appears "enter swarm.city" # Actual behavior when creating new account on iphone, after checking the box to use the private key to be the backup, the button appears "enter swarm.city" doesn't appear. # Steps to reproduce the problem ON IPHONE: create new/ choose username and avatar/ choose pasword/make backup/show keys # Operating system / hardware safari on iphone
test
on iphone button enter swarm city doesn t appear location mykeys expected behavior when creating new account on iphone after checking the box to use the private key to be the backup a button appears enter swarm city actual behavior when creating new account on iphone after checking the box to use the private key to be the backup the button appears enter swarm city doesn t appear steps to reproduce the problem on iphone create new choose username and avatar choose pasword make backup show keys operating system hardware safari on iphone
1
338,309
30,291,860,715
IssuesEvent
2023-07-09 11:36:12
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix blas_and_lapack_ops.test_torch_ger
PyTorch Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5499384391/jobs/10021539111"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5499384391/jobs/10021539111"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5499384391/jobs/10021539111"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5499384391/jobs/10021539111"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5499384391/jobs/10021539111"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix blas_and_lapack_ops.test_torch_ger - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5499384391/jobs/10021539111"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5499384391/jobs/10021539111"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5499384391/jobs/10021539111"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5499384391/jobs/10021539111"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5499384391/jobs/10021539111"><img src=https://img.shields.io/badge/-success-success></a>
test
fix blas and lapack ops test torch ger tensorflow a href src jax a href src numpy a href src torch a href src paddle a href src
1
222,656
17,466,766,700
IssuesEvent
2021-08-06 18:02:25
paritytech/polkadot
https://api.github.com/repos/paritytech/polkadot
closed
Approval Voting unit/integration hybrid tests
F4-tests
In order to adequately test approval voting beyond simple unit tests that validate the behavior of the subsystem, we should create multiple instances of the approval voting subsystem and play them against each other.
1.0
Approval Voting unit/integration hybrid tests - In order to adequately test approval voting beyond simple unit tests that validate the behavior of the subsystem, we should create multiple instances of the approval voting subsystem and play them against each other.
test
approval voting unit integration hybrid tests in order to adequately test approval voting beyond simple unit tests that validate the behavior of the subsystem we should create multiple instances of the approval voting subsystem and play them against each other
1
157,690
12,389,088,909
IssuesEvent
2020-05-20 08:27:22
moment/moment
https://api.github.com/repos/moment/moment
closed
2 tests failed. locale:gu:calendar day (1107.6) locale:x-pseudo:calendar day (2655.6)
DST Unit Test Failed
### Client info ``` Date String : Mon Mar 12 2018 14:40:57 GMT-0700 (Pacific Daylight Time) Locale String : 3/12/2018, 2:40:57 PM Offset : 420 User Agent : Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/64.0.3282.186 Safari/537.36 Moment Version: 2.21.0 ``` ==== ### locale:gu:calendar day (1107.6) yesterday at the same time ```javascript // Expected ગઇકાલે રાત ૨:૦૦ વાગ્યે // Actual ગઇકાલે રાત ૩:૦૦ વાગ્યે "ગઇકાલે રાત ૩:૦૦ વાગ્યે" === "ગઇકાલે રાત ૨:૦૦ વાગ્યે" ``` ==== ### locale:x-pseudo:calendar day (2655.6) yesterday at the same time ```javascript // Expected Ý~ést~érdá~ý át 02:00 // Actual Ý~ést~érdá~ý át 03:00 "Ý~ést~érdá~ý át 03:00" === "Ý~ést~érdá~ý át 02:00" ```
1.0
2 tests failed. locale:gu:calendar day (1107.6) locale:x-pseudo:calendar day (2655.6) - ### Client info ``` Date String : Mon Mar 12 2018 14:40:57 GMT-0700 (Pacific Daylight Time) Locale String : 3/12/2018, 2:40:57 PM Offset : 420 User Agent : Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/64.0.3282.186 Safari/537.36 Moment Version: 2.21.0 ``` ==== ### locale:gu:calendar day (1107.6) yesterday at the same time ```javascript // Expected ગઇકાલે રાત ૨:૦૦ વાગ્યે // Actual ગઇકાલે રાત ૩:૦૦ વાગ્યે "ગઇકાલે રાત ૩:૦૦ વાગ્યે" === "ગઇકાલે રાત ૨:૦૦ વાગ્યે" ``` ==== ### locale:x-pseudo:calendar day (2655.6) yesterday at the same time ```javascript // Expected Ý~ést~érdá~ý át 02:00 // Actual Ý~ést~érdá~ý át 03:00 "Ý~ést~érdá~ý át 03:00" === "Ý~ést~érdá~ý át 02:00" ```
test
tests failed locale gu calendar day locale x pseudo calendar day client info date string mon mar gmt pacific daylight time locale string pm offset user agent mozilla windows nt applewebkit khtml like gecko chrome safari moment version locale gu calendar day yesterday at the same time javascript expected ગઇકાલે રાત વાગ્યે actual ગઇકાલે રાત વાગ્યે ગઇકાલે રાત વાગ્યે ગઇકાલે રાત વાગ્યે locale x pseudo calendar day yesterday at the same time javascript expected ý ést érdá ý át actual ý ést érdá ý át ý ést érdá ý át ý ést érdá ý át
1
615,698
19,273,306,644
IssuesEvent
2021-12-10 08:55:23
googleapis/google-cloud-go
https://api.github.com/repos/googleapis/google-cloud-go
closed
logging: memory leak at high frequency logging
type: bug api: logging priority: p2 lang: go
We have an issue when using the go logging library. We are experiencing a very high memory consumption, potentially a leak, when writing logs at a high frequency. We first observed this issue in one of our Go services hosted in Cloud Run. When receiving a high amount of requests (6000rps+) we can see the memory utilisation increasing until the instances runs our of allocated memory. We write one log entry per request. With the logging disabled the service handles the same load consuming around 10-20% of the total memory. We can replicate this behaviour locally by running a simple for-loop that logs on each iteration. Eventually the client OnError()-function is invoked with the following error: "bundler reached buffered byte limit". We've tried changing the values of: [EntryCountThreshold()](https://pkg.go.dev/cloud.google.com/go/logging#EntryCountThreshold) [ConcurrentWriteLimit()](https://pkg.go.dev/cloud.google.com/go/logging#ConcurrentWriteLimit) This doesn't mitigate the problem. However when adding a 10ms sleep on each iteration the problem is gone. Package version: v1.4.2 Container image: golang:1.17.3-alpine3.14 Local environment/version: go1.17.2 darwin/amd64 Sample code to reproduce the issue: ``` package main import ( "context" "log" "sync" glogging "cloud.google.com/go/logging" "github.com/google/uuid" ) func onInternalStackdriverExporterError(err error) { log.Println(err) } const numOfWorkers = 1 func main() { wg := sync.WaitGroup{} ctx := context.Background() c, _ := glogging.NewClient(ctx, "gcp-project-id") c.OnError = onInternalStackdriverExporterError logr := c.Logger("log-load-test") for i := 0; i < numOfWorkers; i++ { wg.Add(1) go doSomething(logr) } wg.Wait() } func doSomething(logger *glogging.Logger) { for { //time.Sleep(10 * time.Millisecond) //running this Sleep mitigates the memory leak. e := glogging.Entry{ Severity: glogging.Severity(100), Payload: uuid.New().String(), } logger.Log(e) } } ```
1.0
logging: memory leak at high frequency logging - We have an issue when using the go logging library. We are experiencing a very high memory consumption, potentially a leak, when writing logs at a high frequency. We first observed this issue in one of our Go services hosted in Cloud Run. When receiving a high amount of requests (6000rps+) we can see the memory utilisation increasing until the instances runs our of allocated memory. We write one log entry per request. With the logging disabled the service handles the same load consuming around 10-20% of the total memory. We can replicate this behaviour locally by running a simple for-loop that logs on each iteration. Eventually the client OnError()-function is invoked with the following error: "bundler reached buffered byte limit". We've tried changing the values of: [EntryCountThreshold()](https://pkg.go.dev/cloud.google.com/go/logging#EntryCountThreshold) [ConcurrentWriteLimit()](https://pkg.go.dev/cloud.google.com/go/logging#ConcurrentWriteLimit) This doesn't mitigate the problem. However when adding a 10ms sleep on each iteration the problem is gone. Package version: v1.4.2 Container image: golang:1.17.3-alpine3.14 Local environment/version: go1.17.2 darwin/amd64 Sample code to reproduce the issue: ``` package main import ( "context" "log" "sync" glogging "cloud.google.com/go/logging" "github.com/google/uuid" ) func onInternalStackdriverExporterError(err error) { log.Println(err) } const numOfWorkers = 1 func main() { wg := sync.WaitGroup{} ctx := context.Background() c, _ := glogging.NewClient(ctx, "gcp-project-id") c.OnError = onInternalStackdriverExporterError logr := c.Logger("log-load-test") for i := 0; i < numOfWorkers; i++ { wg.Add(1) go doSomething(logr) } wg.Wait() } func doSomething(logger *glogging.Logger) { for { //time.Sleep(10 * time.Millisecond) //running this Sleep mitigates the memory leak. e := glogging.Entry{ Severity: glogging.Severity(100), Payload: uuid.New().String(), } logger.Log(e) } } ```
non_test
logging memory leak at high frequency logging we have an issue when using the go logging library we are experiencing a very high memory consumption potentially a leak when writing logs at a high frequency we first observed this issue in one of our go services hosted in cloud run when receiving a high amount of requests we can see the memory utilisation increasing until the instances runs our of allocated memory we write one log entry per request with the logging disabled the service handles the same load consuming around of the total memory we can replicate this behaviour locally by running a simple for loop that logs on each iteration eventually the client onerror function is invoked with the following error bundler reached buffered byte limit we ve tried changing the values of this doesn t mitigate the problem however when adding a sleep on each iteration the problem is gone package version container image golang local environment version darwin sample code to reproduce the issue package main import context log sync glogging cloud google com go logging github com google uuid func oninternalstackdriverexportererror err error log println err const numofworkers func main wg sync waitgroup ctx context background c glogging newclient ctx gcp project id c onerror oninternalstackdriverexportererror logr c logger log load test for i i numofworkers i wg add go dosomething logr wg wait func dosomething logger glogging logger for time sleep time millisecond running this sleep mitigates the memory leak e glogging entry severity glogging severity payload uuid new string logger log e
0
4,814
2,875,502,626
IssuesEvent
2015-06-09 08:36:24
bpmn-io/bpmn-js
https://api.github.com/repos/bpmn-io/bpmn-js
closed
Investigate: Document our APIs in a user friendly way
documentation in progress
Adding a (self) hosted solution with up to date documentation. [**ReadMe.io**](https://readme.io/) could be a great option since it's a well built, easy to use system and free for open source. ### Primary Use Case Users should learn about our public API, i.e. the [Overlays](https://github.com/bpmn-io/diagram-js/blob/master/lib/features/overlays/Overlays.js) service or the main entry points into bpmn-js. * Only some APIs are public * We want to be able to provide information about these directly in code or as close as possible to it (distance means harder to maintain) * We want documentation to be generated to HTML + published somewhere * Linking to GitHub or other parts of the documentation should work ### Options * Using a library * Parsing stuff manually and feeding it into a page generator, i.e. [kartoffeldruck](https://github.com/nikku/kartoffeldruck) or the like ### Tasks * [ ] generate documentation for `Overlays` + `EventBus` (in `MD` format) * [ ] investigate linking between `Overlays` -> `EventBus` * [ ] create HTML from the MD documentation; focus on readability / public
1.0
Investigate: Document our APIs in a user friendly way - Adding a (self) hosted solution with up to date documentation. [**ReadMe.io**](https://readme.io/) could be a great option since it's a well built, easy to use system and free for open source. ### Primary Use Case Users should learn about our public API, i.e. the [Overlays](https://github.com/bpmn-io/diagram-js/blob/master/lib/features/overlays/Overlays.js) service or the main entry points into bpmn-js. * Only some APIs are public * We want to be able to provide information about these directly in code or as close as possible to it (distance means harder to maintain) * We want documentation to be generated to HTML + published somewhere * Linking to GitHub or other parts of the documentation should work ### Options * Using a library * Parsing stuff manually and feeding it into a page generator, i.e. [kartoffeldruck](https://github.com/nikku/kartoffeldruck) or the like ### Tasks * [ ] generate documentation for `Overlays` + `EventBus` (in `MD` format) * [ ] investigate linking between `Overlays` -> `EventBus` * [ ] create HTML from the MD documentation; focus on readability / public
non_test
investigate document our apis in a user friendly way adding a self hosted solution with up to date documentation could be a great option since it s a well built easy to use system and free for open source primary use case users should learn about our public api i e the service or the main entry points into bpmn js only some apis are public we want to be able to provide information about these directly in code or as close as possible to it distance means harder to maintain we want documentation to be generated to html published somewhere linking to github or other parts of the documentation should work options using a library parsing stuff manually and feeding it into a page generator i e or the like tasks generate documentation for overlays eventbus in md format investigate linking between overlays eventbus create html from the md documentation focus on readability public
0
410,250
11,985,432,924
IssuesEvent
2020-04-07 17:29:51
IpsumCapra/project-3-4
https://api.github.com/repos/IpsumCapra/project-3-4
closed
US-B2 - As a user, I want to be able to quickly choose the amount I want to withdraw.
3 ATM High priority! MUST UI User Story
<h1>Acceptance Criteria</h1> <ul> <li>UI software is present, and running on the Raspberry PI.</li> <li>The Raspberry PI and monitor can communicate correctly.</li> <li>ATM is functional.</li> <li>User can interface with ATM terminal.</li> <li>ATM keypad is communicating with other hardware correctly.</li> </ul> <h1>Technical Requirements</h1> <ul> <li><a href="https://github.com/IpsumCapra/project-3-4/issues/12">US-A2</a></li> <li><a href="https://github.com/IpsumCapra/project-3-4/issues/14">US-A3</a></li> <li><a href="https://github.com/IpsumCapra/project-3-4/issues/16">US-T2</a></li> </ul>
1.0
US-B2 - As a user, I want to be able to quickly choose the amount I want to withdraw. - <h1>Acceptance Criteria</h1> <ul> <li>UI software is present, and running on the Raspberry PI.</li> <li>The Raspberry PI and monitor can communicate correctly.</li> <li>ATM is functional.</li> <li>User can interface with ATM terminal.</li> <li>ATM keypad is communicating with other hardware correctly.</li> </ul> <h1>Technical Requirements</h1> <ul> <li><a href="https://github.com/IpsumCapra/project-3-4/issues/12">US-A2</a></li> <li><a href="https://github.com/IpsumCapra/project-3-4/issues/14">US-A3</a></li> <li><a href="https://github.com/IpsumCapra/project-3-4/issues/16">US-T2</a></li> </ul>
non_test
us as a user i want to be able to quickly choose the amount i want to withdraw acceptance criteria ui software is present and running on the raspberry pi the raspberry pi and monitor can communicate correctly atm is functional user can interface with atm terminal atm keypad is communicating with other hardware correctly technical requirements a href a href a href
0
87,937
8,127,236,613
IssuesEvent
2018-08-17 07:15:36
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
cli: Example_csv_tsv_quoting failed under stress
C-test-failure O-robot X-duplicate
SHA: https://github.com/cockroachdb/cockroach/commits/eccb4a127dd519375d87d5ffd9f6394c37c3a427 Parameters: ``` TAGS= GOFLAGS= ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=842992&tab=buildLog ``` W180817 05:15:21.657719 1 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180817 05:15:21.663213 1 server/server.go:828 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled I180817 05:15:21.663369 1 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180817 05:15:21.663382 1 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180817 05:15:21.676785 1 server/config.go:496 [n?] 1 storage engine initialized I180817 05:15:21.676809 1 server/config.go:499 [n?] RocksDB cache size: 128 MiB I180817 05:15:21.676818 1 server/config.go:499 [n?] store 0: in-memory, size 0 B I180817 05:15:21.679636 1 server/node.go:373 [n?] **** cluster 9ba6fa1c-6fd3-403e-8418-6edd4e925e3e has been created I180817 05:15:21.679668 1 server/server.go:1399 [n?] **** add additional nodes by specifying --join=127.0.0.1:39623 I180817 05:15:21.679796 1 gossip/gossip.go:383 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:39623" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v999" started_at:1534482921679733959 I180817 05:15:21.681670 1 storage/store.go:1506 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available I180817 05:15:21.681856 1 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00} I180817 05:15:21.681926 1 storage/stores.go:242 [n1] read 0 node addresses from persistent storage I180817 05:15:21.682006 1 server/node.go:697 [n1] connecting to gossip network to verify cluster ID... I180817 05:15:21.682030 1 server/node.go:722 [n1] node connected via gossip and verified as part of cluster "9ba6fa1c-6fd3-403e-8418-6edd4e925e3e" I180817 05:15:21.682055 1 server/node.go:546 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes [] I180817 05:15:21.682290 1 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180817 05:15:21.682315 1 server/server.go:1805 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined I180817 05:15:21.682368 1 server/server.go:1536 [n1] starting https server at 127.0.0.1:33795 (use: 127.0.0.1:33795) I180817 05:15:21.682383 1 server/server.go:1538 [n1] starting grpc/postgres server at 127.0.0.1:39623 I180817 05:15:21.682394 1 server/server.go:1539 [n1] advertising CockroachDB node at 127.0.0.1:39623 I180817 05:15:21.682569 30790 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180817 05:15:21.691029 30823 storage/replica_command.go:282 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2] I180817 05:15:21.701391 30806 storage/replica_command.go:282 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3] I180817 05:15:21.707183 30838 storage/replica_command.go:282 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4] I180817 05:15:21.714051 30841 storage/replica_command.go:282 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5] I180817 05:15:21.726470 30814 storage/replica_command.go:282 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6] I180817 05:15:21.734352 30849 storage/replica_command.go:282 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7] I180817 05:15:21.756036 30779 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root} I180817 05:15:21.758697 30576 storage/replica_command.go:282 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8] W180817 05:15:21.766816 30857 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=16a55113 key=/Local/Range/Table/SystemConfigSpan/Start/RangeDescriptor rw=true pri=0.11242717 iso=SERIALIZABLE stat=PENDING epo=0 ts=1534482921.758714102,0 orig=1534482921.758714102,0 max=1534482921.758714102,0 wto=false rop=false seq=1 I180817 05:15:21.769233 30871 storage/replica_command.go:282 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9] I180817 05:15:21.779022 30900 storage/replica_command.go:282 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10] I180817 05:15:21.784849 30853 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:$1 User:root} I180817 05:15:21.793390 30562 storage/replica_command.go:282 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11] I180817 05:15:21.799906 30919 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root} I180817 05:15:21.803606 30946 storage/replica_command.go:282 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12] I180817 05:15:21.817114 30964 storage/replica_command.go:282 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13] I180817 05:15:21.823350 30972 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:gen_random_uuid()::STRING User:root} I180817 05:15:21.832952 30960 storage/replica_command.go:282 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14] I180817 05:15:21.836562 30975 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root} I180817 05:15:21.848127 30925 storage/replica_command.go:282 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15] I180817 05:15:21.854929 31004 storage/replica_command.go:282 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16] I180817 05:15:21.857881 30989 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root} I180817 05:15:21.869594 1 server/server.go:1592 [n1] done ensuring all necessary migrations have run I180817 05:15:21.869629 1 server/server.go:1595 [n1] serving sql connections I180817 05:15:21.885520 30863 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:39623} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v999 StartedAt:1534482921679733959 LocalityAddress:[]} ClusterID:9ba6fa1c-6fd3-403e-8418-6edd4e925e3e StartedAt:1534482921679733959 LastUp:1534482921679733959} I180817 05:15:21.888091 30861 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version I180817 05:15:22.697045 31075 rpc/nodedialer/nodedialer.go:92 [consistencyChecker,n1,s1,r1/1:/{Min-System/}] connection to n1 established I180817 05:15:30.686366 30729 storage/replica_proposal.go:214 [n1,s1,r11/1:/Table/1{4-5}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482930.686040857,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0 I180817 05:15:31.683056 30788 server/status/runtime.go:433 [n1] runtime stats: 188 MiB RSS, 253 goroutines, 15 MiB/38 MiB/71 MiB GO alloc/idle/total, 15 MiB/51 MiB CGO alloc/total, 0.00cgo/sec, 0.00/0.00 %(u/s)time, 0.00 %gc (264x) I180817 05:15:31.686568 30715 storage/replica_proposal.go:214 [n1,s1,r16/1:/{Table/19-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0 I180817 05:15:31.686696 31123 storage/replica_command.go:282 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17] I180817 05:15:31.687259 30736 storage/replica_proposal.go:214 [n1,s1,r15/1:/Table/1{8-9}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686880553,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0 I180817 05:15:31.689132 30735 storage/replica_proposal.go:214 [n1,s1,r7/1:/Table/{SystemCon…-11}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.688758132,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0 I180817 05:15:31.692734 30749 storage/replica_proposal.go:214 [n1,s1,r10/1:/Table/1{3-4}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.692338326,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0 I180817 05:15:31.695746 30739 storage/replica_proposal.go:214 [n1,s1,r16/1:/{Table/19-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 I180817 05:15:31.696105 31014 storage/replica_command.go:282 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18] I180817 05:15:31.702463 30746 storage/replica_proposal.go:214 [n1,s1,r17/1:/{Table/20-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 I180817 05:15:31.702951 31157 storage/replica_command.go:282 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19] I180817 05:15:31.704417 30790 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180817 05:15:31.705092 30752 storage/replica_proposal.go:214 [n1,s1,r5/1:/System/ts{d-e}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.704808789,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0 I180817 05:15:31.711731 30690 storage/replica_proposal.go:214 [n1,s1,r18/1:/{Table/21-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 I180817 05:15:31.712165 31131 storage/replica_command.go:282 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20] I180817 05:15:31.721651 30723 storage/replica_proposal.go:214 [n1,s1,r4/1:/System/{NodeLive…-tsd}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.721254860,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0 I180817 05:15:31.723804 30724 storage/replica_proposal.go:214 [n1,s1,r19/1:/{Table/22-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 I180817 05:15:31.726788 31169 storage/replica_command.go:282 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21] I180817 05:15:31.748832 30706 storage/replica_proposal.go:214 [n1,s1,r20/1:/{Table/23-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 I180817 05:15:31.749392 31188 storage/replica_command.go:282 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22] I180817 05:15:31.756080 30732 storage/replica_proposal.go:214 [n1,s1,r21/1:/{Table/50-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 --- FAIL: Example_csv_tsv_quoting (10.34s) panic: had 16 ranges at startup, expected 22 [recovered] panic: had 16 ranges at startup, expected 22 goroutine 1 [running]: testing.runExample.func2(0xbed5b45a667f7544, 0x7156cb8f0, 0x3adc380, 0xc4200de830, 0xc4200de008, 0xc421acafc0, 0x24f4825, 0x17, 0x2619948, 0x2594e57, ...) /usr/local/go/src/testing/example.go:117 +0x707 panic(0x225d480, 0xc42291bce0) /usr/local/go/src/runtime/panic.go:502 +0x229 github.com/cockroachdb/cockroach/pkg/cli.(*cliTest).fail(0xc421993a80, 0x225d480, 0xc42291bce0) /go/src/github.com/cockroachdb/cockroach/pkg/cli/cli_test.go:79 +0xed github.com/cockroachdb/cockroach/pkg/cli.newCLITest(0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/cli/cli_test.go:133 +0x798 github.com/cockroachdb/cockroach/pkg/cli.Example_csv_tsv_quoting() /go/src/github.com/cockroachdb/cockroach/pkg/cli/cli_test.go:1163 +0x54 testing.runExample(0x24f4825, 0x17, 0x2619948, 0x2594e57, 0x998, 0x0, 0x0) /usr/local/go/src/testing/example.go:122 +0x1f9 testing.runExamples(0xc422f85eb8, 0x3898320, 0x10, 0x10, 0x12acb93810101) /usr/local/go/src/testing/example.go:46 +0x15b testing.(*M).Run(0xc42037d900, 0x0) /usr/local/go/src/testing/testing.go:979 +0x1eb github.com/cockroachdb/cockroach/pkg/cli_test.TestMain(0xc42037d900) /go/src/github.com/cockroachdb/cockroach/pkg/cli/main_test.go:38 +0x88 main.main() _testmain.go:140 +0x151 goroutine 19 [syscall]: os/signal.signal_recv(0x0) /usr/local/go/src/runtime/sigqueue.go:139 +0xa6 os/signal.loop() /usr/local/go/src/os/signal/signal_unix.go:22 +0x22 created by os/signal.init.0 /usr/local/go/src/os/signal/signal_unix.go:28 +0x41 goroutine 22 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.flushDaemon() /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1171 +0xf1 created by github.com/cockroachdb/cockroach/pkg/util/log.init.0 /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:590 +0xf4 goroutine 23 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.signalFlusher() /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:597 +0xab created by github.com/cockroachdb/cockroach/pkg/util/log.init.0 /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:591 +0x10c goroutine 50 [select, locked to thread]: runtime.gopark(0x261f2c8, 0x0, 0x24d594c, 0x6, 0x18, 0x1) /usr/local/go/src/runtime/proc.go:291 +0x11a runtime.selectgo(0xc420532750, 0xc42052a060) /usr/local/go/src/runtime/select.go:392 +0xe50 runtime.ensureSigM.func1() /usr/local/go/src/runtime/signal_unix.go:549 +0x1f4 runtime.goexit() /usr/local/go/src/runtime/asm_amd64.s:2361 +0x1 goroutine 9847 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421045320) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 771 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420267d40) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 770 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420267c20) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 109 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420266d80) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 110 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4202678c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 9846 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421045200) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 3504 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4a7e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 29836 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420bf1d40) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 9003 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421327560) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 15006 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4215a7c20) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 23721 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42356e6c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 11615 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42102cb40) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 2767 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c23200) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 1345 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42142d200) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 19252 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42010cd80) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 10528 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4209125a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 18480 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420bf0900) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 1346 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42142d320) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 6096 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420e27320) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 20391 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420e26900) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 29206 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b7a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 15745 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421326120) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 17110 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b0e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 20392 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420e26a20) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 27154 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421182d80) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 13104 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b9e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 22709 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421727560) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 30597 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4aa20) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 23722 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42356e7e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 24295 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4215a6d80) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 19809 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4214370e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 22710 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421727680) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 29205 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b680) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 25464 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42113a7e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 1960 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c22d80) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 1961 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c22ea0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 27792 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421183560) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 17111 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b200) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 14346 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4217267e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 17791 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4218c6c60) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 14345 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4217266c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 3503 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4a6c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 26015 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42102c000) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 9004 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421327680) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 21548 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42325a5a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 29837 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420bf1e60) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 2768 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c23320) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 4828 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4ac60) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 15007 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4215a7d40) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 13103 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b8c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 8236 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061a900) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 18481 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420bf0a20) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 28436 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4212c6900) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 20972 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061b320) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 28437 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4212c6a20) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 24885 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42356e120) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 30596 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4a900) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 27793 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421183680) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 16442 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4212f5d40) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 27153 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421182c60) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 10527 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420912480) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 19253 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42010cea0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 17792 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4218c6d80) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 5161 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b440) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 5162 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b560) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 15744 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421326000) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 26573 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42090e480) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 3775 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42174d200) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 3776 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42174d320) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 4567 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c23560) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 20971 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061b200) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 6744 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42090f200) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 24884 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42356e000) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 19810 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421437200) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 6745 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42090f320) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 4827 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4ab40) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 7556 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42113a120) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 26016 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42102c120) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 4566 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c23440) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 8235 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061a7e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 24296 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4215a6ea0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 22116 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061a120) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 21547 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42325a480) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 16441 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4212f5c20) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 11616 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42102cc60) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 26574 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42090e5a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 22115 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061a000) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 7555 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42113a000) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 25465 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42113a900) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 6097 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420e27440) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 ERROR: exit status 2 24 runs completed, 1 failures, over 1m53s context canceled ```
1.0
cli: Example_csv_tsv_quoting failed under stress - SHA: https://github.com/cockroachdb/cockroach/commits/eccb4a127dd519375d87d5ffd9f6394c37c3a427 Parameters: ``` TAGS= GOFLAGS= ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=842992&tab=buildLog ``` W180817 05:15:21.657719 1 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180817 05:15:21.663213 1 server/server.go:828 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled I180817 05:15:21.663369 1 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180817 05:15:21.663382 1 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180817 05:15:21.676785 1 server/config.go:496 [n?] 1 storage engine initialized I180817 05:15:21.676809 1 server/config.go:499 [n?] RocksDB cache size: 128 MiB I180817 05:15:21.676818 1 server/config.go:499 [n?] store 0: in-memory, size 0 B I180817 05:15:21.679636 1 server/node.go:373 [n?] **** cluster 9ba6fa1c-6fd3-403e-8418-6edd4e925e3e has been created I180817 05:15:21.679668 1 server/server.go:1399 [n?] **** add additional nodes by specifying --join=127.0.0.1:39623 I180817 05:15:21.679796 1 gossip/gossip.go:383 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:39623" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:11 > build_tag:"v999" started_at:1534482921679733959 I180817 05:15:21.681670 1 storage/store.go:1506 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available I180817 05:15:21.681856 1 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=0, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00} I180817 05:15:21.681926 1 storage/stores.go:242 [n1] read 0 node addresses from persistent storage I180817 05:15:21.682006 1 server/node.go:697 [n1] connecting to gossip network to verify cluster ID... I180817 05:15:21.682030 1 server/node.go:722 [n1] node connected via gossip and verified as part of cluster "9ba6fa1c-6fd3-403e-8418-6edd4e925e3e" I180817 05:15:21.682055 1 server/node.go:546 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes [] I180817 05:15:21.682290 1 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180817 05:15:21.682315 1 server/server.go:1805 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined I180817 05:15:21.682368 1 server/server.go:1536 [n1] starting https server at 127.0.0.1:33795 (use: 127.0.0.1:33795) I180817 05:15:21.682383 1 server/server.go:1538 [n1] starting grpc/postgres server at 127.0.0.1:39623 I180817 05:15:21.682394 1 server/server.go:1539 [n1] advertising CockroachDB node at 127.0.0.1:39623 I180817 05:15:21.682569 30790 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180817 05:15:21.691029 30823 storage/replica_command.go:282 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2] I180817 05:15:21.701391 30806 storage/replica_command.go:282 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3] I180817 05:15:21.707183 30838 storage/replica_command.go:282 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4] I180817 05:15:21.714051 30841 storage/replica_command.go:282 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5] I180817 05:15:21.726470 30814 storage/replica_command.go:282 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6] I180817 05:15:21.734352 30849 storage/replica_command.go:282 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7] I180817 05:15:21.756036 30779 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root} I180817 05:15:21.758697 30576 storage/replica_command.go:282 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8] W180817 05:15:21.766816 30857 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "split" id=16a55113 key=/Local/Range/Table/SystemConfigSpan/Start/RangeDescriptor rw=true pri=0.11242717 iso=SERIALIZABLE stat=PENDING epo=0 ts=1534482921.758714102,0 orig=1534482921.758714102,0 max=1534482921.758714102,0 wto=false rop=false seq=1 I180817 05:15:21.769233 30871 storage/replica_command.go:282 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9] I180817 05:15:21.779022 30900 storage/replica_command.go:282 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10] I180817 05:15:21.784849 30853 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:$1 User:root} I180817 05:15:21.793390 30562 storage/replica_command.go:282 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11] I180817 05:15:21.799906 30919 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root} I180817 05:15:21.803606 30946 storage/replica_command.go:282 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12] I180817 05:15:21.817114 30964 storage/replica_command.go:282 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13] I180817 05:15:21.823350 30972 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:gen_random_uuid()::STRING User:root} I180817 05:15:21.832952 30960 storage/replica_command.go:282 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14] I180817 05:15:21.836562 30975 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root} I180817 05:15:21.848127 30925 storage/replica_command.go:282 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15] I180817 05:15:21.854929 31004 storage/replica_command.go:282 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16] I180817 05:15:21.857881 30989 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root} I180817 05:15:21.869594 1 server/server.go:1592 [n1] done ensuring all necessary migrations have run I180817 05:15:21.869629 1 server/server.go:1595 [n1] serving sql connections I180817 05:15:21.885520 30863 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:39623} Attrs: Locality: ServerVersion:2.0-11 BuildTag:v999 StartedAt:1534482921679733959 LocalityAddress:[]} ClusterID:9ba6fa1c-6fd3-403e-8418-6edd4e925e3e StartedAt:1534482921679733959 LastUp:1534482921679733959} I180817 05:15:21.888091 30861 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version I180817 05:15:22.697045 31075 rpc/nodedialer/nodedialer.go:92 [consistencyChecker,n1,s1,r1/1:/{Min-System/}] connection to n1 established I180817 05:15:30.686366 30729 storage/replica_proposal.go:214 [n1,s1,r11/1:/Table/1{4-5}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482930.686040857,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0 I180817 05:15:31.683056 30788 server/status/runtime.go:433 [n1] runtime stats: 188 MiB RSS, 253 goroutines, 15 MiB/38 MiB/71 MiB GO alloc/idle/total, 15 MiB/51 MiB CGO alloc/total, 0.00cgo/sec, 0.00/0.00 %(u/s)time, 0.00 %gc (264x) I180817 05:15:31.686568 30715 storage/replica_proposal.go:214 [n1,s1,r16/1:/{Table/19-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0 I180817 05:15:31.686696 31123 storage/replica_command.go:282 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17] I180817 05:15:31.687259 30736 storage/replica_proposal.go:214 [n1,s1,r15/1:/Table/1{8-9}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686880553,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0 I180817 05:15:31.689132 30735 storage/replica_proposal.go:214 [n1,s1,r7/1:/Table/{SystemCon…-11}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.688758132,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0 I180817 05:15:31.692734 30749 storage/replica_proposal.go:214 [n1,s1,r10/1:/Table/1{3-4}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.692338326,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0 I180817 05:15:31.695746 30739 storage/replica_proposal.go:214 [n1,s1,r16/1:/{Table/19-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 I180817 05:15:31.696105 31014 storage/replica_command.go:282 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18] I180817 05:15:31.702463 30746 storage/replica_proposal.go:214 [n1,s1,r17/1:/{Table/20-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 I180817 05:15:31.702951 31157 storage/replica_command.go:282 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19] I180817 05:15:31.704417 30790 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180817 05:15:31.705092 30752 storage/replica_proposal.go:214 [n1,s1,r5/1:/System/ts{d-e}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.704808789,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0 I180817 05:15:31.711731 30690 storage/replica_proposal.go:214 [n1,s1,r18/1:/{Table/21-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 I180817 05:15:31.712165 31131 storage/replica_command.go:282 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20] I180817 05:15:31.721651 30723 storage/replica_proposal.go:214 [n1,s1,r4/1:/System/{NodeLive…-tsd}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.721254860,0 following repl=(n1,s1):1 seq=2 start=1534482921.681508117,0 exp=1534482930.681696937,0 pro=1534482921.681754364,0 I180817 05:15:31.723804 30724 storage/replica_proposal.go:214 [n1,s1,r19/1:/{Table/22-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 I180817 05:15:31.726788 31169 storage/replica_command.go:282 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21] I180817 05:15:31.748832 30706 storage/replica_proposal.go:214 [n1,s1,r20/1:/{Table/23-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 I180817 05:15:31.749392 31188 storage/replica_command.go:282 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22] I180817 05:15:31.756080 30732 storage/replica_proposal.go:214 [n1,s1,r21/1:/{Table/50-Max}] new range lease repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 following repl=(n1,s1):1 seq=3 start=1534482921.681508117,0 epo=1 pro=1534482931.686271443,0 --- FAIL: Example_csv_tsv_quoting (10.34s) panic: had 16 ranges at startup, expected 22 [recovered] panic: had 16 ranges at startup, expected 22 goroutine 1 [running]: testing.runExample.func2(0xbed5b45a667f7544, 0x7156cb8f0, 0x3adc380, 0xc4200de830, 0xc4200de008, 0xc421acafc0, 0x24f4825, 0x17, 0x2619948, 0x2594e57, ...) /usr/local/go/src/testing/example.go:117 +0x707 panic(0x225d480, 0xc42291bce0) /usr/local/go/src/runtime/panic.go:502 +0x229 github.com/cockroachdb/cockroach/pkg/cli.(*cliTest).fail(0xc421993a80, 0x225d480, 0xc42291bce0) /go/src/github.com/cockroachdb/cockroach/pkg/cli/cli_test.go:79 +0xed github.com/cockroachdb/cockroach/pkg/cli.newCLITest(0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /go/src/github.com/cockroachdb/cockroach/pkg/cli/cli_test.go:133 +0x798 github.com/cockroachdb/cockroach/pkg/cli.Example_csv_tsv_quoting() /go/src/github.com/cockroachdb/cockroach/pkg/cli/cli_test.go:1163 +0x54 testing.runExample(0x24f4825, 0x17, 0x2619948, 0x2594e57, 0x998, 0x0, 0x0) /usr/local/go/src/testing/example.go:122 +0x1f9 testing.runExamples(0xc422f85eb8, 0x3898320, 0x10, 0x10, 0x12acb93810101) /usr/local/go/src/testing/example.go:46 +0x15b testing.(*M).Run(0xc42037d900, 0x0) /usr/local/go/src/testing/testing.go:979 +0x1eb github.com/cockroachdb/cockroach/pkg/cli_test.TestMain(0xc42037d900) /go/src/github.com/cockroachdb/cockroach/pkg/cli/main_test.go:38 +0x88 main.main() _testmain.go:140 +0x151 goroutine 19 [syscall]: os/signal.signal_recv(0x0) /usr/local/go/src/runtime/sigqueue.go:139 +0xa6 os/signal.loop() /usr/local/go/src/os/signal/signal_unix.go:22 +0x22 created by os/signal.init.0 /usr/local/go/src/os/signal/signal_unix.go:28 +0x41 goroutine 22 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.flushDaemon() /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1171 +0xf1 created by github.com/cockroachdb/cockroach/pkg/util/log.init.0 /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:590 +0xf4 goroutine 23 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.signalFlusher() /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:597 +0xab created by github.com/cockroachdb/cockroach/pkg/util/log.init.0 /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:591 +0x10c goroutine 50 [select, locked to thread]: runtime.gopark(0x261f2c8, 0x0, 0x24d594c, 0x6, 0x18, 0x1) /usr/local/go/src/runtime/proc.go:291 +0x11a runtime.selectgo(0xc420532750, 0xc42052a060) /usr/local/go/src/runtime/select.go:392 +0xe50 runtime.ensureSigM.func1() /usr/local/go/src/runtime/signal_unix.go:549 +0x1f4 runtime.goexit() /usr/local/go/src/runtime/asm_amd64.s:2361 +0x1 goroutine 9847 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421045320) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 771 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420267d40) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 770 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420267c20) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 109 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420266d80) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 110 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4202678c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 9846 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421045200) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 3504 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4a7e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 29836 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420bf1d40) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 9003 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421327560) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 15006 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4215a7c20) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 23721 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42356e6c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 11615 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42102cb40) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 2767 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c23200) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 1345 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42142d200) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 19252 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42010cd80) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 10528 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4209125a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 18480 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420bf0900) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 1346 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42142d320) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 6096 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420e27320) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 20391 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420e26900) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 29206 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b7a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 15745 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421326120) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 17110 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b0e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 20392 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420e26a20) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 27154 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421182d80) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 13104 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b9e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 22709 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421727560) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 30597 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4aa20) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 23722 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42356e7e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 24295 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4215a6d80) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 19809 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4214370e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 22710 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421727680) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 29205 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b680) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 25464 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42113a7e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 1960 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c22d80) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 1961 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c22ea0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 27792 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421183560) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 17111 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b200) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 14346 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4217267e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 17791 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4218c6c60) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 14345 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4217266c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 3503 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4a6c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 26015 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42102c000) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 9004 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421327680) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 21548 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42325a5a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 29837 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420bf1e60) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 2768 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c23320) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 4828 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4ac60) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 15007 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4215a7d40) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 13103 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b8c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 8236 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061a900) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 18481 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420bf0a20) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 28436 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4212c6900) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 20972 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061b320) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 28437 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4212c6a20) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 24885 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42356e120) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 30596 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4a900) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 27793 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421183680) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 16442 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4212f5d40) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 27153 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421182c60) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 10527 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420912480) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 19253 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42010cea0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 17792 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4218c6d80) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 5161 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b440) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 5162 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4b560) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 15744 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421326000) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 26573 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42090e480) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 3775 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42174d200) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 3776 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42174d320) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 4567 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c23560) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 20971 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061b200) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 6744 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42090f200) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 24884 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42356e000) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 19810 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421437200) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 6745 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42090f320) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 4827 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc421a4ab40) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 7556 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42113a120) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 26016 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42102c120) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 4566 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420c23440) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 8235 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061a7e0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 24296 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4215a6ea0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 22116 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061a120) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 21547 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42325a480) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 16441 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc4212f5c20) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 11616 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42102cc60) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 26574 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42090e5a0) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 22115 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42061a000) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 7555 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42113a000) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 25465 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc42113a900) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 goroutine 6097 [chan receive]: github.com/cockroachdb/cockroach/pkg/util/log.(*loggingT).gcDaemon(0xc420e27440) /go/src/github.com/cockroachdb/cockroach/pkg/util/log/clog.go:1223 +0x68 created by github.com/cockroachdb/cockroach/pkg/util/log.NewSecondaryLogger /go/src/github.com/cockroachdb/cockroach/pkg/util/log/secondary_log.go:80 +0x288 ERROR: exit status 2 24 runs completed, 1 failures, over 1m53s context canceled ```
test
cli example csv tsv quoting failed under stress sha parameters tags goflags failed test server status runtime go could not parse build timestamp parsing time as cannot parse as server server go monitoring forward clock jumps based on server clock forward jump check enabled base addr validation go server certificate addresses ip dns localhost local cn node base addr validation go web ui certificate addresses ip dns localhost local cn node server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b server node go cluster has been created server server go add additional nodes by specifying join gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag started at storage store go failed initial metrics computation system config not yet available server node go initialized store disk capacity mib available mib used b logicalbytes kib ranges leases queries writes bytesperreplica pmax writesperreplica pmax storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster server node go node started with engine s and attributes server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go could not start heap profiler worker due to directory to store profiles could not be determined server server go starting https server at use server server go starting grpc postgres server at server server go advertising cockroachdb node at server status recorder go available memory from cgroups eib exceeds system memory gib using system memory storage replica command go initiating a split of this range at key system storage replica command go initiating a split of this range at key system nodeliveness storage replica command go initiating a split of this range at key system nodelivenessmax storage replica command go initiating a split of this range at key system tsd storage replica command go initiating a split of this range at key system tse storage replica command go initiating a split of this range at key table systemconfigspan start sql event log go event set cluster setting target info settingname diagnostics reporting enabled value true user root storage replica command go initiating a split of this range at key table storage intent resolver go failed to push during intent resolution failed to push split id key local range table systemconfigspan start rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname version value user root storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname trace debug enable value false user root storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname cluster secret value gen random uuid string user root storage replica command go initiating a split of this range at key table sql event log go event create database target info databasename defaultdb statement create database if not exists defaultdb user root storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event create database target info databasename postgres statement create database if not exists postgres user root server server go done ensuring all necessary migrations have run server server go serving sql connections sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion buildtag startedat localityaddress clusterid startedat lastup server server update go no need to upgrade cluster already at the newest version rpc nodedialer nodedialer go connection to established storage replica proposal go new range lease repl seq start epo pro following repl seq start exp pro server status runtime go runtime stats mib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total sec u s time gc storage replica proposal go new range lease repl seq start epo pro following repl seq start exp pro storage replica command go initiating a split of this range at key table storage replica proposal go new range lease repl seq start epo pro following repl seq start exp pro storage replica proposal go new range lease repl seq start epo pro following repl seq start exp pro storage replica proposal go new range lease repl seq start epo pro following repl seq start exp pro storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica command go initiating a split of this range at key table storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica command go initiating a split of this range at key table server status recorder go available memory from cgroups eib exceeds system memory gib using system memory storage replica proposal go new range lease repl seq start epo pro following repl seq start exp pro storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica command go initiating a split of this range at key table storage replica proposal go new range lease repl seq start epo pro following repl seq start exp pro storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica command go initiating a split of this range at key table storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro storage replica command go initiating a split of this range at key table storage replica proposal go new range lease repl seq start epo pro following repl seq start epo pro fail example csv tsv quoting panic had ranges at startup expected panic had ranges at startup expected goroutine testing runexample usr local go src testing example go panic usr local go src runtime panic go github com cockroachdb cockroach pkg cli clitest fail go src github com cockroachdb cockroach pkg cli cli test go github com cockroachdb cockroach pkg cli newclitest go src github com cockroachdb cockroach pkg cli cli test go github com cockroachdb cockroach pkg cli example csv tsv quoting go src github com cockroachdb cockroach pkg cli cli test go testing runexample usr local go src testing example go testing runexamples usr local go src testing example go testing m run usr local go src testing testing go github com cockroachdb cockroach pkg cli test testmain go src github com cockroachdb cockroach pkg cli main test go main main testmain go goroutine os signal signal recv usr local go src runtime sigqueue go os signal loop usr local go src os signal signal unix go created by os signal init usr local go src os signal signal unix go goroutine github com cockroachdb cockroach pkg util log flushdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log init go src github com cockroachdb cockroach pkg util log clog go goroutine github com cockroachdb cockroach pkg util log signalflusher go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log init go src github com cockroachdb cockroach pkg util log clog go goroutine runtime gopark usr local go src runtime proc go runtime selectgo usr local go src runtime select go runtime ensuresigm usr local go src runtime signal unix go runtime goexit usr local go src runtime asm s goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go goroutine github com cockroachdb cockroach pkg util log loggingt gcdaemon go src github com cockroachdb cockroach pkg util log clog go created by github com cockroachdb cockroach pkg util log newsecondarylogger go src github com cockroachdb cockroach pkg util log secondary log go error exit status runs completed failures over context canceled
1
103,855
16,610,450,370
IssuesEvent
2021-06-02 10:48:58
Thanraj/OpenSSL_
https://api.github.com/repos/Thanraj/OpenSSL_
opened
CVE-2015-1787 (Low) detected in opensslOpenSSL_1_0_2
security vulnerability
## CVE-2015-1787 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opensslOpenSSL_1_0_2</b></p></summary> <p> <p>TLS/SSL and crypto library</p> <p>Library home page: <a href=https://github.com/openssl/openssl.git>https://github.com/openssl/openssl.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/Thanraj/OpenSSL_/commit/fa3551a60a0e344c1d79971fbfc4e45646f13720">fa3551a60a0e344c1d79971fbfc4e45646f13720</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>OpenSSL_/ssl/s3_srvr.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The ssl3_get_client_key_exchange function in s3_srvr.c in OpenSSL 1.0.2 before 1.0.2a, when client authentication and an ephemeral Diffie-Hellman ciphersuite are enabled, allows remote attackers to cause a denial of service (daemon crash) via a ClientKeyExchange message with a length of zero. <p>Publish Date: 2015-03-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-1787>CVE-2015-1787</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>2.6</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-1787">https://nvd.nist.gov/vuln/detail/CVE-2015-1787</a></p> <p>Release Date: 2015-03-19</p> <p>Fix Resolution: 1.0.2a</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2015-1787 (Low) detected in opensslOpenSSL_1_0_2 - ## CVE-2015-1787 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opensslOpenSSL_1_0_2</b></p></summary> <p> <p>TLS/SSL and crypto library</p> <p>Library home page: <a href=https://github.com/openssl/openssl.git>https://github.com/openssl/openssl.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/Thanraj/OpenSSL_/commit/fa3551a60a0e344c1d79971fbfc4e45646f13720">fa3551a60a0e344c1d79971fbfc4e45646f13720</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>OpenSSL_/ssl/s3_srvr.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The ssl3_get_client_key_exchange function in s3_srvr.c in OpenSSL 1.0.2 before 1.0.2a, when client authentication and an ephemeral Diffie-Hellman ciphersuite are enabled, allows remote attackers to cause a denial of service (daemon crash) via a ClientKeyExchange message with a length of zero. <p>Publish Date: 2015-03-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-1787>CVE-2015-1787</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>2.6</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-1787">https://nvd.nist.gov/vuln/detail/CVE-2015-1787</a></p> <p>Release Date: 2015-03-19</p> <p>Fix Resolution: 1.0.2a</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve low detected in opensslopenssl cve low severity vulnerability vulnerable library opensslopenssl tls ssl and crypto library library home page a href found in head commit a href found in base branch master vulnerable source files openssl ssl srvr c vulnerability details the get client key exchange function in srvr c in openssl before when client authentication and an ephemeral diffie hellman ciphersuite are enabled allows remote attackers to cause a denial of service daemon crash via a clientkeyexchange message with a length of zero publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
69,517
14,992,608,921
IssuesEvent
2021-01-29 10:06:27
KorAP/Kalamar
https://api.github.com/repos/KorAP/Kalamar
closed
Establish Content Security Policy
security
Before integrating the widget plugin mechanism in Kalamar, the server should establish strict rules to limit security risks. Currently there are some violations of basic JS and CSS inline rules, that could easily be fixed.
True
Establish Content Security Policy - Before integrating the widget plugin mechanism in Kalamar, the server should establish strict rules to limit security risks. Currently there are some violations of basic JS and CSS inline rules, that could easily be fixed.
non_test
establish content security policy before integrating the widget plugin mechanism in kalamar the server should establish strict rules to limit security risks currently there are some violations of basic js and css inline rules that could easily be fixed
0
241,367
20,118,266,167
IssuesEvent
2022-02-07 22:06:02
eclipse-openj9/openj9
https://api.github.com/repos/eclipse-openj9/openj9
closed
JDK11 MacOS jdk_imageio_0_FAILED - AWTError: WindowServer is not available & others
comp:vm test failure os:macos
Failure link ------------ From an internal build `Test_openjdk11_j9_extended.openjdk_x86-64_mac/1` ``` 05:23:03 openjdk version "11.0.11" 2021-04-20 05:23:03 OpenJDK Runtime Environment AdoptOpenJDK (build 11.0.11+4) 05:23:03 Eclipse OpenJ9 VM AdoptOpenJDK (build master-f021812fb, JRE 11 Mac OS X amd64-64-Bit Compressed References 20210227_55 (JIT enabled, AOT enabled) 05:23:03 OpenJ9 - f021812fb 05:23:03 OMR - eb4f5a875 05:23:03 JCL - 7222dc7018 based on jdk-11.0.11+4) ``` rebuild the failed tests in one link: 07:09:55 https://hyc-runtimes-jenkins.swg-devops.com/job/Grinder/parambuild/?JDK_VERSION=11&JDK_IMPL=openj9&JDK_VENDOR=adoptopenjdk&BUILD_LIST=openjdk&PLATFORM=x86-64_mac_mixed&TARGET=testList%20TESTLIST=jdk_imageio_0,jdk_imageio_1 Optional info ------------- Failure output (captured from console output) --------------------------------------------- ``` 07:03:29 TEST: javax/imageio/plugins/wbmp/WbmpBigDestinationTest.java 07:03:29 TEST JDK: /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdkbinary/j2sdk-image/Contents/Home/bin/.. 07:03:29 07:03:29 ACTION: build -- Passed. Build successful 07:03:29 REASON: Named class compiled on demand 07:03:29 TIME: 1.184 seconds 07:03:29 messages: 07:03:29 command: build WbmpBigDestinationTest 07:03:29 reason: Named class compiled on demand 07:03:29 Test directory: 07:03:29 compile: WbmpBigDestinationTest 07:03:29 elapsed time (seconds): 1.184 07:03:29 07:03:29 ACTION: compile -- Passed. Compilation successful 07:03:29 REASON: .class file out of date or does not exist 07:03:29 TIME: 1.184 seconds 07:03:29 messages: 07:03:29 command: compile /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.java 07:03:29 reason: .class file out of date or does not exist 07:03:29 Additional options from @modules: --add-modules java.desktop 07:03:29 Mode: othervm 07:03:29 elapsed time (seconds): 1.184 07:03:29 configuration: 07:03:29 javac compilation environment 07:03:29 add modules: java.desktop 07:03:29 source path: /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp 07:03:29 class path: /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp 07:03:29 /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/classes/1/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.d 07:03:29 07:03:29 rerun: 07:03:29 cd /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/javax/imageio/plugins/wbmp/WbmpBigDestinationTest && \ 07:03:29 HOME=/Users/jenkins \ 07:03:29 PATH=/bin:/usr/bin:/usr/sbin \ 07:03:29 /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdkbinary/j2sdk-image/Contents/Home/bin/../bin/javac \ 07:03:29 -J-ea \ 07:03:29 -J-esa \ 07:03:29 -J-Xmx512m \ 07:03:29 -J-XX:+UseCompressedOops \ 07:03:29 -J-Dtest.vm.opts='-ea -esa -Xmx512m -XX:+UseCompressedOops' \ 07:03:29 -J-Dtest.tool.vm.opts='-J-ea -J-esa -J-Xmx512m -J-XX:+UseCompressedOops' \ 07:03:29 -J-Dtest.compiler.opts= \ 07:03:29 -J-Dtest.java.opts= \ 07:03:29 -J-Dtest.jdk=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdkbinary/j2sdk-image/Contents/Home/bin/.. \ 07:03:29 -J-Dcompile.jdk=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdkbinary/j2sdk-image/Contents/Home/bin/.. \ 07:03:29 -J-Dtest.timeout.factor=8.0 \ 07:03:29 -J-Dtest.root=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk \ 07:03:29 -J-Dtest.name=javax/imageio/plugins/wbmp/WbmpBigDestinationTest.java \ 07:03:29 -J-Dtest.file=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.java \ 07:03:29 -J-Dtest.src=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp \ 07:03:29 -J-Dtest.src.path=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp \ 07:03:29 -J-Dtest.classes=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/classes/1/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.d \ 07:03:29 -J-Dtest.class.path=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/classes/1/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.d \ 07:03:29 -J-Dtest.modules=java.desktop \ 07:03:29 --add-modules java.desktop \ 07:03:29 -d /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/classes/1/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.d \ 07:03:29 -sourcepath /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp \ 07:03:29 -classpath /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp:/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/classes/1/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.d /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.java 07:03:29 STDOUT: 07:03:29 STDERR: 07:03:29 07:03:29 ACTION: main -- Failed. Execution failed: `main' threw exception: java.awt.AWTError: WindowServer is not available 07:03:29 REASON: Assumed action based on file name: run main WbmpBigDestinationTest 07:03:29 TIME: 0.349 seconds 07:03:29 messages: 07:03:29 command: main WbmpBigDestinationTest 07:03:29 reason: Assumed action based on file name: run main WbmpBigDestinationTest 07:03:29 Mode: othervm 07:03:29 Additional options from @modules: --add-modules java.desktop 07:03:29 elapsed time (seconds): 0.349 07:03:29 configuration: 07:03:29 Boot Layer 07:03:29 add modules: java.desktop 07:03:29 07:03:29 STDOUT: 07:03:29 STDERR: 07:03:29 java.awt.AWTError: WindowServer is not available 07:03:29 at java.desktop/sun.lwawt.macosx.LWCToolkit.<clinit>(LWCToolkit.java:167) 07:03:29 at java.base/java.lang.Class.forNameImpl(Native Method) 07:03:29 at java.base/java.lang.Class.forName(Class.java:337) 07:03:29 at java.desktop/java.awt.Toolkit$2.run(Toolkit.java:588) 07:03:29 at java.desktop/java.awt.Toolkit$2.run(Toolkit.java:583) 07:03:29 at java.base/java.security.AccessController.doPrivileged(AccessController.java:682) 07:03:29 at java.desktop/java.awt.Toolkit.getDefaultToolkit(Toolkit.java:582) 07:03:29 at java.desktop/sun.awt.CGraphicsEnvironment.<clinit>(CGraphicsEnvironment.java:73) 07:03:29 at java.base/java.lang.Class.forNameImpl(Native Method) 07:03:29 at java.base/java.lang.Class.forName(Class.java:337) 07:03:29 at java.desktop/java.awt.GraphicsEnvironment$LocalGE.createGE(GraphicsEnvironment.java:101) 07:03:29 at java.desktop/java.awt.GraphicsEnvironment$LocalGE.<clinit>(GraphicsEnvironment.java:83) 07:03:29 at java.desktop/java.awt.GraphicsEnvironment.getLocalGraphicsEnvironment(GraphicsEnvironment.java:129) 07:03:29 at java.desktop/java.awt.image.BufferedImage.createGraphics(BufferedImage.java:1181) 07:03:29 at WbmpBigDestinationTest.main(WbmpBigDestinationTest.java:51) 07:03:29 at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 07:03:29 at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 07:03:29 at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) 07:03:29 at java.base/java.lang.reflect.Method.invoke(Method.java:566) 07:03:29 at com.sun.javatest.regtest.agent.MainWrapper$MainThread.run(MainWrapper.java:127) 07:03:29 at java.base/java.lang.Thread.run(Thread.java:836) 07:03:29 07:03:29 JavaTest Message: Test threw exception: java.awt.AWTError: WindowServer is not available 07:03:29 JavaTest Message: shutting down test 07:03:29 07:03:29 STATUS:Failed.`main' threw exception: java.awt.AWTError: WindowServer is not available 07:03:58 -------------------------------------------------- 07:04:27 Test results: passed: 122; failed: 68 07:04:37 Report written to /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/jvmtest/openjdk/report/html/report.html 07:04:37 Results written to /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work 07:04:37 Error: Some tests failed or other problems occurred. 07:04:37 07:04:37 jdk_imageio_0_FAILED ``` It appears related to the test setup.
1.0
JDK11 MacOS jdk_imageio_0_FAILED - AWTError: WindowServer is not available & others - Failure link ------------ From an internal build `Test_openjdk11_j9_extended.openjdk_x86-64_mac/1` ``` 05:23:03 openjdk version "11.0.11" 2021-04-20 05:23:03 OpenJDK Runtime Environment AdoptOpenJDK (build 11.0.11+4) 05:23:03 Eclipse OpenJ9 VM AdoptOpenJDK (build master-f021812fb, JRE 11 Mac OS X amd64-64-Bit Compressed References 20210227_55 (JIT enabled, AOT enabled) 05:23:03 OpenJ9 - f021812fb 05:23:03 OMR - eb4f5a875 05:23:03 JCL - 7222dc7018 based on jdk-11.0.11+4) ``` rebuild the failed tests in one link: 07:09:55 https://hyc-runtimes-jenkins.swg-devops.com/job/Grinder/parambuild/?JDK_VERSION=11&JDK_IMPL=openj9&JDK_VENDOR=adoptopenjdk&BUILD_LIST=openjdk&PLATFORM=x86-64_mac_mixed&TARGET=testList%20TESTLIST=jdk_imageio_0,jdk_imageio_1 Optional info ------------- Failure output (captured from console output) --------------------------------------------- ``` 07:03:29 TEST: javax/imageio/plugins/wbmp/WbmpBigDestinationTest.java 07:03:29 TEST JDK: /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdkbinary/j2sdk-image/Contents/Home/bin/.. 07:03:29 07:03:29 ACTION: build -- Passed. Build successful 07:03:29 REASON: Named class compiled on demand 07:03:29 TIME: 1.184 seconds 07:03:29 messages: 07:03:29 command: build WbmpBigDestinationTest 07:03:29 reason: Named class compiled on demand 07:03:29 Test directory: 07:03:29 compile: WbmpBigDestinationTest 07:03:29 elapsed time (seconds): 1.184 07:03:29 07:03:29 ACTION: compile -- Passed. Compilation successful 07:03:29 REASON: .class file out of date or does not exist 07:03:29 TIME: 1.184 seconds 07:03:29 messages: 07:03:29 command: compile /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.java 07:03:29 reason: .class file out of date or does not exist 07:03:29 Additional options from @modules: --add-modules java.desktop 07:03:29 Mode: othervm 07:03:29 elapsed time (seconds): 1.184 07:03:29 configuration: 07:03:29 javac compilation environment 07:03:29 add modules: java.desktop 07:03:29 source path: /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp 07:03:29 class path: /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp 07:03:29 /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/classes/1/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.d 07:03:29 07:03:29 rerun: 07:03:29 cd /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/javax/imageio/plugins/wbmp/WbmpBigDestinationTest && \ 07:03:29 HOME=/Users/jenkins \ 07:03:29 PATH=/bin:/usr/bin:/usr/sbin \ 07:03:29 /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdkbinary/j2sdk-image/Contents/Home/bin/../bin/javac \ 07:03:29 -J-ea \ 07:03:29 -J-esa \ 07:03:29 -J-Xmx512m \ 07:03:29 -J-XX:+UseCompressedOops \ 07:03:29 -J-Dtest.vm.opts='-ea -esa -Xmx512m -XX:+UseCompressedOops' \ 07:03:29 -J-Dtest.tool.vm.opts='-J-ea -J-esa -J-Xmx512m -J-XX:+UseCompressedOops' \ 07:03:29 -J-Dtest.compiler.opts= \ 07:03:29 -J-Dtest.java.opts= \ 07:03:29 -J-Dtest.jdk=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdkbinary/j2sdk-image/Contents/Home/bin/.. \ 07:03:29 -J-Dcompile.jdk=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdkbinary/j2sdk-image/Contents/Home/bin/.. \ 07:03:29 -J-Dtest.timeout.factor=8.0 \ 07:03:29 -J-Dtest.root=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk \ 07:03:29 -J-Dtest.name=javax/imageio/plugins/wbmp/WbmpBigDestinationTest.java \ 07:03:29 -J-Dtest.file=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.java \ 07:03:29 -J-Dtest.src=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp \ 07:03:29 -J-Dtest.src.path=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp \ 07:03:29 -J-Dtest.classes=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/classes/1/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.d \ 07:03:29 -J-Dtest.class.path=/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/classes/1/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.d \ 07:03:29 -J-Dtest.modules=java.desktop \ 07:03:29 --add-modules java.desktop \ 07:03:29 -d /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/classes/1/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.d \ 07:03:29 -sourcepath /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp \ 07:03:29 -classpath /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp:/Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work/classes/1/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.d /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/openjdk/openjdk-jdk/test/jdk/javax/imageio/plugins/wbmp/WbmpBigDestinationTest.java 07:03:29 STDOUT: 07:03:29 STDERR: 07:03:29 07:03:29 ACTION: main -- Failed. Execution failed: `main' threw exception: java.awt.AWTError: WindowServer is not available 07:03:29 REASON: Assumed action based on file name: run main WbmpBigDestinationTest 07:03:29 TIME: 0.349 seconds 07:03:29 messages: 07:03:29 command: main WbmpBigDestinationTest 07:03:29 reason: Assumed action based on file name: run main WbmpBigDestinationTest 07:03:29 Mode: othervm 07:03:29 Additional options from @modules: --add-modules java.desktop 07:03:29 elapsed time (seconds): 0.349 07:03:29 configuration: 07:03:29 Boot Layer 07:03:29 add modules: java.desktop 07:03:29 07:03:29 STDOUT: 07:03:29 STDERR: 07:03:29 java.awt.AWTError: WindowServer is not available 07:03:29 at java.desktop/sun.lwawt.macosx.LWCToolkit.<clinit>(LWCToolkit.java:167) 07:03:29 at java.base/java.lang.Class.forNameImpl(Native Method) 07:03:29 at java.base/java.lang.Class.forName(Class.java:337) 07:03:29 at java.desktop/java.awt.Toolkit$2.run(Toolkit.java:588) 07:03:29 at java.desktop/java.awt.Toolkit$2.run(Toolkit.java:583) 07:03:29 at java.base/java.security.AccessController.doPrivileged(AccessController.java:682) 07:03:29 at java.desktop/java.awt.Toolkit.getDefaultToolkit(Toolkit.java:582) 07:03:29 at java.desktop/sun.awt.CGraphicsEnvironment.<clinit>(CGraphicsEnvironment.java:73) 07:03:29 at java.base/java.lang.Class.forNameImpl(Native Method) 07:03:29 at java.base/java.lang.Class.forName(Class.java:337) 07:03:29 at java.desktop/java.awt.GraphicsEnvironment$LocalGE.createGE(GraphicsEnvironment.java:101) 07:03:29 at java.desktop/java.awt.GraphicsEnvironment$LocalGE.<clinit>(GraphicsEnvironment.java:83) 07:03:29 at java.desktop/java.awt.GraphicsEnvironment.getLocalGraphicsEnvironment(GraphicsEnvironment.java:129) 07:03:29 at java.desktop/java.awt.image.BufferedImage.createGraphics(BufferedImage.java:1181) 07:03:29 at WbmpBigDestinationTest.main(WbmpBigDestinationTest.java:51) 07:03:29 at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 07:03:29 at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 07:03:29 at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) 07:03:29 at java.base/java.lang.reflect.Method.invoke(Method.java:566) 07:03:29 at com.sun.javatest.regtest.agent.MainWrapper$MainThread.run(MainWrapper.java:127) 07:03:29 at java.base/java.lang.Thread.run(Thread.java:836) 07:03:29 07:03:29 JavaTest Message: Test threw exception: java.awt.AWTError: WindowServer is not available 07:03:29 JavaTest Message: shutting down test 07:03:29 07:03:29 STATUS:Failed.`main' threw exception: java.awt.AWTError: WindowServer is not available 07:03:58 -------------------------------------------------- 07:04:27 Test results: passed: 122; failed: 68 07:04:37 Report written to /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/jvmtest/openjdk/report/html/report.html 07:04:37 Results written to /Users/jenkins/workspace/Test_openjdk11_j9_extended.openjdk_x86-64_mac/openjdk-tests/TKG/output_1614422039661/jdk_imageio_0/work 07:04:37 Error: Some tests failed or other problems occurred. 07:04:37 07:04:37 jdk_imageio_0_FAILED ``` It appears related to the test setup.
test
macos jdk imageio failed awterror windowserver is not available others failure link from an internal build test extended openjdk mac openjdk version openjdk runtime environment adoptopenjdk build eclipse vm adoptopenjdk build master jre mac os x bit compressed references jit enabled aot enabled omr jcl based on jdk rebuild the failed tests in one link optional info failure output captured from console output test javax imageio plugins wbmp wbmpbigdestinationtest java test jdk users jenkins workspace test extended openjdk mac openjdkbinary image contents home bin action build passed build successful reason named class compiled on demand time seconds messages command build wbmpbigdestinationtest reason named class compiled on demand test directory compile wbmpbigdestinationtest elapsed time seconds action compile passed compilation successful reason class file out of date or does not exist time seconds messages command compile users jenkins workspace test extended openjdk mac openjdk tests openjdk openjdk jdk test jdk javax imageio plugins wbmp wbmpbigdestinationtest java reason class file out of date or does not exist additional options from modules add modules java desktop mode othervm elapsed time seconds configuration javac compilation environment add modules java desktop source path users jenkins workspace test extended openjdk mac openjdk tests openjdk openjdk jdk test jdk javax imageio plugins wbmp class path users jenkins workspace test extended openjdk mac openjdk tests openjdk openjdk jdk test jdk javax imageio plugins wbmp users jenkins workspace test extended openjdk mac openjdk tests tkg output jdk imageio work classes javax imageio plugins wbmp wbmpbigdestinationtest d rerun cd users jenkins workspace test extended openjdk mac openjdk tests tkg output jdk imageio work javax imageio plugins wbmp wbmpbigdestinationtest home users jenkins path bin usr bin usr sbin users jenkins workspace test extended openjdk mac openjdkbinary image contents home bin bin javac j ea j esa j j xx usecompressedoops j dtest vm opts ea esa xx usecompressedoops j dtest tool vm opts j ea j esa j j xx usecompressedoops j dtest compiler opts j dtest java opts j dtest jdk users jenkins workspace test extended openjdk mac openjdkbinary image contents home bin j dcompile jdk users jenkins workspace test extended openjdk mac openjdkbinary image contents home bin j dtest timeout factor j dtest root users jenkins workspace test extended openjdk mac openjdk tests openjdk openjdk jdk test jdk j dtest name javax imageio plugins wbmp wbmpbigdestinationtest java j dtest file users jenkins workspace test extended openjdk mac openjdk tests openjdk openjdk jdk test jdk javax imageio plugins wbmp wbmpbigdestinationtest java j dtest src users jenkins workspace test extended openjdk mac openjdk tests openjdk openjdk jdk test jdk javax imageio plugins wbmp j dtest src path users jenkins workspace test extended openjdk mac openjdk tests openjdk openjdk jdk test jdk javax imageio plugins wbmp j dtest classes users jenkins workspace test extended openjdk mac openjdk tests tkg output jdk imageio work classes javax imageio plugins wbmp wbmpbigdestinationtest d j dtest class path users jenkins workspace test extended openjdk mac openjdk tests tkg output jdk imageio work classes javax imageio plugins wbmp wbmpbigdestinationtest d j dtest modules java desktop add modules java desktop d users jenkins workspace test extended openjdk mac openjdk tests tkg output jdk imageio work classes javax imageio plugins wbmp wbmpbigdestinationtest d sourcepath users jenkins workspace test extended openjdk mac openjdk tests openjdk openjdk jdk test jdk javax imageio plugins wbmp classpath users jenkins workspace test extended openjdk mac openjdk tests openjdk openjdk jdk test jdk javax imageio plugins wbmp users jenkins workspace test extended openjdk mac openjdk tests tkg output jdk imageio work classes javax imageio plugins wbmp wbmpbigdestinationtest d users jenkins workspace test extended openjdk mac openjdk tests openjdk openjdk jdk test jdk javax imageio plugins wbmp wbmpbigdestinationtest java stdout stderr action main failed execution failed main threw exception java awt awterror windowserver is not available reason assumed action based on file name run main wbmpbigdestinationtest time seconds messages command main wbmpbigdestinationtest reason assumed action based on file name run main wbmpbigdestinationtest mode othervm additional options from modules add modules java desktop elapsed time seconds configuration boot layer add modules java desktop stdout stderr java awt awterror windowserver is not available at java desktop sun lwawt macosx lwctoolkit lwctoolkit java at java base java lang class fornameimpl native method at java base java lang class forname class java at java desktop java awt toolkit run toolkit java at java desktop java awt toolkit run toolkit java at java base java security accesscontroller doprivileged accesscontroller java at java desktop java awt toolkit getdefaulttoolkit toolkit java at java desktop sun awt cgraphicsenvironment cgraphicsenvironment java at java base java lang class fornameimpl native method at java base java lang class forname class java at java desktop java awt graphicsenvironment localge createge graphicsenvironment java at java desktop java awt graphicsenvironment localge graphicsenvironment java at java desktop java awt graphicsenvironment getlocalgraphicsenvironment graphicsenvironment java at java desktop java awt image bufferedimage creategraphics bufferedimage java at wbmpbigdestinationtest main wbmpbigdestinationtest java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at com sun javatest regtest agent mainwrapper mainthread run mainwrapper java at java base java lang thread run thread java javatest message test threw exception java awt awterror windowserver is not available javatest message shutting down test status failed main threw exception java awt awterror windowserver is not available test results passed failed report written to users jenkins workspace test extended openjdk mac jvmtest openjdk report html report html results written to users jenkins workspace test extended openjdk mac openjdk tests tkg output jdk imageio work error some tests failed or other problems occurred jdk imageio failed it appears related to the test setup
1
145,288
11,683,459,254
IssuesEvent
2020-03-05 03:27:35
creativecommons/cc-chooser
https://api.github.com/repos/creativecommons/cc-chooser
opened
Add unit and e2e tests for the LicenseCopy component
good first issue help wanted test-coverage
Unit and e2e tests need to be written for the LicenseCopy component. Unit tests are done with [Jest](https://jestjs.io/), and e2e tests are done with [nightwatch](https://nightwatchjs.org/). Please remember to test the following things: - That individual parts of the component are present when appropriate. (unit and e2e) - That any computed props and methods work properly, if there are any. (unit) - Any common interactions between the user and component, if there are any. (e2e) - Any other functionality unique to the component being tested! ### Additional Context - [./src/components/LicenseCopy.vue](https://github.com/creativecommons/cc-chooser/blob/master/src/components/LicenseCopy.vue) - [This repo's testing README](https://github.com/creativecommons/cc-chooser/blob/master/tests/README.md) - [Vue's guide on unit testing](https://vuejs.org/v2/guide/unit-testing.html) - [Vue's guide on unit testing with VueX](https://vue-test-utils.vuejs.org/guides/using-with-vuex.html)
1.0
Add unit and e2e tests for the LicenseCopy component - Unit and e2e tests need to be written for the LicenseCopy component. Unit tests are done with [Jest](https://jestjs.io/), and e2e tests are done with [nightwatch](https://nightwatchjs.org/). Please remember to test the following things: - That individual parts of the component are present when appropriate. (unit and e2e) - That any computed props and methods work properly, if there are any. (unit) - Any common interactions between the user and component, if there are any. (e2e) - Any other functionality unique to the component being tested! ### Additional Context - [./src/components/LicenseCopy.vue](https://github.com/creativecommons/cc-chooser/blob/master/src/components/LicenseCopy.vue) - [This repo's testing README](https://github.com/creativecommons/cc-chooser/blob/master/tests/README.md) - [Vue's guide on unit testing](https://vuejs.org/v2/guide/unit-testing.html) - [Vue's guide on unit testing with VueX](https://vue-test-utils.vuejs.org/guides/using-with-vuex.html)
test
add unit and tests for the licensecopy component unit and tests need to be written for the licensecopy component unit tests are done with and tests are done with please remember to test the following things that individual parts of the component are present when appropriate unit and that any computed props and methods work properly if there are any unit any common interactions between the user and component if there are any any other functionality unique to the component being tested additional context
1
75,471
9,855,940,053
IssuesEvent
2019-06-19 20:42:24
ofpinewood/http-exceptions
https://api.github.com/repos/ofpinewood/http-exceptions
closed
Update documentation and sample project
documentation
Update the documentation for the projects and add more sample code to the sample project.
1.0
Update documentation and sample project - Update the documentation for the projects and add more sample code to the sample project.
non_test
update documentation and sample project update the documentation for the projects and add more sample code to the sample project
0
349,467
31,806,832,428
IssuesEvent
2023-09-13 14:24:43
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: cdc/pubsub-sink failed
C-test-failure O-robot O-roachtest branch-master release-blocker T-cdc
roachtest.cdc/pubsub-sink [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/11734739?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/11734739?buildTab=artifacts#/cdc/pubsub-sink) on master @ [0310d119569b59b9efe1b7bbc439c2c604217a4c](https://github.com/cockroachdb/cockroach/commits/0310d119569b59b9efe1b7bbc439c2c604217a4c): ``` (latency_verifier.go:192).assertValid: latency never dropped to acceptable steady level: 1m0s (monitor.go:153).Wait: monitor failure: monitor user task failed: t.Fatal() was called test artifacts and logs in: /artifacts/cdc/pubsub-sink/run_1 ``` <p>Parameters: <code>ROACHTEST_arch=amd64</code> , <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=16</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_fs=ext4</code> , <code>ROACHTEST_localSSD=true</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) See: [Grafana](https://go.crdb.dev/roachtest-grafana/teamcity-11734739/cdc-pubsub-sink/1694515846448/1694517902058) </p> </details> /cc @cockroachdb/cdc <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*cdc/pubsub-sink.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-31422 Epic CRDB-11732
2.0
roachtest: cdc/pubsub-sink failed - roachtest.cdc/pubsub-sink [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/11734739?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/11734739?buildTab=artifacts#/cdc/pubsub-sink) on master @ [0310d119569b59b9efe1b7bbc439c2c604217a4c](https://github.com/cockroachdb/cockroach/commits/0310d119569b59b9efe1b7bbc439c2c604217a4c): ``` (latency_verifier.go:192).assertValid: latency never dropped to acceptable steady level: 1m0s (monitor.go:153).Wait: monitor failure: monitor user task failed: t.Fatal() was called test artifacts and logs in: /artifacts/cdc/pubsub-sink/run_1 ``` <p>Parameters: <code>ROACHTEST_arch=amd64</code> , <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=16</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_fs=ext4</code> , <code>ROACHTEST_localSSD=true</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) See: [Grafana](https://go.crdb.dev/roachtest-grafana/teamcity-11734739/cdc-pubsub-sink/1694515846448/1694517902058) </p> </details> /cc @cockroachdb/cdc <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*cdc/pubsub-sink.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-31422 Epic CRDB-11732
test
roachtest cdc pubsub sink failed roachtest cdc pubsub sink with on master latency verifier go assertvalid latency never dropped to acceptable steady level monitor go wait monitor failure monitor user task failed t fatal was called test artifacts and logs in artifacts cdc pubsub sink run parameters roachtest arch roachtest cloud gce roachtest cpu roachtest encrypted false roachtest fs roachtest localssd true roachtest ssd help see see see cc cockroachdb cdc jira issue crdb epic crdb
1
292,469
25,216,141,519
IssuesEvent
2022-11-14 09:18:07
Joystream/pioneer
https://api.github.com/repos/Joystream/pioneer
closed
Set Working Group Lead Reward
enhancement scope:proposals qa-task qa-tests-failed
[ ] Set Working Group Lead Reward - Same as when updating reward of a worker in group with given inputs, except signer check
1.0
Set Working Group Lead Reward - [ ] Set Working Group Lead Reward - Same as when updating reward of a worker in group with given inputs, except signer check
test
set working group lead reward set working group lead reward same as when updating reward of a worker in group with given inputs except signer check
1
568,767
16,988,328,393
IssuesEvent
2021-06-30 16:55:03
npm/cli
https://api.github.com/repos/npm/cli
closed
[BUG] progress=false is ignored on npm 7
Bug Priority 2 Release 7.x
<!-- Note: Please search to see if an issue already exists for your problem: https://github.com/npm/cli/issues --> ### Current Behavior: Have an .npmrc with progress=false for aesthetic reasons. On npm6 that disables the progress bar, on npm7 it does not although the docs say it still should. This is my full .npmrc if that helps ``` loglevel=http progress=false package-lock=false save=false ``` ### Expected Behavior: Expected the above to disable the progress bar like it does on 6 ### Steps To Reproduce: Save the .npmrc in home and install something. Progress bar is still shown. ### Environment: - OS: macOS big sur - Node: 15.7.0 - npm: 7.4.3
1.0
[BUG] progress=false is ignored on npm 7 - <!-- Note: Please search to see if an issue already exists for your problem: https://github.com/npm/cli/issues --> ### Current Behavior: Have an .npmrc with progress=false for aesthetic reasons. On npm6 that disables the progress bar, on npm7 it does not although the docs say it still should. This is my full .npmrc if that helps ``` loglevel=http progress=false package-lock=false save=false ``` ### Expected Behavior: Expected the above to disable the progress bar like it does on 6 ### Steps To Reproduce: Save the .npmrc in home and install something. Progress bar is still shown. ### Environment: - OS: macOS big sur - Node: 15.7.0 - npm: 7.4.3
non_test
progress false is ignored on npm note please search to see if an issue already exists for your problem current behavior have an npmrc with progress false for aesthetic reasons on that disables the progress bar on it does not although the docs say it still should this is my full npmrc if that helps loglevel http progress false package lock false save false expected behavior expected the above to disable the progress bar like it does on steps to reproduce save the npmrc in home and install something progress bar is still shown environment os macos big sur node npm
0
321,146
27,508,935,704
IssuesEvent
2023-03-06 07:07:08
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
ccl/serverccl: TestServerControllerHTTP failed
C-test-failure O-robot branch-master
ccl/serverccl.TestServerControllerHTTP [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/8933299?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/8933299?buildTab=artifacts#/) on master @ [14b43be03c1c246765be17aa15245493a42c401d](https://github.com/cockroachdb/cockroach/commits/14b43be03c1c246765be17aa15245493a42c401d): ``` === RUN TestServerControllerHTTP test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/c101b7a464a1afc1f5af0cd85792187e/logTestServerControllerHTTP452427513 test_log_scope.go:79: use -show-logs to present logs inline server_controller_test.go:67: Error Trace: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3439/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/ccl/serverccl/serverccl_test_/serverccl_test.runfiles/com_github_cockroachdb_cockroach/pkg/ccl/serverccl/server_controller_test.go:67 Error: Received unexpected error: server stop before start (1) attached stack trace -- stack trace: | github.com/cockroachdb/cockroach/pkg/server.(*serverController).startControlledServer.func3.1 | github.com/cockroachdb/cockroach/pkg/server/server_controller_orchestration.go:257 | github.com/cockroachdb/cockroach/pkg/server.(*serverController).startControlledServer.func3 | github.com/cockroachdb/cockroach/pkg/server/server_controller_orchestration.go:306 | github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTaskEx.func2 | github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:470 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1594 Wraps: (2) server stop before start Error types: (1) *withstack.withStack (2) *errutil.leafError Test: TestServerControllerHTTP panic.go:522: -- test log scope end -- test logs left over in: /artifacts/tmp/_tmp/c101b7a464a1afc1f5af0cd85792187e/logTestServerControllerHTTP452427513 --- FAIL: TestServerControllerHTTP (3.77s) ``` <p>Parameters: <code>TAGS=bazel,gss,deadlock</code> </p> <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/server <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestServerControllerHTTP.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
1.0
ccl/serverccl: TestServerControllerHTTP failed - ccl/serverccl.TestServerControllerHTTP [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/8933299?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/8933299?buildTab=artifacts#/) on master @ [14b43be03c1c246765be17aa15245493a42c401d](https://github.com/cockroachdb/cockroach/commits/14b43be03c1c246765be17aa15245493a42c401d): ``` === RUN TestServerControllerHTTP test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/c101b7a464a1afc1f5af0cd85792187e/logTestServerControllerHTTP452427513 test_log_scope.go:79: use -show-logs to present logs inline server_controller_test.go:67: Error Trace: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3439/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/ccl/serverccl/serverccl_test_/serverccl_test.runfiles/com_github_cockroachdb_cockroach/pkg/ccl/serverccl/server_controller_test.go:67 Error: Received unexpected error: server stop before start (1) attached stack trace -- stack trace: | github.com/cockroachdb/cockroach/pkg/server.(*serverController).startControlledServer.func3.1 | github.com/cockroachdb/cockroach/pkg/server/server_controller_orchestration.go:257 | github.com/cockroachdb/cockroach/pkg/server.(*serverController).startControlledServer.func3 | github.com/cockroachdb/cockroach/pkg/server/server_controller_orchestration.go:306 | github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTaskEx.func2 | github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:470 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1594 Wraps: (2) server stop before start Error types: (1) *withstack.withStack (2) *errutil.leafError Test: TestServerControllerHTTP panic.go:522: -- test log scope end -- test logs left over in: /artifacts/tmp/_tmp/c101b7a464a1afc1f5af0cd85792187e/logTestServerControllerHTTP452427513 --- FAIL: TestServerControllerHTTP (3.77s) ``` <p>Parameters: <code>TAGS=bazel,gss,deadlock</code> </p> <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/server <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestServerControllerHTTP.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
ccl serverccl testservercontrollerhttp failed ccl serverccl testservercontrollerhttp with on master run testservercontrollerhttp test log scope go test logs captured to artifacts tmp tmp test log scope go use show logs to present logs inline server controller test go error trace home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg ccl serverccl serverccl test serverccl test runfiles com github cockroachdb cockroach pkg ccl serverccl server controller test go error received unexpected error server stop before start attached stack trace stack trace github com cockroachdb cockroach pkg server servercontroller startcontrolledserver github com cockroachdb cockroach pkg server server controller orchestration go github com cockroachdb cockroach pkg server servercontroller startcontrolledserver github com cockroachdb cockroach pkg server server controller orchestration go github com cockroachdb cockroach pkg util stop stopper runasynctaskex github com cockroachdb cockroach pkg util stop stopper go runtime goexit goroot src runtime asm s wraps server stop before start error types withstack withstack errutil leaferror test testservercontrollerhttp panic go test log scope end test logs left over in artifacts tmp tmp fail testservercontrollerhttp parameters tags bazel gss deadlock help see also cc cockroachdb server
1
105,759
9,100,592,774
IssuesEvent
2019-02-20 08:58:48
intermine/intermine
https://api.github.com/repos/intermine/intermine
closed
OMIM - failed on bad line
data in-progress please-test
``` Caused by: java.lang.RuntimeException: bad line: '[Beta-glycopyranoside tasting], 617956 (3) {Alcohol dependence, susceptibility to}, 103780 (3) TAS2R16, T2R16, BGLPT 604867 7q31.32' at org.intermine.bio.dataconversion.OmimConverter.processMorbidMapFile(OmimConverter.java:168) at org.intermine.bio.dataconversion.OmimConverter.process(OmimConverter.java:101) at org.intermine.task.DirectoryConverterTask.execute(DirectoryConverterTask.java:104) ... 42 more ```
1.0
OMIM - failed on bad line - ``` Caused by: java.lang.RuntimeException: bad line: '[Beta-glycopyranoside tasting], 617956 (3) {Alcohol dependence, susceptibility to}, 103780 (3) TAS2R16, T2R16, BGLPT 604867 7q31.32' at org.intermine.bio.dataconversion.OmimConverter.processMorbidMapFile(OmimConverter.java:168) at org.intermine.bio.dataconversion.OmimConverter.process(OmimConverter.java:101) at org.intermine.task.DirectoryConverterTask.execute(DirectoryConverterTask.java:104) ... 42 more ```
test
omim failed on bad line caused by java lang runtimeexception bad line alcohol dependence susceptibility to bglpt at org intermine bio dataconversion omimconverter processmorbidmapfile omimconverter java at org intermine bio dataconversion omimconverter process omimconverter java at org intermine task directoryconvertertask execute directoryconvertertask java more
1
93,021
8,391,895,924
IssuesEvent
2018-10-09 16:05:18
kabirbaidhya/boss
https://api.github.com/repos/kabirbaidhya/boss
closed
Write tests for boss api and utils
good first issue hacktoberfest help wanted test
Write tests for: * Modules and functions under `boss.api` * Code under `boss.util` * Any other code that requires tests
1.0
Write tests for boss api and utils - Write tests for: * Modules and functions under `boss.api` * Code under `boss.util` * Any other code that requires tests
test
write tests for boss api and utils write tests for modules and functions under boss api code under boss util any other code that requires tests
1
487,676
14,050,375,076
IssuesEvent
2020-11-02 11:39:48
drashland/dmm
https://api.github.com/repos/drashland/dmm
opened
Support https://raw.githubusercontent.com
Priority: Low Type: Enhancement
## Summary What: Alongside supporting deno.land and x.nest.land, add support for https://raw.githubusercontent Why: Mainly, so we can use dmm to update dependencies pulled from the services repo, but it does have it's own good use ## Acceptance Criteria Below is a list of tasks that must be completed before this issue can be closed. - [ ] Documentation (link) - [ ] Tests - [ ] The feature ## Example Pseudo Code (for implementation) ```typescript // Add example pseudo code for implementation ```
1.0
Support https://raw.githubusercontent.com - ## Summary What: Alongside supporting deno.land and x.nest.land, add support for https://raw.githubusercontent Why: Mainly, so we can use dmm to update dependencies pulled from the services repo, but it does have it's own good use ## Acceptance Criteria Below is a list of tasks that must be completed before this issue can be closed. - [ ] Documentation (link) - [ ] Tests - [ ] The feature ## Example Pseudo Code (for implementation) ```typescript // Add example pseudo code for implementation ```
non_test
support summary what alongside supporting deno land and x nest land add support for why mainly so we can use dmm to update dependencies pulled from the services repo but it does have it s own good use acceptance criteria below is a list of tasks that must be completed before this issue can be closed documentation link tests the feature example pseudo code for implementation typescript add example pseudo code for implementation
0
147,712
23,260,569,060
IssuesEvent
2022-08-04 13:11:33
excalidraw/excalidraw
https://api.github.com/repos/excalidraw/excalidraw
closed
Flip horizontal does not work on linear elements since the redesign (#5501). Flip vertical works.
bug arrow-redesign
ERROR: type should be string, got "\r\nhttps://user-images.githubusercontent.com/14358394/182805730-1ad74bca-da46-4383-b03b-6d89f2a68ae8.mp4\r\n\r\n"
1.0
Flip horizontal does not work on linear elements since the redesign (#5501). Flip vertical works. - https://user-images.githubusercontent.com/14358394/182805730-1ad74bca-da46-4383-b03b-6d89f2a68ae8.mp4
non_test
flip horizontal does not work on linear elements since the redesign flip vertical works
0
293,471
22,059,411,895
IssuesEvent
2022-05-30 15:49:52
ms-club-sliit/ms-meeting-manager
https://api.github.com/repos/ms-club-sliit/ms-meeting-manager
opened
README for the repository
documentation help wanted
Need to create a Readme file for this project. The readme file should contains the following information. * Technologies that used in the project * How to run the project * How to contribute the project * Recent contributors * CI/ CD pipeline status
1.0
README for the repository - Need to create a Readme file for this project. The readme file should contains the following information. * Technologies that used in the project * How to run the project * How to contribute the project * Recent contributors * CI/ CD pipeline status
non_test
readme for the repository need to create a readme file for this project the readme file should contains the following information technologies that used in the project how to run the project how to contribute the project recent contributors ci cd pipeline status
0
120,167
12,060,612,333
IssuesEvent
2020-04-15 21:34:48
deathlyrage/pot-demo-bugs
https://api.github.com/repos/deathlyrage/pot-demo-bugs
closed
Camera spawns in Trees (stuck) and Stuck Daspleto cannot move
documentation needs testing
Spawned in to try AI mode and find my camera unable to move after having been spawned inside the crown of a tree ![unknown](https://user-images.githubusercontent.com/55845529/65700464-0bce0e80-e080-11e9-9343-7dfa0bb69829.png) as well as daspletosaurus being stuck between a rock and a tree the moment it spawned in ![daspleto](https://user-images.githubusercontent.com/55845529/65700630-42a42480-e080-11e9-8344-3c9581f12d69.png)
1.0
Camera spawns in Trees (stuck) and Stuck Daspleto cannot move - Spawned in to try AI mode and find my camera unable to move after having been spawned inside the crown of a tree ![unknown](https://user-images.githubusercontent.com/55845529/65700464-0bce0e80-e080-11e9-9343-7dfa0bb69829.png) as well as daspletosaurus being stuck between a rock and a tree the moment it spawned in ![daspleto](https://user-images.githubusercontent.com/55845529/65700630-42a42480-e080-11e9-8344-3c9581f12d69.png)
non_test
camera spawns in trees stuck and stuck daspleto cannot move spawned in to try ai mode and find my camera unable to move after having been spawned inside the crown of a tree as well as daspletosaurus being stuck between a rock and a tree the moment it spawned in
0
811,365
30,285,275,602
IssuesEvent
2023-07-08 15:43:33
SuffolkLITLab/ALKiln
https://api.github.com/repos/SuffolkLITLab/ALKiln
closed
Restore multi-language tests set by env vars
priority
As per user request, restore running language tests and write internal tests for such (though I'm not sure how we check that the right tests have been run). These should primarily be triggered manually through the workflow dispatch, though allowing an env var for them might still be useful to carry over from previous behavior. Include: - [x] Access language page url as per #360 - [x] Add language to Scenario tags - [x] Internal tests of some kind, maybe using tags in Scenario name section. Check that certain text appears multiple times won't be sufficient as errors can also cause that behavior. [Maybe create our own translation file with different text that we can then look for. Use [MADE](https://github.com/GBLS/docassemble-MAEvictionDefense)'s methods as a model for that as that is our primary user need at the moment?] - [x] ~[If we're creating the language files each time] Ensure InThePlayground deletes the language files during post-test cleanup.~ Should close #360. Edit: - [ ] Add [new issue to document this in the documentation repo](https://github.com/SuffolkLITLab/docassemble-AssemblyLine-documentation/issues/new)
1.0
Restore multi-language tests set by env vars - As per user request, restore running language tests and write internal tests for such (though I'm not sure how we check that the right tests have been run). These should primarily be triggered manually through the workflow dispatch, though allowing an env var for them might still be useful to carry over from previous behavior. Include: - [x] Access language page url as per #360 - [x] Add language to Scenario tags - [x] Internal tests of some kind, maybe using tags in Scenario name section. Check that certain text appears multiple times won't be sufficient as errors can also cause that behavior. [Maybe create our own translation file with different text that we can then look for. Use [MADE](https://github.com/GBLS/docassemble-MAEvictionDefense)'s methods as a model for that as that is our primary user need at the moment?] - [x] ~[If we're creating the language files each time] Ensure InThePlayground deletes the language files during post-test cleanup.~ Should close #360. Edit: - [ ] Add [new issue to document this in the documentation repo](https://github.com/SuffolkLITLab/docassemble-AssemblyLine-documentation/issues/new)
non_test
restore multi language tests set by env vars as per user request restore running language tests and write internal tests for such though i m not sure how we check that the right tests have been run these should primarily be triggered manually through the workflow dispatch though allowing an env var for them might still be useful to carry over from previous behavior include access language page url as per add language to scenario tags internal tests of some kind maybe using tags in scenario name section check that certain text appears multiple times won t be sufficient as errors can also cause that behavior methods as a model for that as that is our primary user need at the moment ensure intheplayground deletes the language files during post test cleanup should close edit add
0
821,730
30,833,468,475
IssuesEvent
2023-08-02 05:02:22
GSM-MSG/SMS-Android
https://api.github.com/repos/GSM-MSG/SMS-Android
opened
Show the snack bar when a screen capture is detected.
0️⃣ Priority: Critical ✨ Feature
### Describe 화면 캡쳐가 감지되었을 때 스낵바를 표시합니다. ### Additional _No response_
1.0
Show the snack bar when a screen capture is detected. - ### Describe 화면 캡쳐가 감지되었을 때 스낵바를 표시합니다. ### Additional _No response_
non_test
show the snack bar when a screen capture is detected describe 화면 캡쳐가 감지되었을 때 스낵바를 표시합니다 additional no response
0
610,596
18,911,904,600
IssuesEvent
2021-11-16 14:53:38
googleapis/google-api-dotnet-client
https://api.github.com/repos/googleapis/google-api-dotnet-client
closed
How to Create Custom Dimension and Custom metrics in google analytics
type: question priority: p2 api: analytics
We have tried below steps to create Custom Dimension and Custom metrics from code side. 1) We have installed Nuget package "Install-Package Google.Apis.Analytics.v3 -Version 1.55.0.1679" into solution and Enabled Google Analytics API's in "Google Analytics Account". 2) We have created service account for google analytics project and Generated ".p12" file for google analytic account authentication and programmatically we are validating the .p12 file with EmailId. Please find the below code. =================================================================== var certificate = new X509Certificate2(keyPath, "notasecret", X509KeyStorageFlags.Exportable); var credentials = new ServiceAccountCredential( new ServiceAccountCredential.Initializer(emailID) { Scopes = new[] { AnalyticsService.Scope.AnalyticsEdit} }.FromCertificate(certificate)); var service = new AnalyticsService(new BaseClientService.Initializer() { HttpClientInitializer = credentials, //ApplicationName = "Analytics API Sample", }); Note: We have tried with Analytics, AnalyticsEdit, and AnalyticsManageUsers Scope. =================================================================== 3) We are creating custom dimension using below code. public static string CreateCustomDimension(AnalyticsService ga, string accountId, string webPropertyId, string dimensionName, string scope, bool active) { List customDimensions = ga.Management.CustomDimensions.List(accountId, webPropertyId).Execute().Items as List; if (!customDimensions.Select(c => c.Name).Contains(dimensionName)) { CustomDimension body = new CustomDimension { AccountId = accountId, WebPropertyId = webPropertyId, Name = dimensionName, Scope = scope, Active = active }; ManagementResource.CustomDimensionsResource.InsertRequest request = ga.Management.CustomDimensions.Insert(body, accountId, webPropertyId); CustomDimension newDimension = request.Execute(); List<CustomDimension> newDimensions = ga.Management.CustomDimensions.List(accountId, webPropertyId).Execute().Items as List<CustomDimension>; if (newDimensions.Select(d => d.Id).Contains(newDimension.Id)) return "Success: Custom Dimension Added"; else return "Failure: Unknown"; } else { return "Custom Dimension Already Exists"; } } we are getting exception at [ga.Management.CustomDimensions.List(accountId, webPropertyId).Execute().Items as List;] in above method. Exception Details: Request had insufficient authentication scopes. [403] Errors [Message[Insufficient Permission] Location[ - ] Reason[insufficientPermissions] Domain[global]]
1.0
How to Create Custom Dimension and Custom metrics in google analytics - We have tried below steps to create Custom Dimension and Custom metrics from code side. 1) We have installed Nuget package "Install-Package Google.Apis.Analytics.v3 -Version 1.55.0.1679" into solution and Enabled Google Analytics API's in "Google Analytics Account". 2) We have created service account for google analytics project and Generated ".p12" file for google analytic account authentication and programmatically we are validating the .p12 file with EmailId. Please find the below code. =================================================================== var certificate = new X509Certificate2(keyPath, "notasecret", X509KeyStorageFlags.Exportable); var credentials = new ServiceAccountCredential( new ServiceAccountCredential.Initializer(emailID) { Scopes = new[] { AnalyticsService.Scope.AnalyticsEdit} }.FromCertificate(certificate)); var service = new AnalyticsService(new BaseClientService.Initializer() { HttpClientInitializer = credentials, //ApplicationName = "Analytics API Sample", }); Note: We have tried with Analytics, AnalyticsEdit, and AnalyticsManageUsers Scope. =================================================================== 3) We are creating custom dimension using below code. public static string CreateCustomDimension(AnalyticsService ga, string accountId, string webPropertyId, string dimensionName, string scope, bool active) { List customDimensions = ga.Management.CustomDimensions.List(accountId, webPropertyId).Execute().Items as List; if (!customDimensions.Select(c => c.Name).Contains(dimensionName)) { CustomDimension body = new CustomDimension { AccountId = accountId, WebPropertyId = webPropertyId, Name = dimensionName, Scope = scope, Active = active }; ManagementResource.CustomDimensionsResource.InsertRequest request = ga.Management.CustomDimensions.Insert(body, accountId, webPropertyId); CustomDimension newDimension = request.Execute(); List<CustomDimension> newDimensions = ga.Management.CustomDimensions.List(accountId, webPropertyId).Execute().Items as List<CustomDimension>; if (newDimensions.Select(d => d.Id).Contains(newDimension.Id)) return "Success: Custom Dimension Added"; else return "Failure: Unknown"; } else { return "Custom Dimension Already Exists"; } } we are getting exception at [ga.Management.CustomDimensions.List(accountId, webPropertyId).Execute().Items as List;] in above method. Exception Details: Request had insufficient authentication scopes. [403] Errors [Message[Insufficient Permission] Location[ - ] Reason[insufficientPermissions] Domain[global]]
non_test
how to create custom dimension and custom metrics in google analytics we have tried below steps to create custom dimension and custom metrics from code side we have installed nuget package install package google apis analytics version into solution and enabled google analytics api s in google analytics account we have created service account for google analytics project and generated file for google analytic account authentication and programmatically we are validating the file with emailid please find the below code var certificate new keypath notasecret exportable var credentials new serviceaccountcredential new serviceaccountcredential initializer emailid scopes new analyticsservice scope analyticsedit fromcertificate certificate var service new analyticsservice new baseclientservice initializer httpclientinitializer credentials applicationname analytics api sample note we have tried with analytics analyticsedit and analyticsmanageusers scope we are creating custom dimension using below code public static string createcustomdimension analyticsservice ga string accountid string webpropertyid string dimensionname string scope bool active list customdimensions ga management customdimensions list accountid webpropertyid execute items as list if customdimensions select c c name contains dimensionname customdimension body new customdimension accountid accountid webpropertyid webpropertyid name dimensionname scope scope active active managementresource customdimensionsresource insertrequest request ga management customdimensions insert body accountid webpropertyid customdimension newdimension request execute list newdimensions ga management customdimensions list accountid webpropertyid execute items as list if newdimensions select d d id contains newdimension id return success custom dimension added else return failure unknown else return custom dimension already exists we are getting exception at in above method exception details request had insufficient authentication scopes errors location reason domain
0
65,592
12,625,364,989
IssuesEvent
2020-06-14 11:32:14
intellij-rust/intellij-rust
https://api.github.com/repos/intellij-rust/intellij-rust
opened
No E0368/E0369 for binary operations
feature subsystem::code insight
<!-- Hello and thank you for the issue! If you would like to report a bug, we have added some points below that you can fill out. Feel free to remove all the irrelevant text to request a new feature. --> ## Environment * **IntelliJ Rust plugin version:** 0.2.125.3158-202-nightly * **Rust toolchain version:** 1.46.0-nightly (0262de554 2020-06-07) x86_64-apple-darwin * **IDE name and version:** IntelliJ IDEA 2020.2 EAP Ultimate Edition (IU-202.5428.22) * **Operating system:** macOS 10.13.6 ## Problem description If the compiler can't find a proper implementation of trait responsible for the particular binary operator, it produces an error. Currently, the plugin says nothing in such cases. ## Steps to reproduce ```rust struct Foo(u8); fn add(lhs: Foo, rhs: Foo) { lhs + rhs; // error[E0369]: cannot add `Foo` to `Foo` } fn munis_assign(mut lhs: Foo, rhs: Foo) { lhs -= rhs; // error[E0368]: binary assignment operation `-=` cannot be applied to type `Foo` } ``` <!-- Please include as much of your codebase as needed to reproduce the error. If the relevant files are large, please provide a link to a public repository or a [Gist](https://gist.github.com/). -->
1.0
No E0368/E0369 for binary operations - <!-- Hello and thank you for the issue! If you would like to report a bug, we have added some points below that you can fill out. Feel free to remove all the irrelevant text to request a new feature. --> ## Environment * **IntelliJ Rust plugin version:** 0.2.125.3158-202-nightly * **Rust toolchain version:** 1.46.0-nightly (0262de554 2020-06-07) x86_64-apple-darwin * **IDE name and version:** IntelliJ IDEA 2020.2 EAP Ultimate Edition (IU-202.5428.22) * **Operating system:** macOS 10.13.6 ## Problem description If the compiler can't find a proper implementation of trait responsible for the particular binary operator, it produces an error. Currently, the plugin says nothing in such cases. ## Steps to reproduce ```rust struct Foo(u8); fn add(lhs: Foo, rhs: Foo) { lhs + rhs; // error[E0369]: cannot add `Foo` to `Foo` } fn munis_assign(mut lhs: Foo, rhs: Foo) { lhs -= rhs; // error[E0368]: binary assignment operation `-=` cannot be applied to type `Foo` } ``` <!-- Please include as much of your codebase as needed to reproduce the error. If the relevant files are large, please provide a link to a public repository or a [Gist](https://gist.github.com/). -->
non_test
no for binary operations hello and thank you for the issue if you would like to report a bug we have added some points below that you can fill out feel free to remove all the irrelevant text to request a new feature environment intellij rust plugin version nightly rust toolchain version nightly apple darwin ide name and version intellij idea eap ultimate edition iu operating system macos problem description if the compiler can t find a proper implementation of trait responsible for the particular binary operator it produces an error currently the plugin says nothing in such cases steps to reproduce rust struct foo fn add lhs foo rhs foo lhs rhs error cannot add foo to foo fn munis assign mut lhs foo rhs foo lhs rhs error binary assignment operation cannot be applied to type foo please include as much of your codebase as needed to reproduce the error if the relevant files are large please provide a link to a public repository or a
0
20,861
6,114,254,682
IssuesEvent
2017-06-22 00:24:09
ganeti/ganeti
https://api.github.com/repos/ganeti/ganeti
closed
burnin instructions & possible bug
imported_from_google_code Status:Invalid
Originally reported of Google Code with ID 106. ``` What software version are you running? Please provide the output of "gnt- cluster --version" and "gnt-cluster version". <b>What distribution are you using?</b> gnt-cluster (ganeti) 2.1.1 software version 2.1.1 internode protocol: 30 configuration format: 201000 os api version: 15 export interface: 0 comments relate to latest debootstrap download (v 0.9) note: doing install from source, download from code.google.com onto Debian Lenny immediately post installation, following the instructions to run burnin (http://ganeti-doc.googlecode.com/svn/ganeti-2.1/html/admin.html#burnin-label) results in an error "can't get the OS list" turns out that the instructions in http://ganeti-doc.googlecode.com/svn/ganeti-2.1/html/install.html#installing-the-operating-system-support-packages say to simply do "make" followed by "make install" - which does not put the operating system in /srv/ganeti/os the instructions in the README file are more specific re. configure options to put files in the proper places following a proper config, make, make install: gnt-os diagnose shows: OS: debootstrap [global status: valid] but, burnin -o debootstrap test now responds with a new error: OS 'debootstrap' not found ``` Originally added on 2010-05-02 11:21:02 +0000 UTC.
1.0
burnin instructions & possible bug - Originally reported of Google Code with ID 106. ``` What software version are you running? Please provide the output of "gnt- cluster --version" and "gnt-cluster version". <b>What distribution are you using?</b> gnt-cluster (ganeti) 2.1.1 software version 2.1.1 internode protocol: 30 configuration format: 201000 os api version: 15 export interface: 0 comments relate to latest debootstrap download (v 0.9) note: doing install from source, download from code.google.com onto Debian Lenny immediately post installation, following the instructions to run burnin (http://ganeti-doc.googlecode.com/svn/ganeti-2.1/html/admin.html#burnin-label) results in an error "can't get the OS list" turns out that the instructions in http://ganeti-doc.googlecode.com/svn/ganeti-2.1/html/install.html#installing-the-operating-system-support-packages say to simply do "make" followed by "make install" - which does not put the operating system in /srv/ganeti/os the instructions in the README file are more specific re. configure options to put files in the proper places following a proper config, make, make install: gnt-os diagnose shows: OS: debootstrap [global status: valid] but, burnin -o debootstrap test now responds with a new error: OS 'debootstrap' not found ``` Originally added on 2010-05-02 11:21:02 +0000 UTC.
non_test
burnin instructions possible bug originally reported of google code with id what software version are you running please provide the output of gnt cluster version and gnt cluster version what distribution are you using gnt cluster ganeti software version internode protocol configuration format os api version export interface comments relate to latest debootstrap download v note doing install from source download from code google com onto debian lenny immediately post installation following the instructions to run burnin results in an error can t get the os list turns out that the instructions in say to simply do make followed by make install which does not put the operating system in srv ganeti os the instructions in the readme file are more specific re configure options to put files in the proper places following a proper config make make install gnt os diagnose shows os debootstrap but burnin o debootstrap test now responds with a new error os debootstrap not found originally added on utc
0
87,624
25,165,008,263
IssuesEvent
2022-11-10 20:00:17
libjxl/libjxl
https://api.github.com/repos/libjxl/libjxl
closed
StoreInterleaved: 2 3 4
building/portability unrelated to 1.0 highway
**Describe the bug** in order to compile `main` branch I had to comment out all lines with `StoreInterleaved2` `StoreInterleaved3` `StoreInterleaved4` (in `dec_group_jpeg.cc` and `stage_write.cc`) **To Reproduce** try to compile `main` branch **Expected behavior** `main` branch compiles successfully **Environment** - OS: Gentoo Linux - Compiler version: gcc-12.2.1 - CPU type: x86_64 - cjxl/djxl version string: JPEG XL encoder v0.8.0 [AVX2] **Additional context** `emerge =media-libs/libjxl-9999` with `-DJXL_HWY_DISABLED_TARGETS_FORCED:BOOL=ON`
1.0
StoreInterleaved: 2 3 4 - **Describe the bug** in order to compile `main` branch I had to comment out all lines with `StoreInterleaved2` `StoreInterleaved3` `StoreInterleaved4` (in `dec_group_jpeg.cc` and `stage_write.cc`) **To Reproduce** try to compile `main` branch **Expected behavior** `main` branch compiles successfully **Environment** - OS: Gentoo Linux - Compiler version: gcc-12.2.1 - CPU type: x86_64 - cjxl/djxl version string: JPEG XL encoder v0.8.0 [AVX2] **Additional context** `emerge =media-libs/libjxl-9999` with `-DJXL_HWY_DISABLED_TARGETS_FORCED:BOOL=ON`
non_test
storeinterleaved describe the bug in order to compile main branch i had to comment out all lines with in dec group jpeg cc and stage write cc to reproduce try to compile main branch expected behavior main branch compiles successfully environment os gentoo linux compiler version gcc cpu type cjxl djxl version string jpeg xl encoder additional context emerge media libs libjxl with djxl hwy disabled targets forced bool on
0
41,907
2,869,088,009
IssuesEvent
2015-06-05 23:14:12
dart-lang/polymer-dart
https://api.github.com/repos/dart-lang/polymer-dart
opened
Provide a way to easily trace async stack traces
bug Priority-Medium
<a href="https://github.com/sigmundch"><img src="https://avatars.githubusercontent.com/u/2049220?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [sigmundch](https://github.com/sigmundch)** _Originally opened as dart-lang/sdk#20322_ ---- maybe a toggle UI like the one we have for logger?
1.0
Provide a way to easily trace async stack traces - <a href="https://github.com/sigmundch"><img src="https://avatars.githubusercontent.com/u/2049220?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [sigmundch](https://github.com/sigmundch)** _Originally opened as dart-lang/sdk#20322_ ---- maybe a toggle UI like the one we have for logger?
non_test
provide a way to easily trace async stack traces issue by originally opened as dart lang sdk maybe a toggle ui like the one we have for logger
0
323,802
27,753,378,396
IssuesEvent
2023-03-15 23:04:15
sanktjodel/cctest1
https://api.github.com/repos/sanktjodel/cctest1
opened
Fix "method_complexity" issue in Signing.java
test1 test2'"><h1>tt
Method `signRequest` has a Cognitive Complexity of 34 (exceeds 5 allowed). Consider refactoring. https://codeclimate.com/github/sanktjodel/cctest1/Signing.java#issue_64124ec1a9c49c0001000012
2.0
Fix "method_complexity" issue in Signing.java - Method `signRequest` has a Cognitive Complexity of 34 (exceeds 5 allowed). Consider refactoring. https://codeclimate.com/github/sanktjodel/cctest1/Signing.java#issue_64124ec1a9c49c0001000012
test
fix method complexity issue in signing java method signrequest has a cognitive complexity of exceeds allowed consider refactoring
1
266,972
8,377,573,801
IssuesEvent
2018-10-06 02:53:54
medic/medic-webapp
https://api.github.com/repos/medic/medic-webapp
closed
Horti crashes on subsequent upgrade
Priority: 1 - High Status: 5 - Ready Type: Bug horticulturalist
Each time an app (api or sentinel) is updated, horti creates a symlink to the previous running version (called `old`). When doing multiple upgrades, this `old` symlink is not removed prior to attempting to write it again resulting in the following fatal error (which crashes horti): ``` horti:debug Primary ddoc written +3s Updating symlinks for changed apps… [ { name: 'medic-api', attachmentName: 'medic-api-0.1.0.tgz', digest: 'md5-wmKqGbAc5bAKugPtyXZdKQ==', deployPath: [Function: deployPath] } ] ********FATAL******** { Error: EEXIST: file already exists, symlink '/home/user/.horticulturalist/deployments/medic-api/md5-vU2CqBOkCe+HMwOM3oCb+w==' -> '/home/user/.horticulturalist/deployments/medic-api/old' at Object.fs.symlinkSync (fs.js:1031:18) at Promise.all.changedApps.map.app (/usr/lib/node_modules/horticulturalist/src/install/deploySteps.js:112:14) at Array.map (<anonymous>) at updateSymlink (/usr/lib/node_modules/horticulturalist/src/install/deploySteps.js:105:36) at Promise.resolve.then.then (/usr/lib/node_modules/horticulturalist/src/install/deploySteps.js:153:25) at <anonymous> at process._tickCallback (internal/process/next_tick.js:182:7) errno: -17, code: 'EEXIST', syscall: 'symlink', path: '/home/user/.horticulturalist/deployments/medic-api/md5-vU2CqBOkCe+HMwOM3oCb+w==', dest: '/home/user/.horticulturalist/deployments/medic-api/old' } ```
1.0
Horti crashes on subsequent upgrade - Each time an app (api or sentinel) is updated, horti creates a symlink to the previous running version (called `old`). When doing multiple upgrades, this `old` symlink is not removed prior to attempting to write it again resulting in the following fatal error (which crashes horti): ``` horti:debug Primary ddoc written +3s Updating symlinks for changed apps… [ { name: 'medic-api', attachmentName: 'medic-api-0.1.0.tgz', digest: 'md5-wmKqGbAc5bAKugPtyXZdKQ==', deployPath: [Function: deployPath] } ] ********FATAL******** { Error: EEXIST: file already exists, symlink '/home/user/.horticulturalist/deployments/medic-api/md5-vU2CqBOkCe+HMwOM3oCb+w==' -> '/home/user/.horticulturalist/deployments/medic-api/old' at Object.fs.symlinkSync (fs.js:1031:18) at Promise.all.changedApps.map.app (/usr/lib/node_modules/horticulturalist/src/install/deploySteps.js:112:14) at Array.map (<anonymous>) at updateSymlink (/usr/lib/node_modules/horticulturalist/src/install/deploySteps.js:105:36) at Promise.resolve.then.then (/usr/lib/node_modules/horticulturalist/src/install/deploySteps.js:153:25) at <anonymous> at process._tickCallback (internal/process/next_tick.js:182:7) errno: -17, code: 'EEXIST', syscall: 'symlink', path: '/home/user/.horticulturalist/deployments/medic-api/md5-vU2CqBOkCe+HMwOM3oCb+w==', dest: '/home/user/.horticulturalist/deployments/medic-api/old' } ```
non_test
horti crashes on subsequent upgrade each time an app api or sentinel is updated horti creates a symlink to the previous running version called old when doing multiple upgrades this old symlink is not removed prior to attempting to write it again resulting in the following fatal error which crashes horti horti debug primary ddoc written updating symlinks for changed apps… name medic api attachmentname medic api tgz digest deploypath fatal error eexist file already exists symlink home user horticulturalist deployments medic api w home user horticulturalist deployments medic api old at object fs symlinksync fs js at promise all changedapps map app usr lib node modules horticulturalist src install deploysteps js at array map at updatesymlink usr lib node modules horticulturalist src install deploysteps js at promise resolve then then usr lib node modules horticulturalist src install deploysteps js at at process tickcallback internal process next tick js errno code eexist syscall symlink path home user horticulturalist deployments medic api w dest home user horticulturalist deployments medic api old
0
158,570
6,031,907,622
IssuesEvent
2017-06-09 01:09:57
chartjs/Chart.js
https://api.github.com/repos/chartjs/Chart.js
closed
First and last bars display problem
Category: Bug Help wanted Inactive: duplicate Priority: p1 Time Scale
I have the following issue: ![image](https://user-images.githubusercontent.com/5883414/26927343-240816e2-4c5a-11e7-8ed1-c20735edcf5b.png) The first bar is not fully displayed and the last one is hidden at all. Any solution to this?
1.0
First and last bars display problem - I have the following issue: ![image](https://user-images.githubusercontent.com/5883414/26927343-240816e2-4c5a-11e7-8ed1-c20735edcf5b.png) The first bar is not fully displayed and the last one is hidden at all. Any solution to this?
non_test
first and last bars display problem i have the following issue the first bar is not fully displayed and the last one is hidden at all any solution to this
0
488,037
14,073,876,166
IssuesEvent
2020-11-04 06:05:17
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.rawstory.com - video or audio doesn't play
browser-fenix engine-gecko ml-needsdiagnosis-false ml-probability-high priority-normal
<!-- @browser: Firefox Mobile 83.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:83.0) Gecko/83.0 Firefox/83.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/61027 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.rawstory.com/2020/11/watch-florida-deputies-caught-on-tape-beating-biden-supporter-at-political-rally/ **Browser / Version**: Firefox Mobile 83.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Video or audio doesn't play **Description**: There is no video **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201025174155</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/11/a2757cf2-64eb-4019-80f2-9be21ee7edc7) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.rawstory.com - video or audio doesn't play - <!-- @browser: Firefox Mobile 83.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:83.0) Gecko/83.0 Firefox/83.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/61027 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.rawstory.com/2020/11/watch-florida-deputies-caught-on-tape-beating-biden-supporter-at-political-rally/ **Browser / Version**: Firefox Mobile 83.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Video or audio doesn't play **Description**: There is no video **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201025174155</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/11/a2757cf2-64eb-4019-80f2-9be21ee7edc7) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
video or audio doesn t play url browser version firefox mobile operating system android tested another browser no problem type video or audio doesn t play description there is no video steps to reproduce browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
115,319
24,746,390,426
IssuesEvent
2022-10-21 10:02:44
arduino-libraries/Arduino_DebugUtils
https://api.github.com/repos/arduino-libraries/Arduino_DebugUtils
closed
Adding Debug.info() style functions
type: enhancement topic: code
Hi, has there been any discussion around adding individual functions for each debug logging level, instead of passing in the level explicitly, similar to how the logging module in python works? I believe something like this should work: ``` void Arduino_DebugUtils::info(const char * fmt, ...) { va_list args; va_start(args, fmt); print(DBG_INFO, fmt, args); va_end(args); } ``` I would be more than happy to open a PR if this feels like something that should be added
1.0
Adding Debug.info() style functions - Hi, has there been any discussion around adding individual functions for each debug logging level, instead of passing in the level explicitly, similar to how the logging module in python works? I believe something like this should work: ``` void Arduino_DebugUtils::info(const char * fmt, ...) { va_list args; va_start(args, fmt); print(DBG_INFO, fmt, args); va_end(args); } ``` I would be more than happy to open a PR if this feels like something that should be added
non_test
adding debug info style functions hi has there been any discussion around adding individual functions for each debug logging level instead of passing in the level explicitly similar to how the logging module in python works i believe something like this should work void arduino debugutils info const char fmt va list args va start args fmt print dbg info fmt args va end args i would be more than happy to open a pr if this feels like something that should be added
0
328,134
28,102,012,833
IssuesEvent
2023-03-30 20:18:25
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: kv/splits/nodes=3/quiesce=false failed
C-test-failure O-robot X-duplicate O-roachtest branch-master release-blocker T-kv
roachtest.kv/splits/nodes=3/quiesce=false [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/9329887?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/9329887?buildTab=artifacts#/kv/splits/nodes=3/quiesce=false) on master @ [1f8024bf14433ca169e5a8c3768c5d223dc5018c](https://github.com/cockroachdb/cockroach/commits/1f8024bf14433ca169e5a8c3768c5d223dc5018c): ``` test artifacts and logs in: /artifacts/kv/splits/nodes=3/quiesce=false/run_1 (cluster.go:1977).Run: output in run_053210.523548575_n4_workload-run-kv-init: ./workload run kv --init --max-ops=1 --concurrency=192 --splits=30000 {pgurl:1-3} returned: COMMAND_PROBLEM: ssh verbose log retained in ssh_053211.295244109_n4_workload-run-kv-init.log: exit status 1 (monitor.go:127).Wait: monitor failure: monitor task failed: t.Fatal() was called ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*kv/splits/nodes=3/quiesce=false.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-26198
2.0
roachtest: kv/splits/nodes=3/quiesce=false failed - roachtest.kv/splits/nodes=3/quiesce=false [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/9329887?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/9329887?buildTab=artifacts#/kv/splits/nodes=3/quiesce=false) on master @ [1f8024bf14433ca169e5a8c3768c5d223dc5018c](https://github.com/cockroachdb/cockroach/commits/1f8024bf14433ca169e5a8c3768c5d223dc5018c): ``` test artifacts and logs in: /artifacts/kv/splits/nodes=3/quiesce=false/run_1 (cluster.go:1977).Run: output in run_053210.523548575_n4_workload-run-kv-init: ./workload run kv --init --max-ops=1 --concurrency=192 --splits=30000 {pgurl:1-3} returned: COMMAND_PROBLEM: ssh verbose log retained in ssh_053211.295244109_n4_workload-run-kv-init.log: exit status 1 (monitor.go:127).Wait: monitor failure: monitor task failed: t.Fatal() was called ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*kv/splits/nodes=3/quiesce=false.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-26198
test
roachtest kv splits nodes quiesce false failed roachtest kv splits nodes quiesce false with on master test artifacts and logs in artifacts kv splits nodes quiesce false run cluster go run output in run workload run kv init workload run kv init max ops concurrency splits pgurl returned command problem ssh verbose log retained in ssh workload run kv init log exit status monitor go wait monitor failure monitor task failed t fatal was called parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see cc cockroachdb kv triage jira issue crdb
1
81,035
7,763,681,590
IssuesEvent
2018-06-01 17:27:32
mozilla-mobile/focus-android
https://api.github.com/repos/mozilla-mobile/focus-android
closed
Create UI Test Script for performance comparision on website loading
size M testing
This is a subtask of #2201 The script will initially load about 10~20 websites in order, where the loading time will be measured and collected. Since the bookends will be hit multiple times, they need to be defined with an iterator inside the client code.
1.0
Create UI Test Script for performance comparision on website loading - This is a subtask of #2201 The script will initially load about 10~20 websites in order, where the loading time will be measured and collected. Since the bookends will be hit multiple times, they need to be defined with an iterator inside the client code.
test
create ui test script for performance comparision on website loading this is a subtask of the script will initially load about websites in order where the loading time will be measured and collected since the bookends will be hit multiple times they need to be defined with an iterator inside the client code
1
106,760
9,181,639,191
IssuesEvent
2019-03-05 10:42:33
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
DeleteExpiredDataIT.testDeleteExpiredData failure
:ml >test-failure
Failed build at https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+master+periodic/813/ Does not reproduce locally. Reproduce block: ``` ./gradlew :x-pack:plugin:ml:qa:native-multi-node-tests:integTestRunner \ -Dtests.seed=26207D14EDCE60BC \ -Dtests.class=org.elasticsearch.xpack.ml.integration.DeleteExpiredDataIT \ -Dtests.method="testDeleteExpiredData" \ -Dtests.security.manager=true \ -Dtests.locale=pl-PL \ -Dtests.timezone=Etc/GMT \ -Dcompiler.java=11 \ -Druntime.java=8 ``` There are a few interesting stack traces here, but this one jumped out at me, and happens a *lot*: ``` [2019-03-01T07:39:15,383][DEBUG][o.e.a.b.TransportShardBulkAction] [node-1] [.ml-state][3] failed to execute bulk item (index) index {[.ml-state-write][doc][non_existing_job_categorizer_state#859], source[{}]} -- java.lang.IllegalArgumentException: Rejecting mapping update to [.ml-state] as the final mapping would have more than 1 type: [_doc, doc] at org.elasticsearch.index.mapper.MapperService.internalMerge(MapperService.java:449) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.index.mapper.MapperService.internalMerge(MapperService.java:398) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.index.mapper.MapperService.merge(MapperService.java:331) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.cluster.metadata.MetaDataMappingService$PutMappingExecutor.applyRequest(MetaDataMappingService.java:315) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.cluster.metadata.MetaDataMappingService$PutMappingExecutor.execute(MetaDataMappingService.java:238) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.cluster.service.MasterService.executeTasks(MasterService.java:687) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.cluster.service.MasterService.calculateTaskOutputs(MasterService.java:310) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.cluster.service.MasterService.runTasks(MasterService.java:210) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.cluster.service.MasterService$Batcher.run(MasterService.java:142) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.cluster.service.TaskBatcher.runIfNotProcessed(TaskBatcher.java:150) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.cluster.service.TaskBatcher$BatchedTask.run(TaskBatcher.java:188) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:681) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.common.util.concurrent.PrioritizedEsThreadPoolExecutor$TieBreakingPrioritizedRunnable.runAndClean(PrioritizedEsThreadPoolExecutor.java:252) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.common.util.concurrent.PrioritizedEsThreadPoolExecutor$TieBreakingPrioritizedRunnable.run(PrioritizedEsThreadPoolExecutor.java:215) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?] at java.lang.Thread.run(Thread.java:834) [?:?] ``` But this is also present: ``` [2019-03-01T07:46:45,108][WARN ][o.e.p.PersistentTasksNodeService] [node-2] task datafeed-realtime-job-given-process-is-killed-datafeed failed with an exception -- org.elasticsearch.xpack.ml.datafeed.DatafeedJob$AnalysisProblemException: ElasticsearchStatusException[Cannot perform requested action because job [realtime-job-given-process-is-killed] is not open] at org.elasticsearch.xpack.ml.datafeed.DatafeedJob.flushJob(DatafeedJob.java:457) ~[x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.ml.datafeed.DatafeedJob.run(DatafeedJob.java:396) ~[x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.ml.datafeed.DatafeedJob.runRealtime(DatafeedJob.java:177) ~[x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.ml.datafeed.DatafeedManager$Holder.executeRealTime(DatafeedManager.java:397) ~[x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.ml.datafeed.DatafeedManager$Holder.access$600(DatafeedManager.java:290) [x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.ml.datafeed.DatafeedManager$3.doRun(DatafeedManager.java:227) [x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingAbstractRunnable.doRun(ThreadContext.java:751) [elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:37) [elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?] at java.lang.Thread.run(Thread.java:834) [?:?] Caused by: org.elasticsearch.ElasticsearchStatusException: Cannot perform requested action because job [realtime-job-given-process-is-killed] is not open at org.elasticsearch.xpack.core.ml.utils.ExceptionsHelper.conflictStatusException(ExceptionsHelper.java:50) ~[?:?] at org.elasticsearch.xpack.ml.action.TransportJobTaskAction.doExecute(TransportJobTaskAction.java:55) ~[?:?] at org.elasticsearch.xpack.ml.action.TransportJobTaskAction.doExecute(TransportJobTaskAction.java:31) ~[?:?] at org.elasticsearch.action.support.TransportAction$RequestFilterChain.proceed(TransportAction.java:145) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$apply$0(SecurityActionFilter.java:86) ~[?:?] at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:61) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$authorizeRequest$4(SecurityActionFilter.java:171) ~[?:?] at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:61) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.security.authz.AuthorizationService.lambda$authorizeAction$4(AuthorizationService.java:237) ~[?:?] at org.elasticsearch.xpack.security.authz.AuthorizationService$AuthorizationResultListener.onResponse(AuthorizationService.java:599) ~[?:?] at org.elasticsearch.xpack.security.authz.AuthorizationService$AuthorizationResultListener.onResponse(AuthorizationService.java:574) ~[?:?] at org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.security.authz.RBACEngine.authorizeClusterAction(RBACEngine.java:140) ~[?:?] at org.elasticsearch.xpack.security.authz.AuthorizationService.authorizeAction(AuthorizationService.java:239) ~[?:?] at org.elasticsearch.xpack.security.authz.AuthorizationService.maybeAuthorizeRunAs(AuthorizationService.java:223) ~[?:?] at org.elasticsearch.xpack.security.authz.AuthorizationService.lambda$authorize$1(AuthorizationService.java:189) ~[?:?] at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:61) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.security.authz.RBACEngine.lambda$resolveAuthorizationInfo$1(RBACEngine.java:113) ~[?:?] at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:61) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.security.authz.store.CompositeRolesStore.getRoles(CompositeRolesStore.java:193) ~[?:?] at org.elasticsearch.xpack.security.authz.RBACEngine.getRoles(RBACEngine.java:119) ~[?:?] at org.elasticsearch.xpack.security.authz.RBACEngine.resolveAuthorizationInfo(RBACEngine.java:107) ~[?:?] at org.elasticsearch.xpack.security.authz.AuthorizationService.authorize(AuthorizationService.java:191) ~[?:?] at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.authorizeRequest(SecurityActionFilter.java:171) ~[?:?] at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$applyInternal$3(SecurityActionFilter.java:157) ~[?:?] at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:61) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$authenticateAsync$2(AuthenticationService.java:245) ~[?:?] at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$lookForExistingAuthentication$6(AuthenticationService.java:305) ~[?:?] at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lookForExistingAuthentication(AuthenticationService.java:316) ~[?:?] at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.authenticateAsync(AuthenticationService.java:243) ~[?:?] at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.access$000(AuthenticationService.java:195) ~[?:?] at org.elasticsearch.xpack.security.authc.AuthenticationService.authenticate(AuthenticationService.java:138) ~[?:?] at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.applyInternal(SecurityActionFilter.java:154) ~[?:?] at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$apply$2(SecurityActionFilter.java:100) ~[?:?] at org.elasticsearch.xpack.core.security.SecurityContext.executeAsUser(SecurityContext.java:97) ~[?:?] at org.elasticsearch.xpack.security.authz.AuthorizationUtils.switchUserBasedOnActionOriginAndExecute(AuthorizationUtils.java:115) ~[?:?] at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.apply(SecurityActionFilter.java:98) ~[?:?] at org.elasticsearch.action.support.TransportAction$RequestFilterChain.proceed(TransportAction.java:143) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.action.support.TransportAction.execute(TransportAction.java:121) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.action.support.TransportAction.execute(TransportAction.java:64) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.client.node.NodeClient.executeLocally(NodeClient.java:83) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.client.node.NodeClient.doExecute(NodeClient.java:72) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.client.support.AbstractClient.execute(AbstractClient.java:394) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.client.support.AbstractClient.execute(AbstractClient.java:383) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.ml.datafeed.DatafeedJob.flushJob(DatafeedJob.java:444) ~[x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] ... 10 more [2019-03-01T07:46:45,112][INFO ][o.e.x.m.d.DatafeedManager] [node-2] [realtime_analysis_error] datafeed [realtime-job-given-process-is-killed-datafeed] for job [realtime-job-given-process-is-killed] has been stopped ```
1.0
DeleteExpiredDataIT.testDeleteExpiredData failure - Failed build at https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+master+periodic/813/ Does not reproduce locally. Reproduce block: ``` ./gradlew :x-pack:plugin:ml:qa:native-multi-node-tests:integTestRunner \ -Dtests.seed=26207D14EDCE60BC \ -Dtests.class=org.elasticsearch.xpack.ml.integration.DeleteExpiredDataIT \ -Dtests.method="testDeleteExpiredData" \ -Dtests.security.manager=true \ -Dtests.locale=pl-PL \ -Dtests.timezone=Etc/GMT \ -Dcompiler.java=11 \ -Druntime.java=8 ``` There are a few interesting stack traces here, but this one jumped out at me, and happens a *lot*: ``` [2019-03-01T07:39:15,383][DEBUG][o.e.a.b.TransportShardBulkAction] [node-1] [.ml-state][3] failed to execute bulk item (index) index {[.ml-state-write][doc][non_existing_job_categorizer_state#859], source[{}]} -- java.lang.IllegalArgumentException: Rejecting mapping update to [.ml-state] as the final mapping would have more than 1 type: [_doc, doc] at org.elasticsearch.index.mapper.MapperService.internalMerge(MapperService.java:449) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.index.mapper.MapperService.internalMerge(MapperService.java:398) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.index.mapper.MapperService.merge(MapperService.java:331) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.cluster.metadata.MetaDataMappingService$PutMappingExecutor.applyRequest(MetaDataMappingService.java:315) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.cluster.metadata.MetaDataMappingService$PutMappingExecutor.execute(MetaDataMappingService.java:238) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.cluster.service.MasterService.executeTasks(MasterService.java:687) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.cluster.service.MasterService.calculateTaskOutputs(MasterService.java:310) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.cluster.service.MasterService.runTasks(MasterService.java:210) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.cluster.service.MasterService$Batcher.run(MasterService.java:142) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.cluster.service.TaskBatcher.runIfNotProcessed(TaskBatcher.java:150) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.cluster.service.TaskBatcher$BatchedTask.run(TaskBatcher.java:188) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:681) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.common.util.concurrent.PrioritizedEsThreadPoolExecutor$TieBreakingPrioritizedRunnable.runAndClean(PrioritizedEsThreadPoolExecutor.java:252) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.common.util.concurrent.PrioritizedEsThreadPoolExecutor$TieBreakingPrioritizedRunnable.run(PrioritizedEsThreadPoolExecutor.java:215) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?] at java.lang.Thread.run(Thread.java:834) [?:?] ``` But this is also present: ``` [2019-03-01T07:46:45,108][WARN ][o.e.p.PersistentTasksNodeService] [node-2] task datafeed-realtime-job-given-process-is-killed-datafeed failed with an exception -- org.elasticsearch.xpack.ml.datafeed.DatafeedJob$AnalysisProblemException: ElasticsearchStatusException[Cannot perform requested action because job [realtime-job-given-process-is-killed] is not open] at org.elasticsearch.xpack.ml.datafeed.DatafeedJob.flushJob(DatafeedJob.java:457) ~[x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.ml.datafeed.DatafeedJob.run(DatafeedJob.java:396) ~[x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.ml.datafeed.DatafeedJob.runRealtime(DatafeedJob.java:177) ~[x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.ml.datafeed.DatafeedManager$Holder.executeRealTime(DatafeedManager.java:397) ~[x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.ml.datafeed.DatafeedManager$Holder.access$600(DatafeedManager.java:290) [x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.ml.datafeed.DatafeedManager$3.doRun(DatafeedManager.java:227) [x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingAbstractRunnable.doRun(ThreadContext.java:751) [elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:37) [elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?] at java.lang.Thread.run(Thread.java:834) [?:?] Caused by: org.elasticsearch.ElasticsearchStatusException: Cannot perform requested action because job [realtime-job-given-process-is-killed] is not open at org.elasticsearch.xpack.core.ml.utils.ExceptionsHelper.conflictStatusException(ExceptionsHelper.java:50) ~[?:?] at org.elasticsearch.xpack.ml.action.TransportJobTaskAction.doExecute(TransportJobTaskAction.java:55) ~[?:?] at org.elasticsearch.xpack.ml.action.TransportJobTaskAction.doExecute(TransportJobTaskAction.java:31) ~[?:?] at org.elasticsearch.action.support.TransportAction$RequestFilterChain.proceed(TransportAction.java:145) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$apply$0(SecurityActionFilter.java:86) ~[?:?] at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:61) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$authorizeRequest$4(SecurityActionFilter.java:171) ~[?:?] at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:61) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.security.authz.AuthorizationService.lambda$authorizeAction$4(AuthorizationService.java:237) ~[?:?] at org.elasticsearch.xpack.security.authz.AuthorizationService$AuthorizationResultListener.onResponse(AuthorizationService.java:599) ~[?:?] at org.elasticsearch.xpack.security.authz.AuthorizationService$AuthorizationResultListener.onResponse(AuthorizationService.java:574) ~[?:?] at org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.security.authz.RBACEngine.authorizeClusterAction(RBACEngine.java:140) ~[?:?] at org.elasticsearch.xpack.security.authz.AuthorizationService.authorizeAction(AuthorizationService.java:239) ~[?:?] at org.elasticsearch.xpack.security.authz.AuthorizationService.maybeAuthorizeRunAs(AuthorizationService.java:223) ~[?:?] at org.elasticsearch.xpack.security.authz.AuthorizationService.lambda$authorize$1(AuthorizationService.java:189) ~[?:?] at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:61) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.security.authz.RBACEngine.lambda$resolveAuthorizationInfo$1(RBACEngine.java:113) ~[?:?] at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:61) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.security.authz.store.CompositeRolesStore.getRoles(CompositeRolesStore.java:193) ~[?:?] at org.elasticsearch.xpack.security.authz.RBACEngine.getRoles(RBACEngine.java:119) ~[?:?] at org.elasticsearch.xpack.security.authz.RBACEngine.resolveAuthorizationInfo(RBACEngine.java:107) ~[?:?] at org.elasticsearch.xpack.security.authz.AuthorizationService.authorize(AuthorizationService.java:191) ~[?:?] at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.authorizeRequest(SecurityActionFilter.java:171) ~[?:?] at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$applyInternal$3(SecurityActionFilter.java:157) ~[?:?] at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:61) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$authenticateAsync$2(AuthenticationService.java:245) ~[?:?] at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lambda$lookForExistingAuthentication$6(AuthenticationService.java:305) ~[?:?] at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.lookForExistingAuthentication(AuthenticationService.java:316) ~[?:?] at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.authenticateAsync(AuthenticationService.java:243) ~[?:?] at org.elasticsearch.xpack.security.authc.AuthenticationService$Authenticator.access$000(AuthenticationService.java:195) ~[?:?] at org.elasticsearch.xpack.security.authc.AuthenticationService.authenticate(AuthenticationService.java:138) ~[?:?] at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.applyInternal(SecurityActionFilter.java:154) ~[?:?] at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.lambda$apply$2(SecurityActionFilter.java:100) ~[?:?] at org.elasticsearch.xpack.core.security.SecurityContext.executeAsUser(SecurityContext.java:97) ~[?:?] at org.elasticsearch.xpack.security.authz.AuthorizationUtils.switchUserBasedOnActionOriginAndExecute(AuthorizationUtils.java:115) ~[?:?] at org.elasticsearch.xpack.security.action.filter.SecurityActionFilter.apply(SecurityActionFilter.java:98) ~[?:?] at org.elasticsearch.action.support.TransportAction$RequestFilterChain.proceed(TransportAction.java:143) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.action.support.TransportAction.execute(TransportAction.java:121) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.action.support.TransportAction.execute(TransportAction.java:64) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.client.node.NodeClient.executeLocally(NodeClient.java:83) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.client.node.NodeClient.doExecute(NodeClient.java:72) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.client.support.AbstractClient.execute(AbstractClient.java:394) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.client.support.AbstractClient.execute(AbstractClient.java:383) ~[elasticsearch-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] at org.elasticsearch.xpack.ml.datafeed.DatafeedJob.flushJob(DatafeedJob.java:444) ~[x-pack-ml-8.0.0-SNAPSHOT.jar:8.0.0-SNAPSHOT] ... 10 more [2019-03-01T07:46:45,112][INFO ][o.e.x.m.d.DatafeedManager] [node-2] [realtime_analysis_error] datafeed [realtime-job-given-process-is-killed-datafeed] for job [realtime-job-given-process-is-killed] has been stopped ```
test
deleteexpireddatait testdeleteexpireddata failure failed build at does not reproduce locally reproduce block gradlew x pack plugin ml qa native multi node tests integtestrunner dtests seed dtests class org elasticsearch xpack ml integration deleteexpireddatait dtests method testdeleteexpireddata dtests security manager true dtests locale pl pl dtests timezone etc gmt dcompiler java druntime java there are a few interesting stack traces here but this one jumped out at me and happens a lot failed to execute bulk item index index source java lang illegalargumentexception rejecting mapping update to as the final mapping would have more than type at org elasticsearch index mapper mapperservice internalmerge mapperservice java at org elasticsearch index mapper mapperservice internalmerge mapperservice java at org elasticsearch index mapper mapperservice merge mapperservice java at org elasticsearch cluster metadata metadatamappingservice putmappingexecutor applyrequest metadatamappingservice java at org elasticsearch cluster metadata metadatamappingservice putmappingexecutor execute metadatamappingservice java at org elasticsearch cluster service masterservice executetasks masterservice java at org elasticsearch cluster service masterservice calculatetaskoutputs masterservice java at org elasticsearch cluster service masterservice runtasks masterservice java at org elasticsearch cluster service masterservice batcher run masterservice java at org elasticsearch cluster service taskbatcher runifnotprocessed taskbatcher java at org elasticsearch cluster service taskbatcher batchedtask run taskbatcher java at org elasticsearch common util concurrent threadcontext contextpreservingrunnable run threadcontext java at org elasticsearch common util concurrent prioritizedesthreadpoolexecutor tiebreakingprioritizedrunnable runandclean prioritizedesthreadpoolexecutor java at org elasticsearch common util concurrent prioritizedesthreadpoolexecutor tiebreakingprioritizedrunnable run prioritizedesthreadpoolexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java but this is also present task datafeed realtime job given process is killed datafeed failed with an exception org elasticsearch xpack ml datafeed datafeedjob analysisproblemexception elasticsearchstatusexception is not open at org elasticsearch xpack ml datafeed datafeedjob flushjob datafeedjob java at org elasticsearch xpack ml datafeed datafeedjob run datafeedjob java at org elasticsearch xpack ml datafeed datafeedjob runrealtime datafeedjob java at org elasticsearch xpack ml datafeed datafeedmanager holder executerealtime datafeedmanager java at org elasticsearch xpack ml datafeed datafeedmanager holder access datafeedmanager java at org elasticsearch xpack ml datafeed datafeedmanager dorun datafeedmanager java at org elasticsearch common util concurrent threadcontext contextpreservingabstractrunnable dorun threadcontext java at org elasticsearch common util concurrent abstractrunnable run abstractrunnable java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java caused by org elasticsearch elasticsearchstatusexception cannot perform requested action because job is not open at org elasticsearch xpack core ml utils exceptionshelper conflictstatusexception exceptionshelper java at org elasticsearch xpack ml action transportjobtaskaction doexecute transportjobtaskaction java at org elasticsearch xpack ml action transportjobtaskaction doexecute transportjobtaskaction java at org elasticsearch action support transportaction requestfilterchain proceed transportaction java at org elasticsearch xpack security action filter securityactionfilter lambda apply securityactionfilter java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch xpack security action filter securityactionfilter lambda authorizerequest securityactionfilter java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch xpack security authz authorizationservice lambda authorizeaction authorizationservice java at org elasticsearch xpack security authz authorizationservice authorizationresultlistener onresponse authorizationservice java at org elasticsearch xpack security authz authorizationservice authorizationresultlistener onresponse authorizationservice java at org elasticsearch action support contextpreservingactionlistener onresponse contextpreservingactionlistener java at org elasticsearch xpack security authz rbacengine authorizeclusteraction rbacengine java at org elasticsearch xpack security authz authorizationservice authorizeaction authorizationservice java at org elasticsearch xpack security authz authorizationservice maybeauthorizerunas authorizationservice java at org elasticsearch xpack security authz authorizationservice lambda authorize authorizationservice java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch action support contextpreservingactionlistener onresponse contextpreservingactionlistener java at org elasticsearch xpack security authz rbacengine lambda resolveauthorizationinfo rbacengine java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch xpack security authz store compositerolesstore getroles compositerolesstore java at org elasticsearch xpack security authz rbacengine getroles rbacengine java at org elasticsearch xpack security authz rbacengine resolveauthorizationinfo rbacengine java at org elasticsearch xpack security authz authorizationservice authorize authorizationservice java at org elasticsearch xpack security action filter securityactionfilter authorizerequest securityactionfilter java at org elasticsearch xpack security action filter securityactionfilter lambda applyinternal securityactionfilter java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch xpack security authc authenticationservice authenticator lambda authenticateasync authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator lambda lookforexistingauthentication authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator lookforexistingauthentication authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator authenticateasync authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticator access authenticationservice java at org elasticsearch xpack security authc authenticationservice authenticate authenticationservice java at org elasticsearch xpack security action filter securityactionfilter applyinternal securityactionfilter java at org elasticsearch xpack security action filter securityactionfilter lambda apply securityactionfilter java at org elasticsearch xpack core security securitycontext executeasuser securitycontext java at org elasticsearch xpack security authz authorizationutils switchuserbasedonactionoriginandexecute authorizationutils java at org elasticsearch xpack security action filter securityactionfilter apply securityactionfilter java at org elasticsearch action support transportaction requestfilterchain proceed transportaction java at org elasticsearch action support transportaction execute transportaction java at org elasticsearch action support transportaction execute transportaction java at org elasticsearch client node nodeclient executelocally nodeclient java at org elasticsearch client node nodeclient doexecute nodeclient java at org elasticsearch client support abstractclient execute abstractclient java at org elasticsearch client support abstractclient execute abstractclient java at org elasticsearch xpack ml datafeed datafeedjob flushjob datafeedjob java more datafeed for job has been stopped
1
188,430
22,046,416,088
IssuesEvent
2022-05-30 02:35:35
DavidSpek/kale
https://api.github.com/repos/DavidSpek/kale
opened
CVE-2022-29204 (Medium) detected in tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl, tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl
security vulnerability
## CVE-2022-29204 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl</b>, <b>tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary> <p> <details><summary><b>tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/7b/c5/a97ed48fcc878e36bb05a3ea700c077360853c0994473a8f6b0ab4c2ddd2/tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/7b/c5/a97ed48fcc878e36bb05a3ea700c077360853c0994473a8f6b0ab4c2ddd2/tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl</a></p> <p>Path to dependency file: /examples/dog-breed-classification/requirements/requirements.txt</p> <p>Path to vulnerable library: /examples/dog-breed-classification/requirements/requirements.txt</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library) </details> <details><summary><b>tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/ef/73/205b5e7f8fe086ffe4165d984acb2c49fa3086f330f03099378753982d2e/tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/ef/73/205b5e7f8fe086ffe4165d984acb2c49fa3086f330f03099378753982d2e/tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: /examples/taxi-cab-classification/requirements.txt</p> <p>Path to vulnerable library: /examples/taxi-cab-classification/requirements.txt</p> <p> Dependency Hierarchy: - tfx_bsl-0.21.4-cp27-cp27mu-manylinux2010_x86_64.whl (Root Library) - :x: **tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an open source platform for machine learning. Prior to versions 2.9.0, 2.8.1, 2.7.2, and 2.6.4, the implementation of `tf.raw_ops.UnsortedSegmentJoin` does not fully validate the input arguments. This results in a `CHECK`-failure which can be used to trigger a denial of service attack. The code assumes `num_segments` is a positive scalar but there is no validation. Since this value is used to allocate the output tensor, a negative value would result in a `CHECK`-failure (assertion failure), as per TFSA-2021-198. Versions 2.9.0, 2.8.1, 2.7.2, and 2.6.4 contain a patch for this issue. <p>Publish Date: 2022-05-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-29204>CVE-2022-29204</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-29204">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-29204</a></p> <p>Release Date: 2022-05-20</p> <p>Fix Resolution: tensorflow - 2.6.4,2.7.2,2.8.1,2.9.0;tensorflow-cpu - 2.6.4,2.7.2,2.8.1,2.9.0;tensorflow-gpu - 2.6.4,2.7.2,2.8.1,2.9.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-29204 (Medium) detected in tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl, tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl - ## CVE-2022-29204 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl</b>, <b>tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary> <p> <details><summary><b>tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/7b/c5/a97ed48fcc878e36bb05a3ea700c077360853c0994473a8f6b0ab4c2ddd2/tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/7b/c5/a97ed48fcc878e36bb05a3ea700c077360853c0994473a8f6b0ab4c2ddd2/tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl</a></p> <p>Path to dependency file: /examples/dog-breed-classification/requirements/requirements.txt</p> <p>Path to vulnerable library: /examples/dog-breed-classification/requirements/requirements.txt</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library) </details> <details><summary><b>tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/ef/73/205b5e7f8fe086ffe4165d984acb2c49fa3086f330f03099378753982d2e/tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/ef/73/205b5e7f8fe086ffe4165d984acb2c49fa3086f330f03099378753982d2e/tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: /examples/taxi-cab-classification/requirements.txt</p> <p>Path to vulnerable library: /examples/taxi-cab-classification/requirements.txt</p> <p> Dependency Hierarchy: - tfx_bsl-0.21.4-cp27-cp27mu-manylinux2010_x86_64.whl (Root Library) - :x: **tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an open source platform for machine learning. Prior to versions 2.9.0, 2.8.1, 2.7.2, and 2.6.4, the implementation of `tf.raw_ops.UnsortedSegmentJoin` does not fully validate the input arguments. This results in a `CHECK`-failure which can be used to trigger a denial of service attack. The code assumes `num_segments` is a positive scalar but there is no validation. Since this value is used to allocate the output tensor, a negative value would result in a `CHECK`-failure (assertion failure), as per TFSA-2021-198. Versions 2.9.0, 2.8.1, 2.7.2, and 2.6.4 contain a patch for this issue. <p>Publish Date: 2022-05-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-29204>CVE-2022-29204</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-29204">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-29204</a></p> <p>Release Date: 2022-05-20</p> <p>Fix Resolution: tensorflow - 2.6.4,2.7.2,2.8.1,2.9.0;tensorflow-cpu - 2.6.4,2.7.2,2.8.1,2.9.0;tensorflow-gpu - 2.6.4,2.7.2,2.8.1,2.9.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in tensorflow whl tensorflow whl cve medium severity vulnerability vulnerable libraries tensorflow whl tensorflow whl tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file examples dog breed classification requirements requirements txt path to vulnerable library examples dog breed classification requirements requirements txt dependency hierarchy x tensorflow whl vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file examples taxi cab classification requirements txt path to vulnerable library examples taxi cab classification requirements txt dependency hierarchy tfx bsl whl root library x tensorflow whl vulnerable library found in base branch master vulnerability details tensorflow is an open source platform for machine learning prior to versions and the implementation of tf raw ops unsortedsegmentjoin does not fully validate the input arguments this results in a check failure which can be used to trigger a denial of service attack the code assumes num segments is a positive scalar but there is no validation since this value is used to allocate the output tensor a negative value would result in a check failure assertion failure as per tfsa versions and contain a patch for this issue publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with mend
0
252,477
21,580,133,159
IssuesEvent
2022-05-02 17:47:26
rancher/dashboard
https://api.github.com/repos/rancher/dashboard
closed
Error messages appear below form buttons
kind/bug [zube]: To Test team/area3
Create a deployment (e.g. name: 'test' and image : 'nginx'). Clone this deployment (from the row action menu for the deployment) and leave the name blank - click 'Create'. The Create button changes to error but the message is not visible - scroll down and you will see it beneath the buttons. The simple fix would be to scroll the page to bottom when there is an error - might be worth just doing that in a PR - that fixes the immediate issue - but I think we need to move the errors, if we want the buttons to stick to the bottom of the page. This will need some UX design.
1.0
Error messages appear below form buttons - Create a deployment (e.g. name: 'test' and image : 'nginx'). Clone this deployment (from the row action menu for the deployment) and leave the name blank - click 'Create'. The Create button changes to error but the message is not visible - scroll down and you will see it beneath the buttons. The simple fix would be to scroll the page to bottom when there is an error - might be worth just doing that in a PR - that fixes the immediate issue - but I think we need to move the errors, if we want the buttons to stick to the bottom of the page. This will need some UX design.
test
error messages appear below form buttons create a deployment e g name test and image nginx clone this deployment from the row action menu for the deployment and leave the name blank click create the create button changes to error but the message is not visible scroll down and you will see it beneath the buttons the simple fix would be to scroll the page to bottom when there is an error might be worth just doing that in a pr that fixes the immediate issue but i think we need to move the errors if we want the buttons to stick to the bottom of the page this will need some ux design
1
109,894
11,661,568,021
IssuesEvent
2020-03-03 07:08:04
ift-gftc/doc.ift.org
https://api.github.com/repos/ift-gftc/doc.ift.org
opened
DOC | How to launch Gatsby server locally?
documentation
## 📖 Document Create documentation on how to run the Gatsby server locally so that future contributors know how to do this when testing branches they are working on.
1.0
DOC | How to launch Gatsby server locally? - ## 📖 Document Create documentation on how to run the Gatsby server locally so that future contributors know how to do this when testing branches they are working on.
non_test
doc how to launch gatsby server locally 📖 document create documentation on how to run the gatsby server locally so that future contributors know how to do this when testing branches they are working on
0
296,735
25,572,508,271
IssuesEvent
2022-11-30 18:55:30
MD-Anderson-Bioinformatics/NG-CHM
https://api.github.com/repos/MD-Anderson-Bioinformatics/NG-CHM
closed
All plugin info not removed when one plugin replaced by another after opening a pane below plot pane.
bug passed retest 2.21.3
To reproduce, Load a map that has extra coordinates for use by 2D or 3D scatterplots (like deprez1000). Replace Summary pane with 3D Scatterplot. (Use default settings). Open pane below 3D Scatterplot pane. Replace 3D Scatterplot pane with 2D Scatterplot. (Use default settings). Save Map. Dialog telling that some parts of Scatterplot can't be saved reports that info for both 2D and 3D Scatterplots even though current map only contains a 2D Scatterplot.
1.0
All plugin info not removed when one plugin replaced by another after opening a pane below plot pane. - To reproduce, Load a map that has extra coordinates for use by 2D or 3D scatterplots (like deprez1000). Replace Summary pane with 3D Scatterplot. (Use default settings). Open pane below 3D Scatterplot pane. Replace 3D Scatterplot pane with 2D Scatterplot. (Use default settings). Save Map. Dialog telling that some parts of Scatterplot can't be saved reports that info for both 2D and 3D Scatterplots even though current map only contains a 2D Scatterplot.
test
all plugin info not removed when one plugin replaced by another after opening a pane below plot pane to reproduce load a map that has extra coordinates for use by or scatterplots like replace summary pane with scatterplot use default settings open pane below scatterplot pane replace scatterplot pane with scatterplot use default settings save map dialog telling that some parts of scatterplot can t be saved reports that info for both and scatterplots even though current map only contains a scatterplot
1
140,100
11,302,418,821
IssuesEvent
2020-01-17 17:37:15
aliasrobotics/RVD
https://api.github.com/repos/aliasrobotics/RVD
opened
CWE-362/CWE-367 (race), This usually indicates a security flaw. If an attacker can change any... @ /canopen_master/objdict.h:525
CWE-362 CWE-367 bug flawfinder flawfinder_level_4 static analysis testing triage
```yaml { "system": "./install/canopen_master/include/canopen_master/objdict.h:525:42", "id": 1, "title": "CWE-362/CWE-367 (race), This usually indicates a security flaw. If an attacker can change any... @ /canopen_master/objdict.h:525", "cwe": [ "CWE-362", "CWE-367" ], "exploitation": { "exploitation-image": "", "exploitation-vector": "", "description": "" }, "keywords": [ "flawfinder", "flawfinder_level_4", "static analysis", "testing", "triage", "CWE-362", "CWE-367", "bug" ], "type": "bug", "cve": "None", "severity": { "rvss-vector": "", "rvss-score": 0, "cvss-score": 0, "severity-description": "", "cvss-vector": "" }, "vendor": null, "links": "", "mitigation": { "pull-request": "", "date-mitigation": "", "description": "Set up the correct permissions (e.g., using setuid()) and try to open the file directly" }, "flaw": { "application": "N/A", "reproducibility": "always", "detected-by": "Alias Robotics", "languages": "None", "package": "N/A", "reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_test_flawfinder/-/jobs/407215108/artifacts/download", "date-detected": "2020-01-17 (17:36)", "issue": "", "specificity": "subject-specific", "phase": "testing", "trace": "(context) template<> String & ObjectStorage::Data::access();", "architectural-location": "application-specific", "reproduction": "See artifacts below (if available)", "reported-by-relationship": "automatic", "date-reported": "2020-01-17 (17:36)", "reported-by": "Alias Robotics", "detected-by-method": "testing static", "subsystem": "N/A" }, "description": "This usually indicates a security flaw. If an attacker can change anything along the path between the call to access() and the files actual use (e.g., by moving files), the attacker can exploit the race condition (CWE-362/CWE-367!). Set up the correct permissions (e.g., using setuid()) and try to open the file directly. . Happening @ .../canopen_master/objdict.h:525" } ```
1.0
CWE-362/CWE-367 (race), This usually indicates a security flaw. If an attacker can change any... @ /canopen_master/objdict.h:525 - ```yaml { "system": "./install/canopen_master/include/canopen_master/objdict.h:525:42", "id": 1, "title": "CWE-362/CWE-367 (race), This usually indicates a security flaw. If an attacker can change any... @ /canopen_master/objdict.h:525", "cwe": [ "CWE-362", "CWE-367" ], "exploitation": { "exploitation-image": "", "exploitation-vector": "", "description": "" }, "keywords": [ "flawfinder", "flawfinder_level_4", "static analysis", "testing", "triage", "CWE-362", "CWE-367", "bug" ], "type": "bug", "cve": "None", "severity": { "rvss-vector": "", "rvss-score": 0, "cvss-score": 0, "severity-description": "", "cvss-vector": "" }, "vendor": null, "links": "", "mitigation": { "pull-request": "", "date-mitigation": "", "description": "Set up the correct permissions (e.g., using setuid()) and try to open the file directly" }, "flaw": { "application": "N/A", "reproducibility": "always", "detected-by": "Alias Robotics", "languages": "None", "package": "N/A", "reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_test_flawfinder/-/jobs/407215108/artifacts/download", "date-detected": "2020-01-17 (17:36)", "issue": "", "specificity": "subject-specific", "phase": "testing", "trace": "(context) template<> String & ObjectStorage::Data::access();", "architectural-location": "application-specific", "reproduction": "See artifacts below (if available)", "reported-by-relationship": "automatic", "date-reported": "2020-01-17 (17:36)", "reported-by": "Alias Robotics", "detected-by-method": "testing static", "subsystem": "N/A" }, "description": "This usually indicates a security flaw. If an attacker can change anything along the path between the call to access() and the files actual use (e.g., by moving files), the attacker can exploit the race condition (CWE-362/CWE-367!). Set up the correct permissions (e.g., using setuid()) and try to open the file directly. . Happening @ .../canopen_master/objdict.h:525" } ```
test
cwe cwe race this usually indicates a security flaw if an attacker can change any canopen master objdict h yaml system install canopen master include canopen master objdict h id title cwe cwe race this usually indicates a security flaw if an attacker can change any canopen master objdict h cwe cwe cwe exploitation exploitation image exploitation vector description keywords flawfinder flawfinder level static analysis testing triage cwe cwe bug type bug cve none severity rvss vector rvss score cvss score severity description cvss vector vendor null links mitigation pull request date mitigation description set up the correct permissions e g using setuid and try to open the file directly flaw application n a reproducibility always detected by alias robotics languages none package n a reproduction image gitlab com aliasrobotics offensive alurity pipelines active pipeline test flawfinder jobs artifacts download date detected issue specificity subject specific phase testing trace context template string objectstorage data access architectural location application specific reproduction see artifacts below if available reported by relationship automatic date reported reported by alias robotics detected by method testing static subsystem n a description this usually indicates a security flaw if an attacker can change anything along the path between the call to access and the files actual use e g by moving files the attacker can exploit the race condition cwe cwe set up the correct permissions e g using setuid and try to open the file directly happening canopen master objdict h
1
23,990
11,996,347,598
IssuesEvent
2020-04-08 16:36:37
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Syntax Error on KEY
Pri2 cognitive-services/svc face-api/subsvc
Hi, I followed your instruction ... KEY = os.environ['FACE_SUBSCRIPTION_KEY'] I replace it something like KEY = os.environ['123567890'] according to the Key value from my Azure portal but I got Key error message. Any suggestions? What's the correct way to set up the Key? My email address is mchou@cdm.depaul.edu Thanks. Regards, Marco --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 9d930c16-d536-f5d3-fed6-8e08a9d4ae19 * Version Independent ID: 505b497d-5bb1-57c9-c400-8305e0a9300b * Content: [Quickstart: Face client library for Python - Azure Cognitive Services](https://docs.microsoft.com/en-us/azure/cognitive-services/face/quickstarts/python-sdk#feedback) * Content Source: [articles/cognitive-services/Face/QuickStarts/python-sdk.md](https://github.com/Microsoft/azure-docs/blob/master/articles/cognitive-services/Face/QuickStarts/python-sdk.md) * Service: **cognitive-services** * Sub-service: **face-api** * GitHub Login: @PatrickFarley * Microsoft Alias: **pafarley**
1.0
Syntax Error on KEY - Hi, I followed your instruction ... KEY = os.environ['FACE_SUBSCRIPTION_KEY'] I replace it something like KEY = os.environ['123567890'] according to the Key value from my Azure portal but I got Key error message. Any suggestions? What's the correct way to set up the Key? My email address is mchou@cdm.depaul.edu Thanks. Regards, Marco --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 9d930c16-d536-f5d3-fed6-8e08a9d4ae19 * Version Independent ID: 505b497d-5bb1-57c9-c400-8305e0a9300b * Content: [Quickstart: Face client library for Python - Azure Cognitive Services](https://docs.microsoft.com/en-us/azure/cognitive-services/face/quickstarts/python-sdk#feedback) * Content Source: [articles/cognitive-services/Face/QuickStarts/python-sdk.md](https://github.com/Microsoft/azure-docs/blob/master/articles/cognitive-services/Face/QuickStarts/python-sdk.md) * Service: **cognitive-services** * Sub-service: **face-api** * GitHub Login: @PatrickFarley * Microsoft Alias: **pafarley**
non_test
syntax error on key hi i followed your instruction key os environ i replace it something like key os environ according to the key value from my azure portal but i got key error message any suggestions what s the correct way to set up the key my email address is mchou cdm depaul edu thanks regards marco document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service cognitive services sub service face api github login patrickfarley microsoft alias pafarley
0
65,353
19,429,147,209
IssuesEvent
2021-12-21 09:54:06
GoldenSoftwareLtd/gedemin
https://api.github.com/repos/GoldenSoftwareLtd/gedemin
opened
На вкладке с множеством в гриде не отображаются колонки
Type-Defect Component-UI Priority-Critical
Сейчас в диалоговых окнах, где есть атрибуты типа множество на вкладке с элементами множества грид по умолчанию отображается со всеми невидимыми колонками. Например, окно группа компаний: ![image](https://user-images.githubusercontent.com/5175764/146909382-cf602484-bfa5-424c-bb57-88ab75784ea4.png) Надо посмотреть по коду. По умолчанию должна быть видима колонка с полем наименование для бизнес-класса. Получается методом getListField.
1.0
На вкладке с множеством в гриде не отображаются колонки - Сейчас в диалоговых окнах, где есть атрибуты типа множество на вкладке с элементами множества грид по умолчанию отображается со всеми невидимыми колонками. Например, окно группа компаний: ![image](https://user-images.githubusercontent.com/5175764/146909382-cf602484-bfa5-424c-bb57-88ab75784ea4.png) Надо посмотреть по коду. По умолчанию должна быть видима колонка с полем наименование для бизнес-класса. Получается методом getListField.
non_test
на вкладке с множеством в гриде не отображаются колонки сейчас в диалоговых окнах где есть атрибуты типа множество на вкладке с элементами множества грид по умолчанию отображается со всеми невидимыми колонками например окно группа компаний надо посмотреть по коду по умолчанию должна быть видима колонка с полем наименование для бизнес класса получается методом getlistfield
0
371,352
10,965,325,437
IssuesEvent
2019-11-28 02:24:59
connect-foundation/2019-21
https://api.github.com/repos/connect-foundation/2019-21
opened
host-app FE 질문 state 에 따른 출력 방식 변경
DB discussion priority:middle react
현재 질문에 대한 state 는 dummyData 를 출력하기 위해 - moderation - newQuestion - popularQuestion - completeQuestion 인데, `DB` 상에는 - moderation - active - completeQuestion 으로 저장되는 상태. DB 에서 가져온 `active` 상태의 question 을 생성 시간에 따라 newQuestion, `like` 수에 따라 popularQuestion 이라는 state 를 붙이고 sorting 해주는 중간 과정이 필요.
1.0
host-app FE 질문 state 에 따른 출력 방식 변경 - 현재 질문에 대한 state 는 dummyData 를 출력하기 위해 - moderation - newQuestion - popularQuestion - completeQuestion 인데, `DB` 상에는 - moderation - active - completeQuestion 으로 저장되는 상태. DB 에서 가져온 `active` 상태의 question 을 생성 시간에 따라 newQuestion, `like` 수에 따라 popularQuestion 이라는 state 를 붙이고 sorting 해주는 중간 과정이 필요.
non_test
host app fe 질문 state 에 따른 출력 방식 변경 현재 질문에 대한 state 는 dummydata 를 출력하기 위해 moderation newquestion popularquestion completequestion 인데 db 상에는 moderation active completequestion 으로 저장되는 상태 db 에서 가져온 active 상태의 question 을 생성 시간에 따라 newquestion like 수에 따라 popularquestion 이라는 state 를 붙이고 sorting 해주는 중간 과정이 필요
0
3,353
2,610,061,139
IssuesEvent
2015-02-26 18:17:59
chrsmith/jsjsj122
https://api.github.com/repos/chrsmith/jsjsj122
opened
路桥看不育去哪里最好
auto-migrated Priority-Medium Type-Defect
``` 路桥看不育去哪里最好【台州五洲生殖医院】24小时健康咨询 热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州市 椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108、1 18、198及椒江一金清公交车直达枫南小区,乘坐107、105、109、 112、901、 902公交车到星星广场下车,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 ``` ----- Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:19
1.0
路桥看不育去哪里最好 - ``` 路桥看不育去哪里最好【台州五洲生殖医院】24小时健康咨询 热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州市 椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108、1 18、198及椒江一金清公交车直达枫南小区,乘坐107、105、109、 112、901、 902公交车到星星广场下车,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 ``` ----- Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:19
non_test
路桥看不育去哪里最好 路桥看不育去哪里最好【台州五洲生殖医院】 热线 微信号tzwzszyy 医院地址 台州市 (枫南大转盘旁)乘车线路 、 、 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
0
237,635
19,662,298,332
IssuesEvent
2022-01-10 18:18:30
RocketChat/Rocket.Chat.Electron
https://api.github.com/repos/RocketChat/Rocket.Chat.Electron
closed
Updates on macOS do not install
type: bug plat: osx stat: needs testing
## My Setup - Operating System: 10.13.6 - App Version: 2.14.0 (previous version also has this issue) - Installation type: dmg <!-- Answer questions by putting x in box, e.g. [x] --> - [ x] I have tested with the latest version - [ x] I can simulate the issue easily ### Current Behavior 0. Click on rocket chat icon. 1. New update is available! 2. Click install update ... blah blah. you will be notified when its ready... 3. Click install now. 4. Program closes. 5. Click on rocket chat icon. 6. Goto step #1. (repeat forever). ### Expected Behavior Not an endless loop as described above. I did figure out a workaround. Basically if you run chat from the command line, a zip file is listed. You can download the zip file and uncompress/install manually which installs the update!! Yay!
1.0
Updates on macOS do not install - ## My Setup - Operating System: 10.13.6 - App Version: 2.14.0 (previous version also has this issue) - Installation type: dmg <!-- Answer questions by putting x in box, e.g. [x] --> - [ x] I have tested with the latest version - [ x] I can simulate the issue easily ### Current Behavior 0. Click on rocket chat icon. 1. New update is available! 2. Click install update ... blah blah. you will be notified when its ready... 3. Click install now. 4. Program closes. 5. Click on rocket chat icon. 6. Goto step #1. (repeat forever). ### Expected Behavior Not an endless loop as described above. I did figure out a workaround. Basically if you run chat from the command line, a zip file is listed. You can download the zip file and uncompress/install manually which installs the update!! Yay!
test
updates on macos do not install my setup operating system app version previous version also has this issue installation type dmg i have tested with the latest version i can simulate the issue easily current behavior click on rocket chat icon new update is available click install update blah blah you will be notified when its ready click install now program closes click on rocket chat icon goto step repeat forever expected behavior not an endless loop as described above i did figure out a workaround basically if you run chat from the command line a zip file is listed you can download the zip file and uncompress install manually which installs the update yay
1
795,696
28,082,759,214
IssuesEvent
2023-03-30 07:44:57
AY2223S2-CS2103-F11-3/tp
https://api.github.com/repos/AY2223S2-CS2103-F11-3/tp
closed
As a user, I can enter relative time as dates
type.Story priority.Low
... so that I need not enter the entire date. Improve date parser method in `ParserUtil` to allow relative times. * `today` - today at 0000. * If now is 2023-05-03 0455, `today` will parse to 2023-05-0455 0000 * `hhmm` - earliest time from now to the time hhmm that has not yet passed. * If now is 2023-05-03 0455, `0400` will parse to 2023-05-04 0400
1.0
As a user, I can enter relative time as dates - ... so that I need not enter the entire date. Improve date parser method in `ParserUtil` to allow relative times. * `today` - today at 0000. * If now is 2023-05-03 0455, `today` will parse to 2023-05-0455 0000 * `hhmm` - earliest time from now to the time hhmm that has not yet passed. * If now is 2023-05-03 0455, `0400` will parse to 2023-05-04 0400
non_test
as a user i can enter relative time as dates so that i need not enter the entire date improve date parser method in parserutil to allow relative times today today at if now is today will parse to hhmm earliest time from now to the time hhmm that has not yet passed if now is will parse to
0
89,120
3,789,897,007
IssuesEvent
2016-03-21 19:31:54
NuGet/Home
https://api.github.com/repos/NuGet/Home
closed
All sources "browse" can hit a null ref, and continually retry
NuGet Visual Studio UI Priority:1 ShipRoom Check
My configuration is two Klondike servers, one accessed via localhost and the other via remote request to a VM. Both of them have ~5 different packages, with a few versions in common, etc... Browse via All sources...and you'll see an error with Null Ref in output for package manager. @alpaix is investigating.
1.0
All sources "browse" can hit a null ref, and continually retry - My configuration is two Klondike servers, one accessed via localhost and the other via remote request to a VM. Both of them have ~5 different packages, with a few versions in common, etc... Browse via All sources...and you'll see an error with Null Ref in output for package manager. @alpaix is investigating.
non_test
all sources browse can hit a null ref and continually retry my configuration is two klondike servers one accessed via localhost and the other via remote request to a vm both of them have different packages with a few versions in common etc browse via all sources and you ll see an error with null ref in output for package manager alpaix is investigating
0
641,311
20,823,790,961
IssuesEvent
2022-03-18 18:10:44
rathena/rathena
https://api.github.com/repos/rathena/rathena
closed
4th Class Windhawk can attack normally when riding the wolf?
status:confirmed component:core priority:low mode:renewal type:bug
<!-- NOTE: Anything within these brackets will be hidden on the preview of the Issue. --> * **rAthena Hash**: latest <!-- Please specify the rAthena [GitHub hash](https://help.github.com/articles/autolinked-references-and-urls/#commit-shas) on which you encountered this issue. How to get your GitHub Hash: 1. cd your/rAthena/directory/ 2. git rev-parse --short HEAD 3. Copy the resulting hash. --> * **Client Date**: 20211103 <!-- Please specify the client date you used. --> * **Server Mode**: Re <!-- Which mode does your server use: Pre-Renewal or Renewal? --> * **Description of Issue**: Windhawk can attack normally when riding the wolf, is this the offical behaviour? * Result: <!-- Describe the issue that you experienced in detail. --> * Expected Result: <!-- Describe what you would expect to happen in detail. --> * How to Reproduce: <!-- If you have not stated in the description of the result already, please give us a short guide how we can reproduce your issue. -->4th class Windhawk, use WUGRIDER skill, then attack target * Official Information: <!-- If possible, provide information from official servers (kRO or other sources) which prove that the result is wrong. Please take into account that iRO (especially iRO Wiki) is not always the same as kRO. --> <!-- * _NOTE: Make sure you quote ``` `@atcommands` ``` just like this so that you do not tag uninvolved GitHub users!_ --> * **Modifications that may affect results**: <!-- * Please provide any information that could influence the expected result. --> <!-- * This can be either configurations you changed, database values you changed, or even external source modifications. -->
1.0
4th Class Windhawk can attack normally when riding the wolf? - <!-- NOTE: Anything within these brackets will be hidden on the preview of the Issue. --> * **rAthena Hash**: latest <!-- Please specify the rAthena [GitHub hash](https://help.github.com/articles/autolinked-references-and-urls/#commit-shas) on which you encountered this issue. How to get your GitHub Hash: 1. cd your/rAthena/directory/ 2. git rev-parse --short HEAD 3. Copy the resulting hash. --> * **Client Date**: 20211103 <!-- Please specify the client date you used. --> * **Server Mode**: Re <!-- Which mode does your server use: Pre-Renewal or Renewal? --> * **Description of Issue**: Windhawk can attack normally when riding the wolf, is this the offical behaviour? * Result: <!-- Describe the issue that you experienced in detail. --> * Expected Result: <!-- Describe what you would expect to happen in detail. --> * How to Reproduce: <!-- If you have not stated in the description of the result already, please give us a short guide how we can reproduce your issue. -->4th class Windhawk, use WUGRIDER skill, then attack target * Official Information: <!-- If possible, provide information from official servers (kRO or other sources) which prove that the result is wrong. Please take into account that iRO (especially iRO Wiki) is not always the same as kRO. --> <!-- * _NOTE: Make sure you quote ``` `@atcommands` ``` just like this so that you do not tag uninvolved GitHub users!_ --> * **Modifications that may affect results**: <!-- * Please provide any information that could influence the expected result. --> <!-- * This can be either configurations you changed, database values you changed, or even external source modifications. -->
non_test
class windhawk can attack normally when riding the wolf rathena hash latest please specify the rathena on which you encountered this issue how to get your github hash cd your rathena directory git rev parse short head copy the resulting hash client date server mode re description of issue windhawk can attack normally when riding the wolf is this the offical behaviour result expected result how to reproduce class windhawk use wugrider skill then attack target official information modifications that may affect results
0
162,086
12,619,360,474
IssuesEvent
2020-06-13 00:14:42
avidrucker/autofocus-exp
https://api.github.com/repos/avidrucker/autofocus-exp
closed
User can focus in on current todo
awaiting tests clarify no-issue-activity question user story
ala the zen design of https://momentumdash.com/ Also called "focus mode", the user sees simply the current most want to do item, and a prompt (or button) to leave the screen / progress forward. Note: This is not a "screen" per se, but simply a clearing of the console, a display of the current in-progress todo item, and a prompt to hit the enter key to answer the question, "Is there further work left on this todo item? (y/n, or, hit 'q' then 'enter' to quit without saving progress) " where the next "screen" (display) is the return to main menu (prompt). To sum it up: - [x] 1. User sees the console cleared - [x] 2. User sees current in-progress todo item (the cmwtd) - [ ] 3. User can hit any key to indicate they are "done working for now" - [ ] 4A. User can answer whether there is work left to do in the future (y/n/quit) - [ ] 4B. User can quit to indicate that the cmwtd shouldn't be marked as complete - [ ] 5. User is returned to the main menu
1.0
User can focus in on current todo - ala the zen design of https://momentumdash.com/ Also called "focus mode", the user sees simply the current most want to do item, and a prompt (or button) to leave the screen / progress forward. Note: This is not a "screen" per se, but simply a clearing of the console, a display of the current in-progress todo item, and a prompt to hit the enter key to answer the question, "Is there further work left on this todo item? (y/n, or, hit 'q' then 'enter' to quit without saving progress) " where the next "screen" (display) is the return to main menu (prompt). To sum it up: - [x] 1. User sees the console cleared - [x] 2. User sees current in-progress todo item (the cmwtd) - [ ] 3. User can hit any key to indicate they are "done working for now" - [ ] 4A. User can answer whether there is work left to do in the future (y/n/quit) - [ ] 4B. User can quit to indicate that the cmwtd shouldn't be marked as complete - [ ] 5. User is returned to the main menu
test
user can focus in on current todo ala the zen design of also called focus mode the user sees simply the current most want to do item and a prompt or button to leave the screen progress forward note this is not a screen per se but simply a clearing of the console a display of the current in progress todo item and a prompt to hit the enter key to answer the question is there further work left on this todo item y n or hit q then enter to quit without saving progress where the next screen display is the return to main menu prompt to sum it up user sees the console cleared user sees current in progress todo item the cmwtd user can hit any key to indicate they are done working for now user can answer whether there is work left to do in the future y n quit user can quit to indicate that the cmwtd shouldn t be marked as complete user is returned to the main menu
1