Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
101,557
4,119,995,008
IssuesEvent
2016-06-08 16:27:38
buildo/github-workflow-pal
https://api.github.com/repos/buildo/github-workflow-pal
closed
automatize milestone change workflow
feature request low priority
## requirements - our milestone workflow is as follows: - issues to focus on stay in `this week` - all issues that have to be considered soon stay in `next week` - all other issues do not have a milestone assigned - the milestone switch is a mechanical procedure that executes these steps: - move all open issues from `this week` to `next week` - rename `this week` to `2016-02-14` (the "due date" of the milestone) and close it - rename `next week` to `this week` - create a new milestone `next week` with due date set to this week's due date + 7 days ## misc open milestones: ![image](https://cloud.githubusercontent.com/assets/554092/12577356/1c5e94a8-c41a-11e5-9b01-cfe89bc35d88.png) closed milestones: ![image](https://cloud.githubusercontent.com/assets/554092/12577353/17790ff4-c41a-11e5-86b0-bf16a8860d16.png)
1.0
automatize milestone change workflow - ## requirements - our milestone workflow is as follows: - issues to focus on stay in `this week` - all issues that have to be considered soon stay in `next week` - all other issues do not have a milestone assigned - the milestone switch is a mechanical procedure that executes these steps: - move all open issues from `this week` to `next week` - rename `this week` to `2016-02-14` (the "due date" of the milestone) and close it - rename `next week` to `this week` - create a new milestone `next week` with due date set to this week's due date + 7 days ## misc open milestones: ![image](https://cloud.githubusercontent.com/assets/554092/12577356/1c5e94a8-c41a-11e5-9b01-cfe89bc35d88.png) closed milestones: ![image](https://cloud.githubusercontent.com/assets/554092/12577353/17790ff4-c41a-11e5-86b0-bf16a8860d16.png)
non_test
automatize milestone change workflow requirements our milestone workflow is as follows issues to focus on stay in this week all issues that have to be considered soon stay in next week all other issues do not have a milestone assigned the milestone switch is a mechanical procedure that executes these steps move all open issues from this week to next week rename this week to the due date of the milestone and close it rename next week to this week create a new milestone next week with due date set to this week s due date days misc open milestones closed milestones
0
774,854
27,214,708,679
IssuesEvent
2023-02-20 20:14:34
phetsims/papyrus
https://api.github.com/repos/phetsims/papyrus
closed
Add support for creating relationships between programs/papers
priority:1-top
We will need a way to set up relationships between programs/papers. For example, one program represents a value and a second paper controls that value.
1.0
Add support for creating relationships between programs/papers - We will need a way to set up relationships between programs/papers. For example, one program represents a value and a second paper controls that value.
non_test
add support for creating relationships between programs papers we will need a way to set up relationships between programs papers for example one program represents a value and a second paper controls that value
0
111,889
11,745,153,424
IssuesEvent
2020-03-12 09:17:13
inf112-v20/error_brain_not_found
https://api.github.com/repos/inf112-v20/error_brain_not_found
closed
Move one-time methods to BoardLayer
documentation
Move methods that we use once to the boardLayer. Board are starting to get big.
1.0
Move one-time methods to BoardLayer - Move methods that we use once to the boardLayer. Board are starting to get big.
non_test
move one time methods to boardlayer move methods that we use once to the boardlayer board are starting to get big
0
224,861
17,778,752,488
IssuesEvent
2021-08-30 23:33:26
apache/trafficserver
https://api.github.com/repos/apache/trafficserver
closed
microDNS crash causes failures in redirect_actions autest
Tests
Redirect_actions fails with fairly high frequency. Different cases will fail. The immediate cause is a gold file mismatch due to an unexpected HTTP response. Finally tracked it down in an autest failure on my PR this evening that the micro DNS crashed after responding to the last successful test case. Then subsequent DNS requests fail, and ATS cannot perform the appropriate redirect action for the type of address. DNS stream.all.out for this evening's failing case. ``` Domain name: iwillredirect.test. [127.0.0.1] Domain name: testdomain10_0_0_1.test. [10.0.0.1] Domain name: testdomain_fc00__1_.test. [fc00::1] Domain name: testdomain127_1_2_3.test. [127.1.2.3] Domain name: testdomain224_1_2_3.test. [224.1.2.3] Domain name: testdomain_ff42___.test. [ff42::] Domain name: testdomain169_254_0_1.test. [169.254.0.1] Domain name: testdomain_fe80___.test. [fe80::] Domain name: testdomain72_30_35_10.test. [72.30.35.10] Domain name: testdomain_2001_4998_58_1836__10_.test. [2001:4998:58:1836::10] Domain name: testdomain10_217_6_28.test. [10.217.6.28] Domain name: testdomain64_202_188_14.test. [64.202.188.14] Starting DNS on address 127.0.0.1 port 61696... UDP request 2019-08-07 22:21:49.411352 (127.0.0.1 44708): ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 65077 ;; flags: rd; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;iwillredirect.test. IN A ---- Reply: ---- ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 65077 ;; flags: qr aa rd ra; QUERY: 1, ANSWER: 1, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;iwillredirect.test. IN A ;; ANSWER SECTION: iwillredirect.test. 300 IN A 127.0.0.1 UDP request 2019-08-07 22:21:49.436376 (127.0.0.1 44708): ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 55033 ;; flags: rd; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain10_0_0_1.test. IN A ---- Reply: ---- ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 55033 ;; flags: qr aa rd ra; QUERY: 1, ANSWER: 1, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain10_0_0_1.test. IN A ;; ANSWER SECTION: testdomain10_0_0_1.test. 300 IN A 10.0.0.1 UDP request 2019-08-07 22:21:49.576167 (127.0.0.1 44708): ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 24737 ;; flags: rd; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain_fc00__1_.test. IN A ---- Reply: ---- ;; ->>HEADER<<- opcode: QUERY, status: NXDOMAIN, id: 24737 ;; flags: qr aa rd ra; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain_fc00__1_.test. IN A UDP request 2019-08-07 22:21:49.596277 (127.0.0.1 44708): ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 6597 ;; flags: rd; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain_fc00__1_.test. IN AAAA ---- Reply: ---- ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 6597 ;; flags: qr aa rd ra; QUERY: 1, ANSWER: 1, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain_fc00__1_.test. IN AAAA ;; ANSWER SECTION: testdomain_fc00__1_.test. 300 IN AAAA fc00::1 UDP request 2019-08-07 22:21:49.688272 (127.0.0.1 44708): ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 33872 ;; flags: rd; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain127_1_2_3.test. IN A ---- Reply: ---- ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 33872 ;; flags: qr aa rd ra; QUERY: 1, ANSWER: 1, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain127_1_2_3.test. IN A ;; ANSWER SECTION: testdomain127_1_2_3.test. 300 IN A 127.1.2.3 UDP request 2019-08-07 22:21:49.828581 (127.0.0.1 44708): ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 43881 ;; flags: rd; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain224_1_2_3.test. IN A ---- Reply: ---- ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 43881 ;; flags: qr aa rd ra; QUERY: 1, ANSWER: 1, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain224_1_2_3.test. IN A ;; ANSWER SECTION: testdomain224_1_2_3.test. 300 IN A 224.1.2.3 UDP request 2019-08-07 22:21:49.917403 (127.0.0.1 44708): ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 43035 ;; flags: rd; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain_ff42___.test. IN A ---- Reply: ---- ;; ->>HEADER<<- opcode: QUERY, status: NXDOMAIN, id: 43035 ;; flags: qr aa rd ra; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain_ff42___.test. IN A UDP request 2019-08-07 22:21:49.938068 (127.0.0.1 44708): ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 12032 ;; flags: rd; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain_ff42___.test. IN AAAA ---- Reply: ---- ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 12032 ;; flags: qr aa rd ra; QUERY: 1, ANSWER: 1, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain_ff42___.test. IN AAAA ;; ANSWER SECTION: testdomain_ff42___.test. 300 IN AAAA ff42:: UDP request 2019-08-07 22:21:50.039013 (127.0.0.1 44708): ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 60928 ;; flags: rd; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain169_254_0_1.test. IN A ---- Reply: ---- ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 60928 ;; flags: qr aa rd ra; QUERY: 1, ANSWER: 1, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain169_254_0_1.test. IN A ;; ANSWER SECTION: testdomain169_254_0_1.test. 300 IN A 169.254.0.1 Traceback (most recent call last): File "/usr/local/lib/python3.6/site-packages/dnslib/dns.py", line 678, in parse qname = buffer.decode_name() File "/usr/local/lib/python3.6/site-packages/dnslib/label.py", line 255, in decode_name l = self.get(length) File "/usr/local/lib/python3.6/site-packages/dnslib/buffer.py", line 64, in get (self.offset,self.remaining(),length)) dnslib.buffer.BufferError: Not enough bytes [offset=13,remaining=27,requested=116] During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.6/site-packages/microdns/microdns.py", line 59, in handle self.send_data(self.dns_response(data)) File "/usr/local/lib/python3.6/site-packages/microdns/microdns.py", line 70, in dns_response request = DNSRecord.parse(data) File "/usr/local/lib/python3.6/site-packages/dnslib/dns.py", line 106, in parse questions.append(DNSQuestion.parse(buffer)) File "/usr/local/lib/python3.6/site-packages/dnslib/dns.py", line 683, in parse buffer.offset,e)) dnslib.dns.DNSError: Error unpacking DNSQuestion [offset=13]: Not enough bytes [offset=13,remaining=27,requested=116] UDP request 2019-08-07 22:21:50.169075 (127.0.0.1 44708): Traceback (most recent call last): File "/usr/local/lib/python3.6/site-packages/dnslib/dns.py", line 678, in parse qname = buffer.decode_name() File "/usr/local/lib/python3.6/site-packages/dnslib/label.py", line 255, in decode_name l = self.get(length) File "/usr/local/lib/python3.6/site-packages/dnslib/buffer.py", line 64, in get (self.offset,self.remaining(),length)) dnslib.buffer.BufferError: Not enough bytes [offset=13,remaining=27,requested=116] During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.6/site-packages/microdns/microdns.py", line 59, in handle self.send_data(self.dns_response(data)) File "/usr/local/lib/python3.6/site-packages/microdns/microdns.py", line 70, in dns_response request = DNSRecord.parse(data) File "/usr/local/lib/python3.6/site-packages/dnslib/dns.py", line 106, in parse questions.append(DNSQuestion.parse(buffer)) File "/usr/local/lib/python3.6/site-packages/dnslib/dns.py", line 683, in parse buffer.offset,e)) dnslib.dns.DNSError: Error unpacking DNSQuestion [offset=13]: Not enough bytes [offset=13,remaining=27,requested=116] UDP request 2019-08-07 22:22:10.166592 (127.0.0.1 44708): === ^C received, shutting down microDNS === ```
1.0
microDNS crash causes failures in redirect_actions autest - Redirect_actions fails with fairly high frequency. Different cases will fail. The immediate cause is a gold file mismatch due to an unexpected HTTP response. Finally tracked it down in an autest failure on my PR this evening that the micro DNS crashed after responding to the last successful test case. Then subsequent DNS requests fail, and ATS cannot perform the appropriate redirect action for the type of address. DNS stream.all.out for this evening's failing case. ``` Domain name: iwillredirect.test. [127.0.0.1] Domain name: testdomain10_0_0_1.test. [10.0.0.1] Domain name: testdomain_fc00__1_.test. [fc00::1] Domain name: testdomain127_1_2_3.test. [127.1.2.3] Domain name: testdomain224_1_2_3.test. [224.1.2.3] Domain name: testdomain_ff42___.test. [ff42::] Domain name: testdomain169_254_0_1.test. [169.254.0.1] Domain name: testdomain_fe80___.test. [fe80::] Domain name: testdomain72_30_35_10.test. [72.30.35.10] Domain name: testdomain_2001_4998_58_1836__10_.test. [2001:4998:58:1836::10] Domain name: testdomain10_217_6_28.test. [10.217.6.28] Domain name: testdomain64_202_188_14.test. [64.202.188.14] Starting DNS on address 127.0.0.1 port 61696... UDP request 2019-08-07 22:21:49.411352 (127.0.0.1 44708): ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 65077 ;; flags: rd; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;iwillredirect.test. IN A ---- Reply: ---- ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 65077 ;; flags: qr aa rd ra; QUERY: 1, ANSWER: 1, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;iwillredirect.test. IN A ;; ANSWER SECTION: iwillredirect.test. 300 IN A 127.0.0.1 UDP request 2019-08-07 22:21:49.436376 (127.0.0.1 44708): ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 55033 ;; flags: rd; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain10_0_0_1.test. IN A ---- Reply: ---- ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 55033 ;; flags: qr aa rd ra; QUERY: 1, ANSWER: 1, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain10_0_0_1.test. IN A ;; ANSWER SECTION: testdomain10_0_0_1.test. 300 IN A 10.0.0.1 UDP request 2019-08-07 22:21:49.576167 (127.0.0.1 44708): ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 24737 ;; flags: rd; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain_fc00__1_.test. IN A ---- Reply: ---- ;; ->>HEADER<<- opcode: QUERY, status: NXDOMAIN, id: 24737 ;; flags: qr aa rd ra; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain_fc00__1_.test. IN A UDP request 2019-08-07 22:21:49.596277 (127.0.0.1 44708): ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 6597 ;; flags: rd; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain_fc00__1_.test. IN AAAA ---- Reply: ---- ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 6597 ;; flags: qr aa rd ra; QUERY: 1, ANSWER: 1, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain_fc00__1_.test. IN AAAA ;; ANSWER SECTION: testdomain_fc00__1_.test. 300 IN AAAA fc00::1 UDP request 2019-08-07 22:21:49.688272 (127.0.0.1 44708): ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 33872 ;; flags: rd; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain127_1_2_3.test. IN A ---- Reply: ---- ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 33872 ;; flags: qr aa rd ra; QUERY: 1, ANSWER: 1, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain127_1_2_3.test. IN A ;; ANSWER SECTION: testdomain127_1_2_3.test. 300 IN A 127.1.2.3 UDP request 2019-08-07 22:21:49.828581 (127.0.0.1 44708): ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 43881 ;; flags: rd; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain224_1_2_3.test. IN A ---- Reply: ---- ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 43881 ;; flags: qr aa rd ra; QUERY: 1, ANSWER: 1, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain224_1_2_3.test. IN A ;; ANSWER SECTION: testdomain224_1_2_3.test. 300 IN A 224.1.2.3 UDP request 2019-08-07 22:21:49.917403 (127.0.0.1 44708): ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 43035 ;; flags: rd; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain_ff42___.test. IN A ---- Reply: ---- ;; ->>HEADER<<- opcode: QUERY, status: NXDOMAIN, id: 43035 ;; flags: qr aa rd ra; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain_ff42___.test. IN A UDP request 2019-08-07 22:21:49.938068 (127.0.0.1 44708): ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 12032 ;; flags: rd; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain_ff42___.test. IN AAAA ---- Reply: ---- ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 12032 ;; flags: qr aa rd ra; QUERY: 1, ANSWER: 1, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain_ff42___.test. IN AAAA ;; ANSWER SECTION: testdomain_ff42___.test. 300 IN AAAA ff42:: UDP request 2019-08-07 22:21:50.039013 (127.0.0.1 44708): ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 60928 ;; flags: rd; QUERY: 1, ANSWER: 0, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain169_254_0_1.test. IN A ---- Reply: ---- ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 60928 ;; flags: qr aa rd ra; QUERY: 1, ANSWER: 1, AUTHORITY: 0, ADDITIONAL: 0 ;; QUESTION SECTION: ;testdomain169_254_0_1.test. IN A ;; ANSWER SECTION: testdomain169_254_0_1.test. 300 IN A 169.254.0.1 Traceback (most recent call last): File "/usr/local/lib/python3.6/site-packages/dnslib/dns.py", line 678, in parse qname = buffer.decode_name() File "/usr/local/lib/python3.6/site-packages/dnslib/label.py", line 255, in decode_name l = self.get(length) File "/usr/local/lib/python3.6/site-packages/dnslib/buffer.py", line 64, in get (self.offset,self.remaining(),length)) dnslib.buffer.BufferError: Not enough bytes [offset=13,remaining=27,requested=116] During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.6/site-packages/microdns/microdns.py", line 59, in handle self.send_data(self.dns_response(data)) File "/usr/local/lib/python3.6/site-packages/microdns/microdns.py", line 70, in dns_response request = DNSRecord.parse(data) File "/usr/local/lib/python3.6/site-packages/dnslib/dns.py", line 106, in parse questions.append(DNSQuestion.parse(buffer)) File "/usr/local/lib/python3.6/site-packages/dnslib/dns.py", line 683, in parse buffer.offset,e)) dnslib.dns.DNSError: Error unpacking DNSQuestion [offset=13]: Not enough bytes [offset=13,remaining=27,requested=116] UDP request 2019-08-07 22:21:50.169075 (127.0.0.1 44708): Traceback (most recent call last): File "/usr/local/lib/python3.6/site-packages/dnslib/dns.py", line 678, in parse qname = buffer.decode_name() File "/usr/local/lib/python3.6/site-packages/dnslib/label.py", line 255, in decode_name l = self.get(length) File "/usr/local/lib/python3.6/site-packages/dnslib/buffer.py", line 64, in get (self.offset,self.remaining(),length)) dnslib.buffer.BufferError: Not enough bytes [offset=13,remaining=27,requested=116] During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.6/site-packages/microdns/microdns.py", line 59, in handle self.send_data(self.dns_response(data)) File "/usr/local/lib/python3.6/site-packages/microdns/microdns.py", line 70, in dns_response request = DNSRecord.parse(data) File "/usr/local/lib/python3.6/site-packages/dnslib/dns.py", line 106, in parse questions.append(DNSQuestion.parse(buffer)) File "/usr/local/lib/python3.6/site-packages/dnslib/dns.py", line 683, in parse buffer.offset,e)) dnslib.dns.DNSError: Error unpacking DNSQuestion [offset=13]: Not enough bytes [offset=13,remaining=27,requested=116] UDP request 2019-08-07 22:22:10.166592 (127.0.0.1 44708): === ^C received, shutting down microDNS === ```
test
microdns crash causes failures in redirect actions autest redirect actions fails with fairly high frequency different cases will fail the immediate cause is a gold file mismatch due to an unexpected http response finally tracked it down in an autest failure on my pr this evening that the micro dns crashed after responding to the last successful test case then subsequent dns requests fail and ats cannot perform the appropriate redirect action for the type of address dns stream all out for this evening s failing case domain name iwillredirect test domain name test domain name testdomain test domain name test domain name test domain name testdomain test domain name test domain name testdomain test domain name test domain name testdomain test domain name test domain name test starting dns on address port udp request header opcode query status noerror id flags rd query answer authority additional question section iwillredirect test in a reply header opcode query status noerror id flags qr aa rd ra query answer authority additional question section iwillredirect test in a answer section iwillredirect test in a udp request header opcode query status noerror id flags rd query answer authority additional question section test in a reply header opcode query status noerror id flags qr aa rd ra query answer authority additional question section test in a answer section test in a udp request header opcode query status noerror id flags rd query answer authority additional question section testdomain test in a reply header opcode query status nxdomain id flags qr aa rd ra query answer authority additional question section testdomain test in a udp request header opcode query status noerror id flags rd query answer authority additional question section testdomain test in aaaa reply header opcode query status noerror id flags qr aa rd ra query answer authority additional question section testdomain test in aaaa answer section testdomain test in aaaa udp request header opcode query status noerror id flags rd query answer authority additional question section test in a reply header opcode query status noerror id flags qr aa rd ra query answer authority additional question section test in a answer section test in a udp request header opcode query status noerror id flags rd query answer authority additional question section test in a reply header opcode query status noerror id flags qr aa rd ra query answer authority additional question section test in a answer section test in a udp request header opcode query status noerror id flags rd query answer authority additional question section testdomain test in a reply header opcode query status nxdomain id flags qr aa rd ra query answer authority additional question section testdomain test in a udp request header opcode query status noerror id flags rd query answer authority additional question section testdomain test in aaaa reply header opcode query status noerror id flags qr aa rd ra query answer authority additional question section testdomain test in aaaa answer section testdomain test in aaaa udp request header opcode query status noerror id flags rd query answer authority additional question section test in a reply header opcode query status noerror id flags qr aa rd ra query answer authority additional question section test in a answer section test in a traceback most recent call last file usr local lib site packages dnslib dns py line in parse qname buffer decode name file usr local lib site packages dnslib label py line in decode name l self get length file usr local lib site packages dnslib buffer py line in get self offset self remaining length dnslib buffer buffererror not enough bytes during handling of the above exception another exception occurred traceback most recent call last file usr local lib site packages microdns microdns py line in handle self send data self dns response data file usr local lib site packages microdns microdns py line in dns response request dnsrecord parse data file usr local lib site packages dnslib dns py line in parse questions append dnsquestion parse buffer file usr local lib site packages dnslib dns py line in parse buffer offset e dnslib dns dnserror error unpacking dnsquestion not enough bytes udp request traceback most recent call last file usr local lib site packages dnslib dns py line in parse qname buffer decode name file usr local lib site packages dnslib label py line in decode name l self get length file usr local lib site packages dnslib buffer py line in get self offset self remaining length dnslib buffer buffererror not enough bytes during handling of the above exception another exception occurred traceback most recent call last file usr local lib site packages microdns microdns py line in handle self send data self dns response data file usr local lib site packages microdns microdns py line in dns response request dnsrecord parse data file usr local lib site packages dnslib dns py line in parse questions append dnsquestion parse buffer file usr local lib site packages dnslib dns py line in parse buffer offset e dnslib dns dnserror error unpacking dnsquestion not enough bytes udp request c received shutting down microdns
1
804,383
29,485,449,813
IssuesEvent
2023-06-02 09:22:51
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
opened
[Bug]: Invalid cyclic ref error for module level function
Type/Bug Priority/High Team/CompilerFE
### Description When there is a function call for another module and if there is a function definition in same name, the compiler detects a cyclic ref error. ### Steps to Reproduce ```ballerina import ballerina/edi; type EdiDeserialize function (); public function fromEdi835String() { json|error res = edi:fromEdiString("ediText", {delimiters: {segment: "", 'field: "", component: ""}, name: ""}); } final readonly & map<EdiDeserialize> ediDeserializers = { "835": fromEdi835String }; public isolated function fromEdiString() { // function same to `fromEdiString` EdiDeserialize? ediDeserialize = ediDeserializers["ediName"]; if ediDeserialize is () { return (); } _ = ediDeserialize(); } ``` ### Affected Version(s) U5 ### OS, DB, other environment details and versions _No response_ ### Related area -> Compilation ### Related issue(s) (optional) _No response_ ### Suggested label(s) (optional) _No response_ ### Suggested assignee(s) (optional) _No response_
1.0
[Bug]: Invalid cyclic ref error for module level function - ### Description When there is a function call for another module and if there is a function definition in same name, the compiler detects a cyclic ref error. ### Steps to Reproduce ```ballerina import ballerina/edi; type EdiDeserialize function (); public function fromEdi835String() { json|error res = edi:fromEdiString("ediText", {delimiters: {segment: "", 'field: "", component: ""}, name: ""}); } final readonly & map<EdiDeserialize> ediDeserializers = { "835": fromEdi835String }; public isolated function fromEdiString() { // function same to `fromEdiString` EdiDeserialize? ediDeserialize = ediDeserializers["ediName"]; if ediDeserialize is () { return (); } _ = ediDeserialize(); } ``` ### Affected Version(s) U5 ### OS, DB, other environment details and versions _No response_ ### Related area -> Compilation ### Related issue(s) (optional) _No response_ ### Suggested label(s) (optional) _No response_ ### Suggested assignee(s) (optional) _No response_
non_test
invalid cyclic ref error for module level function description when there is a function call for another module and if there is a function definition in same name the compiler detects a cyclic ref error steps to reproduce ballerina import ballerina edi type edideserialize function public function json error res edi fromedistring editext delimiters segment field component name final readonly map edideserializers public isolated function fromedistring function same to fromedistring edideserialize edideserialize edideserializers if edideserialize is return edideserialize affected version s os db other environment details and versions no response related area compilation related issue s optional no response suggested label s optional no response suggested assignee s optional no response
0
24,040
12,016,533,018
IssuesEvent
2020-04-10 16:18:28
edgexfoundry/edgex-go
https://api.github.com/repos/edgexfoundry/edgex-go
closed
Update Services to Disable Remote Logging
1-low core-services enhancement geneva
# 🚀 Feature Request ### Relevant Package <!-- Can you pin-point one or more EdgeX services the are relevant for this feature request? --> <!-- ✍️edit: --> All core and supporting services ### Description Update the core and supporting services to avoid using remote logging and only support standard out by default. ### Describe the solution you'd like Service's configuration to not use remote logging. This can be done by updating the service's configuration `logging` section to something like: ```toml [Logging] EnableRemote = false File = '' ``` We may be able to do this by removing the `Logging` section completely which would allow us to comment the section out with a comment. Just a thought. ### Describe alternatives you've considered N\A
1.0
Update Services to Disable Remote Logging - # 🚀 Feature Request ### Relevant Package <!-- Can you pin-point one or more EdgeX services the are relevant for this feature request? --> <!-- ✍️edit: --> All core and supporting services ### Description Update the core and supporting services to avoid using remote logging and only support standard out by default. ### Describe the solution you'd like Service's configuration to not use remote logging. This can be done by updating the service's configuration `logging` section to something like: ```toml [Logging] EnableRemote = false File = '' ``` We may be able to do this by removing the `Logging` section completely which would allow us to comment the section out with a comment. Just a thought. ### Describe alternatives you've considered N\A
non_test
update services to disable remote logging 🚀 feature request relevant package all core and supporting services description update the core and supporting services to avoid using remote logging and only support standard out by default describe the solution you d like service s configuration to not use remote logging this can be done by updating the service s configuration logging section to something like toml enableremote false file we may be able to do this by removing the logging section completely which would allow us to comment the section out with a comment just a thought describe alternatives you ve considered n a
0
37,632
6,625,202,467
IssuesEvent
2017-09-22 14:37:10
spring-cloud/spring-cloud-stream-binder-kafka
https://api.github.com/repos/spring-cloud/spring-cloud-stream-binder-kafka
opened
Investigate why "Kafka Metrics" section is missing from the generated docs
bug documentation ready
As a user, I'm looking for ["Kafka Metrics"](https://github.com/spring-cloud/spring-cloud-stream-binder-kafka/blob/790b141799fb658e2489f3dee1a090eaa9c1625b/spring-cloud-stream-binder-kafka-docs/src/main/asciidoc/metrics.adoc) section under "Apache Kafka Binder", but I don't see it in Ditmar's reference guides (_snapshot, m1, m2 .._ ). It appears to be included in the [`index.adoc`](https://github.com/spring-cloud/spring-cloud-stream-binder-kafka/blob/790b141799fb658e2489f3dee1a090eaa9c1625b/spring-cloud-stream-binder-kafka-docs/src/main/asciidoc/index.adoc#L29).
1.0
Investigate why "Kafka Metrics" section is missing from the generated docs - As a user, I'm looking for ["Kafka Metrics"](https://github.com/spring-cloud/spring-cloud-stream-binder-kafka/blob/790b141799fb658e2489f3dee1a090eaa9c1625b/spring-cloud-stream-binder-kafka-docs/src/main/asciidoc/metrics.adoc) section under "Apache Kafka Binder", but I don't see it in Ditmar's reference guides (_snapshot, m1, m2 .._ ). It appears to be included in the [`index.adoc`](https://github.com/spring-cloud/spring-cloud-stream-binder-kafka/blob/790b141799fb658e2489f3dee1a090eaa9c1625b/spring-cloud-stream-binder-kafka-docs/src/main/asciidoc/index.adoc#L29).
non_test
investigate why kafka metrics section is missing from the generated docs as a user i m looking for section under apache kafka binder but i don t see it in ditmar s reference guides snapshot it appears to be included in the
0
247,404
26,706,929,377
IssuesEvent
2023-01-27 19:05:06
billmcchesney1/smart-test-framework
https://api.github.com/repos/billmcchesney1/smart-test-framework
closed
CVE-2022-40153 (High) detected in woodstox-core-6.2.0.jar - autoclosed
security vulnerability
## CVE-2022-40153 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>woodstox-core-6.2.0.jar</b></p></summary> <p>Woodstox is a high-performance XML processor that implements Stax (JSR-173), SAX2 and Stax2 APIs</p> <p>Library home page: <a href="https://github.com/FasterXML/woodstox">https://github.com/FasterXML/woodstox</a></p> <p>Path to dependency file: /stf-misc-lib/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.0/woodstox-core-6.2.0.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.0/woodstox-core-6.2.0.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.0/woodstox-core-6.2.0.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.0/woodstox-core-6.2.0.jar</p> <p> Dependency Hierarchy: - stf-misc-1.3-SNAPSHOT.jar (Root Library) - jackson-dataformat-xml-2.11.0.jar - :x: **woodstox-core-6.2.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/smart-test-framework/commit/9c749e8542c6194fbfda15186a19f7be23142106">9c749e8542c6194fbfda15186a19f7be23142106</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Those using Xstream to seralize XML data may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow. This effect may support a denial of service attack. <p>Publish Date: 2022-09-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-40153>CVE-2022-40153</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-09-16</p> <p>Fix Resolution: com.fasterxml.woodstox:woodstox-core:5.4.0,6.4.0</p> </p> </details> <p></p>
True
CVE-2022-40153 (High) detected in woodstox-core-6.2.0.jar - autoclosed - ## CVE-2022-40153 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>woodstox-core-6.2.0.jar</b></p></summary> <p>Woodstox is a high-performance XML processor that implements Stax (JSR-173), SAX2 and Stax2 APIs</p> <p>Library home page: <a href="https://github.com/FasterXML/woodstox">https://github.com/FasterXML/woodstox</a></p> <p>Path to dependency file: /stf-misc-lib/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.0/woodstox-core-6.2.0.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.0/woodstox-core-6.2.0.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.0/woodstox-core-6.2.0.jar,/home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/6.2.0/woodstox-core-6.2.0.jar</p> <p> Dependency Hierarchy: - stf-misc-1.3-SNAPSHOT.jar (Root Library) - jackson-dataformat-xml-2.11.0.jar - :x: **woodstox-core-6.2.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/smart-test-framework/commit/9c749e8542c6194fbfda15186a19f7be23142106">9c749e8542c6194fbfda15186a19f7be23142106</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Those using Xstream to seralize XML data may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow. This effect may support a denial of service attack. <p>Publish Date: 2022-09-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-40153>CVE-2022-40153</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-09-16</p> <p>Fix Resolution: com.fasterxml.woodstox:woodstox-core:5.4.0,6.4.0</p> </p> </details> <p></p>
non_test
cve high detected in woodstox core jar autoclosed cve high severity vulnerability vulnerable library woodstox core jar woodstox is a high performance xml processor that implements stax jsr and apis library home page a href path to dependency file stf misc lib pom xml path to vulnerable library home wss scanner repository com fasterxml woodstox woodstox core woodstox core jar home wss scanner repository com fasterxml woodstox woodstox core woodstox core jar home wss scanner repository com fasterxml woodstox woodstox core woodstox core jar home wss scanner repository com fasterxml woodstox woodstox core woodstox core jar dependency hierarchy stf misc snapshot jar root library jackson dataformat xml jar x woodstox core jar vulnerable library found in head commit a href found in base branch master vulnerability details those using xstream to seralize xml data may be vulnerable to denial of service attacks dos if the parser is running on user supplied input an attacker may supply content that causes the parser to crash by stackoverflow this effect may support a denial of service attack publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution com fasterxml woodstox woodstox core
0
333,739
24,389,390,302
IssuesEvent
2022-10-04 14:11:18
fastapi-mvc/fastapi-mvc
https://api.github.com/repos/fastapi-mvc/fastapi-mvc
closed
Update documentation after moving project template to separate repository
documentation
<!-- Checked checkbox should look like this: [x] --> - [x] I have searched the [issues](https://github.com/fastapi-mvc/fastapi-mvc/issues) of this repo and believe that this is not a duplicate. ## Feature Request <!-- Now feel free to write your idea for improvement. Thanks again 🙌 ❤️ --> Todo: * Update congributing * Document new feature of using a different version or even template repo for `fastapi-mvc new ...` * Document new feature of declaring template as a remote repo in generators * Document how to use Nix with this project * ?
1.0
Update documentation after moving project template to separate repository - <!-- Checked checkbox should look like this: [x] --> - [x] I have searched the [issues](https://github.com/fastapi-mvc/fastapi-mvc/issues) of this repo and believe that this is not a duplicate. ## Feature Request <!-- Now feel free to write your idea for improvement. Thanks again 🙌 ❤️ --> Todo: * Update congributing * Document new feature of using a different version or even template repo for `fastapi-mvc new ...` * Document new feature of declaring template as a remote repo in generators * Document how to use Nix with this project * ?
non_test
update documentation after moving project template to separate repository i have searched the of this repo and believe that this is not a duplicate feature request todo update congributing document new feature of using a different version or even template repo for fastapi mvc new document new feature of declaring template as a remote repo in generators document how to use nix with this project
0
344,828
30,764,109,498
IssuesEvent
2023-07-30 04:11:32
It-s-Spring/Together
https://api.github.com/repos/It-s-Spring/Together
opened
채팅 관련 XML Mapper 작성, DB CRUD 테스트
feature test
## Intro - XML Mapper 세부 작성 ## Description - ChatMemberFileMapper.xml - ChatMemberMessageMapper.xml - ChatOperatorFileMapper.xml - ChatOperatorMessageMapper.xml - ChatRoomMapper.xml ## Linked Issues or Members - 설계: #1
1.0
채팅 관련 XML Mapper 작성, DB CRUD 테스트 - ## Intro - XML Mapper 세부 작성 ## Description - ChatMemberFileMapper.xml - ChatMemberMessageMapper.xml - ChatOperatorFileMapper.xml - ChatOperatorMessageMapper.xml - ChatRoomMapper.xml ## Linked Issues or Members - 설계: #1
test
채팅 관련 xml mapper 작성 db crud 테스트 intro xml mapper 세부 작성 description chatmemberfilemapper xml chatmembermessagemapper xml chatoperatorfilemapper xml chatoperatormessagemapper xml chatroommapper xml linked issues or members 설계
1
103,603
8,923,201,232
IssuesEvent
2019-01-21 15:03:44
symfony/symfony-docs
https://api.github.com/repos/symfony/symfony-docs
closed
Documentation Invalid for Testing Host Matched Routes
Actionable Bug Testing
The section [Testing your Controllers](http://symfony.com/doc/2.8/components/routing/hostname_pattern.html#testing-your-controllers) in the **How to Match a Route Based on the Host** document no longer appears to be valid from Symfony 2.8. The documentation indicates to set the Host HTTP header on the request: ``` $crawler = $client->request( 'GET', '/homepage', array(), array(), array('HTTP_HOST' => 'm.' . $client->getContainer()->getParameter('domain')) ); ``` After recently updating my application from 2.7 to 2.8, I found that this technique no longer works, and seems to have been invalided by the fix for issue [#15398](https://github.com/symfony/symfony/issues/15398). As per the amended test class `src/Symfony/Component/BrowserKit/Tests/ClientTest.php` the correct method now appears to be to provide the full uri to the `$client->request()` method. The updated test request might instead look like this: ``` $uri = 'm.' . $client->getContainer()->getParameter('domain')) . '/homepage'; $crawler = $client->request('GET', $uri); ```
1.0
Documentation Invalid for Testing Host Matched Routes - The section [Testing your Controllers](http://symfony.com/doc/2.8/components/routing/hostname_pattern.html#testing-your-controllers) in the **How to Match a Route Based on the Host** document no longer appears to be valid from Symfony 2.8. The documentation indicates to set the Host HTTP header on the request: ``` $crawler = $client->request( 'GET', '/homepage', array(), array(), array('HTTP_HOST' => 'm.' . $client->getContainer()->getParameter('domain')) ); ``` After recently updating my application from 2.7 to 2.8, I found that this technique no longer works, and seems to have been invalided by the fix for issue [#15398](https://github.com/symfony/symfony/issues/15398). As per the amended test class `src/Symfony/Component/BrowserKit/Tests/ClientTest.php` the correct method now appears to be to provide the full uri to the `$client->request()` method. The updated test request might instead look like this: ``` $uri = 'm.' . $client->getContainer()->getParameter('domain')) . '/homepage'; $crawler = $client->request('GET', $uri); ```
test
documentation invalid for testing host matched routes the section in the how to match a route based on the host document no longer appears to be valid from symfony the documentation indicates to set the host http header on the request crawler client request get homepage array array array http host m client getcontainer getparameter domain after recently updating my application from to i found that this technique no longer works and seems to have been invalided by the fix for issue as per the amended test class src symfony component browserkit tests clienttest php the correct method now appears to be to provide the full uri to the client request method the updated test request might instead look like this uri m client getcontainer getparameter domain homepage crawler client request get uri
1
312,347
26,858,377,337
IssuesEvent
2023-02-03 16:18:05
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix view_tensor.test_view_tensor_arcsin_
PyTorch Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3993969412/jobs/6851165216" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/3993969412/jobs/6851165216" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3993969412/jobs/6851165216" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/3993969412/jobs/6851165216" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_view_tensor.py::test_view_tensor_arcsin_[cpu-ivy.functional.backends.torch-False-False]</summary> 2023-01-24T07:58:12.1335291Z E TypeError: reshape(): argument 'shape' (position 2) must be tuple of ints, not Tensor 2023-01-24T07:58:12.1393503Z E ivy.exceptions.IvyBackendException: torch: reshape: reshape(): argument 'shape' (position 2) must be tuple of ints, not Tensor </details>
1.0
Fix view_tensor.test_view_tensor_arcsin_ - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3993969412/jobs/6851165216" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/3993969412/jobs/6851165216" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3993969412/jobs/6851165216" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/3993969412/jobs/6851165216" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_view_tensor.py::test_view_tensor_arcsin_[cpu-ivy.functional.backends.torch-False-False]</summary> 2023-01-24T07:58:12.1335291Z E TypeError: reshape(): argument 'shape' (position 2) must be tuple of ints, not Tensor 2023-01-24T07:58:12.1393503Z E ivy.exceptions.IvyBackendException: torch: reshape: reshape(): argument 'shape' (position 2) must be tuple of ints, not Tensor </details>
test
fix view tensor test view tensor arcsin tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test frontends test torch test view tensor py test view tensor arcsin e typeerror reshape argument shape position must be tuple of ints not tensor e ivy exceptions ivybackendexception torch reshape reshape argument shape position must be tuple of ints not tensor
1
288,501
24,908,858,372
IssuesEvent
2022-10-29 16:01:46
FrancoSer/ISPC_2do_semestre_2022
https://api.github.com/repos/FrancoSer/ISPC_2do_semestre_2022
reopened
Registrarme como paciente. #US02
scripting User Story Testing
#US02 **Como** paciente **quiero** poder registrarme **para** poder acceder a mi historial clínico - [x] Generar registro de paciente usando CUIL. - [ ] Validación de que los datos sean correctos. (Perfil de paciente > #20 )
1.0
Registrarme como paciente. #US02 - #US02 **Como** paciente **quiero** poder registrarme **para** poder acceder a mi historial clínico - [x] Generar registro de paciente usando CUIL. - [ ] Validación de que los datos sean correctos. (Perfil de paciente > #20 )
test
registrarme como paciente como paciente quiero poder registrarme para poder acceder a mi historial clínico generar registro de paciente usando cuil validación de que los datos sean correctos perfil de paciente
1
234,072
19,093,380,473
IssuesEvent
2021-11-29 14:25:33
pharo-project/pharo
https://api.github.com/repos/pharo-project/pharo
closed
Fleaky FT2 test
Bug Fleaky-test
**Describe the bug** I've seen a fleaky test related to FT2 and taskbar creation **To Reproduce** **Stack trace in the CI** ``` FT2Error: Error embolding glyph [error 20][invalid outline] [64] FreeTypeFace(FT2Face)>>emboldenOutline: [64] FreeTypeSubPixelAntiAliasedGlyphRenderer(FreeTypeGlyphRenderer)>>renderGlyph:depth:subpixelPosition:font: [64] FreeTypeSubPixelAntiAliasedGlyphRenderer(FreeTypeGlyphRenderer)>>glyphOf:colorValue:mono:subpixelPosition:font: [64] [ [64] FreeTypeGlyphRenderer current [64] glyphOf: aCharacter [64] colorValue: aColorValue [64] mono: monoBoolean [64] subpixelPosition: sub [64] font: self] in FreeTypeFont>>glyphOf:colorValue:mono:subpixelPosition: in Block: [... [64] [ [64] charCodeTable := fontTable at: aFreeTypeFont ifAbsentPut:[self dictionaryClass new: 60]. [64] typeTable := charCodeTable at: charCodeInteger ifAbsentPut:[self dictionaryClass new: 10]. [64] entry := typeTable at: typeFlag ifAbsent:[]. [64] entry [64] ifNotNil:[ [64] fifo moveDown: entry. [64] ^entry object]. [64] v := aBlock value. [64] vSize := self sizeOf: v. [64] (maximumSize notNil and:[vSize > maximumSize]) [64] ifTrue:[^v]. [64] used := used + vSize. [64] entry := (self fifoEntryClass new [64] font: aFreeTypeFont; [64] charCode: charCodeInteger; [64] type: typeFlag; [64] object: v; [64] yourself). [64] typeTable at: typeFlag put: entry. [64] fifo addLast: entry. [64] maximumSize ifNotNil:[self shrinkTo: maximumSize]. [64] ^v ] in FreeTypeCache>>atFont:charCode:type:ifAbsentPut: in Block: [ ... [64] BlockClosure>>on:do: [64] [ [64] caught := true. [64] self wait. [64] blockValue := mutuallyExcludedBlock [64] on: Exception [64] do: [ :e | caught ifTrue: [self signal]. [64] caught := false. [64] e pass.]. [64] [64] ] in Semaphore>>criticalReleasingOnError: in Block: [... [64] BlockClosure>>ensure: [64] Semaphore>>criticalReleasingOnError: [64] FreeTypeCache>>atFont:charCode:type:ifAbsentPut: [64] FreeTypeFont>>glyphOf:colorValue:mono:subpixelPosition: [64] FreeTypeFont>>widthOfString:from:to: [64] FreeTypeFont(AbstractFont)>>widthOfString: [64] LogicalFont>>widthOfString: [64] LabelMorph(StringMorph)>>measureContents [64] LabelMorph(StringMorph)>>fitContents [64] LabelMorph(StringMorph)>>font:emphasis: [64] LabelMorph(StringMorph)>>emphasis: [64] TaskbarItemMorph>>addEmphaseTo: [64] TaskbarItemMorph>>initializeFor: [64] PharoLightTheme(UITheme)>>newTaskbarButtonIn:for: [64] StandardWindow(SystemWindow)>>taskbarButtonFor: [64] TaskbarTask>>taskbarButtonFor: [64] [:t | | button | [64] button := t taskbarButtonFor: self. [64] button ifNotNil: [self addMorphBack: button]] in TaskbarMorph>>updateTaskButtons in Block: [:t | | button |... [64] OrderedCollection>>do: [64] TaskbarMorph>>updateTaskButtons [64] TaskbarMorph>>updateTasks [64] [self updateBounds. [64] self updateTasks] in TaskbarMorph>>ownerChanged in Block: [self updateBounds.... [64] BlockClosure>>on:do: [64] TaskbarMorph>>ownerChanged ``` **Version information (please complete the following information, you can get it from Help->about):** - OS: CI failed in unix - Version: Pharo 9.0
1.0
Fleaky FT2 test - **Describe the bug** I've seen a fleaky test related to FT2 and taskbar creation **To Reproduce** **Stack trace in the CI** ``` FT2Error: Error embolding glyph [error 20][invalid outline] [64] FreeTypeFace(FT2Face)>>emboldenOutline: [64] FreeTypeSubPixelAntiAliasedGlyphRenderer(FreeTypeGlyphRenderer)>>renderGlyph:depth:subpixelPosition:font: [64] FreeTypeSubPixelAntiAliasedGlyphRenderer(FreeTypeGlyphRenderer)>>glyphOf:colorValue:mono:subpixelPosition:font: [64] [ [64] FreeTypeGlyphRenderer current [64] glyphOf: aCharacter [64] colorValue: aColorValue [64] mono: monoBoolean [64] subpixelPosition: sub [64] font: self] in FreeTypeFont>>glyphOf:colorValue:mono:subpixelPosition: in Block: [... [64] [ [64] charCodeTable := fontTable at: aFreeTypeFont ifAbsentPut:[self dictionaryClass new: 60]. [64] typeTable := charCodeTable at: charCodeInteger ifAbsentPut:[self dictionaryClass new: 10]. [64] entry := typeTable at: typeFlag ifAbsent:[]. [64] entry [64] ifNotNil:[ [64] fifo moveDown: entry. [64] ^entry object]. [64] v := aBlock value. [64] vSize := self sizeOf: v. [64] (maximumSize notNil and:[vSize > maximumSize]) [64] ifTrue:[^v]. [64] used := used + vSize. [64] entry := (self fifoEntryClass new [64] font: aFreeTypeFont; [64] charCode: charCodeInteger; [64] type: typeFlag; [64] object: v; [64] yourself). [64] typeTable at: typeFlag put: entry. [64] fifo addLast: entry. [64] maximumSize ifNotNil:[self shrinkTo: maximumSize]. [64] ^v ] in FreeTypeCache>>atFont:charCode:type:ifAbsentPut: in Block: [ ... [64] BlockClosure>>on:do: [64] [ [64] caught := true. [64] self wait. [64] blockValue := mutuallyExcludedBlock [64] on: Exception [64] do: [ :e | caught ifTrue: [self signal]. [64] caught := false. [64] e pass.]. [64] [64] ] in Semaphore>>criticalReleasingOnError: in Block: [... [64] BlockClosure>>ensure: [64] Semaphore>>criticalReleasingOnError: [64] FreeTypeCache>>atFont:charCode:type:ifAbsentPut: [64] FreeTypeFont>>glyphOf:colorValue:mono:subpixelPosition: [64] FreeTypeFont>>widthOfString:from:to: [64] FreeTypeFont(AbstractFont)>>widthOfString: [64] LogicalFont>>widthOfString: [64] LabelMorph(StringMorph)>>measureContents [64] LabelMorph(StringMorph)>>fitContents [64] LabelMorph(StringMorph)>>font:emphasis: [64] LabelMorph(StringMorph)>>emphasis: [64] TaskbarItemMorph>>addEmphaseTo: [64] TaskbarItemMorph>>initializeFor: [64] PharoLightTheme(UITheme)>>newTaskbarButtonIn:for: [64] StandardWindow(SystemWindow)>>taskbarButtonFor: [64] TaskbarTask>>taskbarButtonFor: [64] [:t | | button | [64] button := t taskbarButtonFor: self. [64] button ifNotNil: [self addMorphBack: button]] in TaskbarMorph>>updateTaskButtons in Block: [:t | | button |... [64] OrderedCollection>>do: [64] TaskbarMorph>>updateTaskButtons [64] TaskbarMorph>>updateTasks [64] [self updateBounds. [64] self updateTasks] in TaskbarMorph>>ownerChanged in Block: [self updateBounds.... [64] BlockClosure>>on:do: [64] TaskbarMorph>>ownerChanged ``` **Version information (please complete the following information, you can get it from Help->about):** - OS: CI failed in unix - Version: Pharo 9.0
test
fleaky test describe the bug i ve seen a fleaky test related to and taskbar creation to reproduce stack trace in the ci error embolding glyph freetypeface emboldenoutline freetypesubpixelantialiasedglyphrenderer freetypeglyphrenderer renderglyph depth subpixelposition font freetypesubpixelantialiasedglyphrenderer freetypeglyphrenderer glyphof colorvalue mono subpixelposition font freetypeglyphrenderer current glyphof acharacter colorvalue acolorvalue mono monoboolean subpixelposition sub font self in freetypefont glyphof colorvalue mono subpixelposition in block charcodetable fonttable at afreetypefont ifabsentput typetable charcodetable at charcodeinteger ifabsentput entry typetable at typeflag ifabsent entry ifnotnil fifo movedown entry entry object v ablock value vsize self sizeof v maximumsize notnil and iftrue used used vsize entry self fifoentryclass new font afreetypefont charcode charcodeinteger type typeflag object v yourself typetable at typeflag put entry fifo addlast entry maximumsize ifnotnil v in freetypecache atfont charcode type ifabsentput in block blockclosure on do caught true self wait blockvalue mutuallyexcludedblock on exception do caught false e pass in semaphore criticalreleasingonerror in block blockclosure ensure semaphore criticalreleasingonerror freetypecache atfont charcode type ifabsentput freetypefont glyphof colorvalue mono subpixelposition freetypefont widthofstring from to freetypefont abstractfont widthofstring logicalfont widthofstring labelmorph stringmorph measurecontents labelmorph stringmorph fitcontents labelmorph stringmorph font emphasis labelmorph stringmorph emphasis taskbaritemmorph addemphaseto taskbaritemmorph initializefor pharolighttheme uitheme newtaskbarbuttonin for standardwindow systemwindow taskbarbuttonfor taskbartask taskbarbuttonfor t button button t taskbarbuttonfor self button ifnotnil in taskbarmorph updatetaskbuttons in block t button orderedcollection do taskbarmorph updatetaskbuttons taskbarmorph updatetasks self updatebounds self updatetasks in taskbarmorph ownerchanged in block self updatebounds blockclosure on do taskbarmorph ownerchanged version information please complete the following information you can get it from help about os ci failed in unix version pharo
1
90,284
8,231,169,616
IssuesEvent
2018-09-07 15:06:23
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
cmd/cockroach: (unknown) failed under stress
C-test-failure O-robot
SHA: https://github.com/cockroachdb/cockroach/commits/517663fa1489a2f36a593dc9722384a50bfcdbb5 Parameters: ``` TAGS= GOFLAGS= ``` Stress build found a failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=891059&tab=buildLog ``` yes checking whether compiler supports -g3... github.com/cockroachdb/cockroach/vendor/github.com/armon/go-radix github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/gps/paths touch bin/.cpp_protobuf_sources github.com/cockroachdb/cockroach/vendor/github.com/Masterminds/semver github.com/cockroachdb/cockroach/vendor/github.com/boltdb/bolt github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/gps/internal/pb github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/fs github.com/cockroachdb/cockroach/vendor/github.com/nightlyone/lockfile yes github.com/cockroachdb/cockroach/vendor/github.com/sdboyer/constext checking how to run the C preprocessor... github.com/cockroachdb/cockroach/vendor/github.com/Masterminds/vcs github.com/cockroachdb/cockroach/pkg/cmd/protoc-gen-gogoroach github.com/cockroachdb/cockroach/vendor/github.com/pelletier/go-toml github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/gps/pkgtree github.com/cockroachdb/cockroach/vendor/github.com/go-yaml/yaml text/scanner github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/crlfmt cc -E github.com/cockroachdb/cockroach/vendor/github.com/client9/misspell Scanning dependencies of target roach github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/stress github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/go/gcimporter15 [ 2%] Building CXX object CMakeFiles/roach.dir/comparator.cc.o [ 5%] Building CXX object CMakeFiles/roach.dir/batch.cc.o [ 8%] Building CXX object CMakeFiles/roach.dir/db.cc.o [ 11%] Building CXX object CMakeFiles/roach.dir/chunked_buffer.cc.o [ 14%] Building CXX object CMakeFiles/roach.dir/encoding.cc.o github.com/cockroachdb/cockroach/vendor/github.com/jmank88/nuts github.com/cockroachdb/cockroach/vendor/github.com/google/pprof checking for grep that handles long lines and -e... github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/gps /bin/grep checking for egrep... github.com/cockroachdb/cockroach/vendor/github.com/golang/glog /bin/grep -E checking for ANSI C header files... github.com/cockroachdb/cockroach/vendor/github.com/golang/protobuf/protoc-gen-go/plugin github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/go/gcexportdata github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/httprule github.com/cockroachdb/cockroach/vendor/github.com/jteeuwen/go-bindata github.com/cockroachdb/cockroach/vendor/github.com/kisielk/errcheck/internal/errcheck github.com/cockroachdb/cockroach/vendor/github.com/golang/protobuf/protoc-gen-go/generator github.com/cockroachdb/cockroach/vendor/github.com/golang/lint github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cover github.com/cockroachdb/cockroach/vendor/github.com/jteeuwen/go-bindata/go-bindata github.com/cockroachdb/cockroach/vendor/github.com/kisielk/errcheck github.com/cockroachdb/cockroach/vendor/github.com/mattn/goveralls github.com/cockroachdb/cockroach/vendor/github.com/client9/misspell/cmd/misspell github.com/cockroachdb/cockroach/vendor/github.com/golang/lint/golint yes checking for sys/types.h... yes checking for sys/stat.h... find ./pkg -name node_modules -prune -o -type f -name '*.pb.go' -exec rm {} + github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/suffixtree set -e; for dir in ./pkg/acceptance/cluster/ ./pkg/build/ ./pkg/ccl/backupccl/ ./pkg/ccl/baseccl/ ./pkg/ccl/storageccl/engineccl/enginepbccl/ ./pkg/ccl/utilccl/licenseccl/ ./pkg/config/ ./pkg/gossip/ ./pkg/internal/client/ ./pkg/roachpb/ ./pkg/rpc/ ./pkg/server/diagnosticspb/ ./pkg/server/serverpb/ ./pkg/server/status/ ./pkg/settings/cluster/ ./pkg/sql/distsqlrun/ ./pkg/sql/jobs/ ./pkg/sql/pgwire/pgerror/ ./pkg/sql/sqlbase/ ./pkg/sql/stats/ ./pkg/storage/ ./pkg/storage/engine/enginepb/ ./pkg/storage/storagebase/ ./pkg/ts/tspb/ ./pkg/util/ ./pkg/util/hlc/ ./pkg/util/log/ ./pkg/util/protoutil/ ./pkg/util/tracing/; do \ build/werror.sh /go/native/x86_64-pc-linux-gnu/protobuf/protoc -I./pkg:./vendor/github.com/gogo/protobuf:./vendor/github.com/gogo/protobuf/protobuf:./vendor/github.com/coreos:./vendor/github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis --plugin=bin/protoc-gen-gogoroach --gogoroach_out=Mgoogle/api/annotations.proto=github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis/google/api,Mgoogle/protobuf/timestamp.proto=github.com/gogo/protobuf/types,,plugins=grpc,import_prefix=github.com/cockroachdb/cockroach/pkg/:./pkg $dir/*.proto; \ done github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/syntax yes github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/descriptor github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/syntax/golang checking for stdlib.h... github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/job yes github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/output checking for string.h... github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/generator github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/internal/stats github.com/cockroachdb/cockroach/vendor/github.com/wadey/gocovmerge github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/gengateway yes github.com/cockroachdb/cockroach/vendor/github.com/golang/dep github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/feedback checking for memory.h... github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/storage/benchfmt yes github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/benchstat github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway checking for strings.h... github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/imports github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/base github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/cmd/benchstat yes github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cmd/goyacc checking for inttypes.h... github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/glide go/internal/gccgoimporter github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/glock github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/godep yes github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/govend checking for stdint.h... github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/govendor github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/gvt github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/vndr go/internal/gcimporter go/internal/srcimporter yes github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers checking for unistd.h... github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/cmd/dep go/importer yes checking whether byte ordering is bigendian... github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cmd/stringer github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cmd/goimports [ 17%] Building CXX object CMakeFiles/roach.dir/engine.cc.o no checking size of void *... 8 checking size of int... [ 20%] Building CXX object CMakeFiles/roach.dir/merge.cc.o 4 checking size of long... [ 23%] Building CXX object CMakeFiles/roach.dir/mvcc.cc.o 8 checking size of long long... 8 checking size of intmax_t... 8 checking build system type... x86_64-pc-linux-gnu checking host system type... x86_64-pc-linux-gnu checking whether pause instruction is compilable... [ 26%] Building CXX object CMakeFiles/roach.dir/options.cc.o yes checking for ar... ar checking malloc.h usability... yes checking malloc.h presence... yes checking for malloc.h... yes checking whether malloc_usable_size definition can use const argument... no checking for library containing log... -lm checking whether __attribute__ syntax is compilable... yes checking whether compiler supports -fvisibility=hidden... yes checking whether compiler supports -Werror... touch bin/.bootstrap [ 29%] Building CXX object CMakeFiles/roach.dir/snapshot.cc.o yes checking whether compiler supports -herror_on_warning... no checking whether tls_model attribute is compilable... yes checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether alloc_size attribute is compilable... no checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether format(gnu_printf, ...) attribute is compilable... no checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether format(printf, ...) attribute is compilable... yes checking for a BSD-compatible install... /usr/bin/install -c checking for ranlib... ranlib checking for ld... /usr/bin/ld checking for autoconf... /usr/bin/autoconf checking for memalign... yes checking for valloc... yes checking whether compiler supports -O3... yes checking whether compiler supports -funroll-loops... yes checking unwind.h usability... yes checking unwind.h presence... yes checking for unwind.h... yes checking for _Unwind_Backtrace in -lgcc... yes checking configured backtracing method... libgcc checking for sbrk... [ 32%] Building CXX object CMakeFiles/roach.dir/protos/roachpb/data.pb.cc.o yes checking whether utrace(2) is compilable... no checking whether valgrind is compilable... no checking whether a program using __builtin_unreachable is compilable... yes checking whether a program using __builtin_ffsl is compilable... yes checking LG_PAGE... 12 Missing VERSION file, and unable to generate it; creating bogus VERSION checking pthread.h usability... [ 35%] Building CXX object CMakeFiles/roach.dir/protos/roachpb/internal.pb.cc.o yes checking pthread.h presence... yes checking for pthread.h... yes checking for pthread_create in -lpthread... yes checking whether pthread_atfork(3) is compilable... yes checking for library containing clock_gettime... none required checking whether clock_gettime(CLOCK_MONOTONIC_COARSE, ...) is compilable... yes checking whether clock_gettime(CLOCK_MONOTONIC, ...) is compilable... yes checking whether mach_absolute_time() is compilable... no checking whether compiler supports -Werror... yes checking whether syscall(2) is compilable... yes checking for secure_getenv... [ 38%] Building CXX object CMakeFiles/roach.dir/protos/roachpb/metadata.pb.cc.o [ 41%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/mvcc.pb.cc.o yes checking for issetugid... no checking for _malloc_thread_cleanup... no checking for _pthread_mutex_init_calloc_cb... no checking for TLS... yes checking whether C11 atomics is compilable... yes checking whether atomic(9) is compilable... [ 44%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/mvcc3.pb.cc.o [ 47%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/file_registry.pb.cc.o no checking whether Darwin OSAtomic*() is compilable... no checking whether madvise(2) is compilable... yes checking whether madvise(..., MADV_FREE) is compilable... no checking whether madvise(..., MADV_DONTNEED) is compilable... yes checking whether madvise(..., MADV_[NO]HUGEPAGE) is compilable... yes checking whether to force 32-bit __sync_{add,sub}_and_fetch()... no checking whether to force 64-bit __sync_{add,sub}_and_fetch()... no checking for __builtin_clz... [ 50%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/rocksdb.pb.cc.o yes checking whether Darwin os_unfair_lock_*() is compilable... no checking whether Darwin OSSpin*() is compilable... no checking whether glibc malloc hook is compilable... yes checking whether glibc memalign hook is compilable... yes checking whether pthreads adaptive mutexes is compilable... yes checking for stdbool.h that conforms to C99... yes checking for _Bool... yes configure: creating ./config.status config.status: creating Makefile config.status: creating jemalloc.pc config.status: creating doc/html.xsl config.status: creating doc/manpages.xsl [ 52%] Building CXX object CMakeFiles/roach.dir/protos/util/hlc/legacy_timestamp.pb.cc.o config.status: creating doc/jemalloc.xml config.status: creating include/jemalloc/jemalloc_macros.h config.status: creating include/jemalloc/jemalloc_protos.h config.status: creating include/jemalloc/jemalloc_typedefs.h config.status: creating include/jemalloc/internal/jemalloc_internal.h config.status: creating test/test.sh config.status: creating test/include/test/jemalloc_test.h config.status: creating config.stamp config.status: creating bin/jemalloc-config config.status: creating bin/jemalloc.sh config.status: creating bin/jeprof config.status: creating include/jemalloc/jemalloc_defs.h config.status: creating include/jemalloc/internal/jemalloc_internal_defs.h config.status: creating test/include/test/jemalloc_test_defs.h config.status: executing include/jemalloc/internal/private_namespace.h commands [ 55%] Building CXX object CMakeFiles/roach.dir/protos/util/hlc/timestamp.pb.cc.o config.status: executing include/jemalloc/internal/private_unnamespace.h commands [ 58%] Building CXX object CMakeFiles/roach.dir/protos/util/unresolved_addr.pb.cc.o config.status: executing include/jemalloc/internal/public_symbols.txt commands config.status: executing include/jemalloc/internal/public_namespace.h commands config.status: executing include/jemalloc/internal/public_unnamespace.h commands config.status: executing include/jemalloc/internal/size_classes.h commands set -euo pipefail; \ ret=$(cd ./pkg/sql/parser/gen && goyacc -p sql -o sql.go.tmp sql.y); \ if expr "$ret" : ".*conflicts" >/dev/null; then \ echo "$ret"; exit 1; \ fi find ./pkg -name node_modules -prune -o -type f -name '*.pb.gw.go' -exec rm {} + build/werror.sh /go/native/x86_64-pc-linux-gnu/protobuf/protoc -I./pkg:./vendor/github.com/gogo/protobuf:./vendor/github.com/gogo/protobuf/protobuf:./vendor/github.com/coreos:./vendor/github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis --grpc-gateway_out=logtostderr=true,request_context=true:./pkg ./pkg/server/serverpb/admin.proto ./pkg/server/serverpb/status.proto ./pkg/server/serverpb/authentication.proto config.status: executing include/jemalloc/jemalloc_protos_jet.h commands config.status: executing include/jemalloc/jemalloc_rename.h commands build/werror.sh /go/native/x86_64-pc-linux-gnu/protobuf/protoc -I./pkg:./vendor/github.com/gogo/protobuf:./vendor/github.com/gogo/protobuf/protobuf:./vendor/github.com/coreos:./vendor/github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis --grpc-gateway_out=logtostderr=true,request_context=true:./pkg ./pkg/ts/tspb/timeseries.proto sed -i -E 's!golang.org/x/net/context!context!g' ./pkg/server/serverpb/admin.pb.gw.go ./pkg/server/serverpb/status.pb.gw.go ./pkg/server/serverpb/authentication.pb.gw.go ./pkg/ts/tspb/timeseries.pb.gw.go gofmt -s -w ./pkg/server/serverpb/admin.pb.gw.go ./pkg/server/serverpb/status.pb.gw.go ./pkg/server/serverpb/authentication.pb.gw.go ./pkg/ts/tspb/timeseries.pb.gw.go config.status: executing include/jemalloc/jemalloc_mangle.h commands goimports -w ./pkg/server/serverpb/admin.pb.gw.go ./pkg/server/serverpb/status.pb.gw.go ./pkg/server/serverpb/authentication.pb.gw.go ./pkg/ts/tspb/timeseries.pb.gw.go config.status: executing include/jemalloc/jemalloc_mangle_jet.h commands config.status: executing include/jemalloc/jemalloc.h commands =============================================================================== jemalloc version : 0.0.0-0-g0000000000000000000000000000000000000000 library revision : 2 CONFIG : --enable-prof CC : cc CONFIGURE_CFLAGS : -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops SPECIFIED_CFLAGS : EXTRA_CFLAGS : CPPFLAGS : -D_GNU_SOURCE -D_REENTRANT LDFLAGS : EXTRA_LDFLAGS : LIBS : -lm -lgcc -lm -lpthread RPATH_EXTRA : XSLTPROC : false XSLROOT : PREFIX : /usr/local BINDIR : /usr/local/bin DATADIR : /usr/local/share INCLUDEDIR : /usr/local/include LIBDIR : /usr/local/lib MANDIR : /usr/local/share/man touch bin/.gw_protobuf_sources srcroot : /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/ abs_srcroot : /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/ objroot : abs_objroot : /go/native/x86_64-pc-linux-gnu/jemalloc/ JEMALLOC_PREFIX : JEMALLOC_PRIVATE_NAMESPACE : je_ install_suffix : malloc_conf : autogen : 0 cc-silence : 1 debug : 0 code-coverage : 0 stats : 1 prof : 1 prof-libunwind : 0 prof-libgcc : 1 prof-gcc : 0 tcache : 1 thp : 1 fill : 1 utrace : 0 valgrind : 0 xmalloc : 0 munmap : 0 lazy_lock : 0 tls : 1 cache-oblivious : 1 =============================================================================== [ 61%] Linking CXX static library libroach.a [ 82%] Built target roach Scanning dependencies of target roachccl [ 85%] Building CXX object CMakeFiles/roachccl.dir/ccl/db.cc.o sed -i '/import _/d' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go sed -i -E 's!import (fmt|math) "github.com/cockroachdb/cockroach/pkg/(fmt|math)"! !g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go sed -i -E 's!cockroachdb/cockroach/pkg/(etcd)!coreos/\1!g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/arena.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/arena.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/atomic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/atomic.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/jemalloc.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/jemalloc.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/base.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/base.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/bitmap.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/bitmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_dss.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_dss.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_mmap.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_mmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ckh.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ckh.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ctl.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ctl.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/extent.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/extent.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/hash.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/hash.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hiddesed -i -E 's!github.com/cockroachdb/cockroach/pkg/(bytes|encoding/binary|errors|fmt|io|math|github\.com|(google\.)?golang\.org)!\1!g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go sed -i -E 's!golang.org/x/net/context!context!g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go gofmt -s -w ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go n -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/huge.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/huge.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mb.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mb.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mutex.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mutex.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/nstime.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/nstime.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/pages.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/pages.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prng.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prng.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prof.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prof.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/quarantine.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/quarantine.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/rtree.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/rtree.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/stats.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/stats.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/spin.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/spin.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tcache.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tcache.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/t[ 88%] Building CXX object CMakeFiles/roachccl.dir/ccl/key_manager.cc.o [ 91%] Building CXX object CMakeFiles/roachccl.dir/protosccl/ccl/baseccl/encryption_options.pb.cc.o (echo "// Code generated by goyacc. DO NOT EDIT."; \ echo "// GENERATED FILE DO NOT EDIT"; \ cat pkg/sql/parser/gen/sql.go.tmp | \ sed -E 's/^const ([A-Z][_A-Z0-9]*) =.*$/const \1 = lex.\1/g') > pkg/sql/parser/sql.go (echo "// Code generated by make. DO NOT EDIT."; \ echo "// GENERATED FILE DO NOT EDIT"; \ echo; \ echo "package lex"; \ echo; \ grep '^const [A-Z][_A-Z0-9]* ' pkg/sql/parser/gen/sql.go.tmp) > pkg/sql/lex/tokens.go touch bin/.go_protobuf_sources icker.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ticker.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tsd.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tsd.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/util.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/util.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/witness.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/witness.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/jemalloc.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/jemalloc.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/arena.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/arena.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/atomic.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/atomic.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/base.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/base.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/bitmap.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/bitmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_dss.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_dss.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_mmap.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_mmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemal[ 94%] Building CXX object CMakeFiles/roachccl.dir/protosccl/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.cc.o go run pkg/cmd/docgen/{main,funcs}.go functions docs/generated/sql --quiet loc/include -Iinclude -o src/ckh.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ckh.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ctl.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ctl.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/extent.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/extent.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/hash.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/hash.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/huge.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/huge.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mb.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mb.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mutex.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mutex.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/nstime.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/nstime.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/pages.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/pages.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prng.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prng.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prof.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prof.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/quarantine.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/quarantine.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/sar: `u' modifier ignored since `D' is the default (see `U') ar: `u' modifier ignored since `D' is the default (see `U') rc/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/rtree.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/rtree.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/stats.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/stats.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/spin.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/spin.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tcache.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tcache.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ticker.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ticker.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tsd.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tsd.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/util.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/util.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/witness.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/witness.c ar crus lib/libjemalloc.a src/jemalloc.o src/arena.o src/atomic.o src/base.o src/bitmap.o src/chunk.o src/chunk_dss.o src/chunk_mmap.o src/ckh.o src/ctl.o src/extent.o src/hash.o src/huge.o src/mb.o src/mutex.o src/nstime.o src/pages.o src/prng.o src/prof.o src/quarantine.o src/rtree.o src/stats.o src/spin.o src/tcache.o src/ticker.o src/tsd.o src/util.o src/witness.o ar crus lib/libjemalloc_pic.a src/jemalloc.pic.o src/arena.pic.o src/atomic.pic.o src/base.pic.o src/bitmap.pic.o src/chunk.pic.o src/chunk_dss.pic.o src/chunk_mmap.pic.o src/ckh.pic.o src/ctl.pic.o src/extent.pic.o src/hash.pic.o src/huge.pic.o src/mb.pic.o src/mutex.pic.o src/nstime.pic.o src/pages.pic.o src/prng.pic.o src/prof.pic.o src/quarantine.pic.o src/rtree.pic.o src/stats.pic.o src/spin.pic.o src/tcache.pic.o src/ticker.pic.o src/tsd.pic.o src/util.pic.o src/witness.pic.o [ 0%] Built target build_version Scanning dependencies of target rocksdb [ 0%] Building CXX object CMakeFiles/rocksdb.dir/cache/clock_cache.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/cache/sharded_cache.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/cache/lru_cache.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/db/builder.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/db/c.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/db/column_family.cc.o [ 97%] Linking CXX static library libroachccl.a [100%] Built target roachccl [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compacted_db_impl.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_iterator.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_job.cc.o touch bin/.docgen_functions [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_picker.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_picker_universal.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/convenience.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_filesnapshot.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_write.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_compaction_flush.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_files.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_open.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_debug.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_experimental.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_readonly.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/db_info_dumper.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/db_iter.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/dbformat.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/event_helpers.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/experimental.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/external_sst_file_ingestion_job.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/file_indexer.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/flush_job.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/flush_scheduler.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/forward_iterator.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/internal_stats.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/log_reader.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/log_writer.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/malloc_stats.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/managed_iterator.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/memtable.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/memtable_list.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/merge_helper.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/range_del_aggregator.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/repair.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/table_cache.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/table_properties_collector.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/transaction_log_impl.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/version_builder.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/version_edit.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/version_set.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/db/wal_manager.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/db/write_batch.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/db/write_thread.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/env/env.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/env/env_encryption.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/env/mock_env.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/alloc_tracker.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/hash_cuckoo_rep.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/hash_linklist_rep.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/hash_skiplist_rep.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/skiplistrep.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/memtable/vectorrep.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/memtable/write_buffer_manager.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/histogram.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/instrumented_mutex.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/iostats_context.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/perf_context.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/perf_level.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/statistics.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/thread_status_updater.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/thread_status_util.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/thread_status_util_debug.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/options/cf_options.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/db_options.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/options.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/options_helper.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/options_parser.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/table/adaptive_table_factory.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/table/block.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_filter_block.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_table_builder.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_table_factory.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_table_reader.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_builder.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_prefix_index.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/bloom_block.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/cuckoo_table_builder.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/cuckoo_table_factory.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/cuckoo_table_reader.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/format.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/full_filter_block.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/get_context.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/index_builder.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/iterator.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/merging_iterator.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/meta_blocks.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/partitioned_filter_block.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/persistent_cache_helper.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_builder.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_factory.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_index.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_key_coding.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_reader.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/sst_file_writer.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/table_properties.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/two_level_iterator.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/tools/db_bench_tool.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/tools/dump/db_dump_tool.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/tools/ldb_cmd.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/tools/sst_dump_tool.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/util/arena.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/util/auto_roll_logger.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/util/bloom.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/coding.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/comparator.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/concurrent_arena.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/crc32c.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/delete_scheduler.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/dynamic_bloom.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/event_logger.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/file_reader_writer.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/file_util.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/filename.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/hash.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/log_buffer.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/random.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/rate_limiter.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/sst_file_manager_impl.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/status.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/sync_point.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/testutil.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/thread_local.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/threadpool_imp.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/transaction_test_util.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/utilities/backupable/backupable_db.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_db.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_db_impl.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_dump_tool.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_file.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_log_reader.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_log_writer.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_log_format.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/ttl_extractor.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/checkpoint/checkpoint_impl.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/col_buf_decoder.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/col_buf_encoder.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/column_aware_encoding_util.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/date_tiered/date_tiered_db_impl.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/debug.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/document/document_db.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/document/json_document.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/env_timed.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/geodb/geodb_impl.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/memory/memory_util.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/merge_operators/uint64add.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/options/options_util.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/block_cache_tier.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/block_cache_tier_file.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/block_cache_tier_metadata.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/volatile_tier_impl.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/redis/redis_lists.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/simulator_cache/sim_cache.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/spatialdb/spatial_db.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/optimistic_transaction_db_impl.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/optimistic_transaction.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/pessimistic_transaction.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/pessimistic_transaction_db.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/snapshot_checker.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/transaction_base.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/transaction_lock_mgr.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/transaction_util.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/write_prepared_txn.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/ttl/db_ttl_impl.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/write_batch_with_index/write_batch_with_index.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/write_batch_with_index/write_batch_with_index_internal.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/port/port_posix.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/env/env_posix.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/env/io_posix.cc.o [ 89%] Linking CXX static library librocksdb.a [100%] Built target rocksdb go test -v -tags ' make x86_64_pc_linux_gnu' -ldflags ' -X github.com/cockroachdb/cockroach/pkg/build.typ=development -X "github.com/cockroachdb/cockroach/pkg/build.tag=v2.0.5-34-g517663f-dirty" -X "github.com/cockroachdb/cockroach/pkg/build.utcTime=2018/09/07 07:36:16" -X "github.com/cockroachdb/cockroach/pkg/build.rev=517663fa1489a2f36a593dc9722384a50bfcdbb5" -X "github.com/cockroachdb/cockroach/pkg/build.cgoTargetTriple=x86_64-pc-linux-gnu" ' -i github.com/cockroachdb/cockroach/pkg/cmd/cockroach can't load package: package github.com/cockroachdb/cockroach/pkg/cmd/cockroach: cannot find package "github.com/cockroachdb/cockroach/pkg/cmd/cockroach" in any of: /usr/local/go/src/github.com/cockroachdb/cockroach/pkg/cmd/cockroach (from $GOROOT) /go/src/github.com/cockroachdb/cockroach/pkg/cmd/cockroach (from $GOPATH) make: *** [gotestdashi] Error 1 Makefile:756: recipe for target 'gotestdashi' failed ```
1.0
cmd/cockroach: (unknown) failed under stress - SHA: https://github.com/cockroachdb/cockroach/commits/517663fa1489a2f36a593dc9722384a50bfcdbb5 Parameters: ``` TAGS= GOFLAGS= ``` Stress build found a failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=891059&tab=buildLog ``` yes checking whether compiler supports -g3... github.com/cockroachdb/cockroach/vendor/github.com/armon/go-radix github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/gps/paths touch bin/.cpp_protobuf_sources github.com/cockroachdb/cockroach/vendor/github.com/Masterminds/semver github.com/cockroachdb/cockroach/vendor/github.com/boltdb/bolt github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/gps/internal/pb github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/fs github.com/cockroachdb/cockroach/vendor/github.com/nightlyone/lockfile yes github.com/cockroachdb/cockroach/vendor/github.com/sdboyer/constext checking how to run the C preprocessor... github.com/cockroachdb/cockroach/vendor/github.com/Masterminds/vcs github.com/cockroachdb/cockroach/pkg/cmd/protoc-gen-gogoroach github.com/cockroachdb/cockroach/vendor/github.com/pelletier/go-toml github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/gps/pkgtree github.com/cockroachdb/cockroach/vendor/github.com/go-yaml/yaml text/scanner github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/crlfmt cc -E github.com/cockroachdb/cockroach/vendor/github.com/client9/misspell Scanning dependencies of target roach github.com/cockroachdb/cockroach/vendor/github.com/cockroachdb/stress github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/go/gcimporter15 [ 2%] Building CXX object CMakeFiles/roach.dir/comparator.cc.o [ 5%] Building CXX object CMakeFiles/roach.dir/batch.cc.o [ 8%] Building CXX object CMakeFiles/roach.dir/db.cc.o [ 11%] Building CXX object CMakeFiles/roach.dir/chunked_buffer.cc.o [ 14%] Building CXX object CMakeFiles/roach.dir/encoding.cc.o github.com/cockroachdb/cockroach/vendor/github.com/jmank88/nuts github.com/cockroachdb/cockroach/vendor/github.com/google/pprof checking for grep that handles long lines and -e... github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/gps /bin/grep checking for egrep... github.com/cockroachdb/cockroach/vendor/github.com/golang/glog /bin/grep -E checking for ANSI C header files... github.com/cockroachdb/cockroach/vendor/github.com/golang/protobuf/protoc-gen-go/plugin github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/go/gcexportdata github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/httprule github.com/cockroachdb/cockroach/vendor/github.com/jteeuwen/go-bindata github.com/cockroachdb/cockroach/vendor/github.com/kisielk/errcheck/internal/errcheck github.com/cockroachdb/cockroach/vendor/github.com/golang/protobuf/protoc-gen-go/generator github.com/cockroachdb/cockroach/vendor/github.com/golang/lint github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cover github.com/cockroachdb/cockroach/vendor/github.com/jteeuwen/go-bindata/go-bindata github.com/cockroachdb/cockroach/vendor/github.com/kisielk/errcheck github.com/cockroachdb/cockroach/vendor/github.com/mattn/goveralls github.com/cockroachdb/cockroach/vendor/github.com/client9/misspell/cmd/misspell github.com/cockroachdb/cockroach/vendor/github.com/golang/lint/golint yes checking for sys/types.h... yes checking for sys/stat.h... find ./pkg -name node_modules -prune -o -type f -name '*.pb.go' -exec rm {} + github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/suffixtree set -e; for dir in ./pkg/acceptance/cluster/ ./pkg/build/ ./pkg/ccl/backupccl/ ./pkg/ccl/baseccl/ ./pkg/ccl/storageccl/engineccl/enginepbccl/ ./pkg/ccl/utilccl/licenseccl/ ./pkg/config/ ./pkg/gossip/ ./pkg/internal/client/ ./pkg/roachpb/ ./pkg/rpc/ ./pkg/server/diagnosticspb/ ./pkg/server/serverpb/ ./pkg/server/status/ ./pkg/settings/cluster/ ./pkg/sql/distsqlrun/ ./pkg/sql/jobs/ ./pkg/sql/pgwire/pgerror/ ./pkg/sql/sqlbase/ ./pkg/sql/stats/ ./pkg/storage/ ./pkg/storage/engine/enginepb/ ./pkg/storage/storagebase/ ./pkg/ts/tspb/ ./pkg/util/ ./pkg/util/hlc/ ./pkg/util/log/ ./pkg/util/protoutil/ ./pkg/util/tracing/; do \ build/werror.sh /go/native/x86_64-pc-linux-gnu/protobuf/protoc -I./pkg:./vendor/github.com/gogo/protobuf:./vendor/github.com/gogo/protobuf/protobuf:./vendor/github.com/coreos:./vendor/github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis --plugin=bin/protoc-gen-gogoroach --gogoroach_out=Mgoogle/api/annotations.proto=github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis/google/api,Mgoogle/protobuf/timestamp.proto=github.com/gogo/protobuf/types,,plugins=grpc,import_prefix=github.com/cockroachdb/cockroach/pkg/:./pkg $dir/*.proto; \ done github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/syntax yes github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/descriptor github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/syntax/golang checking for stdlib.h... github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/job yes github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/output checking for string.h... github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/generator github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/internal/stats github.com/cockroachdb/cockroach/vendor/github.com/wadey/gocovmerge github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/gengateway yes github.com/cockroachdb/cockroach/vendor/github.com/golang/dep github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/feedback checking for memory.h... github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/storage/benchfmt yes github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/benchstat github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway checking for strings.h... github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/imports github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/base github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/cmd/benchstat yes github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cmd/goyacc checking for inttypes.h... github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/glide go/internal/gccgoimporter github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/glock github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/godep yes github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/govend checking for stdint.h... github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/govendor github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/gvt github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/vndr go/internal/gcimporter go/internal/srcimporter yes github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers checking for unistd.h... github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/cmd/dep go/importer yes checking whether byte ordering is bigendian... github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cmd/stringer github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cmd/goimports [ 17%] Building CXX object CMakeFiles/roach.dir/engine.cc.o no checking size of void *... 8 checking size of int... [ 20%] Building CXX object CMakeFiles/roach.dir/merge.cc.o 4 checking size of long... [ 23%] Building CXX object CMakeFiles/roach.dir/mvcc.cc.o 8 checking size of long long... 8 checking size of intmax_t... 8 checking build system type... x86_64-pc-linux-gnu checking host system type... x86_64-pc-linux-gnu checking whether pause instruction is compilable... [ 26%] Building CXX object CMakeFiles/roach.dir/options.cc.o yes checking for ar... ar checking malloc.h usability... yes checking malloc.h presence... yes checking for malloc.h... yes checking whether malloc_usable_size definition can use const argument... no checking for library containing log... -lm checking whether __attribute__ syntax is compilable... yes checking whether compiler supports -fvisibility=hidden... yes checking whether compiler supports -Werror... touch bin/.bootstrap [ 29%] Building CXX object CMakeFiles/roach.dir/snapshot.cc.o yes checking whether compiler supports -herror_on_warning... no checking whether tls_model attribute is compilable... yes checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether alloc_size attribute is compilable... no checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether format(gnu_printf, ...) attribute is compilable... no checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether format(printf, ...) attribute is compilable... yes checking for a BSD-compatible install... /usr/bin/install -c checking for ranlib... ranlib checking for ld... /usr/bin/ld checking for autoconf... /usr/bin/autoconf checking for memalign... yes checking for valloc... yes checking whether compiler supports -O3... yes checking whether compiler supports -funroll-loops... yes checking unwind.h usability... yes checking unwind.h presence... yes checking for unwind.h... yes checking for _Unwind_Backtrace in -lgcc... yes checking configured backtracing method... libgcc checking for sbrk... [ 32%] Building CXX object CMakeFiles/roach.dir/protos/roachpb/data.pb.cc.o yes checking whether utrace(2) is compilable... no checking whether valgrind is compilable... no checking whether a program using __builtin_unreachable is compilable... yes checking whether a program using __builtin_ffsl is compilable... yes checking LG_PAGE... 12 Missing VERSION file, and unable to generate it; creating bogus VERSION checking pthread.h usability... [ 35%] Building CXX object CMakeFiles/roach.dir/protos/roachpb/internal.pb.cc.o yes checking pthread.h presence... yes checking for pthread.h... yes checking for pthread_create in -lpthread... yes checking whether pthread_atfork(3) is compilable... yes checking for library containing clock_gettime... none required checking whether clock_gettime(CLOCK_MONOTONIC_COARSE, ...) is compilable... yes checking whether clock_gettime(CLOCK_MONOTONIC, ...) is compilable... yes checking whether mach_absolute_time() is compilable... no checking whether compiler supports -Werror... yes checking whether syscall(2) is compilable... yes checking for secure_getenv... [ 38%] Building CXX object CMakeFiles/roach.dir/protos/roachpb/metadata.pb.cc.o [ 41%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/mvcc.pb.cc.o yes checking for issetugid... no checking for _malloc_thread_cleanup... no checking for _pthread_mutex_init_calloc_cb... no checking for TLS... yes checking whether C11 atomics is compilable... yes checking whether atomic(9) is compilable... [ 44%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/mvcc3.pb.cc.o [ 47%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/file_registry.pb.cc.o no checking whether Darwin OSAtomic*() is compilable... no checking whether madvise(2) is compilable... yes checking whether madvise(..., MADV_FREE) is compilable... no checking whether madvise(..., MADV_DONTNEED) is compilable... yes checking whether madvise(..., MADV_[NO]HUGEPAGE) is compilable... yes checking whether to force 32-bit __sync_{add,sub}_and_fetch()... no checking whether to force 64-bit __sync_{add,sub}_and_fetch()... no checking for __builtin_clz... [ 50%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/rocksdb.pb.cc.o yes checking whether Darwin os_unfair_lock_*() is compilable... no checking whether Darwin OSSpin*() is compilable... no checking whether glibc malloc hook is compilable... yes checking whether glibc memalign hook is compilable... yes checking whether pthreads adaptive mutexes is compilable... yes checking for stdbool.h that conforms to C99... yes checking for _Bool... yes configure: creating ./config.status config.status: creating Makefile config.status: creating jemalloc.pc config.status: creating doc/html.xsl config.status: creating doc/manpages.xsl [ 52%] Building CXX object CMakeFiles/roach.dir/protos/util/hlc/legacy_timestamp.pb.cc.o config.status: creating doc/jemalloc.xml config.status: creating include/jemalloc/jemalloc_macros.h config.status: creating include/jemalloc/jemalloc_protos.h config.status: creating include/jemalloc/jemalloc_typedefs.h config.status: creating include/jemalloc/internal/jemalloc_internal.h config.status: creating test/test.sh config.status: creating test/include/test/jemalloc_test.h config.status: creating config.stamp config.status: creating bin/jemalloc-config config.status: creating bin/jemalloc.sh config.status: creating bin/jeprof config.status: creating include/jemalloc/jemalloc_defs.h config.status: creating include/jemalloc/internal/jemalloc_internal_defs.h config.status: creating test/include/test/jemalloc_test_defs.h config.status: executing include/jemalloc/internal/private_namespace.h commands [ 55%] Building CXX object CMakeFiles/roach.dir/protos/util/hlc/timestamp.pb.cc.o config.status: executing include/jemalloc/internal/private_unnamespace.h commands [ 58%] Building CXX object CMakeFiles/roach.dir/protos/util/unresolved_addr.pb.cc.o config.status: executing include/jemalloc/internal/public_symbols.txt commands config.status: executing include/jemalloc/internal/public_namespace.h commands config.status: executing include/jemalloc/internal/public_unnamespace.h commands config.status: executing include/jemalloc/internal/size_classes.h commands set -euo pipefail; \ ret=$(cd ./pkg/sql/parser/gen && goyacc -p sql -o sql.go.tmp sql.y); \ if expr "$ret" : ".*conflicts" >/dev/null; then \ echo "$ret"; exit 1; \ fi find ./pkg -name node_modules -prune -o -type f -name '*.pb.gw.go' -exec rm {} + build/werror.sh /go/native/x86_64-pc-linux-gnu/protobuf/protoc -I./pkg:./vendor/github.com/gogo/protobuf:./vendor/github.com/gogo/protobuf/protobuf:./vendor/github.com/coreos:./vendor/github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis --grpc-gateway_out=logtostderr=true,request_context=true:./pkg ./pkg/server/serverpb/admin.proto ./pkg/server/serverpb/status.proto ./pkg/server/serverpb/authentication.proto config.status: executing include/jemalloc/jemalloc_protos_jet.h commands config.status: executing include/jemalloc/jemalloc_rename.h commands build/werror.sh /go/native/x86_64-pc-linux-gnu/protobuf/protoc -I./pkg:./vendor/github.com/gogo/protobuf:./vendor/github.com/gogo/protobuf/protobuf:./vendor/github.com/coreos:./vendor/github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis --grpc-gateway_out=logtostderr=true,request_context=true:./pkg ./pkg/ts/tspb/timeseries.proto sed -i -E 's!golang.org/x/net/context!context!g' ./pkg/server/serverpb/admin.pb.gw.go ./pkg/server/serverpb/status.pb.gw.go ./pkg/server/serverpb/authentication.pb.gw.go ./pkg/ts/tspb/timeseries.pb.gw.go gofmt -s -w ./pkg/server/serverpb/admin.pb.gw.go ./pkg/server/serverpb/status.pb.gw.go ./pkg/server/serverpb/authentication.pb.gw.go ./pkg/ts/tspb/timeseries.pb.gw.go config.status: executing include/jemalloc/jemalloc_mangle.h commands goimports -w ./pkg/server/serverpb/admin.pb.gw.go ./pkg/server/serverpb/status.pb.gw.go ./pkg/server/serverpb/authentication.pb.gw.go ./pkg/ts/tspb/timeseries.pb.gw.go config.status: executing include/jemalloc/jemalloc_mangle_jet.h commands config.status: executing include/jemalloc/jemalloc.h commands =============================================================================== jemalloc version : 0.0.0-0-g0000000000000000000000000000000000000000 library revision : 2 CONFIG : --enable-prof CC : cc CONFIGURE_CFLAGS : -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops SPECIFIED_CFLAGS : EXTRA_CFLAGS : CPPFLAGS : -D_GNU_SOURCE -D_REENTRANT LDFLAGS : EXTRA_LDFLAGS : LIBS : -lm -lgcc -lm -lpthread RPATH_EXTRA : XSLTPROC : false XSLROOT : PREFIX : /usr/local BINDIR : /usr/local/bin DATADIR : /usr/local/share INCLUDEDIR : /usr/local/include LIBDIR : /usr/local/lib MANDIR : /usr/local/share/man touch bin/.gw_protobuf_sources srcroot : /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/ abs_srcroot : /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/ objroot : abs_objroot : /go/native/x86_64-pc-linux-gnu/jemalloc/ JEMALLOC_PREFIX : JEMALLOC_PRIVATE_NAMESPACE : je_ install_suffix : malloc_conf : autogen : 0 cc-silence : 1 debug : 0 code-coverage : 0 stats : 1 prof : 1 prof-libunwind : 0 prof-libgcc : 1 prof-gcc : 0 tcache : 1 thp : 1 fill : 1 utrace : 0 valgrind : 0 xmalloc : 0 munmap : 0 lazy_lock : 0 tls : 1 cache-oblivious : 1 =============================================================================== [ 61%] Linking CXX static library libroach.a [ 82%] Built target roach Scanning dependencies of target roachccl [ 85%] Building CXX object CMakeFiles/roachccl.dir/ccl/db.cc.o sed -i '/import _/d' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go sed -i -E 's!import (fmt|math) "github.com/cockroachdb/cockroach/pkg/(fmt|math)"! !g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go sed -i -E 's!cockroachdb/cockroach/pkg/(etcd)!coreos/\1!g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/arena.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/arena.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/atomic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/atomic.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/jemalloc.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/jemalloc.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/base.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/base.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/bitmap.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/bitmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_dss.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_dss.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_mmap.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_mmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ckh.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ckh.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ctl.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ctl.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/extent.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/extent.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/hash.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/hash.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hiddesed -i -E 's!github.com/cockroachdb/cockroach/pkg/(bytes|encoding/binary|errors|fmt|io|math|github\.com|(google\.)?golang\.org)!\1!g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go sed -i -E 's!golang.org/x/net/context!context!g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go gofmt -s -w ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go n -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/huge.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/huge.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mb.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mb.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mutex.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mutex.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/nstime.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/nstime.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/pages.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/pages.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prng.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prng.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prof.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prof.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/quarantine.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/quarantine.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/rtree.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/rtree.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/stats.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/stats.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/spin.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/spin.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tcache.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tcache.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/t[ 88%] Building CXX object CMakeFiles/roachccl.dir/ccl/key_manager.cc.o [ 91%] Building CXX object CMakeFiles/roachccl.dir/protosccl/ccl/baseccl/encryption_options.pb.cc.o (echo "// Code generated by goyacc. DO NOT EDIT."; \ echo "// GENERATED FILE DO NOT EDIT"; \ cat pkg/sql/parser/gen/sql.go.tmp | \ sed -E 's/^const ([A-Z][_A-Z0-9]*) =.*$/const \1 = lex.\1/g') > pkg/sql/parser/sql.go (echo "// Code generated by make. DO NOT EDIT."; \ echo "// GENERATED FILE DO NOT EDIT"; \ echo; \ echo "package lex"; \ echo; \ grep '^const [A-Z][_A-Z0-9]* ' pkg/sql/parser/gen/sql.go.tmp) > pkg/sql/lex/tokens.go touch bin/.go_protobuf_sources icker.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ticker.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tsd.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tsd.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/util.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/util.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/witness.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/witness.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/jemalloc.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/jemalloc.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/arena.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/arena.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/atomic.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/atomic.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/base.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/base.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/bitmap.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/bitmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_dss.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_dss.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_mmap.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_mmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemal[ 94%] Building CXX object CMakeFiles/roachccl.dir/protosccl/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.cc.o go run pkg/cmd/docgen/{main,funcs}.go functions docs/generated/sql --quiet loc/include -Iinclude -o src/ckh.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ckh.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ctl.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ctl.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/extent.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/extent.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/hash.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/hash.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/huge.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/huge.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mb.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mb.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mutex.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mutex.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/nstime.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/nstime.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/pages.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/pages.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prng.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prng.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prof.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prof.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/quarantine.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/quarantine.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/sar: `u' modifier ignored since `D' is the default (see `U') ar: `u' modifier ignored since `D' is the default (see `U') rc/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/rtree.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/rtree.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/stats.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/stats.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/spin.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/spin.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tcache.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tcache.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ticker.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ticker.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tsd.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tsd.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/util.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/util.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/witness.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/witness.c ar crus lib/libjemalloc.a src/jemalloc.o src/arena.o src/atomic.o src/base.o src/bitmap.o src/chunk.o src/chunk_dss.o src/chunk_mmap.o src/ckh.o src/ctl.o src/extent.o src/hash.o src/huge.o src/mb.o src/mutex.o src/nstime.o src/pages.o src/prng.o src/prof.o src/quarantine.o src/rtree.o src/stats.o src/spin.o src/tcache.o src/ticker.o src/tsd.o src/util.o src/witness.o ar crus lib/libjemalloc_pic.a src/jemalloc.pic.o src/arena.pic.o src/atomic.pic.o src/base.pic.o src/bitmap.pic.o src/chunk.pic.o src/chunk_dss.pic.o src/chunk_mmap.pic.o src/ckh.pic.o src/ctl.pic.o src/extent.pic.o src/hash.pic.o src/huge.pic.o src/mb.pic.o src/mutex.pic.o src/nstime.pic.o src/pages.pic.o src/prng.pic.o src/prof.pic.o src/quarantine.pic.o src/rtree.pic.o src/stats.pic.o src/spin.pic.o src/tcache.pic.o src/ticker.pic.o src/tsd.pic.o src/util.pic.o src/witness.pic.o [ 0%] Built target build_version Scanning dependencies of target rocksdb [ 0%] Building CXX object CMakeFiles/rocksdb.dir/cache/clock_cache.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/cache/sharded_cache.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/cache/lru_cache.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/db/builder.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/db/c.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/db/column_family.cc.o [ 97%] Linking CXX static library libroachccl.a [100%] Built target roachccl [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compacted_db_impl.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_iterator.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_job.cc.o touch bin/.docgen_functions [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_picker.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_picker_universal.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/convenience.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_filesnapshot.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_write.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_compaction_flush.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_files.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_open.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_debug.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_experimental.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_readonly.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/db_info_dumper.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/db_iter.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/dbformat.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/event_helpers.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/experimental.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/external_sst_file_ingestion_job.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/file_indexer.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/flush_job.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/flush_scheduler.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/forward_iterator.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/internal_stats.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/log_reader.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/log_writer.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/malloc_stats.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/managed_iterator.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/memtable.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/memtable_list.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/merge_helper.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/range_del_aggregator.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/repair.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/table_cache.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/table_properties_collector.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/transaction_log_impl.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/version_builder.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/version_edit.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/version_set.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/db/wal_manager.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/db/write_batch.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/db/write_thread.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/env/env.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/env/env_encryption.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/env/mock_env.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/alloc_tracker.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/hash_cuckoo_rep.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/hash_linklist_rep.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/hash_skiplist_rep.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/skiplistrep.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/memtable/vectorrep.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/memtable/write_buffer_manager.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/histogram.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/instrumented_mutex.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/iostats_context.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/perf_context.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/perf_level.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/statistics.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/thread_status_updater.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/thread_status_util.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/thread_status_util_debug.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/options/cf_options.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/db_options.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/options.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/options_helper.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/options_parser.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/table/adaptive_table_factory.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/table/block.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_filter_block.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_table_builder.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_table_factory.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_table_reader.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_builder.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_prefix_index.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/bloom_block.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/cuckoo_table_builder.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/cuckoo_table_factory.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/cuckoo_table_reader.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/format.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/full_filter_block.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/get_context.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/index_builder.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/iterator.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/merging_iterator.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/meta_blocks.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/partitioned_filter_block.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/persistent_cache_helper.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_builder.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_factory.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_index.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_key_coding.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_reader.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/sst_file_writer.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/table_properties.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/two_level_iterator.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/tools/db_bench_tool.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/tools/dump/db_dump_tool.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/tools/ldb_cmd.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/tools/sst_dump_tool.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/util/arena.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/util/auto_roll_logger.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/util/bloom.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/coding.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/comparator.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/concurrent_arena.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/crc32c.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/delete_scheduler.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/dynamic_bloom.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/event_logger.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/file_reader_writer.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/file_util.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/filename.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/hash.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/log_buffer.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/random.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/rate_limiter.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/sst_file_manager_impl.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/status.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/sync_point.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/testutil.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/thread_local.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/threadpool_imp.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/transaction_test_util.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/utilities/backupable/backupable_db.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_db.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_db_impl.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_dump_tool.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_file.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_log_reader.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_log_writer.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_log_format.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/ttl_extractor.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/checkpoint/checkpoint_impl.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/col_buf_decoder.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/col_buf_encoder.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/column_aware_encoding_util.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/date_tiered/date_tiered_db_impl.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/debug.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/document/document_db.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/document/json_document.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/env_timed.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/geodb/geodb_impl.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/memory/memory_util.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/merge_operators/uint64add.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/options/options_util.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/block_cache_tier.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/block_cache_tier_file.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/block_cache_tier_metadata.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/volatile_tier_impl.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/redis/redis_lists.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/simulator_cache/sim_cache.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/spatialdb/spatial_db.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/optimistic_transaction_db_impl.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/optimistic_transaction.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/pessimistic_transaction.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/pessimistic_transaction_db.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/snapshot_checker.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/transaction_base.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/transaction_lock_mgr.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/transaction_util.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/write_prepared_txn.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/ttl/db_ttl_impl.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/write_batch_with_index/write_batch_with_index.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/write_batch_with_index/write_batch_with_index_internal.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/port/port_posix.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/env/env_posix.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/env/io_posix.cc.o [ 89%] Linking CXX static library librocksdb.a [100%] Built target rocksdb go test -v -tags ' make x86_64_pc_linux_gnu' -ldflags ' -X github.com/cockroachdb/cockroach/pkg/build.typ=development -X "github.com/cockroachdb/cockroach/pkg/build.tag=v2.0.5-34-g517663f-dirty" -X "github.com/cockroachdb/cockroach/pkg/build.utcTime=2018/09/07 07:36:16" -X "github.com/cockroachdb/cockroach/pkg/build.rev=517663fa1489a2f36a593dc9722384a50bfcdbb5" -X "github.com/cockroachdb/cockroach/pkg/build.cgoTargetTriple=x86_64-pc-linux-gnu" ' -i github.com/cockroachdb/cockroach/pkg/cmd/cockroach can't load package: package github.com/cockroachdb/cockroach/pkg/cmd/cockroach: cannot find package "github.com/cockroachdb/cockroach/pkg/cmd/cockroach" in any of: /usr/local/go/src/github.com/cockroachdb/cockroach/pkg/cmd/cockroach (from $GOROOT) /go/src/github.com/cockroachdb/cockroach/pkg/cmd/cockroach (from $GOPATH) make: *** [gotestdashi] Error 1 Makefile:756: recipe for target 'gotestdashi' failed ```
test
cmd cockroach unknown failed under stress sha parameters tags goflags stress build found a failed test yes checking whether compiler supports github com cockroachdb cockroach vendor github com armon go radix github com cockroachdb cockroach vendor github com golang dep gps paths touch bin cpp protobuf sources github com cockroachdb cockroach vendor github com masterminds semver github com cockroachdb cockroach vendor github com boltdb bolt github com cockroachdb cockroach vendor github com golang dep gps internal pb github com cockroachdb cockroach vendor github com golang dep internal fs github com cockroachdb cockroach vendor github com nightlyone lockfile yes github com cockroachdb cockroach vendor github com sdboyer constext checking how to run the c preprocessor github com cockroachdb cockroach vendor github com masterminds vcs github com cockroachdb cockroach pkg cmd protoc gen gogoroach github com cockroachdb cockroach vendor github com pelletier go toml github com cockroachdb cockroach vendor github com golang dep gps pkgtree github com cockroachdb cockroach vendor github com go yaml yaml text scanner github com cockroachdb cockroach vendor github com cockroachdb crlfmt cc e github com cockroachdb cockroach vendor github com misspell scanning dependencies of target roach github com cockroachdb cockroach vendor github com cockroachdb stress github com cockroachdb cockroach vendor golang org x tools go building cxx object cmakefiles roach dir comparator cc o building cxx object cmakefiles roach dir batch cc o building cxx object cmakefiles roach dir db cc o building cxx object cmakefiles roach dir chunked buffer cc o building cxx object cmakefiles roach dir encoding cc o github com cockroachdb cockroach vendor github com nuts github com cockroachdb cockroach vendor github com google pprof checking for grep that handles long lines and e github com cockroachdb cockroach vendor github com golang dep gps bin grep checking for egrep github com cockroachdb cockroach vendor github com golang glog bin grep e checking for ansi c header files github com cockroachdb cockroach vendor github com golang protobuf protoc gen go plugin github com cockroachdb cockroach vendor golang org x tools go gcexportdata github com cockroachdb cockroach vendor github com grpc ecosystem grpc gateway protoc gen grpc gateway httprule github com cockroachdb cockroach vendor github com jteeuwen go bindata github com cockroachdb cockroach vendor github com kisielk errcheck internal errcheck github com cockroachdb cockroach vendor github com golang protobuf protoc gen go generator github com cockroachdb cockroach vendor github com golang lint github com cockroachdb cockroach vendor golang org x tools cover github com cockroachdb cockroach vendor github com jteeuwen go bindata go bindata github com cockroachdb cockroach vendor github com kisielk errcheck github com cockroachdb cockroach vendor github com mattn goveralls github com cockroachdb cockroach vendor github com misspell cmd misspell github com cockroachdb cockroach vendor github com golang lint golint yes checking for sys types h yes checking for sys stat h find pkg name node modules prune o type f name pb go exec rm github com cockroachdb cockroach vendor github com mibk dupl suffixtree set e for dir in pkg acceptance cluster pkg build pkg ccl backupccl pkg ccl baseccl pkg ccl storageccl engineccl enginepbccl pkg ccl utilccl licenseccl pkg config pkg gossip pkg internal client pkg roachpb pkg rpc pkg server diagnosticspb pkg server serverpb pkg server status pkg settings cluster pkg sql distsqlrun pkg sql jobs pkg sql pgwire pgerror pkg sql sqlbase pkg sql stats pkg storage pkg storage engine enginepb pkg storage storagebase pkg ts tspb pkg util pkg util hlc pkg util log pkg util protoutil pkg util tracing do build werror sh go native pc linux gnu protobuf protoc i pkg vendor github com gogo protobuf vendor github com gogo protobuf protobuf vendor github com coreos vendor github com grpc ecosystem grpc gateway third party googleapis plugin bin protoc gen gogoroach gogoroach out mgoogle api annotations proto github com grpc ecosystem grpc gateway third party googleapis google api mgoogle protobuf timestamp proto github com gogo protobuf types plugins grpc import prefix github com cockroachdb cockroach pkg pkg dir proto done github com cockroachdb cockroach vendor github com mibk dupl syntax yes github com cockroachdb cockroach vendor github com grpc ecosystem grpc gateway protoc gen grpc gateway descriptor github com cockroachdb cockroach vendor github com mibk dupl syntax golang checking for stdlib h github com cockroachdb cockroach vendor github com mibk dupl job yes github com cockroachdb cockroach vendor github com mibk dupl output checking for string h github com cockroachdb cockroach vendor github com grpc ecosystem grpc gateway protoc gen grpc gateway generator github com cockroachdb cockroach vendor golang org x perf internal stats github com cockroachdb cockroach vendor github com wadey gocovmerge github com cockroachdb cockroach vendor github com mibk dupl github com cockroachdb cockroach vendor github com grpc ecosystem grpc gateway protoc gen grpc gateway gengateway yes github com cockroachdb cockroach vendor github com golang dep github com cockroachdb cockroach vendor github com golang dep internal feedback checking for memory h github com cockroachdb cockroach vendor golang org x perf storage benchfmt yes github com cockroachdb cockroach vendor golang org x perf benchstat github com cockroachdb cockroach vendor github com grpc ecosystem grpc gateway protoc gen grpc gateway checking for strings h github com cockroachdb cockroach vendor golang org x tools imports github com cockroachdb cockroach vendor github com golang dep internal importers base github com cockroachdb cockroach vendor golang org x perf cmd benchstat yes github com cockroachdb cockroach vendor golang org x tools cmd goyacc checking for inttypes h github com cockroachdb cockroach vendor github com golang dep internal importers glide go internal gccgoimporter github com cockroachdb cockroach vendor github com golang dep internal importers glock github com cockroachdb cockroach vendor github com golang dep internal importers godep yes github com cockroachdb cockroach vendor github com golang dep internal importers govend checking for stdint h github com cockroachdb cockroach vendor github com golang dep internal importers govendor github com cockroachdb cockroach vendor github com golang dep internal importers gvt github com cockroachdb cockroach vendor github com golang dep internal importers vndr go internal gcimporter go internal srcimporter yes github com cockroachdb cockroach vendor github com golang dep internal importers checking for unistd h github com cockroachdb cockroach vendor github com golang dep cmd dep go importer yes checking whether byte ordering is bigendian github com cockroachdb cockroach vendor golang org x tools cmd stringer github com cockroachdb cockroach vendor golang org x tools cmd goimports building cxx object cmakefiles roach dir engine cc o no checking size of void checking size of int building cxx object cmakefiles roach dir merge cc o checking size of long building cxx object cmakefiles roach dir mvcc cc o checking size of long long checking size of intmax t checking build system type pc linux gnu checking host system type pc linux gnu checking whether pause instruction is compilable building cxx object cmakefiles roach dir options cc o yes checking for ar ar checking malloc h usability yes checking malloc h presence yes checking for malloc h yes checking whether malloc usable size definition can use const argument no checking for library containing log lm checking whether attribute syntax is compilable yes checking whether compiler supports fvisibility hidden yes checking whether compiler supports werror touch bin bootstrap building cxx object cmakefiles roach dir snapshot cc o yes checking whether compiler supports herror on warning no checking whether tls model attribute is compilable yes checking whether compiler supports werror yes checking whether compiler supports herror on warning no checking whether alloc size attribute is compilable no checking whether compiler supports werror yes checking whether compiler supports herror on warning no checking whether format gnu printf attribute is compilable no checking whether compiler supports werror yes checking whether compiler supports herror on warning no checking whether format printf attribute is compilable yes checking for a bsd compatible install usr bin install c checking for ranlib ranlib checking for ld usr bin ld checking for autoconf usr bin autoconf checking for memalign yes checking for valloc yes checking whether compiler supports yes checking whether compiler supports funroll loops yes checking unwind h usability yes checking unwind h presence yes checking for unwind h yes checking for unwind backtrace in lgcc yes checking configured backtracing method libgcc checking for sbrk building cxx object cmakefiles roach dir protos roachpb data pb cc o yes checking whether utrace is compilable no checking whether valgrind is compilable no checking whether a program using builtin unreachable is compilable yes checking whether a program using builtin ffsl is compilable yes checking lg page missing version file and unable to generate it creating bogus version checking pthread h usability building cxx object cmakefiles roach dir protos roachpb internal pb cc o yes checking pthread h presence yes checking for pthread h yes checking for pthread create in lpthread yes checking whether pthread atfork is compilable yes checking for library containing clock gettime none required checking whether clock gettime clock monotonic coarse is compilable yes checking whether clock gettime clock monotonic is compilable yes checking whether mach absolute time is compilable no checking whether compiler supports werror yes checking whether syscall is compilable yes checking for secure getenv building cxx object cmakefiles roach dir protos roachpb metadata pb cc o building cxx object cmakefiles roach dir protos storage engine enginepb mvcc pb cc o yes checking for issetugid no checking for malloc thread cleanup no checking for pthread mutex init calloc cb no checking for tls yes checking whether atomics is compilable yes checking whether atomic is compilable building cxx object cmakefiles roach dir protos storage engine enginepb pb cc o building cxx object cmakefiles roach dir protos storage engine enginepb file registry pb cc o no checking whether darwin osatomic is compilable no checking whether madvise is compilable yes checking whether madvise madv free is compilable no checking whether madvise madv dontneed is compilable yes checking whether madvise madv hugepage is compilable yes checking whether to force bit sync add sub and fetch no checking whether to force bit sync add sub and fetch no checking for builtin clz building cxx object cmakefiles roach dir protos storage engine enginepb rocksdb pb cc o yes checking whether darwin os unfair lock is compilable no checking whether darwin osspin is compilable no checking whether glibc malloc hook is compilable yes checking whether glibc memalign hook is compilable yes checking whether pthreads adaptive mutexes is compilable yes checking for stdbool h that conforms to yes checking for bool yes configure creating config status config status creating makefile config status creating jemalloc pc config status creating doc html xsl config status creating doc manpages xsl building cxx object cmakefiles roach dir protos util hlc legacy timestamp pb cc o config status creating doc jemalloc xml config status creating include jemalloc jemalloc macros h config status creating include jemalloc jemalloc protos h config status creating include jemalloc jemalloc typedefs h config status creating include jemalloc internal jemalloc internal h config status creating test test sh config status creating test include test jemalloc test h config status creating config stamp config status creating bin jemalloc config config status creating bin jemalloc sh config status creating bin jeprof config status creating include jemalloc jemalloc defs h config status creating include jemalloc internal jemalloc internal defs h config status creating test include test jemalloc test defs h config status executing include jemalloc internal private namespace h commands building cxx object cmakefiles roach dir protos util hlc timestamp pb cc o config status executing include jemalloc internal private unnamespace h commands building cxx object cmakefiles roach dir protos util unresolved addr pb cc o config status executing include jemalloc internal public symbols txt commands config status executing include jemalloc internal public namespace h commands config status executing include jemalloc internal public unnamespace h commands config status executing include jemalloc internal size classes h commands set euo pipefail ret cd pkg sql parser gen goyacc p sql o sql go tmp sql y if expr ret conflicts dev null then echo ret exit fi find pkg name node modules prune o type f name pb gw go exec rm build werror sh go native pc linux gnu protobuf protoc i pkg vendor github com gogo protobuf vendor github com gogo protobuf protobuf vendor github com coreos vendor github com grpc ecosystem grpc gateway third party googleapis grpc gateway out logtostderr true request context true pkg pkg server serverpb admin proto pkg server serverpb status proto pkg server serverpb authentication proto config status executing include jemalloc jemalloc protos jet h commands config status executing include jemalloc jemalloc rename h commands build werror sh go native pc linux gnu protobuf protoc i pkg vendor github com gogo protobuf vendor github com gogo protobuf protobuf vendor github com coreos vendor github com grpc ecosystem grpc gateway third party googleapis grpc gateway out logtostderr true request context true pkg pkg ts tspb timeseries proto sed i e s golang org x net context context g pkg server serverpb admin pb gw go pkg server serverpb status pb gw go pkg server serverpb authentication pb gw go pkg ts tspb timeseries pb gw go gofmt s w pkg server serverpb admin pb gw go pkg server serverpb status pb gw go pkg server serverpb authentication pb gw go pkg ts tspb timeseries pb gw go config status executing include jemalloc jemalloc mangle h commands goimports w pkg server serverpb admin pb gw go pkg server serverpb status pb gw go pkg server serverpb authentication pb gw go pkg ts tspb timeseries pb gw go config status executing include jemalloc jemalloc mangle jet h commands config status executing include jemalloc jemalloc h commands jemalloc version library revision config enable prof cc cc configure cflags std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops specified cflags extra cflags cppflags d gnu source d reentrant ldflags extra ldflags libs lm lgcc lm lpthread rpath extra xsltproc false xslroot prefix usr local bindir usr local bin datadir usr local share includedir usr local include libdir usr local lib mandir usr local share man touch bin gw protobuf sources srcroot go src github com cockroachdb cockroach c deps jemalloc abs srcroot go src github com cockroachdb cockroach c deps jemalloc objroot abs objroot go native pc linux gnu jemalloc jemalloc prefix jemalloc private namespace je install suffix malloc conf autogen cc silence debug code coverage stats prof prof libunwind prof libgcc prof gcc tcache thp fill utrace valgrind xmalloc munmap lazy lock tls cache oblivious linking cxx static library libroach a built target roach scanning dependencies of target roachccl building cxx object cmakefiles roachccl dir ccl db cc o sed i import d pkg acceptance cluster testconfig pb go pkg build info pb go pkg ccl backupccl backup pb go pkg ccl baseccl encryption options pb go pkg ccl storageccl engineccl enginepbccl key registry pb go pkg ccl utilccl licenseccl license pb go pkg config system pb go pkg config zone pb go pkg gossip gossip pb go pkg internal client lease pb go pkg roachpb api pb go pkg roachpb app stats pb go pkg roachpb csv pb go pkg roachpb data pb go pkg roachpb errors pb go pkg roachpb internal pb go pkg roachpb internal raft pb go pkg roachpb metadata pb go pkg rpc heartbeat pb go pkg server diagnosticspb diagnostics pb go pkg server serverpb admin pb go pkg server serverpb authentication pb go pkg server serverpb init pb go pkg server serverpb status pb go pkg server status status pb go pkg settings cluster cluster version pb go pkg sql distsqlrun api pb go pkg sql distsqlrun data pb go pkg sql distsqlrun processors pb go pkg sql jobs jobs pb go pkg sql pgwire pgerror errors pb go pkg sql sqlbase encoded datum pb go pkg sql sqlbase join type pb go pkg sql sqlbase privilege pb go pkg sql sqlbase structured pb go pkg sql stats histogram pb go pkg storage api pb go pkg storage engine enginepb file registry pb go pkg storage engine enginepb mvcc pb go pkg storage engine enginepb pb go pkg storage engine enginepb rocksdb pb go pkg storage lease status pb go pkg storage liveness pb go pkg storage log pb go pkg storage raft pb go pkg storage storagebase proposer kv pb go pkg storage storagebase state pb go pkg ts tspb timeseries pb go pkg util hlc legacy timestamp pb go pkg util hlc timestamp pb go pkg util log log pb go pkg util protoutil clone pb go pkg util tracing recorded span pb go pkg util unresolved addr pb go sed i e s import fmt math github com cockroachdb cockroach pkg fmt math g pkg acceptance cluster testconfig pb go pkg build info pb go pkg ccl backupccl backup pb go pkg ccl baseccl encryption options pb go pkg ccl storageccl engineccl enginepbccl key registry pb go pkg ccl utilccl licenseccl license pb go pkg config system pb go pkg config zone pb go pkg gossip gossip pb go pkg internal client lease pb go pkg roachpb api pb go pkg roachpb app stats pb go pkg roachpb csv pb go pkg roachpb data pb go pkg roachpb errors pb go pkg roachpb internal pb go pkg roachpb internal raft pb go pkg roachpb metadata pb go pkg rpc heartbeat pb go pkg server diagnosticspb diagnostics pb go pkg server serverpb admin pb go pkg server serverpb authentication pb go pkg server serverpb init pb go pkg server serverpb status pb go pkg server status status pb go pkg settings cluster cluster version pb go pkg sql distsqlrun api pb go pkg sql distsqlrun data pb go pkg sql distsqlrun processors pb go pkg sql jobs jobs pb go pkg sql pgwire pgerror errors pb go pkg sql sqlbase encoded datum pb go pkg sql sqlbase join type pb go pkg sql sqlbase privilege pb go pkg sql sqlbase structured pb go pkg sql stats histogram pb go pkg storage api pb go pkg storage engine enginepb file registry pb go pkg storage engine enginepb mvcc pb go pkg storage engine enginepb pb go pkg storage engine enginepb rocksdb pb go pkg storage lease status pb go pkg storage liveness pb go pkg storage log pb go pkg storage raft pb go pkg storage storagebase proposer kv pb go pkg storage storagebase state pb go pkg ts tspb timeseries pb go pkg util hlc legacy timestamp pb go pkg util hlc timestamp pb go pkg util log log pb go pkg util protoutil clone pb go pkg util tracing recorded span pb go pkg util unresolved addr pb go sed i e s cockroachdb cockroach pkg etcd coreos g pkg acceptance cluster testconfig pb go pkg build info pb go pkg ccl backupccl backup pb go pkg ccl baseccl encryption options pb go pkg ccl storageccl engineccl enginepbccl key registry pb go pkg ccl utilccl licenseccl license pb go pkg config system pb go pkg config zone pb go pkg gossip gossip pb go pkg internal client lease pb go pkg roachpb api pb go pkg roachpb app stats pb go pkg roachpb csv pb go pkg roachpb data pb go pkg roachpb errors pb go pkg roachpb internal pb go pkg roachpb internal raft pb go pkg roachpb metadata pb go pkg rpc heartbeat pb go pkg server diagnosticspb diagnostics pb go pkg server serverpb admin pb go pkg server serverpb authentication pb go pkg server serverpb init pb go pkg server serverpb status pb go pkg server status status pb go pkg settings cluster cluster version pb go pkg sql distsqlrun api pb go pkg sql distsqlrun data pb go pkg sql distsqlrun processors pb go pkg sql jobs jobs pb go pkg sql pgwire pgerror errors pb go pkg sql sqlbase encoded datum pb go pkg sql sqlbase join type pb go pkg sql sqlbase privilege pb go pkg sql sqlbase structured pb go pkg sql stats histogram pb go pkg storage api pb go pkg storage engine enginepb file registry pb go pkg storage engine enginepb mvcc pb go pkg storage engine enginepb pb go pkg storage engine enginepb rocksdb pb go pkg storage lease status pb go pkg storage liveness pb go pkg storage log pb go pkg storage raft pb go pkg storage storagebase proposer kv pb go pkg storage storagebase state pb go pkg ts tspb timeseries pb go pkg util hlc legacy timestamp pb go pkg util hlc timestamp pb go pkg util log log pb go pkg util protoutil clone pb go pkg util tracing recorded span pb go pkg util unresolved addr pb go cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src arena o go src github com cockroachdb cockroach c deps jemalloc src arena c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src atomic o go src github com cockroachdb cockroach c deps jemalloc src atomic c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src jemalloc o go src github com cockroachdb cockroach c deps jemalloc src jemalloc c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src base o go src github com cockroachdb cockroach c deps jemalloc src base c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src bitmap o go src github com cockroachdb cockroach c deps jemalloc src bitmap c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src chunk o go src github com cockroachdb cockroach c deps jemalloc src chunk c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src chunk dss o go src github com cockroachdb cockroach c deps jemalloc src chunk dss c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src chunk mmap o go src github com cockroachdb cockroach c deps jemalloc src chunk mmap c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src ckh o go src github com cockroachdb cockroach c deps jemalloc src ckh c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src ctl o go src github com cockroachdb cockroach c deps jemalloc src ctl c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src extent o go src github com cockroachdb cockroach c deps jemalloc src extent c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src hash o go src github com cockroachdb cockroach c deps jemalloc src hash c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hiddesed i e s github com cockroachdb cockroach pkg bytes encoding binary errors fmt io math github com google golang org g pkg acceptance cluster testconfig pb go pkg build info pb go pkg ccl backupccl backup pb go pkg ccl baseccl encryption options pb go pkg ccl storageccl engineccl enginepbccl key registry pb go pkg ccl utilccl licenseccl license pb go pkg config system pb go pkg config zone pb go pkg gossip gossip pb go pkg internal client lease pb go pkg roachpb api pb go pkg roachpb app stats pb go pkg roachpb csv pb go pkg roachpb data pb go pkg roachpb errors pb go pkg roachpb internal pb go pkg roachpb internal raft pb go pkg roachpb metadata pb go pkg rpc heartbeat pb go pkg server diagnosticspb diagnostics pb go pkg server serverpb admin pb go pkg server serverpb authentication pb go pkg server serverpb init pb go pkg server serverpb status pb go pkg server status status pb go pkg settings cluster cluster version pb go pkg sql distsqlrun api pb go pkg sql distsqlrun data pb go pkg sql distsqlrun processors pb go pkg sql jobs jobs pb go pkg sql pgwire pgerror errors pb go pkg sql sqlbase encoded datum pb go pkg sql sqlbase join type pb go pkg sql sqlbase privilege pb go pkg sql sqlbase structured pb go pkg sql stats histogram pb go pkg storage api pb go pkg storage engine enginepb file registry pb go pkg storage engine enginepb mvcc pb go pkg storage engine enginepb pb go pkg storage engine enginepb rocksdb pb go pkg storage lease status pb go pkg storage liveness pb go pkg storage log pb go pkg storage raft pb go pkg storage storagebase proposer kv pb go pkg storage storagebase state pb go pkg ts tspb timeseries pb go pkg util hlc legacy timestamp pb go pkg util hlc timestamp pb go pkg util log log pb go pkg util protoutil clone pb go pkg util tracing recorded span pb go pkg util unresolved addr pb go sed i e s golang org x net context context g pkg acceptance cluster testconfig pb go pkg build info pb go pkg ccl backupccl backup pb go pkg ccl baseccl encryption options pb go pkg ccl storageccl engineccl enginepbccl key registry pb go pkg ccl utilccl licenseccl license pb go pkg config system pb go pkg config zone pb go pkg gossip gossip pb go pkg internal client lease pb go pkg roachpb api pb go pkg roachpb app stats pb go pkg roachpb csv pb go pkg roachpb data pb go pkg roachpb errors pb go pkg roachpb internal pb go pkg roachpb internal raft pb go pkg roachpb metadata pb go pkg rpc heartbeat pb go pkg server diagnosticspb diagnostics pb go pkg server serverpb admin pb go pkg server serverpb authentication pb go pkg server serverpb init pb go pkg server serverpb status pb go pkg server status status pb go pkg settings cluster cluster version pb go pkg sql distsqlrun api pb go pkg sql distsqlrun data pb go pkg sql distsqlrun processors pb go pkg sql jobs jobs pb go pkg sql pgwire pgerror errors pb go pkg sql sqlbase encoded datum pb go pkg sql sqlbase join type pb go pkg sql sqlbase privilege pb go pkg sql sqlbase structured pb go pkg sql stats histogram pb go pkg storage api pb go pkg storage engine enginepb file registry pb go pkg storage engine enginepb mvcc pb go pkg storage engine enginepb pb go pkg storage engine enginepb rocksdb pb go pkg storage lease status pb go pkg storage liveness pb go pkg storage log pb go pkg storage raft pb go pkg storage storagebase proposer kv pb go pkg storage storagebase state pb go pkg ts tspb timeseries pb go pkg util hlc legacy timestamp pb go pkg util hlc timestamp pb go pkg util log log pb go pkg util protoutil clone pb go pkg util tracing recorded span pb go pkg util unresolved addr pb go gofmt s w pkg acceptance cluster testconfig pb go pkg build info pb go pkg ccl backupccl backup pb go pkg ccl baseccl encryption options pb go pkg ccl storageccl engineccl enginepbccl key registry pb go pkg ccl utilccl licenseccl license pb go pkg config system pb go pkg config zone pb go pkg gossip gossip pb go pkg internal client lease pb go pkg roachpb api pb go pkg roachpb app stats pb go pkg roachpb csv pb go pkg roachpb data pb go pkg roachpb errors pb go pkg roachpb internal pb go pkg roachpb internal raft pb go pkg roachpb metadata pb go pkg rpc heartbeat pb go pkg server diagnosticspb diagnostics pb go pkg server serverpb admin pb go pkg server serverpb authentication pb go pkg server serverpb init pb go pkg server serverpb status pb go pkg server status status pb go pkg settings cluster cluster version pb go pkg sql distsqlrun api pb go pkg sql distsqlrun data pb go pkg sql distsqlrun processors pb go pkg sql jobs jobs pb go pkg sql pgwire pgerror errors pb go pkg sql sqlbase encoded datum pb go pkg sql sqlbase join type pb go pkg sql sqlbase privilege pb go pkg sql sqlbase structured pb go pkg sql stats histogram pb go pkg storage api pb go pkg storage engine enginepb file registry pb go pkg storage engine enginepb mvcc pb go pkg storage engine enginepb pb go pkg storage engine enginepb rocksdb pb go pkg storage lease status pb go pkg storage liveness pb go pkg storage log pb go pkg storage raft pb go pkg storage storagebase proposer kv pb go pkg storage storagebase state pb go pkg ts tspb timeseries pb go pkg util hlc legacy timestamp pb go pkg util hlc timestamp pb go pkg util log log pb go pkg util protoutil clone pb go pkg util tracing recorded span pb go pkg util unresolved addr pb go n funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src huge o go src github com cockroachdb cockroach c deps jemalloc src huge c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src mb o go src github com cockroachdb cockroach c deps jemalloc src mb c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src mutex o go src github com cockroachdb cockroach c deps jemalloc src mutex c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src nstime o go src github com cockroachdb cockroach c deps jemalloc src nstime c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src pages o go src github com cockroachdb cockroach c deps jemalloc src pages c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src prng o go src github com cockroachdb cockroach c deps jemalloc src prng c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src prof o go src github com cockroachdb cockroach c deps jemalloc src prof c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src quarantine o go src github com cockroachdb cockroach c deps jemalloc src quarantine c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src rtree o go src github com cockroachdb cockroach c deps jemalloc src rtree c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src stats o go src github com cockroachdb cockroach c deps jemalloc src stats c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src spin o go src github com cockroachdb cockroach c deps jemalloc src spin c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src tcache o go src github com cockroachdb cockroach c deps jemalloc src tcache c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src t building cxx object cmakefiles roachccl dir ccl key manager cc o building cxx object cmakefiles roachccl dir protosccl ccl baseccl encryption options pb cc o echo code generated by goyacc do not edit echo generated file do not edit cat pkg sql parser gen sql go tmp sed e s const const lex g pkg sql parser sql go echo code generated by make do not edit echo generated file do not edit echo echo package lex echo grep const pkg sql parser gen sql go tmp pkg sql lex tokens go touch bin go protobuf sources icker o go src github com cockroachdb cockroach c deps jemalloc src ticker c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src tsd o go src github com cockroachdb cockroach c deps jemalloc src tsd c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src util o go src github com cockroachdb cockroach c deps jemalloc src util c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src witness o go src github com cockroachdb cockroach c deps jemalloc src witness c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src jemalloc pic o go src github com cockroachdb cockroach c deps jemalloc src jemalloc c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src arena pic o go src github com cockroachdb cockroach c deps jemalloc src arena c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src atomic pic o go src github com cockroachdb cockroach c deps jemalloc src atomic c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src base pic o go src github com cockroachdb cockroach c deps jemalloc src base c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src bitmap pic o go src github com cockroachdb cockroach c deps jemalloc src bitmap c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src chunk pic o go src github com cockroachdb cockroach c deps jemalloc src chunk c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src chunk dss pic o go src github com cockroachdb cockroach c deps jemalloc src chunk dss c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src chunk mmap pic o go src github com cockroachdb cockroach c deps jemalloc src chunk mmap c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemal building cxx object cmakefiles roachccl dir protosccl ccl storageccl engineccl enginepbccl key registry pb cc o go run pkg cmd docgen main funcs go functions docs generated sql quiet loc include iinclude o src ckh pic o go src github com cockroachdb cockroach c deps jemalloc src ckh c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src ctl pic o go src github com cockroachdb cockroach c deps jemalloc src ctl c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src extent pic o go src github com cockroachdb cockroach c deps jemalloc src extent c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src hash pic o go src github com cockroachdb cockroach c deps jemalloc src hash c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src huge pic o go src github com cockroachdb cockroach c deps jemalloc src huge c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src mb pic o go src github com cockroachdb cockroach c deps jemalloc src mb c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src mutex pic o go src github com cockroachdb cockroach c deps jemalloc src mutex c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src nstime pic o go src github com cockroachdb cockroach c deps jemalloc src nstime c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src pages pic o go src github com cockroachdb cockroach c deps jemalloc src pages c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src prng pic o go src github com cockroachdb cockroach c deps jemalloc src prng c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src prof pic o go src github com cockroachdb cockroach c deps jemalloc src prof c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src quarantine pic o go src github com cockroachdb cockroach c deps jemalloc src quarantine c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go sar u modifier ignored since d is the default see u ar u modifier ignored since d is the default see u rc github com cockroachdb cockroach c deps jemalloc include iinclude o src rtree pic o go src github com cockroachdb cockroach c deps jemalloc src rtree c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src stats pic o go src github com cockroachdb cockroach c deps jemalloc src stats c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src spin pic o go src github com cockroachdb cockroach c deps jemalloc src spin c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src tcache pic o go src github com cockroachdb cockroach c deps jemalloc src tcache c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src ticker pic o go src github com cockroachdb cockroach c deps jemalloc src ticker c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src tsd pic o go src github com cockroachdb cockroach c deps jemalloc src tsd c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src util pic o go src github com cockroachdb cockroach c deps jemalloc src util c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src witness pic o go src github com cockroachdb cockroach c deps jemalloc src witness c ar crus lib libjemalloc a src jemalloc o src arena o src atomic o src base o src bitmap o src chunk o src chunk dss o src chunk mmap o src ckh o src ctl o src extent o src hash o src huge o src mb o src mutex o src nstime o src pages o src prng o src prof o src quarantine o src rtree o src stats o src spin o src tcache o src ticker o src tsd o src util o src witness o ar crus lib libjemalloc pic a src jemalloc pic o src arena pic o src atomic pic o src base pic o src bitmap pic o src chunk pic o src chunk dss pic o src chunk mmap pic o src ckh pic o src ctl pic o src extent pic o src hash pic o src huge pic o src mb pic o src mutex pic o src nstime pic o src pages pic o src prng pic o src prof pic o src quarantine pic o src rtree pic o src stats pic o src spin pic o src tcache pic o src ticker pic o src tsd pic o src util pic o src witness pic o built target build version scanning dependencies of target rocksdb building cxx object cmakefiles rocksdb dir cache clock cache cc o building cxx object cmakefiles rocksdb dir cache sharded cache cc o building cxx object cmakefiles rocksdb dir cache lru cache cc o building cxx object cmakefiles rocksdb dir db builder cc o building cxx object cmakefiles rocksdb dir db c cc o building cxx object cmakefiles rocksdb dir db column family cc o linking cxx static library libroachccl a built target roachccl building cxx object cmakefiles rocksdb dir db compacted db impl cc o building cxx object cmakefiles rocksdb dir db compaction cc o building cxx object cmakefiles rocksdb dir db compaction iterator cc o building cxx object cmakefiles rocksdb dir db compaction job cc o touch bin docgen functions building cxx object cmakefiles rocksdb dir db compaction picker cc o building cxx object cmakefiles rocksdb dir db compaction picker universal cc o building cxx object cmakefiles rocksdb dir db convenience cc o building cxx object cmakefiles rocksdb dir db db filesnapshot cc o building cxx object cmakefiles rocksdb dir db db impl cc o building cxx object cmakefiles rocksdb dir db db impl write cc o building cxx object cmakefiles rocksdb dir db db impl compaction flush cc o building cxx object cmakefiles rocksdb dir db db impl files cc o building cxx object cmakefiles rocksdb dir db db impl open cc o building cxx object cmakefiles rocksdb dir db db impl debug cc o building cxx object cmakefiles rocksdb dir db db impl experimental cc o building cxx object cmakefiles rocksdb dir db db impl readonly cc o building cxx object cmakefiles rocksdb dir db db info dumper cc o building cxx object cmakefiles rocksdb dir db db iter cc o building cxx object cmakefiles rocksdb dir db dbformat cc o building cxx object cmakefiles rocksdb dir db event helpers cc o building cxx object cmakefiles rocksdb dir db experimental cc o building cxx object cmakefiles rocksdb dir db external sst file ingestion job cc o building cxx object cmakefiles rocksdb dir db file indexer cc o building cxx object cmakefiles rocksdb dir db flush job cc o building cxx object cmakefiles rocksdb dir db flush scheduler cc o building cxx object cmakefiles rocksdb dir db forward iterator cc o building cxx object cmakefiles rocksdb dir db internal stats cc o building cxx object cmakefiles rocksdb dir db log reader cc o building cxx object cmakefiles rocksdb dir db log writer cc o building cxx object cmakefiles rocksdb dir db malloc stats cc o building cxx object cmakefiles rocksdb dir db managed iterator cc o building cxx object cmakefiles rocksdb dir db memtable cc o building cxx object cmakefiles rocksdb dir db memtable list cc o building cxx object cmakefiles rocksdb dir db merge helper cc o building cxx object cmakefiles rocksdb dir db range del aggregator cc o building cxx object cmakefiles rocksdb dir db repair cc o building cxx object cmakefiles rocksdb dir db table cache cc o building cxx object cmakefiles rocksdb dir db table properties collector cc o building cxx object cmakefiles rocksdb dir db transaction log impl cc o building cxx object cmakefiles rocksdb dir db version builder cc o building cxx object cmakefiles rocksdb dir db version edit cc o building cxx object cmakefiles rocksdb dir db version set cc o building cxx object cmakefiles rocksdb dir db wal manager cc o building cxx object cmakefiles rocksdb dir db write batch cc o building cxx object cmakefiles rocksdb dir db write thread cc o building cxx object cmakefiles rocksdb dir env env cc o building cxx object cmakefiles rocksdb dir env env encryption cc o building cxx object cmakefiles rocksdb dir env mock env cc o building cxx object cmakefiles rocksdb dir memtable alloc tracker cc o building cxx object cmakefiles rocksdb dir memtable hash cuckoo rep cc o building cxx object cmakefiles rocksdb dir memtable hash linklist rep cc o building cxx object cmakefiles rocksdb dir memtable hash skiplist rep cc o building cxx object cmakefiles rocksdb dir memtable skiplistrep cc o building cxx object cmakefiles rocksdb dir memtable vectorrep cc o building cxx object cmakefiles rocksdb dir memtable write buffer manager cc o building cxx object cmakefiles rocksdb dir monitoring histogram cc o building cxx object cmakefiles rocksdb dir monitoring instrumented mutex cc o building cxx object cmakefiles rocksdb dir monitoring iostats context cc o building cxx object cmakefiles rocksdb dir monitoring perf context cc o building cxx object cmakefiles rocksdb dir monitoring perf level cc o building cxx object cmakefiles rocksdb dir monitoring statistics cc o building cxx object cmakefiles rocksdb dir monitoring thread status updater cc o building cxx object cmakefiles rocksdb dir monitoring thread status util cc o building cxx object cmakefiles rocksdb dir monitoring thread status util debug cc o building cxx object cmakefiles rocksdb dir options cf options cc o building cxx object cmakefiles rocksdb dir options db options cc o building cxx object cmakefiles rocksdb dir options options cc o building cxx object cmakefiles rocksdb dir options options helper cc o building cxx object cmakefiles rocksdb dir options options parser cc o building cxx object cmakefiles rocksdb dir table adaptive table factory cc o building cxx object cmakefiles rocksdb dir table block cc o building cxx object cmakefiles rocksdb dir table block based filter block cc o building cxx object cmakefiles rocksdb dir table block based table builder cc o building cxx object cmakefiles rocksdb dir table block based table factory cc o building cxx object cmakefiles rocksdb dir table block based table reader cc o building cxx object cmakefiles rocksdb dir table block builder cc o building cxx object cmakefiles rocksdb dir table block prefix index cc o building cxx object cmakefiles rocksdb dir table bloom block cc o building cxx object cmakefiles rocksdb dir table cuckoo table builder cc o building cxx object cmakefiles rocksdb dir table cuckoo table factory cc o building cxx object cmakefiles rocksdb dir table cuckoo table reader cc o building cxx object cmakefiles rocksdb dir table format cc o building cxx object cmakefiles rocksdb dir table full filter block cc o building cxx object cmakefiles rocksdb dir table get context cc o building cxx object cmakefiles rocksdb dir table index builder cc o building cxx object cmakefiles rocksdb dir table iterator cc o building cxx object cmakefiles rocksdb dir table merging iterator cc o building cxx object cmakefiles rocksdb dir table meta blocks cc o building cxx object cmakefiles rocksdb dir table partitioned filter block cc o building cxx object cmakefiles rocksdb dir table persistent cache helper cc o building cxx object cmakefiles rocksdb dir table plain table builder cc o building cxx object cmakefiles rocksdb dir table plain table factory cc o building cxx object cmakefiles rocksdb dir table plain table index cc o building cxx object cmakefiles rocksdb dir table plain table key coding cc o building cxx object cmakefiles rocksdb dir table plain table reader cc o building cxx object cmakefiles rocksdb dir table sst file writer cc o building cxx object cmakefiles rocksdb dir table table properties cc o building cxx object cmakefiles rocksdb dir table two level iterator cc o building cxx object cmakefiles rocksdb dir tools db bench tool cc o building cxx object cmakefiles rocksdb dir tools dump db dump tool cc o building cxx object cmakefiles rocksdb dir tools ldb cmd cc o building cxx object cmakefiles rocksdb dir tools sst dump tool cc o building cxx object cmakefiles rocksdb dir util arena cc o building cxx object cmakefiles rocksdb dir util auto roll logger cc o building cxx object cmakefiles rocksdb dir util bloom cc o building cxx object cmakefiles rocksdb dir util coding cc o building cxx object cmakefiles rocksdb dir util comparator cc o building cxx object cmakefiles rocksdb dir util concurrent arena cc o building cxx object cmakefiles rocksdb dir util cc o building cxx object cmakefiles rocksdb dir util delete scheduler cc o building cxx object cmakefiles rocksdb dir util dynamic bloom cc o building cxx object cmakefiles rocksdb dir util event logger cc o building cxx object cmakefiles rocksdb dir util file reader writer cc o building cxx object cmakefiles rocksdb dir util file util cc o building cxx object cmakefiles rocksdb dir util filename cc o building cxx object cmakefiles rocksdb dir util hash cc o building cxx object cmakefiles rocksdb dir util log buffer cc o building cxx object cmakefiles rocksdb dir util random cc o building cxx object cmakefiles rocksdb dir util rate limiter cc o building cxx object cmakefiles rocksdb dir util sst file manager impl cc o building cxx object cmakefiles rocksdb dir util status cc o building cxx object cmakefiles rocksdb dir util sync point cc o building cxx object cmakefiles rocksdb dir util testutil cc o building cxx object cmakefiles rocksdb dir util thread local cc o building cxx object cmakefiles rocksdb dir util threadpool imp cc o building cxx object cmakefiles rocksdb dir util transaction test util cc o building cxx object cmakefiles rocksdb dir utilities backupable backupable db cc o building cxx object cmakefiles rocksdb dir utilities blob db blob db cc o building cxx object cmakefiles rocksdb dir utilities blob db blob db impl cc o building cxx object cmakefiles rocksdb dir utilities blob db blob dump tool cc o building cxx object cmakefiles rocksdb dir utilities blob db blob file cc o building cxx object cmakefiles rocksdb dir utilities blob db blob log reader cc o building cxx object cmakefiles rocksdb dir utilities blob db blob log writer cc o building cxx object cmakefiles rocksdb dir utilities blob db blob log format cc o building cxx object cmakefiles rocksdb dir utilities blob db ttl extractor cc o building cxx object cmakefiles rocksdb dir utilities checkpoint checkpoint impl cc o building cxx object cmakefiles rocksdb dir utilities col buf decoder cc o building cxx object cmakefiles rocksdb dir utilities col buf encoder cc o building cxx object cmakefiles rocksdb dir utilities column aware encoding util cc o building cxx object cmakefiles rocksdb dir utilities date tiered date tiered db impl cc o building cxx object cmakefiles rocksdb dir utilities debug cc o building cxx object cmakefiles rocksdb dir utilities document document db cc o building cxx object cmakefiles rocksdb dir utilities document json document cc o building cxx object cmakefiles rocksdb dir utilities env timed cc o building cxx object cmakefiles rocksdb dir utilities geodb geodb impl cc o building cxx object cmakefiles rocksdb dir utilities memory memory util cc o building cxx object cmakefiles rocksdb dir utilities merge operators cc o building cxx object cmakefiles rocksdb dir utilities options options util cc o building cxx object cmakefiles rocksdb dir utilities persistent cache block cache tier cc o building cxx object cmakefiles rocksdb dir utilities persistent cache block cache tier file cc o building cxx object cmakefiles rocksdb dir utilities persistent cache block cache tier metadata cc o building cxx object cmakefiles rocksdb dir utilities persistent cache volatile tier impl cc o building cxx object cmakefiles rocksdb dir utilities redis redis lists cc o building cxx object cmakefiles rocksdb dir utilities simulator cache sim cache cc o building cxx object cmakefiles rocksdb dir utilities spatialdb spatial db cc o building cxx object cmakefiles rocksdb dir utilities transactions optimistic transaction db impl cc o building cxx object cmakefiles rocksdb dir utilities transactions optimistic transaction cc o building cxx object cmakefiles rocksdb dir utilities transactions pessimistic transaction cc o building cxx object cmakefiles rocksdb dir utilities transactions pessimistic transaction db cc o building cxx object cmakefiles rocksdb dir utilities transactions snapshot checker cc o building cxx object cmakefiles rocksdb dir utilities transactions transaction base cc o building cxx object cmakefiles rocksdb dir utilities transactions transaction lock mgr cc o building cxx object cmakefiles rocksdb dir utilities transactions transaction util cc o building cxx object cmakefiles rocksdb dir utilities transactions write prepared txn cc o building cxx object cmakefiles rocksdb dir utilities ttl db ttl impl cc o building cxx object cmakefiles rocksdb dir utilities write batch with index write batch with index cc o building cxx object cmakefiles rocksdb dir utilities write batch with index write batch with index internal cc o building cxx object cmakefiles rocksdb dir port port posix cc o building cxx object cmakefiles rocksdb dir env env posix cc o building cxx object cmakefiles rocksdb dir env io posix cc o linking cxx static library librocksdb a built target rocksdb go test v tags make pc linux gnu ldflags x github com cockroachdb cockroach pkg build typ development x github com cockroachdb cockroach pkg build tag dirty x github com cockroachdb cockroach pkg build utctime x github com cockroachdb cockroach pkg build rev x github com cockroachdb cockroach pkg build cgotargettriple pc linux gnu i github com cockroachdb cockroach pkg cmd cockroach can t load package package github com cockroachdb cockroach pkg cmd cockroach cannot find package github com cockroachdb cockroach pkg cmd cockroach in any of usr local go src github com cockroachdb cockroach pkg cmd cockroach from goroot go src github com cockroachdb cockroach pkg cmd cockroach from gopath make error makefile recipe for target gotestdashi failed
1
423,532
28,605,915,678
IssuesEvent
2023-04-24 00:21:03
CarsOk/ADSO2698056PROJECT1
https://api.github.com/repos/CarsOk/ADSO2698056PROJECT1
closed
visita-al-empresario
documentation
## NOSOTROS como ## APRENDICES Sena ## QUEREMOS hacer una vista al empresario con el fin de analizar, evaluar y deducir las dificultades de la empresa ## REQUERIMIENTO DE ACEPTACION - [ ] análisis de la empresa - [ ] encuesta - [ ] evidencia de asistencia
1.0
visita-al-empresario - ## NOSOTROS como ## APRENDICES Sena ## QUEREMOS hacer una vista al empresario con el fin de analizar, evaluar y deducir las dificultades de la empresa ## REQUERIMIENTO DE ACEPTACION - [ ] análisis de la empresa - [ ] encuesta - [ ] evidencia de asistencia
non_test
visita al empresario nosotros como aprendices sena queremos hacer una vista al empresario con el fin de analizar evaluar y deducir las dificultades de la empresa requerimiento de aceptacion análisis de la empresa encuesta evidencia de asistencia
0
75,705
7,480,964,322
IssuesEvent
2018-04-04 19:04:37
curationexperts/epigaea
https://api.github.com/repos/curationexperts/epigaea
closed
Acceptance testing: Normalize space on deposit.
acceptance testing bug in progress
End users are going to be copy and pasting titles, descriptions, etc into the fields, and this will result in strange spacing if it is not normalized. As a result, catalogers will need to spend extra time cleaning it up. Normalizing space on deposit should occur automatically. See title field in: w9505044z ![screen shot 2017-10-04 at 8 27 29 am](https://user-images.githubusercontent.com/4700092/31175687-1b03e7e8-a8de-11e7-9eed-535ed3602d20.png)
1.0
Acceptance testing: Normalize space on deposit. - End users are going to be copy and pasting titles, descriptions, etc into the fields, and this will result in strange spacing if it is not normalized. As a result, catalogers will need to spend extra time cleaning it up. Normalizing space on deposit should occur automatically. See title field in: w9505044z ![screen shot 2017-10-04 at 8 27 29 am](https://user-images.githubusercontent.com/4700092/31175687-1b03e7e8-a8de-11e7-9eed-535ed3602d20.png)
test
acceptance testing normalize space on deposit end users are going to be copy and pasting titles descriptions etc into the fields and this will result in strange spacing if it is not normalized as a result catalogers will need to spend extra time cleaning it up normalizing space on deposit should occur automatically see title field in
1
71,941
7,273,339,757
IssuesEvent
2018-02-21 04:21:52
ppy/osu-framework
https://api.github.com/repos/ppy/osu-framework
closed
After exiting dynamically compiled test there's no way to test it again
bug testing
When you switch to another test you can't switch back to dynamically compiled test.
1.0
After exiting dynamically compiled test there's no way to test it again - When you switch to another test you can't switch back to dynamically compiled test.
test
after exiting dynamically compiled test there s no way to test it again when you switch to another test you can t switch back to dynamically compiled test
1
248,464
21,034,617,417
IssuesEvent
2022-03-31 06:27:44
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: schemachange/mixed-versions failed
C-test-failure O-robot O-roachtest release-blocker branch-release-22.1
roachtest.schemachange/mixed-versions [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4741011&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4741011&tab=artifacts#/schemachange/mixed-versions) on release-22.1 @ [a88367d7bd168a3614b7946cd7b77bb9c6374a30](https://github.com/cockroachdb/cockroach/commits/a88367d7bd168a3614b7946cd7b77bb9c6374a30): ``` | main.(*testRunner).runTest.func2 | main/pkg/cmd/roachtest/test_runner.go:866 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1581 Wraps: (2) output in run_052724.245330906_n3_workload_run_schemachange Wraps: (3) ./workload run schemachange --verbose=1 --tolerate-errors=true --max-ops 100 --concurrency 5 {pgurl:1-4} returned | stderr: | <... some data truncated by circular buffer; go to artifacts for details ...> | elpers.go:79 [-] 3536 pgx logger [error]: Query logParams=map[args:[] err:ERROR: relation "table247" (231): primary index column "col247_261" cannot be virtual (SQLSTATE XXUUU) sql: | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 +SELECT max(regexp_extract(name, '[0-9]+$')::INT8) | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + FROM ( | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + SELECT name | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + FROM ( | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + (SELECT table_name FROM [SHOW TABLES]) UNION | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + (SELECT sequence_name FROM [SHOW SEQUENCES]) UNION | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + (SELECT name FROM [SHOW ENUMS]) UNION | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + (SELECT schema_name FROM [SHOW SCHEMAS]) UNION | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + (SELECT column_name FROM information_schema.columns) UNION | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + (SELECT index_name FROM information_schema.statistics) | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + ) AS obj (name) | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + ) | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + WHERE name ~ '^(table|view|seq|enum|schema)[0-9]+$' | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + OR name ~ '^(col|index)[0-9]+_[0-9]+$';] | W220331 06:27:23.690961 1 workload/cli/run.go:423 [-] 3537 retrying after error while creating load: failed to initialize the load generator: ERROR: relation "table247" (231): primary index column "col247_261" cannot be virtual (SQLSTATE XXUUU) | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 pgx logger [error]: Query logParams=map[args:[] err:ERROR: relation "table247" (231): primary index column "col247_261" cannot be virtual (SQLSTATE XXUUU) sql: | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 +SELECT max(regexp_extract(name, '[0-9]+$')::INT8) | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + FROM ( | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + SELECT name | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + FROM ( | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + (SELECT table_name FROM [SHOW TABLES]) UNION | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + (SELECT sequence_name FROM [SHOW SEQUENCES]) UNION | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + (SELECT name FROM [SHOW ENUMS]) UNION | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + (SELECT schema_name FROM [SHOW SCHEMAS]) UNION | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + (SELECT column_name FROM information_schema.columns) UNION | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + (SELECT index_name FROM information_schema.statistics) | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + ) AS obj (name) | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + ) | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + WHERE name ~ '^(table|view|seq|enum|schema)[0-9]+$' | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + OR name ~ '^(col|index)[0-9]+_[0-9]+$';] | E220331 06:27:25.839340 1 workload/cli/run.go:437 [-] 3539 Attempt to create load generator failed. It's been more than 1h0m0s since we started trying to create the load generator so we're giving up. Last failure: failed to initialize the load generator: ERROR: relation "table247" (231): primary index column "col247_261" cannot be virtual (SQLSTATE XXUUU) | Error: failed to initialize the load generator: ERROR: relation "table247" (231): primary index column "col247_261" cannot be virtual (SQLSTATE XXUUU) | | stdout: Wraps: (4) COMMAND_PROBLEM Wraps: (5) Node 3. Command with error: | `````` | ./workload run schemachange --verbose=1 --tolerate-errors=true --max-ops 100 --concurrency 5 {pgurl:1-4} | `````` Wraps: (6) exit status 1 Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *cluster.WithCommandDetails (4) errors.Cmd (5) *hintdetail.withDetail (6) *exec.ExitError ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #78500 roachtest: schemachange/mixed-versions failed [C-test-failure O-roachtest O-robot T-sql-schema branch-release-21.2] - #64687 roachtest: schemachange/mixed-versions failed [C-test-failure O-roachtest O-robot T-sql-schema branch-master] - #62158 roachtest: schemachange/mixed-versions failed [C-test-failure O-roachtest O-robot T-sql-schema branch-release-21.1] </p> </details> /cc @cockroachdb/sql-schema <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*schemachange/mixed-versions.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: schemachange/mixed-versions failed - roachtest.schemachange/mixed-versions [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4741011&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4741011&tab=artifacts#/schemachange/mixed-versions) on release-22.1 @ [a88367d7bd168a3614b7946cd7b77bb9c6374a30](https://github.com/cockroachdb/cockroach/commits/a88367d7bd168a3614b7946cd7b77bb9c6374a30): ``` | main.(*testRunner).runTest.func2 | main/pkg/cmd/roachtest/test_runner.go:866 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1581 Wraps: (2) output in run_052724.245330906_n3_workload_run_schemachange Wraps: (3) ./workload run schemachange --verbose=1 --tolerate-errors=true --max-ops 100 --concurrency 5 {pgurl:1-4} returned | stderr: | <... some data truncated by circular buffer; go to artifacts for details ...> | elpers.go:79 [-] 3536 pgx logger [error]: Query logParams=map[args:[] err:ERROR: relation "table247" (231): primary index column "col247_261" cannot be virtual (SQLSTATE XXUUU) sql: | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 +SELECT max(regexp_extract(name, '[0-9]+$')::INT8) | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + FROM ( | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + SELECT name | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + FROM ( | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + (SELECT table_name FROM [SHOW TABLES]) UNION | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + (SELECT sequence_name FROM [SHOW SEQUENCES]) UNION | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + (SELECT name FROM [SHOW ENUMS]) UNION | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + (SELECT schema_name FROM [SHOW SCHEMAS]) UNION | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + (SELECT column_name FROM information_schema.columns) UNION | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + (SELECT index_name FROM information_schema.statistics) | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + ) AS obj (name) | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + ) | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + WHERE name ~ '^(table|view|seq|enum|schema)[0-9]+$' | I220331 06:27:21.792399 1 workload/pgx_helpers.go:79 [-] 3536 + OR name ~ '^(col|index)[0-9]+_[0-9]+$';] | W220331 06:27:23.690961 1 workload/cli/run.go:423 [-] 3537 retrying after error while creating load: failed to initialize the load generator: ERROR: relation "table247" (231): primary index column "col247_261" cannot be virtual (SQLSTATE XXUUU) | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 pgx logger [error]: Query logParams=map[args:[] err:ERROR: relation "table247" (231): primary index column "col247_261" cannot be virtual (SQLSTATE XXUUU) sql: | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 +SELECT max(regexp_extract(name, '[0-9]+$')::INT8) | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + FROM ( | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + SELECT name | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + FROM ( | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + (SELECT table_name FROM [SHOW TABLES]) UNION | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + (SELECT sequence_name FROM [SHOW SEQUENCES]) UNION | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + (SELECT name FROM [SHOW ENUMS]) UNION | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + (SELECT schema_name FROM [SHOW SCHEMAS]) UNION | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + (SELECT column_name FROM information_schema.columns) UNION | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + (SELECT index_name FROM information_schema.statistics) | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + ) AS obj (name) | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + ) | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + WHERE name ~ '^(table|view|seq|enum|schema)[0-9]+$' | I220331 06:27:23.729482 1 workload/pgx_helpers.go:79 [-] 3538 + OR name ~ '^(col|index)[0-9]+_[0-9]+$';] | E220331 06:27:25.839340 1 workload/cli/run.go:437 [-] 3539 Attempt to create load generator failed. It's been more than 1h0m0s since we started trying to create the load generator so we're giving up. Last failure: failed to initialize the load generator: ERROR: relation "table247" (231): primary index column "col247_261" cannot be virtual (SQLSTATE XXUUU) | Error: failed to initialize the load generator: ERROR: relation "table247" (231): primary index column "col247_261" cannot be virtual (SQLSTATE XXUUU) | | stdout: Wraps: (4) COMMAND_PROBLEM Wraps: (5) Node 3. Command with error: | `````` | ./workload run schemachange --verbose=1 --tolerate-errors=true --max-ops 100 --concurrency 5 {pgurl:1-4} | `````` Wraps: (6) exit status 1 Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *cluster.WithCommandDetails (4) errors.Cmd (5) *hintdetail.withDetail (6) *exec.ExitError ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #78500 roachtest: schemachange/mixed-versions failed [C-test-failure O-roachtest O-robot T-sql-schema branch-release-21.2] - #64687 roachtest: schemachange/mixed-versions failed [C-test-failure O-roachtest O-robot T-sql-schema branch-master] - #62158 roachtest: schemachange/mixed-versions failed [C-test-failure O-roachtest O-robot T-sql-schema branch-release-21.1] </p> </details> /cc @cockroachdb/sql-schema <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*schemachange/mixed-versions.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
roachtest schemachange mixed versions failed roachtest schemachange mixed versions with on release main testrunner runtest main pkg cmd roachtest test runner go runtime goexit goroot src runtime asm s wraps output in run workload run schemachange wraps workload run schemachange verbose tolerate errors true max ops concurrency pgurl returned stderr elpers go pgx logger query logparams map err error relation primary index column cannot be virtual sqlstate xxuuu sql workload pgx helpers go select max regexp extract name workload pgx helpers go from workload pgx helpers go select name workload pgx helpers go from workload pgx helpers go select table name from union workload pgx helpers go select sequence name from union workload pgx helpers go select name from union workload pgx helpers go select schema name from union workload pgx helpers go select column name from information schema columns union workload pgx helpers go select index name from information schema statistics workload pgx helpers go as obj name workload pgx helpers go workload pgx helpers go where name table view seq enum schema workload pgx helpers go or name col index workload cli run go retrying after error while creating load failed to initialize the load generator error relation primary index column cannot be virtual sqlstate xxuuu workload pgx helpers go pgx logger query logparams map err error relation primary index column cannot be virtual sqlstate xxuuu sql workload pgx helpers go select max regexp extract name workload pgx helpers go from workload pgx helpers go select name workload pgx helpers go from workload pgx helpers go select table name from union workload pgx helpers go select sequence name from union workload pgx helpers go select name from union workload pgx helpers go select schema name from union workload pgx helpers go select column name from information schema columns union workload pgx helpers go select index name from information schema statistics workload pgx helpers go as obj name workload pgx helpers go workload pgx helpers go where name table view seq enum schema workload pgx helpers go or name col index workload cli run go attempt to create load generator failed it s been more than since we started trying to create the load generator so we re giving up last failure failed to initialize the load generator error relation primary index column cannot be virtual sqlstate xxuuu error failed to initialize the load generator error relation primary index column cannot be virtual sqlstate xxuuu stdout wraps command problem wraps node command with error workload run schemachange verbose tolerate errors true max ops concurrency pgurl wraps exit status error types withstack withstack errutil withprefix cluster withcommanddetails errors cmd hintdetail withdetail exec exiterror help see see same failure on other branches roachtest schemachange mixed versions failed roachtest schemachange mixed versions failed roachtest schemachange mixed versions failed cc cockroachdb sql schema
1
254,623
21,801,233,860
IssuesEvent
2022-05-16 05:37:05
StoneCypher/fsl
https://api.github.com/repos/StoneCypher/fsl
closed
Better error object
Test suite Ease of use List issues Underway
We'd like a better error object, which will assist in improved debugging. This error object should offer: 1. [ ] #755 2. [ ] #740 3. [ ] #741 4. [ ] #742 5. [ ] #743 6. [ ] #744 7. [ ] #745 8. [ ] #746 9. [ ] #747 10. [ ] #748 11. [ ] #749 12. [ ] #750
1.0
Better error object - We'd like a better error object, which will assist in improved debugging. This error object should offer: 1. [ ] #755 2. [ ] #740 3. [ ] #741 4. [ ] #742 5. [ ] #743 6. [ ] #744 7. [ ] #745 8. [ ] #746 9. [ ] #747 10. [ ] #748 11. [ ] #749 12. [ ] #750
test
better error object we d like a better error object which will assist in improved debugging this error object should offer
1
258,634
27,565,514,689
IssuesEvent
2023-03-08 03:12:41
rvvergara/bookstore_rails_app
https://api.github.com/repos/rvvergara/bookstore_rails_app
closed
CVE-2022-24836 (High) detected in nokogiri-1.10.8.gem - autoclosed
Mend: dependency security vulnerability
## CVE-2022-24836 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nokogiri-1.10.8.gem</b></p></summary> <p>Nokogiri (鋸) is an HTML, XML, SAX, and Reader parser. Among Nokogiri's many features is the ability to search documents via XPath or CSS3 selectors.</p> <p>Library home page: <a href="https://rubygems.org/gems/nokogiri-1.10.8.gem">https://rubygems.org/gems/nokogiri-1.10.8.gem</a></p> <p> Dependency Hierarchy: - rails-5.2.3.gem (Root Library) - actionview-5.2.3.gem - rails-dom-testing-2.0.3.gem - :x: **nokogiri-1.10.8.gem** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/rvvergara/bookstore_rails_app/commit/e8aa1b3e0f79399ea66767ed9ce2ffeb0b8e7f45">e8aa1b3e0f79399ea66767ed9ce2ffeb0b8e7f45</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Nokogiri is an open source XML and HTML library for Ruby. Nokogiri `< v1.13.4` contains an inefficient regular expression that is susceptible to excessive backtracking when attempting to detect encoding in HTML documents. Users are advised to upgrade to Nokogiri `>= 1.13.4`. There are no known workarounds for this issue. <p>Publish Date: 2022-04-11 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-24836>CVE-2022-24836</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sparklemotion/nokogiri/security/advisories/GHSA-crjr-9rc5-ghw8">https://github.com/sparklemotion/nokogiri/security/advisories/GHSA-crjr-9rc5-ghw8</a></p> <p>Release Date: 2022-04-11</p> <p>Fix Resolution: nokogiri - 1.13.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-24836 (High) detected in nokogiri-1.10.8.gem - autoclosed - ## CVE-2022-24836 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nokogiri-1.10.8.gem</b></p></summary> <p>Nokogiri (鋸) is an HTML, XML, SAX, and Reader parser. Among Nokogiri's many features is the ability to search documents via XPath or CSS3 selectors.</p> <p>Library home page: <a href="https://rubygems.org/gems/nokogiri-1.10.8.gem">https://rubygems.org/gems/nokogiri-1.10.8.gem</a></p> <p> Dependency Hierarchy: - rails-5.2.3.gem (Root Library) - actionview-5.2.3.gem - rails-dom-testing-2.0.3.gem - :x: **nokogiri-1.10.8.gem** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/rvvergara/bookstore_rails_app/commit/e8aa1b3e0f79399ea66767ed9ce2ffeb0b8e7f45">e8aa1b3e0f79399ea66767ed9ce2ffeb0b8e7f45</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Nokogiri is an open source XML and HTML library for Ruby. Nokogiri `< v1.13.4` contains an inefficient regular expression that is susceptible to excessive backtracking when attempting to detect encoding in HTML documents. Users are advised to upgrade to Nokogiri `>= 1.13.4`. There are no known workarounds for this issue. <p>Publish Date: 2022-04-11 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-24836>CVE-2022-24836</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sparklemotion/nokogiri/security/advisories/GHSA-crjr-9rc5-ghw8">https://github.com/sparklemotion/nokogiri/security/advisories/GHSA-crjr-9rc5-ghw8</a></p> <p>Release Date: 2022-04-11</p> <p>Fix Resolution: nokogiri - 1.13.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in nokogiri gem autoclosed cve high severity vulnerability vulnerable library nokogiri gem nokogiri 鋸 is an html xml sax and reader parser among nokogiri s many features is the ability to search documents via xpath or selectors library home page a href dependency hierarchy rails gem root library actionview gem rails dom testing gem x nokogiri gem vulnerable library found in head commit a href found in base branch master vulnerability details nokogiri is an open source xml and html library for ruby nokogiri there are no known workarounds for this issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution nokogiri step up your open source security game with mend
0
264,370
23,114,892,638
IssuesEvent
2022-07-27 15:48:26
tarantool/expirationd
https://api.github.com/repos/tarantool/expirationd
closed
Add test for reloading the module with external iteration state via Cartridge hotreload
teamE tests 3sp
Use case: A user creates an expiration task, for which the expiration check callback or the iteration function saves some state into the globals or `cartridge.vars` (e.g. iterator state or the last tuple value, etc). In case of reloading the module in a Cartridge role, the user must have the ability to preserve the state for continuing the iteration with it after the reload. AC: New integration test added for the above scenario and is passing in the CI
1.0
Add test for reloading the module with external iteration state via Cartridge hotreload - Use case: A user creates an expiration task, for which the expiration check callback or the iteration function saves some state into the globals or `cartridge.vars` (e.g. iterator state or the last tuple value, etc). In case of reloading the module in a Cartridge role, the user must have the ability to preserve the state for continuing the iteration with it after the reload. AC: New integration test added for the above scenario and is passing in the CI
test
add test for reloading the module with external iteration state via cartridge hotreload use case a user creates an expiration task for which the expiration check callback or the iteration function saves some state into the globals or cartridge vars e g iterator state or the last tuple value etc in case of reloading the module in a cartridge role the user must have the ability to preserve the state for continuing the iteration with it after the reload ac new integration test added for the above scenario and is passing in the ci
1
7,072
3,077,653,295
IssuesEvent
2015-08-21 02:42:23
golang/go
https://api.github.com/repos/golang/go
closed
encoding/base64: documentation typo
Documentation
In the docs for the base64 package in the [variables](http://golang.org/pkg/encoding/base64/#URLEncoding) section there is a typo. The description under `var RawURLEncoding = URLEncoding.WithPadding(NoPadding)` uses `URLEncoding` to describe the above variable when it should be using `RawURLEncoding`.
1.0
encoding/base64: documentation typo - In the docs for the base64 package in the [variables](http://golang.org/pkg/encoding/base64/#URLEncoding) section there is a typo. The description under `var RawURLEncoding = URLEncoding.WithPadding(NoPadding)` uses `URLEncoding` to describe the above variable when it should be using `RawURLEncoding`.
non_test
encoding documentation typo in the docs for the package in the section there is a typo the description under var rawurlencoding urlencoding withpadding nopadding uses urlencoding to describe the above variable when it should be using rawurlencoding
0
116,926
9,888,796,515
IssuesEvent
2019-06-25 12:27:09
wazuh/wazuh-splunk
https://api.github.com/repos/wazuh/wazuh-splunk
opened
Testing v3.9.3-7.3.0-rc3
testing
Tag: [v3.9.3-7.3.0-rc3](https://github.com/wazuh/wazuh-splunk/tree/v3.9.3-7.3.0-rc3) Package: [v3.9.3_7.3.0-rc3.tar.gz](https://packages-dev.wazuh.com/staging/app/splunk/v3.9.3_7.3.0-rc3.tar.gz) Revision: **31** Put extra attention to the next points: - [ ] Inventory tab https://github.com/wazuh/wazuh-splunk/issues/744 - [ ] Bugs fixed from RC2 https://github.com/wazuh/wazuh-splunk/issues/765 - [ ] No "Invalid Date" in Chrome, Firefox, Safari, Edge
1.0
Testing v3.9.3-7.3.0-rc3 - Tag: [v3.9.3-7.3.0-rc3](https://github.com/wazuh/wazuh-splunk/tree/v3.9.3-7.3.0-rc3) Package: [v3.9.3_7.3.0-rc3.tar.gz](https://packages-dev.wazuh.com/staging/app/splunk/v3.9.3_7.3.0-rc3.tar.gz) Revision: **31** Put extra attention to the next points: - [ ] Inventory tab https://github.com/wazuh/wazuh-splunk/issues/744 - [ ] Bugs fixed from RC2 https://github.com/wazuh/wazuh-splunk/issues/765 - [ ] No "Invalid Date" in Chrome, Firefox, Safari, Edge
test
testing tag package revision put extra attention to the next points inventory tab bugs fixed from no invalid date in chrome firefox safari edge
1
155,252
12,244,429,015
IssuesEvent
2020-05-05 11:06:55
enonic/app-contentstudio
https://api.github.com/repos/enonic/app-contentstudio
closed
Add ui-tests to verify issue#1570
Test
Details Panel should be reset after switching to another project #1570
1.0
Add ui-tests to verify issue#1570 - Details Panel should be reset after switching to another project #1570
test
add ui tests to verify issue details panel should be reset after switching to another project
1
449,697
31,857,330,389
IssuesEvent
2023-09-15 08:26:01
nf-core/website
https://api.github.com/repos/nf-core/website
closed
Adding a new pipeline docs contains out of date git commands
bug documentation hackathon infrastructure
https://nf-co.re/developers/adding_pipelines#push-to-github Says `git push`, whereas ```bash nf-core create ``` Says `git push --all origin` (which is correct, as we need `dev` and `template`.
1.0
Adding a new pipeline docs contains out of date git commands - https://nf-co.re/developers/adding_pipelines#push-to-github Says `git push`, whereas ```bash nf-core create ``` Says `git push --all origin` (which is correct, as we need `dev` and `template`.
non_test
adding a new pipeline docs contains out of date git commands says git push whereas bash nf core create says git push all origin which is correct as we need dev and template
0
166,685
6,310,001,567
IssuesEvent
2017-07-23 05:25:08
Rsl1122/Plan-PlayerAnalytics
https://api.github.com/repos/Rsl1122/Plan-PlayerAnalytics
closed
NullPointerException cache.queue.ClearConsumer
Bug Priority: HIGH
Plan 3.5.2. http://paste.ubuntu.com/25147094/ ``` [Jul 22, 08:55:30] main.java.com.djrapitops.plan.data.cache.queue.ClearConsumer Caught java.lang.NullPointerException [Jul 22, 08:55:30] java.util.Objects.requireNonNull(Objects.java:203) [Jul 22, 08:55:30] java.util.Optional.<init>(Optional.java:96) [Jul 22, 08:55:30] java.util.Optional.of(Optional.java:108) [Jul 22, 08:55:30] com.djrapitops.plugin.utilities.player.Fetch.getPlayer(Fetch.java:100) [Jul 22, 08:55:30] com.djrapitops.plugin.utilities.player.Fetch.isOnline(Fetch.java:51) [Jul 22, 08:55:30] main.java.com.djrapitops.plan.data.cache.DataCacheHandler.clearFromCache(DataCacheHandler.java:415) [Jul 22, 08:55:30] main.java.com.djrapitops.plan.data.cache.queue.ClearConsumer.consume(DataCacheClearQueue.java:79) [Jul 22, 08:55:30] main.java.com.djrapitops.plan.data.cache.queue.ClearConsumer.consume(DataCacheClearQueue.java:61) [Jul 22, 08:55:30] main.java.com.djrapitops.plan.data.cache.queue.Consumer.run(Consumer.java:33) [Jul 22, 08:55:30] com.djrapitops.plugin.task.RunnableFactory$1.run(RunnableFactory.java:36) [Jul 22, 08:55:30] org.bukkit.craftbukkit.v1_12_R1.scheduler.CraftTask.run(CraftTask.java:58) [Jul 22, 08:55:30] org.bukkit.craftbukkit.v1_12_R1.scheduler.CraftAsyncTask.run(CraftAsyncTask.java:52) [Jul 22, 08:55:30] com.destroystokyo.paper.ServerSchedulerReportingWrapper.run(ServerSchedulerReportingWrapper.java:22) [Jul 22, 08:55:30] java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [Jul 22, 08:55:30] java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [Jul 22, 08:55:30] java.lang.Thread.run(Thread.java:748) [Jul 22, 08:55:30] ```
1.0
NullPointerException cache.queue.ClearConsumer - Plan 3.5.2. http://paste.ubuntu.com/25147094/ ``` [Jul 22, 08:55:30] main.java.com.djrapitops.plan.data.cache.queue.ClearConsumer Caught java.lang.NullPointerException [Jul 22, 08:55:30] java.util.Objects.requireNonNull(Objects.java:203) [Jul 22, 08:55:30] java.util.Optional.<init>(Optional.java:96) [Jul 22, 08:55:30] java.util.Optional.of(Optional.java:108) [Jul 22, 08:55:30] com.djrapitops.plugin.utilities.player.Fetch.getPlayer(Fetch.java:100) [Jul 22, 08:55:30] com.djrapitops.plugin.utilities.player.Fetch.isOnline(Fetch.java:51) [Jul 22, 08:55:30] main.java.com.djrapitops.plan.data.cache.DataCacheHandler.clearFromCache(DataCacheHandler.java:415) [Jul 22, 08:55:30] main.java.com.djrapitops.plan.data.cache.queue.ClearConsumer.consume(DataCacheClearQueue.java:79) [Jul 22, 08:55:30] main.java.com.djrapitops.plan.data.cache.queue.ClearConsumer.consume(DataCacheClearQueue.java:61) [Jul 22, 08:55:30] main.java.com.djrapitops.plan.data.cache.queue.Consumer.run(Consumer.java:33) [Jul 22, 08:55:30] com.djrapitops.plugin.task.RunnableFactory$1.run(RunnableFactory.java:36) [Jul 22, 08:55:30] org.bukkit.craftbukkit.v1_12_R1.scheduler.CraftTask.run(CraftTask.java:58) [Jul 22, 08:55:30] org.bukkit.craftbukkit.v1_12_R1.scheduler.CraftAsyncTask.run(CraftAsyncTask.java:52) [Jul 22, 08:55:30] com.destroystokyo.paper.ServerSchedulerReportingWrapper.run(ServerSchedulerReportingWrapper.java:22) [Jul 22, 08:55:30] java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [Jul 22, 08:55:30] java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [Jul 22, 08:55:30] java.lang.Thread.run(Thread.java:748) [Jul 22, 08:55:30] ```
non_test
nullpointerexception cache queue clearconsumer plan main java com djrapitops plan data cache queue clearconsumer caught java lang nullpointerexception java util objects requirenonnull objects java java util optional optional java java util optional of optional java com djrapitops plugin utilities player fetch getplayer fetch java com djrapitops plugin utilities player fetch isonline fetch java main java com djrapitops plan data cache datacachehandler clearfromcache datacachehandler java main java com djrapitops plan data cache queue clearconsumer consume datacacheclearqueue java main java com djrapitops plan data cache queue clearconsumer consume datacacheclearqueue java main java com djrapitops plan data cache queue consumer run consumer java com djrapitops plugin task runnablefactory run runnablefactory java org bukkit craftbukkit scheduler crafttask run crafttask java org bukkit craftbukkit scheduler craftasynctask run craftasynctask java com destroystokyo paper serverschedulerreportingwrapper run serverschedulerreportingwrapper java java util concurrent threadpoolexecutor runworker threadpoolexecutor java java util concurrent threadpoolexecutor worker run threadpoolexecutor java java lang thread run thread java
0
92,997
11,730,415,872
IssuesEvent
2020-03-10 21:19:55
ibm-openbmc/dev
https://api.github.com/repos/ibm-openbmc/dev
closed
GUI : Design : Accommodate multiple logo sizes
GUI Milestone Tgt UI Design prio_low
## Overview Community members have expressed concern with the height of the header. The small height of the header will not accommodate company logos above 100px in height (roughly). We need to consider this and determine how to support logos of varying sizes (including width). There are various possibilities, but we will also want to design with the expand/collapse navigation in mind.
1.0
GUI : Design : Accommodate multiple logo sizes - ## Overview Community members have expressed concern with the height of the header. The small height of the header will not accommodate company logos above 100px in height (roughly). We need to consider this and determine how to support logos of varying sizes (including width). There are various possibilities, but we will also want to design with the expand/collapse navigation in mind.
non_test
gui design accommodate multiple logo sizes overview community members have expressed concern with the height of the header the small height of the header will not accommodate company logos above in height roughly we need to consider this and determine how to support logos of varying sizes including width there are various possibilities but we will also want to design with the expand collapse navigation in mind
0
298,224
22,470,696,760
IssuesEvent
2022-06-22 07:51:56
GroupePSA/sonarqube-cleanup-cli
https://api.github.com/repos/GroupePSA/sonarqube-cleanup-cli
opened
Details about projects deleted (analysis date & branch)
documentation
https://github.com/GroupePSA/sonarqube-cleanup-cli/blob/08e77f635043cc725edf7ab5dfc31cad789938e7/src/main/java/org/psa/sonarqube/cleanup/rest/SonarQubeClient.java#L58 Currently the program **takes the most old projects by date analysis** ; of the **main branch**. A warning should be written about branch concept. In case of last analysis on main branch is old (and part of deletion) but some branches are "active" (recent analyze), the project will be deleted. ----- Another approach could perhaps to check branches for project which will be deleted, and avoid deletion if "recent" branches analysis. But this algorithm should be detailed, what is a "recent" branch in this case ...
1.0
Details about projects deleted (analysis date & branch) - https://github.com/GroupePSA/sonarqube-cleanup-cli/blob/08e77f635043cc725edf7ab5dfc31cad789938e7/src/main/java/org/psa/sonarqube/cleanup/rest/SonarQubeClient.java#L58 Currently the program **takes the most old projects by date analysis** ; of the **main branch**. A warning should be written about branch concept. In case of last analysis on main branch is old (and part of deletion) but some branches are "active" (recent analyze), the project will be deleted. ----- Another approach could perhaps to check branches for project which will be deleted, and avoid deletion if "recent" branches analysis. But this algorithm should be detailed, what is a "recent" branch in this case ...
non_test
details about projects deleted analysis date branch currently the program takes the most old projects by date analysis of the main branch a warning should be written about branch concept in case of last analysis on main branch is old and part of deletion but some branches are active recent analyze the project will be deleted another approach could perhaps to check branches for project which will be deleted and avoid deletion if recent branches analysis but this algorithm should be detailed what is a recent branch in this case
0
68,413
17,273,234,903
IssuesEvent
2021-07-22 23:43:10
o3de/o3de
https://api.github.com/repos/o3de/o3de
closed
LY_3RDPARTY_PATH, LY_PACKAGE_DOWNLOAD_CACHE_LOCATION and LY_PACKAGE_UNPACK_LOCATION
sig/build
**Is your feature request related to a problem? Please describe.** Currently, `LY_PACKAGE_DOWNLOAD_CACHE_LOCATION` and `LY_PACKAGE_UNPACK_LOCATION` default to `@LY_3RDPARTY_PATH@/downloaded_packages` and `@LY_3RDPARTY_PATH@/packages` respectively. Once the project is generated, changing `LY_3RDPARTY_PATH` does not change `LY_PACKAGE_DOWNLOAD_CACHE_LOCATION` nor `LY_PACKAGE_UNPACK_LOCATION`. **Describe the solution you'd like** The problem is that the ${LY_3RDPARTY_PATH} gets expanded the first time and then their value is a cache variable. I think it would be neat if the default is `@LY_3RDPARTY_PATH@/downloaded_packages` and `@LY_3RDPARTY_PATH@/packages` and then we run a `string(CONFIGURE` to expand the variable. That way, if the user wants to set a different `LY_PACKAGE_DOWNLOAD_CACHE_LOCATION` and `LY_PACKAGE_UNPACK_LOCATION` they can, (and make them, e.g. absolute paths pointing to other locations). If they dont change it. changing ${LY_3RDPARTY_PATH} will affect the other two because the `sting(CONFIGURE` will run and expand the value again.
1.0
LY_3RDPARTY_PATH, LY_PACKAGE_DOWNLOAD_CACHE_LOCATION and LY_PACKAGE_UNPACK_LOCATION - **Is your feature request related to a problem? Please describe.** Currently, `LY_PACKAGE_DOWNLOAD_CACHE_LOCATION` and `LY_PACKAGE_UNPACK_LOCATION` default to `@LY_3RDPARTY_PATH@/downloaded_packages` and `@LY_3RDPARTY_PATH@/packages` respectively. Once the project is generated, changing `LY_3RDPARTY_PATH` does not change `LY_PACKAGE_DOWNLOAD_CACHE_LOCATION` nor `LY_PACKAGE_UNPACK_LOCATION`. **Describe the solution you'd like** The problem is that the ${LY_3RDPARTY_PATH} gets expanded the first time and then their value is a cache variable. I think it would be neat if the default is `@LY_3RDPARTY_PATH@/downloaded_packages` and `@LY_3RDPARTY_PATH@/packages` and then we run a `string(CONFIGURE` to expand the variable. That way, if the user wants to set a different `LY_PACKAGE_DOWNLOAD_CACHE_LOCATION` and `LY_PACKAGE_UNPACK_LOCATION` they can, (and make them, e.g. absolute paths pointing to other locations). If they dont change it. changing ${LY_3RDPARTY_PATH} will affect the other two because the `sting(CONFIGURE` will run and expand the value again.
non_test
ly path ly package download cache location and ly package unpack location is your feature request related to a problem please describe currently ly package download cache location and ly package unpack location default to ly path downloaded packages and ly path packages respectively once the project is generated changing ly path does not change ly package download cache location nor ly package unpack location describe the solution you d like the problem is that the ly path gets expanded the first time and then their value is a cache variable i think it would be neat if the default is ly path downloaded packages and ly path packages and then we run a string configure to expand the variable that way if the user wants to set a different ly package download cache location and ly package unpack location they can and make them e g absolute paths pointing to other locations if they dont change it changing ly path will affect the other two because the sting configure will run and expand the value again
0
299,022
25,875,329,653
IssuesEvent
2022-12-14 07:23:24
zephyrproject-rtos/test_results
https://api.github.com/repos/zephyrproject-rtos/test_results
closed
tests-ci : net: virtual: tunnel.ip6ip test No Console Output(Timeout)
bug area: Tests
**Describe the bug** tunnel.ip6ip test is No Console Output(Timeout) on zephyr-v3.2.0-2490-ga1b4896efe46 on mimxrt1170_evk_cm7 see logs for details **To Reproduce** 1. ``` scripts/twister --device-testing --device-serial /dev/ttyACM0 -p mimxrt1170_evk_cm7 --sub-test net.virtual ``` 2. See error **Expected behavior** test pass **Impact** **Logs and console output** ``` E: ***** USAGE FAULT ***** E: Illegal use of the EPSR E: r0/a1: 0x5eb4a428 r1/a2: 0xe000ed00 r2/a3: 0xffffffe8 E: r3/a4: 0x8000544c r12/ip: 0x40460000 r14/lr: 0x3000efd5 E: xpsr: 0x60000000 E: Faulting instruction address (r15/pc): 0x00000000 E: >>> ZEPHYR FATAL ERROR 0: CPU exception on CPU 0 E: Current thread: 0x800019b8 (main) E: Halting system ``` **Environment (please complete the following information):** - OS: (e.g. Linux ) - Toolchain (e.g Zephyr SDK) - Commit SHA or Version used: zephyr-v3.2.0-2490-ga1b4896efe46
1.0
tests-ci : net: virtual: tunnel.ip6ip test No Console Output(Timeout) - **Describe the bug** tunnel.ip6ip test is No Console Output(Timeout) on zephyr-v3.2.0-2490-ga1b4896efe46 on mimxrt1170_evk_cm7 see logs for details **To Reproduce** 1. ``` scripts/twister --device-testing --device-serial /dev/ttyACM0 -p mimxrt1170_evk_cm7 --sub-test net.virtual ``` 2. See error **Expected behavior** test pass **Impact** **Logs and console output** ``` E: ***** USAGE FAULT ***** E: Illegal use of the EPSR E: r0/a1: 0x5eb4a428 r1/a2: 0xe000ed00 r2/a3: 0xffffffe8 E: r3/a4: 0x8000544c r12/ip: 0x40460000 r14/lr: 0x3000efd5 E: xpsr: 0x60000000 E: Faulting instruction address (r15/pc): 0x00000000 E: >>> ZEPHYR FATAL ERROR 0: CPU exception on CPU 0 E: Current thread: 0x800019b8 (main) E: Halting system ``` **Environment (please complete the following information):** - OS: (e.g. Linux ) - Toolchain (e.g Zephyr SDK) - Commit SHA or Version used: zephyr-v3.2.0-2490-ga1b4896efe46
test
tests ci net virtual tunnel test no console output timeout describe the bug tunnel test is no console output timeout on zephyr on evk see logs for details to reproduce scripts twister device testing device serial dev p evk sub test net virtual see error expected behavior test pass impact logs and console output e usage fault e illegal use of the epsr e e ip lr e xpsr e faulting instruction address pc e zephyr fatal error cpu exception on cpu e current thread main e halting system environment please complete the following information os e g linux toolchain e g zephyr sdk commit sha or version used zephyr
1
532,028
15,528,856,526
IssuesEvent
2021-03-13 12:50:37
unitystation/unitystation
https://api.github.com/repos/unitystation/unitystation
closed
Put in placeholder lavaland NPC mobs and add mob spawners to lavaland
Priority: Low Scene Changes Type: Feature
## Description This is a simple request that's come up several times on the discord, mainly by Waffielz; add some hostile mobs to the game and place some mob spawners in lavaland so that miners who go there have just a _little_ bit more to do. - [x] add placeholder hostile NPC mobs to lavaland
1.0
Put in placeholder lavaland NPC mobs and add mob spawners to lavaland - ## Description This is a simple request that's come up several times on the discord, mainly by Waffielz; add some hostile mobs to the game and place some mob spawners in lavaland so that miners who go there have just a _little_ bit more to do. - [x] add placeholder hostile NPC mobs to lavaland
non_test
put in placeholder lavaland npc mobs and add mob spawners to lavaland description this is a simple request that s come up several times on the discord mainly by waffielz add some hostile mobs to the game and place some mob spawners in lavaland so that miners who go there have just a little bit more to do add placeholder hostile npc mobs to lavaland
0
80,595
7,751,827,411
IssuesEvent
2018-05-30 18:19:08
Microsoft/AzureStorageExplorer
https://api.github.com/repos/Microsoft/AzureStorageExplorer
closed
No 'Access Tier'/'AccessTierChangeTime' columns display on container's editor for blob account and V2 account
:white_check_mark: merged testing
**Storage Explorer Version**: <!-- Enter the version of Storage Explorer in which you encountered your issue (see About dialog). -->1.1.0 **Platform**: <!-- Enter the platform (see About dialog). -->Windows **Architecture**: <!-- Enter the architecture of the app (see About dialog). -->x64 **Build Number**: <!-- Enter the build number (see About dialog). -->20180510.2 **Commit**: <!-- Enter the commit ID (see About dialog). -->d271505a **Regression From**: <!-- Was this working on a previous version? If so, which one? -->Previous release 1.1.0 #### Steps to Reproduce: #### 1. <!-- Describe the steps necessary to reproduce your issue. -->Launch Storage Explorer and select a V2 account. 2. <!-- Be as detailed as you can. -->Open one blob container and check whether 'Access Tier'/'AccessTierChangeTime' columns display well or not. #### Expected Experience: #### <!-- Describe the outcome you expect when performing the above steps. -->'Access Tier'/'AccessTierChangeTime' columns display well. #### Actual Experience: #### <!-- Describe the outcome that actually occurs when performing the above steps. -->No 'Access Tier'/'AccessTierChangeTime' columns display on the container's editor. ![columns](https://user-images.githubusercontent.com/34729022/39985314-c7c9e9fa-574d-11e8-8e97-958af3044afa.png)
1.0
No 'Access Tier'/'AccessTierChangeTime' columns display on container's editor for blob account and V2 account - **Storage Explorer Version**: <!-- Enter the version of Storage Explorer in which you encountered your issue (see About dialog). -->1.1.0 **Platform**: <!-- Enter the platform (see About dialog). -->Windows **Architecture**: <!-- Enter the architecture of the app (see About dialog). -->x64 **Build Number**: <!-- Enter the build number (see About dialog). -->20180510.2 **Commit**: <!-- Enter the commit ID (see About dialog). -->d271505a **Regression From**: <!-- Was this working on a previous version? If so, which one? -->Previous release 1.1.0 #### Steps to Reproduce: #### 1. <!-- Describe the steps necessary to reproduce your issue. -->Launch Storage Explorer and select a V2 account. 2. <!-- Be as detailed as you can. -->Open one blob container and check whether 'Access Tier'/'AccessTierChangeTime' columns display well or not. #### Expected Experience: #### <!-- Describe the outcome you expect when performing the above steps. -->'Access Tier'/'AccessTierChangeTime' columns display well. #### Actual Experience: #### <!-- Describe the outcome that actually occurs when performing the above steps. -->No 'Access Tier'/'AccessTierChangeTime' columns display on the container's editor. ![columns](https://user-images.githubusercontent.com/34729022/39985314-c7c9e9fa-574d-11e8-8e97-958af3044afa.png)
test
no access tier accesstierchangetime columns display on container s editor for blob account and account storage explorer version platform windows architecture build number commit regression from previous release steps to reproduce launch storage explorer and select a account open one blob container and check whether access tier accesstierchangetime columns display well or not expected experience access tier accesstierchangetime columns display well actual experience no access tier accesstierchangetime columns display on the container s editor
1
31,541
2,733,777,718
IssuesEvent
2015-04-17 15:49:07
sympy/sympy
https://api.github.com/repos/sympy/sympy
closed
Missing eigenvalues
bug duplicate imported matrices polys Priority: Medium solvers valid
``` When calculating eigenvalues for some matricies I'm working with I get fewer than there should be. For example: In [10]: W Out[10]: Matrix([ [0, 5/8, 3/8, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 4/5, 1/5, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 1/3, 2/3, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1/4, 0, 0, 3/4, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 1/2, 1/2, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1/3, 0, 0, 0, 0, 2/3, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0], [0, 0, 0, 0, 1/2, 0, 0, 0, 0, 0, 0, 1/2, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1], [0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0]]) In [11]: W.eigenvals() Out[11]: {0: 6, 1: 1} In [12]: Wnp = np.array([[np.float(i) for i in row] for row in W.tolist()]) In [13]: np.linalg.eigvals(Wnp) Out[13]: array([ 0.00000000+0.j , 0.00000000+0.j , 0.00000000+0.j , 0.00000000+0.j , 0.00000000+0.j , 0.00000000+0.j , 1.00000000+0.j , 0.64438276+0.69026752j, 0.64438276-0.69026752j, -0.13703613+0.86913433j, -0.13703613-0.86913433j, -0.75734663+0.38822447j, -0.75734663-0.38822447j]) In [14]: sympy.roots(W.charpoly()) Out[14]: {0: 6, 1: 1} In [15]: notice that I only get 7 eigenvalues from the 13x13 matrix. numpy reports all 13. furthermore, I only get 7 roots from the characteristic polynomial. ``` Original issue for #7070: http://code.google.com/p/sympy/issues/detail?id=3971 Original author: https://code.google.com/u/102053718705531995255/
1.0
Missing eigenvalues - ``` When calculating eigenvalues for some matricies I'm working with I get fewer than there should be. For example: In [10]: W Out[10]: Matrix([ [0, 5/8, 3/8, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 4/5, 1/5, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 1/3, 2/3, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1/4, 0, 0, 3/4, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 1/2, 1/2, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1/3, 0, 0, 0, 0, 2/3, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0], [0, 0, 0, 0, 1/2, 0, 0, 0, 0, 0, 0, 1/2, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1], [0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0]]) In [11]: W.eigenvals() Out[11]: {0: 6, 1: 1} In [12]: Wnp = np.array([[np.float(i) for i in row] for row in W.tolist()]) In [13]: np.linalg.eigvals(Wnp) Out[13]: array([ 0.00000000+0.j , 0.00000000+0.j , 0.00000000+0.j , 0.00000000+0.j , 0.00000000+0.j , 0.00000000+0.j , 1.00000000+0.j , 0.64438276+0.69026752j, 0.64438276-0.69026752j, -0.13703613+0.86913433j, -0.13703613-0.86913433j, -0.75734663+0.38822447j, -0.75734663-0.38822447j]) In [14]: sympy.roots(W.charpoly()) Out[14]: {0: 6, 1: 1} In [15]: notice that I only get 7 eigenvalues from the 13x13 matrix. numpy reports all 13. furthermore, I only get 7 roots from the characteristic polynomial. ``` Original issue for #7070: http://code.google.com/p/sympy/issues/detail?id=3971 Original author: https://code.google.com/u/102053718705531995255/
non_test
missing eigenvalues when calculating eigenvalues for some matricies i m working with i get fewer than there should be for example in w out matrix in w eigenvals out in wnp np array for row in w tolist in np linalg eigvals wnp out array j j j j j j j in sympy roots w charpoly out in notice that i only get eigenvalues from the matrix numpy reports all furthermore i only get roots from the characteristic polynomial original issue for original author
0
191,068
14,593,236,606
IssuesEvent
2020-12-19 21:40:20
valadaa48/retroroller
https://api.github.com/repos/valadaa48/retroroller
closed
EmuElec - ability to modify Advanced Game Options in ES
please test
This is handy if you want to change ROM specific options such as core. MAME ROMs can use this feature.
1.0
EmuElec - ability to modify Advanced Game Options in ES - This is handy if you want to change ROM specific options such as core. MAME ROMs can use this feature.
test
emuelec ability to modify advanced game options in es this is handy if you want to change rom specific options such as core mame roms can use this feature
1
61,663
6,744,834,264
IssuesEvent
2017-10-20 17:08:59
yrahul3910/video-sharing-site
https://api.github.com/repos/yrahul3910/video-sharing-site
closed
Validation isn't performed yet
needs testing
The usernames and names aren't validated for junk values and special characters.
1.0
Validation isn't performed yet - The usernames and names aren't validated for junk values and special characters.
test
validation isn t performed yet the usernames and names aren t validated for junk values and special characters
1
250,872
21,367,377,703
IssuesEvent
2022-04-20 04:14:56
dotnet/fsharp
https://api.github.com/repos/dotnet/fsharp
closed
Add unit tests for automatic brace completion
Area-Testing Needs-Test cases
Will need to add unit tests for automatic brace completion, from here: https://github.com/Microsoft/visualfsharp/pull/4049
2.0
Add unit tests for automatic brace completion - Will need to add unit tests for automatic brace completion, from here: https://github.com/Microsoft/visualfsharp/pull/4049
test
add unit tests for automatic brace completion will need to add unit tests for automatic brace completion from here
1
72,502
31,768,920,835
IssuesEvent
2023-09-12 10:28:48
gauravrs18/issue_onboarding
https://api.github.com/repos/gauravrs18/issue_onboarding
closed
dev-angular-integration-account-services-new-connection-component-connect-component -consumer-details-component -activate-component -reject-button-component
CX-account-services
dev-angular-integration-account-services-new-connection-component-connect-component -consumer-details-component -activate-component -reject-button-component
1.0
dev-angular-integration-account-services-new-connection-component-connect-component -consumer-details-component -activate-component -reject-button-component - dev-angular-integration-account-services-new-connection-component-connect-component -consumer-details-component -activate-component -reject-button-component
non_test
dev angular integration account services new connection component connect component consumer details component activate component reject button component dev angular integration account services new connection component connect component consumer details component activate component reject button component
0
28,850
13,853,721,479
IssuesEvent
2020-10-15 08:32:20
hlissner/doom-emacs
https://api.github.com/repos/hlissner/doom-emacs
reopened
open c file with lsp could lead to emacs freeze
:lang cc is:regression re:performance re:stability status:resolved
**What did you expect to happen?** open a c/c++ project with ccls doesn't lead to emacs freeze **What actually happened?** when opening a c/c++ project with ccls enabled, the emacs freeze **Steps to reproduce:** 1. open a c/c++ project **System information:** <details> ``` SYSTEM type gnu/linux config x86_64-pc-linux-gnu shell /usr/bin/zsh uname Linux 5.4.67-1-MANJARO #1 SMP PREEMPT Wed Sep 23 14:20:18 UTC 2020 x86_64 path (~/.local/bin /usr/local/bin /usr/bin /bin /usr/local/sbin /opt/android-ndk /usr/lib/jvm/default/bin /usr/bin/site_perl /usr/bin/vendor_perl /usr/bin/core_perl /usr/lib/emacs/27.1/x86_64-pc-linux-gnu) EMACS dir ~/.emacs.d/ version 27.1 build 8月 29, 2020 buildopts --prefix=/usr --sysconfdir=/etc --libexecdir=/usr/lib --localstatedir=/var --with-x-toolkit=gtk3 --with-xft --with-wide-int --with-modules --with-cairo --with-harfbuzz 'CFLAGS=-march=x86-64 -mtune=generic -O2 -pipe -fno-plt' CPPFLAGS=-D_FORTIFY_SOURCE=2 LDFLAGS=-Wl,-O1,--sort-common,--as-needed,-z,relro,-z,now features XPM JPEG TIFF GIF PNG RSVG CAIRO SOUND GPM DBUS GSETTINGS GLIB NOTIFY INOTIFY ACL GNUTLS LIBXML2 FREETYPE HARFBUZZ M17N_FLT LIBOTF ZLIB TOOLKIT_SCROLL_BARS GTK3 X11 XDBE XIM MODULES THREADS LIBSYSTEMD JSON PDUMPER LCMS2 GMP traits (server-running) DOOM dir ~/.doom.d/ version 2.0.9 build HEAD -> upstream, zero/develop, origin/upstream a776664eb 2020-10-14 01:40:11 -0400 elc-files 0 modules (:completion company (ivy +fuzzy) :ui deft doom doom-dashboard doom-quit hl-todo modeline nav-flash ophints (popup +defaults) treemacs vc-gutter vi-tilde-fringe window-select workspaces :editor file-templates fold format snippets :emacs dired electric undo vc :term vterm :checkers syntax :tools (eval +overlay) (lookup +docsets) lsp magit :lang (cc +lsp) data emacs-lisp lua markdown org (python +lsp) (rust +lsp) (sh +zsh) (web +html +css) :email mu4e :config (default +bindings +smartparens) :private (&user prow-edit) (&user prow-c-c++) (&user prow-misc) (&user gtags)) packages ((cnfonts) (jinja2-mode) (mu4e-maildirs-extension :disable t) (org-mu4e :disable t) (glsl-mode :disable t) (opencl-mode :disable t) (disaster :disable t) (cuda-mode :disable t) (demangle-mode :disable t)) unpin (n/a) elpa (n/a) ``` </details>
True
open c file with lsp could lead to emacs freeze - **What did you expect to happen?** open a c/c++ project with ccls doesn't lead to emacs freeze **What actually happened?** when opening a c/c++ project with ccls enabled, the emacs freeze **Steps to reproduce:** 1. open a c/c++ project **System information:** <details> ``` SYSTEM type gnu/linux config x86_64-pc-linux-gnu shell /usr/bin/zsh uname Linux 5.4.67-1-MANJARO #1 SMP PREEMPT Wed Sep 23 14:20:18 UTC 2020 x86_64 path (~/.local/bin /usr/local/bin /usr/bin /bin /usr/local/sbin /opt/android-ndk /usr/lib/jvm/default/bin /usr/bin/site_perl /usr/bin/vendor_perl /usr/bin/core_perl /usr/lib/emacs/27.1/x86_64-pc-linux-gnu) EMACS dir ~/.emacs.d/ version 27.1 build 8月 29, 2020 buildopts --prefix=/usr --sysconfdir=/etc --libexecdir=/usr/lib --localstatedir=/var --with-x-toolkit=gtk3 --with-xft --with-wide-int --with-modules --with-cairo --with-harfbuzz 'CFLAGS=-march=x86-64 -mtune=generic -O2 -pipe -fno-plt' CPPFLAGS=-D_FORTIFY_SOURCE=2 LDFLAGS=-Wl,-O1,--sort-common,--as-needed,-z,relro,-z,now features XPM JPEG TIFF GIF PNG RSVG CAIRO SOUND GPM DBUS GSETTINGS GLIB NOTIFY INOTIFY ACL GNUTLS LIBXML2 FREETYPE HARFBUZZ M17N_FLT LIBOTF ZLIB TOOLKIT_SCROLL_BARS GTK3 X11 XDBE XIM MODULES THREADS LIBSYSTEMD JSON PDUMPER LCMS2 GMP traits (server-running) DOOM dir ~/.doom.d/ version 2.0.9 build HEAD -> upstream, zero/develop, origin/upstream a776664eb 2020-10-14 01:40:11 -0400 elc-files 0 modules (:completion company (ivy +fuzzy) :ui deft doom doom-dashboard doom-quit hl-todo modeline nav-flash ophints (popup +defaults) treemacs vc-gutter vi-tilde-fringe window-select workspaces :editor file-templates fold format snippets :emacs dired electric undo vc :term vterm :checkers syntax :tools (eval +overlay) (lookup +docsets) lsp magit :lang (cc +lsp) data emacs-lisp lua markdown org (python +lsp) (rust +lsp) (sh +zsh) (web +html +css) :email mu4e :config (default +bindings +smartparens) :private (&user prow-edit) (&user prow-c-c++) (&user prow-misc) (&user gtags)) packages ((cnfonts) (jinja2-mode) (mu4e-maildirs-extension :disable t) (org-mu4e :disable t) (glsl-mode :disable t) (opencl-mode :disable t) (disaster :disable t) (cuda-mode :disable t) (demangle-mode :disable t)) unpin (n/a) elpa (n/a) ``` </details>
non_test
open c file with lsp could lead to emacs freeze what did you expect to happen open a c c project with ccls doesn t lead to emacs freeze what actually happened when opening a c c project with ccls enabled the emacs freeze steps to reproduce open a c c project system information system type gnu linux config pc linux gnu shell usr bin zsh uname linux manjaro smp preempt wed sep utc path local bin usr local bin usr bin bin usr local sbin opt android ndk usr lib jvm default bin usr bin site perl usr bin vendor perl usr bin core perl usr lib emacs pc linux gnu emacs dir emacs d version build buildopts prefix usr sysconfdir etc libexecdir usr lib localstatedir var with x toolkit with xft with wide int with modules with cairo with harfbuzz cflags march mtune generic pipe fno plt cppflags d fortify source ldflags wl sort common as needed z relro z now features xpm jpeg tiff gif png rsvg cairo sound gpm dbus gsettings glib notify inotify acl gnutls freetype harfbuzz flt libotf zlib toolkit scroll bars xdbe xim modules threads libsystemd json pdumper gmp traits server running doom dir doom d version build head upstream zero develop origin upstream elc files modules completion company ivy fuzzy ui deft doom doom dashboard doom quit hl todo modeline nav flash ophints popup defaults treemacs vc gutter vi tilde fringe window select workspaces editor file templates fold format snippets emacs dired electric undo vc term vterm checkers syntax tools eval overlay lookup docsets lsp magit lang cc lsp data emacs lisp lua markdown org python lsp rust lsp sh zsh web html css email config default bindings smartparens private user prow edit user prow c c user prow misc user gtags packages cnfonts mode maildirs extension disable t org disable t glsl mode disable t opencl mode disable t disaster disable t cuda mode disable t demangle mode disable t unpin n a elpa n a
0
395,019
27,050,298,438
IssuesEvent
2023-02-13 12:47:51
learntocloud/cloud-dictionary
https://api.github.com/repos/learntocloud/cloud-dictionary
closed
it is network oriented virtual storage space and processing services by internet
documentation good first issue
- Word [REQUIRED]: The word you are defining. Check [this URL](https://clouddictionary.azurewebsites.net/api/GetWords) for all words we currently have. - Content [REQUIRED]: The definition. No more than 3 sentences. - Author Name [REQUIRED]: Your name. - Author Link [REQUIRED]: The URL you want your name to link to. - Learn More URL [REQUIRED]: Page where people can visit to learn more about the word. - Category [REQUIRED, select one]: Tech category the word fits in. Options: compute, security, service, general, analytics, developer tools, web, networking, database, storage, devops, ai/ml, identiy, iot, monitoring. - Abbreviation [OPTIONAL]: If the word is commonly abbreviated, please provide it. For example, command line interface is often abbreviated as CLI.
1.0
it is network oriented virtual storage space and processing services by internet - - Word [REQUIRED]: The word you are defining. Check [this URL](https://clouddictionary.azurewebsites.net/api/GetWords) for all words we currently have. - Content [REQUIRED]: The definition. No more than 3 sentences. - Author Name [REQUIRED]: Your name. - Author Link [REQUIRED]: The URL you want your name to link to. - Learn More URL [REQUIRED]: Page where people can visit to learn more about the word. - Category [REQUIRED, select one]: Tech category the word fits in. Options: compute, security, service, general, analytics, developer tools, web, networking, database, storage, devops, ai/ml, identiy, iot, monitoring. - Abbreviation [OPTIONAL]: If the word is commonly abbreviated, please provide it. For example, command line interface is often abbreviated as CLI.
non_test
it is network oriented virtual storage space and processing services by internet word the word you are defining check for all words we currently have content the definition no more than sentences author name your name author link the url you want your name to link to learn more url page where people can visit to learn more about the word category tech category the word fits in options compute security service general analytics developer tools web networking database storage devops ai ml identiy iot monitoring abbreviation if the word is commonly abbreviated please provide it for example command line interface is often abbreviated as cli
0
54,404
13,907,082,705
IssuesEvent
2020-10-20 12:09:17
three11/email-template
https://api.github.com/repos/three11/email-template
closed
CVE-2020-7733 (High) detected in ua-parser-js-0.7.17.tgz
security vulnerability
## CVE-2020-7733 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ua-parser-js-0.7.17.tgz</b></p></summary> <p>Lightweight JavaScript-based user-agent string parser</p> <p>Library home page: <a href="https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz">https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz</a></p> <p>Path to dependency file: email-template/package.json</p> <p>Path to vulnerable library: email-template/node_modules/ua-parser-js/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.26.7.tgz (Root Library) - :x: **ua-parser-js-0.7.17.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/three11/email-template/commit/77f963fc45d92d0604bfc6cb13509271f9a2233b">77f963fc45d92d0604bfc6cb13509271f9a2233b</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package ua-parser-js before 0.7.22 are vulnerable to Regular Expression Denial of Service (ReDoS) via the regex for Redmi Phones and Mi Pad Tablets UA. <p>Publish Date: 2020-09-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7733>CVE-2020-7733</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: 0.7.22</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7733 (High) detected in ua-parser-js-0.7.17.tgz - ## CVE-2020-7733 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ua-parser-js-0.7.17.tgz</b></p></summary> <p>Lightweight JavaScript-based user-agent string parser</p> <p>Library home page: <a href="https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz">https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz</a></p> <p>Path to dependency file: email-template/package.json</p> <p>Path to vulnerable library: email-template/node_modules/ua-parser-js/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.26.7.tgz (Root Library) - :x: **ua-parser-js-0.7.17.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/three11/email-template/commit/77f963fc45d92d0604bfc6cb13509271f9a2233b">77f963fc45d92d0604bfc6cb13509271f9a2233b</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package ua-parser-js before 0.7.22 are vulnerable to Regular Expression Denial of Service (ReDoS) via the regex for Redmi Phones and Mi Pad Tablets UA. <p>Publish Date: 2020-09-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7733>CVE-2020-7733</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: 0.7.22</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in ua parser js tgz cve high severity vulnerability vulnerable library ua parser js tgz lightweight javascript based user agent string parser library home page a href path to dependency file email template package json path to vulnerable library email template node modules ua parser js package json dependency hierarchy browser sync tgz root library x ua parser js tgz vulnerable library found in head commit a href found in base branch main vulnerability details the package ua parser js before are vulnerable to regular expression denial of service redos via the regex for redmi phones and mi pad tablets ua publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
51,656
13,635,153,786
IssuesEvent
2020-09-25 02:03:15
nasifimtiazohi/openmrs-module-metadatamapping-1.3.4
https://api.github.com/repos/nasifimtiazohi/openmrs-module-metadatamapping-1.3.4
opened
CVE-2018-16487 (Medium) detected in lodash-3.10.1.tgz
security vulnerability
## CVE-2018-16487 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-3.10.1.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p> <p>Path to dependency file: openmrs-module-metadatamapping-1.3.4/owa/package.json</p> <p>Path to vulnerable library: openmrs-module-metadatamapping-1.3.4/owa/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - babel-core-6.2.1.tgz (Root Library) - :x: **lodash-3.10.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/nasifimtiazohi/openmrs-module-metadatamapping-1.3.4/commit/dbf14247c8c0a7b64ae301a8ab42df19cc87107e">dbf14247c8c0a7b64ae301a8ab42df19cc87107e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A prototype pollution vulnerability was found in lodash <4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype. <p>Publish Date: 2019-02-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487>CVE-2018-16487</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487</a></p> <p>Release Date: 2019-02-01</p> <p>Fix Resolution: 4.17.11</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-16487 (Medium) detected in lodash-3.10.1.tgz - ## CVE-2018-16487 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-3.10.1.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p> <p>Path to dependency file: openmrs-module-metadatamapping-1.3.4/owa/package.json</p> <p>Path to vulnerable library: openmrs-module-metadatamapping-1.3.4/owa/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - babel-core-6.2.1.tgz (Root Library) - :x: **lodash-3.10.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/nasifimtiazohi/openmrs-module-metadatamapping-1.3.4/commit/dbf14247c8c0a7b64ae301a8ab42df19cc87107e">dbf14247c8c0a7b64ae301a8ab42df19cc87107e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A prototype pollution vulnerability was found in lodash <4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype. <p>Publish Date: 2019-02-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487>CVE-2018-16487</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487</a></p> <p>Release Date: 2019-02-01</p> <p>Fix Resolution: 4.17.11</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in lodash tgz cve medium severity vulnerability vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file openmrs module metadatamapping owa package json path to vulnerable library openmrs module metadatamapping owa node modules lodash package json dependency hierarchy babel core tgz root library x lodash tgz vulnerable library found in head commit a href found in base branch master vulnerability details a prototype pollution vulnerability was found in lodash where the functions merge mergewith and defaultsdeep can be tricked into adding or modifying properties of object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
67,474
20,963,543,006
IssuesEvent
2022-03-28 02:39:16
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
opened
Unable to compile on ARM64 with clang
Type: Defect
<!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Lenovo c630 w/Snapdragon sdm850 cortex.a75-cortex.a55 --- | --- Distribution Name | Arch Distribution Version | Kernel Version | 5.17.0 Architecture | ARM64 OpenZFS Version | Latest git + stable release <!-- Command to find OpenZFS version: zfs version Commands to find kernel version: uname -r # Linux freebsd-version -r # FreeBSD --> ### Describe the problem you're observing While building under GCC works fine, I am unable to with clang+lto or -lto ### Describe how to reproduce the problem Compile linux kernel with ZFS +clang or compile ZFS as a dkms module for a kernel compiled with clang ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` --> LTO [M] drivers/bluetooth/btrsi.lto.o LTO [M] drivers/target/target_core_user.lto.o GEN .version CHK include/generated/compile.h GEN .tmp_initcalls.lds GEN .tmp_symversions.lds LTO vmlinux.o ld.lld: error: couldn't allocate output register for constraint '{v0}' at line 2150099659 ld.lld: error: couldn't allocate output register for constraint '{v7}' at line 2150100117 ld.lld: error: couldn't allocate input reg for constraint '{v1}' at line 2150100832 ld.lld: error: couldn't allocate output register for constraint '{v0}' at line 2150101251 ld.lld: error: couldn't allocate output register for constraint '{v7}' at line 2150101709 ld.lld: error: couldn't allocate input reg for constraint '{v1}' at line 2150102484 ld.lld: error: couldn't allocate output register for constraint '{v0}' at line 2157696494 ld.lld: error: couldn't allocate input reg for constraint '{v0}' at line 2157700699 ld.lld: error: couldn't allocate output register for constraint '{v0}' at line 2157704887 ld.lld: error: couldn't allocate input reg for constraint '{v0}' at line 2157709206 ld.lld: error: couldn't allocate output register for constraint '{v0}' at line 2157714137 ld.lld: error: couldn't allocate output register for constraint '{v0}' at line 2157718730 ld.lld: error: couldn't allocate input reg for constraint '{v0}' at line 2157724053 ld.lld: error: couldn't allocate output register for constraint '{v0}' at line 2157728241 ld.lld: error: couldn't allocate output register for constraint '{v0}' at line 2157732948 ld.lld: error: couldn't allocate input reg for constraint '{v0}' at line 2157738385 ld.lld: error: couldn't allocate output register for constraint '{v16}' at line 2157781218 ld.lld: error: couldn't allocate output register for constraint '{v0}' at line 2157782350 ld.lld: error: couldn't allocate output register for constraint '{v4}' at line 2157786517 ld.lld: error: couldn't allocate output register for constraint '{v4}' at line 2157789877 ld.lld: error: too many errors emitted, stopping now (use -error-limit=0 to see all errors) make: *** [Makefile:1156: vmlinux] Error 1 ==> ERROR: A failure occurred in build(). Aborting... ``` ``` More detail is obtained when not using LTO, here is output without LTO and using with "make v=1" : CC fs/proc/inode.o fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:151:2: error: couldn't allocate output register for constraint '{v0}' NEON_INIT_LOOP(); ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:75:6: note: expanded from macro 'NEON_INIT_LOOP' asm("eor %[ZERO].16b,%[ZERO].16b,%[ZERO].16b\n" \ ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:154:3: error: couldn't allocate output register for constraint '{v7}' NEON_MAIN_LOOP(NEON_DONT_REVERSE); ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:93:6: note: expanded from macro 'NEON_MAIN_LOOP' asm("ld1 { %[SRC].4s }, %[IP]\n" \ ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:157:2: error: couldn't allocate input reg for constraint '{v1}' NEON_FINI_LOOP(); ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:112:6: note: expanded from macro 'NEON_FINI_LOOP' asm("st1 { %[ACC0].4s },%[DST0]\n" \ ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:190:2: error: couldn't allocate output register for constraint '{v0}' NEON_INIT_LOOP(); ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:75:6: note: expanded from macro 'NEON_INIT_LOOP' asm("eor %[ZERO].16b,%[ZERO].16b,%[ZERO].16b\n" \ ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:193:3: error: couldn't allocate output register for constraint '{v7}' NEON_MAIN_LOOP(NEON_DO_REVERSE); ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:93:6: note: expanded from macro 'NEON_MAIN_LOOP' asm("ld1 { %[SRC].4s }, %[IP]\n" \ ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:196:2: error: couldn't allocate input reg for constraint '{v1}' NEON_FINI_LOOP(); ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:112:6: note: expanded from macro 'NEON_FINI_LOOP' asm("st1 { %[ACC0].4s },%[DST0]\n" \ ^ 6 errors generated. make[3]: *** [scripts/Makefile.build:288: fs/zfs/zcommon/zfs_fletcher_aarch64_neon.o] Error 1 make[2]: *** [scripts/Makefile.build:550: fs/zfs/zcommon] Error 2 make[1]: *** [scripts/Makefile.build:550: fs/zfs] Error 2 make[1]: *** Waiting for unfinished jobs.... ``` ```
1.0
Unable to compile on ARM64 with clang - <!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Lenovo c630 w/Snapdragon sdm850 cortex.a75-cortex.a55 --- | --- Distribution Name | Arch Distribution Version | Kernel Version | 5.17.0 Architecture | ARM64 OpenZFS Version | Latest git + stable release <!-- Command to find OpenZFS version: zfs version Commands to find kernel version: uname -r # Linux freebsd-version -r # FreeBSD --> ### Describe the problem you're observing While building under GCC works fine, I am unable to with clang+lto or -lto ### Describe how to reproduce the problem Compile linux kernel with ZFS +clang or compile ZFS as a dkms module for a kernel compiled with clang ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` --> LTO [M] drivers/bluetooth/btrsi.lto.o LTO [M] drivers/target/target_core_user.lto.o GEN .version CHK include/generated/compile.h GEN .tmp_initcalls.lds GEN .tmp_symversions.lds LTO vmlinux.o ld.lld: error: couldn't allocate output register for constraint '{v0}' at line 2150099659 ld.lld: error: couldn't allocate output register for constraint '{v7}' at line 2150100117 ld.lld: error: couldn't allocate input reg for constraint '{v1}' at line 2150100832 ld.lld: error: couldn't allocate output register for constraint '{v0}' at line 2150101251 ld.lld: error: couldn't allocate output register for constraint '{v7}' at line 2150101709 ld.lld: error: couldn't allocate input reg for constraint '{v1}' at line 2150102484 ld.lld: error: couldn't allocate output register for constraint '{v0}' at line 2157696494 ld.lld: error: couldn't allocate input reg for constraint '{v0}' at line 2157700699 ld.lld: error: couldn't allocate output register for constraint '{v0}' at line 2157704887 ld.lld: error: couldn't allocate input reg for constraint '{v0}' at line 2157709206 ld.lld: error: couldn't allocate output register for constraint '{v0}' at line 2157714137 ld.lld: error: couldn't allocate output register for constraint '{v0}' at line 2157718730 ld.lld: error: couldn't allocate input reg for constraint '{v0}' at line 2157724053 ld.lld: error: couldn't allocate output register for constraint '{v0}' at line 2157728241 ld.lld: error: couldn't allocate output register for constraint '{v0}' at line 2157732948 ld.lld: error: couldn't allocate input reg for constraint '{v0}' at line 2157738385 ld.lld: error: couldn't allocate output register for constraint '{v16}' at line 2157781218 ld.lld: error: couldn't allocate output register for constraint '{v0}' at line 2157782350 ld.lld: error: couldn't allocate output register for constraint '{v4}' at line 2157786517 ld.lld: error: couldn't allocate output register for constraint '{v4}' at line 2157789877 ld.lld: error: too many errors emitted, stopping now (use -error-limit=0 to see all errors) make: *** [Makefile:1156: vmlinux] Error 1 ==> ERROR: A failure occurred in build(). Aborting... ``` ``` More detail is obtained when not using LTO, here is output without LTO and using with "make v=1" : CC fs/proc/inode.o fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:151:2: error: couldn't allocate output register for constraint '{v0}' NEON_INIT_LOOP(); ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:75:6: note: expanded from macro 'NEON_INIT_LOOP' asm("eor %[ZERO].16b,%[ZERO].16b,%[ZERO].16b\n" \ ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:154:3: error: couldn't allocate output register for constraint '{v7}' NEON_MAIN_LOOP(NEON_DONT_REVERSE); ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:93:6: note: expanded from macro 'NEON_MAIN_LOOP' asm("ld1 { %[SRC].4s }, %[IP]\n" \ ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:157:2: error: couldn't allocate input reg for constraint '{v1}' NEON_FINI_LOOP(); ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:112:6: note: expanded from macro 'NEON_FINI_LOOP' asm("st1 { %[ACC0].4s },%[DST0]\n" \ ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:190:2: error: couldn't allocate output register for constraint '{v0}' NEON_INIT_LOOP(); ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:75:6: note: expanded from macro 'NEON_INIT_LOOP' asm("eor %[ZERO].16b,%[ZERO].16b,%[ZERO].16b\n" \ ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:193:3: error: couldn't allocate output register for constraint '{v7}' NEON_MAIN_LOOP(NEON_DO_REVERSE); ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:93:6: note: expanded from macro 'NEON_MAIN_LOOP' asm("ld1 { %[SRC].4s }, %[IP]\n" \ ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:196:2: error: couldn't allocate input reg for constraint '{v1}' NEON_FINI_LOOP(); ^ fs/zfs/zcommon/zfs_fletcher_aarch64_neon.c:112:6: note: expanded from macro 'NEON_FINI_LOOP' asm("st1 { %[ACC0].4s },%[DST0]\n" \ ^ 6 errors generated. make[3]: *** [scripts/Makefile.build:288: fs/zfs/zcommon/zfs_fletcher_aarch64_neon.o] Error 1 make[2]: *** [scripts/Makefile.build:550: fs/zfs/zcommon] Error 2 make[1]: *** [scripts/Makefile.build:550: fs/zfs] Error 2 make[1]: *** Waiting for unfinished jobs.... ``` ```
non_test
unable to compile on with clang thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type lenovo w snapdragon cortex cortex distribution name arch distribution version kernel version architecture openzfs version latest git stable release command to find openzfs version zfs version commands to find kernel version uname r linux freebsd version r freebsd describe the problem you re observing while building under gcc works fine i am unable to with clang lto or lto describe how to reproduce the problem compile linux kernel with zfs clang or compile zfs as a dkms module for a kernel compiled with clang include any warning errors backtraces from the system logs important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with lto drivers bluetooth btrsi lto o lto drivers target target core user lto o gen version chk include generated compile h gen tmp initcalls lds gen tmp symversions lds lto vmlinux o ld lld error couldn t allocate output register for constraint at line ld lld error couldn t allocate output register for constraint at line ld lld error couldn t allocate input reg for constraint at line ld lld error couldn t allocate output register for constraint at line ld lld error couldn t allocate output register for constraint at line ld lld error couldn t allocate input reg for constraint at line ld lld error couldn t allocate output register for constraint at line ld lld error couldn t allocate input reg for constraint at line ld lld error couldn t allocate output register for constraint at line ld lld error couldn t allocate input reg for constraint at line ld lld error couldn t allocate output register for constraint at line ld lld error couldn t allocate output register for constraint at line ld lld error couldn t allocate input reg for constraint at line ld lld error couldn t allocate output register for constraint at line ld lld error couldn t allocate output register for constraint at line ld lld error couldn t allocate input reg for constraint at line ld lld error couldn t allocate output register for constraint at line ld lld error couldn t allocate output register for constraint at line ld lld error couldn t allocate output register for constraint at line ld lld error couldn t allocate output register for constraint at line ld lld error too many errors emitted stopping now use error limit to see all errors make error error a failure occurred in build aborting more detail is obtained when not using lto here is output without lto and using with make v cc fs proc inode o fs zfs zcommon zfs fletcher neon c error couldn t allocate output register for constraint neon init loop fs zfs zcommon zfs fletcher neon c note expanded from macro neon init loop asm eor n fs zfs zcommon zfs fletcher neon c error couldn t allocate output register for constraint neon main loop neon dont reverse fs zfs zcommon zfs fletcher neon c note expanded from macro neon main loop asm n fs zfs zcommon zfs fletcher neon c error couldn t allocate input reg for constraint neon fini loop fs zfs zcommon zfs fletcher neon c note expanded from macro neon fini loop asm n fs zfs zcommon zfs fletcher neon c error couldn t allocate output register for constraint neon init loop fs zfs zcommon zfs fletcher neon c note expanded from macro neon init loop asm eor n fs zfs zcommon zfs fletcher neon c error couldn t allocate output register for constraint neon main loop neon do reverse fs zfs zcommon zfs fletcher neon c note expanded from macro neon main loop asm n fs zfs zcommon zfs fletcher neon c error couldn t allocate input reg for constraint neon fini loop fs zfs zcommon zfs fletcher neon c note expanded from macro neon fini loop asm n errors generated make error make error make error make waiting for unfinished jobs
0
478,799
13,785,731,473
IssuesEvent
2020-10-08 23:41:19
googlemaps/js-api-loader
https://api.github.com/repos/googlemaps/js-api-loader
closed
Documentation page disappeared?
priority: p1 type: bug
Hi, the [link](https://googlemaps.github.io/js-api-loader/docs/index.html) to the documentation page in the Readme looks dead. Where did they go?
1.0
Documentation page disappeared? - Hi, the [link](https://googlemaps.github.io/js-api-loader/docs/index.html) to the documentation page in the Readme looks dead. Where did they go?
non_test
documentation page disappeared hi the to the documentation page in the readme looks dead where did they go
0
38,196
5,168,685,572
IssuesEvent
2017-01-17 22:16:03
cryptonomex/graphene
https://api.github.com/repos/cryptonomex/graphene
closed
Implement and test virtual operations
needs testing
Some big fat TODO's in `db_market.cpp` here: https://github.com/cryptonomex/graphene/blob/eeeab17477a635fcff7f2d294dff8ecf25449b10/libraries/chain/db_market.cpp#L89-L116 This is basically to generate a virtual cancel op to inform account history listeners that orders have been cancelled due to certain asset management actions.
1.0
Implement and test virtual operations - Some big fat TODO's in `db_market.cpp` here: https://github.com/cryptonomex/graphene/blob/eeeab17477a635fcff7f2d294dff8ecf25449b10/libraries/chain/db_market.cpp#L89-L116 This is basically to generate a virtual cancel op to inform account history listeners that orders have been cancelled due to certain asset management actions.
test
implement and test virtual operations some big fat todo s in db market cpp here this is basically to generate a virtual cancel op to inform account history listeners that orders have been cancelled due to certain asset management actions
1
655,519
21,693,642,098
IssuesEvent
2022-05-09 17:46:15
rich-iannone/pointblank
https://api.github.com/repos/rich-iannone/pointblank
closed
BigQuery support
Difficulty: [3] Advanced Effort: [3] High Type: ★ Enhancement Priority: [3] High
### Discussed in https://github.com/rich-iannone/pointblank/discussions/404 Thanks for your comment. I'd be happy to help test this, and thanks for the link. I'll if I can get it running with one of the public big query datasets, and report back to you how I fare. --- <div type='discussions-op-text'> <sup>Originally posted by **good-marketing** April 25, 2022</sup> Hi, Thanks for creating pointblank, it is a next level R package, in terms of documentation, approach and execution. [ Hat tip]. I work in GCP environment, and was wondering if support for BigQuery (https://bigrquery.r-dbi.org/) is something you might consider. Cheers, Bob</div>
1.0
BigQuery support - ### Discussed in https://github.com/rich-iannone/pointblank/discussions/404 Thanks for your comment. I'd be happy to help test this, and thanks for the link. I'll if I can get it running with one of the public big query datasets, and report back to you how I fare. --- <div type='discussions-op-text'> <sup>Originally posted by **good-marketing** April 25, 2022</sup> Hi, Thanks for creating pointblank, it is a next level R package, in terms of documentation, approach and execution. [ Hat tip]. I work in GCP environment, and was wondering if support for BigQuery (https://bigrquery.r-dbi.org/) is something you might consider. Cheers, Bob</div>
non_test
bigquery support discussed in thanks for your comment i d be happy to help test this and thanks for the link i ll if i can get it running with one of the public big query datasets and report back to you how i fare originally posted by good marketing april hi thanks for creating pointblank it is a next level r package in terms of documentation approach and execution i work in gcp environment and was wondering if support for bigquery is something you might consider cheers bob
0
4,839
17,702,497,286
IssuesEvent
2021-08-25 00:54:58
MinaProtocol/mina
https://api.github.com/repos/MinaProtocol/mina
closed
Integration Test: Archive Node
acceptance-automation
Write a new test which deploys a network with an archive node and a few block producers. Let the network run for a bit while ensuring the archive node does not throw any errors (the log engine automatically accumulates errors from nodes to fail the test, but the archive node is currently not deployed or subscribed to). Additionally, once the test it complete, have the test export data from the archive node and save this information to the local filesystem. In the near future, we will run further tests against this exported data to test the archive recovery functionality. The exported data should include a psql dump of the database as well as a copy of the block log. Implementing this will meet the acceptance feature requirement for archive node exporting.
1.0
Integration Test: Archive Node - Write a new test which deploys a network with an archive node and a few block producers. Let the network run for a bit while ensuring the archive node does not throw any errors (the log engine automatically accumulates errors from nodes to fail the test, but the archive node is currently not deployed or subscribed to). Additionally, once the test it complete, have the test export data from the archive node and save this information to the local filesystem. In the near future, we will run further tests against this exported data to test the archive recovery functionality. The exported data should include a psql dump of the database as well as a copy of the block log. Implementing this will meet the acceptance feature requirement for archive node exporting.
non_test
integration test archive node write a new test which deploys a network with an archive node and a few block producers let the network run for a bit while ensuring the archive node does not throw any errors the log engine automatically accumulates errors from nodes to fail the test but the archive node is currently not deployed or subscribed to additionally once the test it complete have the test export data from the archive node and save this information to the local filesystem in the near future we will run further tests against this exported data to test the archive recovery functionality the exported data should include a psql dump of the database as well as a copy of the block log implementing this will meet the acceptance feature requirement for archive node exporting
0
351,655
32,019,355,350
IssuesEvent
2023-09-22 02:11:52
opensearch-project/dashboards-visualizations
https://api.github.com/repos/opensearch-project/dashboards-visualizations
closed
[AUTOCUT] Integration Test failed for ganttChartDashboards: 2.10.0 tar distribution
untriaged autocut integ-test-failure v2.10.0
The integration test failed at distribution level for component ganttChartDashboards<br>Version: 2.10.0<br>Distribution: tar<br>Architecture: x64<br>Platform: linux<br><br>Please check the logs: https://build.ci.opensearch.org/job/integ-test-opensearch-dashboards/4181/display/redirect<br><br> * Test-report manifest:*<br> - https://ci.opensearch.org/ci/dbc/integ-test-opensearch-dashboards/2.10.0/6612/linux/x64/tar/test-results/4181/integ-test/test-report.yml <br><br> _Note: Steps to reproduce, additional logs and other files can be found within the above test-report manifest. <br>Instructions of this test-report manifest can be found [here](https://github.com/opensearch-project/opensearch-build/tree/main/src/report_workflow#guide-on-test-report-manifest-from-ci)._
1.0
[AUTOCUT] Integration Test failed for ganttChartDashboards: 2.10.0 tar distribution - The integration test failed at distribution level for component ganttChartDashboards<br>Version: 2.10.0<br>Distribution: tar<br>Architecture: x64<br>Platform: linux<br><br>Please check the logs: https://build.ci.opensearch.org/job/integ-test-opensearch-dashboards/4181/display/redirect<br><br> * Test-report manifest:*<br> - https://ci.opensearch.org/ci/dbc/integ-test-opensearch-dashboards/2.10.0/6612/linux/x64/tar/test-results/4181/integ-test/test-report.yml <br><br> _Note: Steps to reproduce, additional logs and other files can be found within the above test-report manifest. <br>Instructions of this test-report manifest can be found [here](https://github.com/opensearch-project/opensearch-build/tree/main/src/report_workflow#guide-on-test-report-manifest-from-ci)._
test
integration test failed for ganttchartdashboards tar distribution the integration test failed at distribution level for component ganttchartdashboards version distribution tar architecture platform linux please check the logs test report manifest note steps to reproduce additional logs and other files can be found within the above test report manifest instructions of this test report manifest can be found
1
258,906
27,591,588,178
IssuesEvent
2023-03-09 01:04:15
L00171222/easybuggy4django
https://api.github.com/repos/L00171222/easybuggy4django
opened
psutil-5.4.3.tar.gz: 1 vulnerabilities (highest severity is: 7.5)
Mend: dependency security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>psutil-5.4.3.tar.gz</b></p></summary> <p>Cross-platform lib for process and system monitoring in Python.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/e2/e1/600326635f97fee89bf8426fef14c5c29f4849c79f68fd79f433d8c1bd96/psutil-5.4.3.tar.gz">https://files.pythonhosted.org/packages/e2/e1/600326635f97fee89bf8426fef14c5c29f4849c79f68fd79f433d8c1bd96/psutil-5.4.3.tar.gz</a></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (psutil version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2019-18874](https://www.mend.io/vulnerability-database/CVE-2019-18874) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | psutil-5.4.3.tar.gz | Direct | 5.6.6 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-18874</summary> ### Vulnerable Library - <b>psutil-5.4.3.tar.gz</b></p> <p>Cross-platform lib for process and system monitoring in Python.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/e2/e1/600326635f97fee89bf8426fef14c5c29f4849c79f68fd79f433d8c1bd96/psutil-5.4.3.tar.gz">https://files.pythonhosted.org/packages/e2/e1/600326635f97fee89bf8426fef14c5c29f4849c79f68fd79f433d8c1bd96/psutil-5.4.3.tar.gz</a></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> Dependency Hierarchy: - :x: **psutil-5.4.3.tar.gz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> psutil (aka python-psutil) through 5.6.5 can have a double free. This occurs because of refcount mishandling within a while or for loop that converts system data into a Python object. <p>Publish Date: 2019-11-12 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-18874>CVE-2019-18874</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-18874">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-18874</a></p> <p>Release Date: 2019-11-12</p> <p>Fix Resolution: 5.6.6</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
True
psutil-5.4.3.tar.gz: 1 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>psutil-5.4.3.tar.gz</b></p></summary> <p>Cross-platform lib for process and system monitoring in Python.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/e2/e1/600326635f97fee89bf8426fef14c5c29f4849c79f68fd79f433d8c1bd96/psutil-5.4.3.tar.gz">https://files.pythonhosted.org/packages/e2/e1/600326635f97fee89bf8426fef14c5c29f4849c79f68fd79f433d8c1bd96/psutil-5.4.3.tar.gz</a></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (psutil version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2019-18874](https://www.mend.io/vulnerability-database/CVE-2019-18874) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | psutil-5.4.3.tar.gz | Direct | 5.6.6 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-18874</summary> ### Vulnerable Library - <b>psutil-5.4.3.tar.gz</b></p> <p>Cross-platform lib for process and system monitoring in Python.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/e2/e1/600326635f97fee89bf8426fef14c5c29f4849c79f68fd79f433d8c1bd96/psutil-5.4.3.tar.gz">https://files.pythonhosted.org/packages/e2/e1/600326635f97fee89bf8426fef14c5c29f4849c79f68fd79f433d8c1bd96/psutil-5.4.3.tar.gz</a></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> Dependency Hierarchy: - :x: **psutil-5.4.3.tar.gz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> psutil (aka python-psutil) through 5.6.5 can have a double free. This occurs because of refcount mishandling within a while or for loop that converts system data into a Python object. <p>Publish Date: 2019-11-12 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-18874>CVE-2019-18874</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-18874">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-18874</a></p> <p>Release Date: 2019-11-12</p> <p>Fix Resolution: 5.6.6</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
non_test
psutil tar gz vulnerabilities highest severity is vulnerable library psutil tar gz cross platform lib for process and system monitoring in python library home page a href path to dependency file requirements txt path to vulnerable library requirements txt vulnerabilities cve severity cvss dependency type fixed in psutil version remediation available high psutil tar gz direct details cve vulnerable library psutil tar gz cross platform lib for process and system monitoring in python library home page a href path to dependency file requirements txt path to vulnerable library requirements txt dependency hierarchy x psutil tar gz vulnerable library found in base branch master vulnerability details psutil aka python psutil through can have a double free this occurs because of refcount mishandling within a while or for loop that converts system data into a python object publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
25,121
4,146,812,144
IssuesEvent
2016-06-15 02:28:48
Prof-Calebe/substituicao
https://api.github.com/repos/Prof-Calebe/substituicao
closed
Configurar o codady para usar 120 caracteres por linha de code convention
enhancement environment tests & QA
O codacy define um padrão denominado _Checks for long lines_, que é uma opção do [Checklist ](http://checkstyle.sourceforge.net/checks.html) chamado [LineLength](http://checkstyle.sourceforge.net/config_sizes.html#LineLength).
1.0
Configurar o codady para usar 120 caracteres por linha de code convention - O codacy define um padrão denominado _Checks for long lines_, que é uma opção do [Checklist ](http://checkstyle.sourceforge.net/checks.html) chamado [LineLength](http://checkstyle.sourceforge.net/config_sizes.html#LineLength).
test
configurar o codady para usar caracteres por linha de code convention o codacy define um padrão denominado checks for long lines que é uma opção do chamado
1
322,288
9,815,921,489
IssuesEvent
2019-06-13 13:40:47
python/mypy
https://api.github.com/repos/python/mypy
opened
New semantic analyzer: some assignments aren't recognized as types if incomplete
false-positive new-semantic-analyzer priority-1-normal
In the following test case the forward reference to `In` in the class body doesn't work: ``` [case testNewAnalyzerNamedTupleCallNested] from typing import NamedTuple o: C.Out i: C.In reveal_type(o) # E: Revealed type is 'Tuple[Tuple[builtins.str, __main__.C.Other, fallback=__main__.C.In], __main__.C.Other, fallback=__main__.C.Out]' reveal_type(o.x) # E: Revealed type is 'Tuple[builtins.str, __main__.C.Other, fallback=__main__.C.In]' reveal_type(o.y) # E: Revealed type is '__main__.C.Other' reveal_type(o.x.t) # E: Revealed type is '__main__.C.Other' reveal_type(i.t) # E: Revealed type is '__main__.C.Other' class C: Out = NamedTuple('Out', [('x', In), ('y', Other)]) # 'In' undefined <<--- should be ok? In = NamedTuple('In', [('s', str), ('t', Other)]) class Other: pass ``` The root cause is that if the rvalue of an assignment is incomplete, we add a `PlaceholderNode` that doesn't have `becomes_typeinfo` set. This may also affect other things such as assignment-based TypedDicts, type aliases, type variables and NewTypes (haven't checked). Not sure what would be the best way to fix this. One idea would be to treat every placeholder as a possible type, at least if it's generated from an incomplete assignment. A second idea would be to recognize possible types by looking at an incomplete rvalue, but this doesn't seem always possible. This is follow-up to #6983.
1.0
New semantic analyzer: some assignments aren't recognized as types if incomplete - In the following test case the forward reference to `In` in the class body doesn't work: ``` [case testNewAnalyzerNamedTupleCallNested] from typing import NamedTuple o: C.Out i: C.In reveal_type(o) # E: Revealed type is 'Tuple[Tuple[builtins.str, __main__.C.Other, fallback=__main__.C.In], __main__.C.Other, fallback=__main__.C.Out]' reveal_type(o.x) # E: Revealed type is 'Tuple[builtins.str, __main__.C.Other, fallback=__main__.C.In]' reveal_type(o.y) # E: Revealed type is '__main__.C.Other' reveal_type(o.x.t) # E: Revealed type is '__main__.C.Other' reveal_type(i.t) # E: Revealed type is '__main__.C.Other' class C: Out = NamedTuple('Out', [('x', In), ('y', Other)]) # 'In' undefined <<--- should be ok? In = NamedTuple('In', [('s', str), ('t', Other)]) class Other: pass ``` The root cause is that if the rvalue of an assignment is incomplete, we add a `PlaceholderNode` that doesn't have `becomes_typeinfo` set. This may also affect other things such as assignment-based TypedDicts, type aliases, type variables and NewTypes (haven't checked). Not sure what would be the best way to fix this. One idea would be to treat every placeholder as a possible type, at least if it's generated from an incomplete assignment. A second idea would be to recognize possible types by looking at an incomplete rvalue, but this doesn't seem always possible. This is follow-up to #6983.
non_test
new semantic analyzer some assignments aren t recognized as types if incomplete in the following test case the forward reference to in in the class body doesn t work from typing import namedtuple o c out i c in reveal type o e revealed type is tuple main c other fallback main c out reveal type o x e revealed type is tuple reveal type o y e revealed type is main c other reveal type o x t e revealed type is main c other reveal type i t e revealed type is main c other class c out namedtuple out in undefined should be ok in namedtuple in class other pass the root cause is that if the rvalue of an assignment is incomplete we add a placeholdernode that doesn t have becomes typeinfo set this may also affect other things such as assignment based typeddicts type aliases type variables and newtypes haven t checked not sure what would be the best way to fix this one idea would be to treat every placeholder as a possible type at least if it s generated from an incomplete assignment a second idea would be to recognize possible types by looking at an incomplete rvalue but this doesn t seem always possible this is follow up to
0
174,015
13,452,854,081
IssuesEvent
2020-09-08 23:14:47
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Catalogs are broken in Dashboard in latest master-head
[zube]: To Test kind/bug-qa status/blocker
bug on a downstream cluster in Rancher single install setup Steps: - run `master-4fbd822ceb7fbf030722bf5622911fef47d7984e-head` single install - proviosion a cluster with EC2 nodes - go to dashbord -> apps results: no apps show in the list, although it shows 2 repo existing. ![image (1)](https://user-images.githubusercontent.com/6218999/92527114-5e460380-f1db-11ea-9d7f-5bca65f44102.png) ![image](https://user-images.githubusercontent.com/6218999/92527116-5ede9a00-f1db-11ea-86b4-7b4dfec8753a.png) ![image](https://user-images.githubusercontent.com/6218999/92527957-c34e2900-f1dc-11ea-8e77-1cf5ddf866f0.png) <img width="1848" alt="Screen Shot 2020-09-08 at 2 18 14 PM" src="https://user-images.githubusercontent.com/6218999/92528883-694e6300-f1de-11ea-9cda-8b61f8f07286.png"> More info: This does not happen in the `local` cluster, it also does not happen in the downstream cluster in Rancher HA setup Update: This is confirmed happening on the downstream cluster in Rancher single install setup only
1.0
Catalogs are broken in Dashboard in latest master-head - bug on a downstream cluster in Rancher single install setup Steps: - run `master-4fbd822ceb7fbf030722bf5622911fef47d7984e-head` single install - proviosion a cluster with EC2 nodes - go to dashbord -> apps results: no apps show in the list, although it shows 2 repo existing. ![image (1)](https://user-images.githubusercontent.com/6218999/92527114-5e460380-f1db-11ea-9d7f-5bca65f44102.png) ![image](https://user-images.githubusercontent.com/6218999/92527116-5ede9a00-f1db-11ea-86b4-7b4dfec8753a.png) ![image](https://user-images.githubusercontent.com/6218999/92527957-c34e2900-f1dc-11ea-8e77-1cf5ddf866f0.png) <img width="1848" alt="Screen Shot 2020-09-08 at 2 18 14 PM" src="https://user-images.githubusercontent.com/6218999/92528883-694e6300-f1de-11ea-9cda-8b61f8f07286.png"> More info: This does not happen in the `local` cluster, it also does not happen in the downstream cluster in Rancher HA setup Update: This is confirmed happening on the downstream cluster in Rancher single install setup only
test
catalogs are broken in dashboard in latest master head bug on a downstream cluster in rancher single install setup steps run master head single install proviosion a cluster with nodes go to dashbord apps results no apps show in the list although it shows repo existing img width alt screen shot at pm src more info this does not happen in the local cluster it also does not happen in the downstream cluster in rancher ha setup update this is confirmed happening on the downstream cluster in rancher single install setup only
1
711,928
24,479,828,600
IssuesEvent
2022-10-08 17:23:09
michaelrsweet/pappl
https://api.github.com/repos/michaelrsweet/pappl
closed
HEIC/HEIF support for AirPrint/iOS
enhancement priority-low platform issue
**Is your feature request related to a problem? Please describe.** So by default doing AirPrint of a photo to an image printer using a newer iOS device the HEIC images get ballooned from 3 MB to a 36+ MB PDF file. Super slow on low bandwidth WIFI as well as IOT printer server devices. To make matters worse, Android devices appear to print MUCH quicker on the same print system only because they only deal with image/jpeg types. **Describe the solution you'd like** Support image/heic since that is the path ALL Apple devices are moving to for image printing. **Describe alternatives you've considered** I don't have any good options here. The _ipps txt record var of pdl shows the different types the printer supports. If the phone has the image type as HEIC then Apple's implementation will ALWAYS force HEIC->PDF conversion before transmit due to the 'Driverless' printing contract. Apple's going to force EVERYONE to leave image/jpeg and has chosen to burn this bridge and make things difficult for novice users just wanting to print a photo image from their phone to a photo printer via AirPrint. **Additional context** I'm happy to test and setup up a test framework to make this happen. I wrongly put this as a bug against CUPS which is not the case. That ticket and more details is located here: https://github.com/apple/cups/issues/5849
1.0
HEIC/HEIF support for AirPrint/iOS - **Is your feature request related to a problem? Please describe.** So by default doing AirPrint of a photo to an image printer using a newer iOS device the HEIC images get ballooned from 3 MB to a 36+ MB PDF file. Super slow on low bandwidth WIFI as well as IOT printer server devices. To make matters worse, Android devices appear to print MUCH quicker on the same print system only because they only deal with image/jpeg types. **Describe the solution you'd like** Support image/heic since that is the path ALL Apple devices are moving to for image printing. **Describe alternatives you've considered** I don't have any good options here. The _ipps txt record var of pdl shows the different types the printer supports. If the phone has the image type as HEIC then Apple's implementation will ALWAYS force HEIC->PDF conversion before transmit due to the 'Driverless' printing contract. Apple's going to force EVERYONE to leave image/jpeg and has chosen to burn this bridge and make things difficult for novice users just wanting to print a photo image from their phone to a photo printer via AirPrint. **Additional context** I'm happy to test and setup up a test framework to make this happen. I wrongly put this as a bug against CUPS which is not the case. That ticket and more details is located here: https://github.com/apple/cups/issues/5849
non_test
heic heif support for airprint ios is your feature request related to a problem please describe so by default doing airprint of a photo to an image printer using a newer ios device the heic images get ballooned from mb to a mb pdf file super slow on low bandwidth wifi as well as iot printer server devices to make matters worse android devices appear to print much quicker on the same print system only because they only deal with image jpeg types describe the solution you d like support image heic since that is the path all apple devices are moving to for image printing describe alternatives you ve considered i don t have any good options here the ipps txt record var of pdl shows the different types the printer supports if the phone has the image type as heic then apple s implementation will always force heic pdf conversion before transmit due to the driverless printing contract apple s going to force everyone to leave image jpeg and has chosen to burn this bridge and make things difficult for novice users just wanting to print a photo image from their phone to a photo printer via airprint additional context i m happy to test and setup up a test framework to make this happen i wrongly put this as a bug against cups which is not the case that ticket and more details is located here
0
324,166
9,885,147,597
IssuesEvent
2019-06-25 01:12:10
CentOS-PaaS-SIG/linchpin
https://api.github.com/repos/CentOS-PaaS-SIG/linchpin
closed
Attaching additional_volumes to an os_server fails not able to find the system
bug community high priority openstack
This is a contintuation of https://github.com/CentOS-PaaS-SIG/linchpin/issues/1137 using linchpin 1.7.4.1. Where once I patched the schema locally, I re-ran the pinfile and this time it is able to provision the VM, create the volume, but fails to attach it to the system. The error from the ansible task is below as well as the pinfile. The problem seems to be that the server name is *db2_ci_test_client_a* but linchpin is appending a number based on count, db2_ci_test_client_a**1**. Which causes the os_server_volume module to fail because it can find the system. It seems to stem from this filter plugin https://github.com/CentOS-PaaS-SIG/linchpin/blob/develop/linchpin/provision/filter_plugins/render_os_server_insts.py This conflicts with what the rest of the provision_os_server.yml playbook is doing when it provisions the instance. Might make more sense to remove the system naming logic from the filter plugin and then just move the following tasks in provision_os_server.yml BEFORE performing the rendering so that the filter plugin just creates the required server dict using the provided names from the two tasks below and volume info from the res_def? ``` - name: generate res_def_names set_fact: res_def_names: "{{ res_def_names + [os_resource_name + default_delimiter + inst_c] }}" with_sequence: start=0 end={{ res_def['count']|int - 1 }} when: res_def['count'] > 1 loop_control: loop_var: inst_c - name: generate res_def_names set_fact: res_def_names: "{{ res_def_names + [os_resource_name] }}" when: res_def['count'] == 1 ``` ``` -- os-server-addl-vols: topology: topology_name: os-server-addl-vols resource_groups: - resource_group_name: os-server-addl-vols resource_group_type: openstack resource_definitions: - name: "db2_ci_test_client_a" role: os_server flavor: m1.small image: rhel-7.4-server-x86_64-released count: 1 keypair: db2-test additional_volumes: - name: "test_vol" size: 2 device_name: "/dev/vdb" networks: - provider_net_cci_4 credentials: filename: clouds.yaml profile: ci-rhos layout: inventory_layout: vars: hostname: __IP__ hosts: addl-vols-node: count: 2 host_groups: - db2_ci_test_client_a ``` ``` 2019-05-23 16:00:51,823 p=16631 u=dbaez | TASK [openstack : Attach addtional volumes] *********************************************************************************************************************************************************************** 2019-05-23 16:00:54,020 p=16631 u=dbaez | An exception occurred during task execution. To see the full traceback, use -vvv. The error was: AttributeError: 'NoneType' object has no attribute 'id' 2019-05-23 16:00:54,021 p=16631 u=dbaez | failed: [localhost] (item={'device_name': u'/dev/vdb', 'name': u'test_vol-db2_ci_test_client_a1', 'server_name': u'db2_ci_test_client_a1', 'size': 2}) => {"add_vol": {"device_name": "/dev/vdb", "name": "test_vol-db2_ci_test_client_a1", "server_name": "db2_ci_test_client_a1", "size": 2}, "changed": false, "module_stderr": "Traceback (most recent call last):\n File \"/home/dbaez/.ansible/tmp/ansible-tmp-1558641651.89-96864677547322/AnsiballZ_os_server_volume.py\", line 113, in <module>\n _ansiballz_main()\n File \"/home/dbaez/.ansible/tmp/ansible-tmp-1558641651.89-96864677547322/AnsiballZ_os_server_volume.py\", line 105, in _ansiballz_main\n invoke_module(zipped_mod, temp_path, ANSIBALLZ_PARAMS)\n File \"/home/dbaez/.ansible/tmp/ansible-tmp-1558641651.89-96864677547322/AnsiballZ_os_server_volume.py\", line 48, in invoke_module\n imp.load_module('__main__', mod, module, MOD_DESC)\n File \"/tmp/ansible_os_server_volume_payload_Waz3Ge/__main__.py\", line 147, in <module>\n File \"/tmp/ansible_os_server_volume_payload_Waz3Ge/__main__.py\", line 107, in main\nAttributeError: 'NoneType' object has no attribute 'id'\n", "module_stdout": "", "msg": "MODULE FAILURE\nSee stdout/stderr for the exact error", "rc": 1} ```
1.0
Attaching additional_volumes to an os_server fails not able to find the system - This is a contintuation of https://github.com/CentOS-PaaS-SIG/linchpin/issues/1137 using linchpin 1.7.4.1. Where once I patched the schema locally, I re-ran the pinfile and this time it is able to provision the VM, create the volume, but fails to attach it to the system. The error from the ansible task is below as well as the pinfile. The problem seems to be that the server name is *db2_ci_test_client_a* but linchpin is appending a number based on count, db2_ci_test_client_a**1**. Which causes the os_server_volume module to fail because it can find the system. It seems to stem from this filter plugin https://github.com/CentOS-PaaS-SIG/linchpin/blob/develop/linchpin/provision/filter_plugins/render_os_server_insts.py This conflicts with what the rest of the provision_os_server.yml playbook is doing when it provisions the instance. Might make more sense to remove the system naming logic from the filter plugin and then just move the following tasks in provision_os_server.yml BEFORE performing the rendering so that the filter plugin just creates the required server dict using the provided names from the two tasks below and volume info from the res_def? ``` - name: generate res_def_names set_fact: res_def_names: "{{ res_def_names + [os_resource_name + default_delimiter + inst_c] }}" with_sequence: start=0 end={{ res_def['count']|int - 1 }} when: res_def['count'] > 1 loop_control: loop_var: inst_c - name: generate res_def_names set_fact: res_def_names: "{{ res_def_names + [os_resource_name] }}" when: res_def['count'] == 1 ``` ``` -- os-server-addl-vols: topology: topology_name: os-server-addl-vols resource_groups: - resource_group_name: os-server-addl-vols resource_group_type: openstack resource_definitions: - name: "db2_ci_test_client_a" role: os_server flavor: m1.small image: rhel-7.4-server-x86_64-released count: 1 keypair: db2-test additional_volumes: - name: "test_vol" size: 2 device_name: "/dev/vdb" networks: - provider_net_cci_4 credentials: filename: clouds.yaml profile: ci-rhos layout: inventory_layout: vars: hostname: __IP__ hosts: addl-vols-node: count: 2 host_groups: - db2_ci_test_client_a ``` ``` 2019-05-23 16:00:51,823 p=16631 u=dbaez | TASK [openstack : Attach addtional volumes] *********************************************************************************************************************************************************************** 2019-05-23 16:00:54,020 p=16631 u=dbaez | An exception occurred during task execution. To see the full traceback, use -vvv. The error was: AttributeError: 'NoneType' object has no attribute 'id' 2019-05-23 16:00:54,021 p=16631 u=dbaez | failed: [localhost] (item={'device_name': u'/dev/vdb', 'name': u'test_vol-db2_ci_test_client_a1', 'server_name': u'db2_ci_test_client_a1', 'size': 2}) => {"add_vol": {"device_name": "/dev/vdb", "name": "test_vol-db2_ci_test_client_a1", "server_name": "db2_ci_test_client_a1", "size": 2}, "changed": false, "module_stderr": "Traceback (most recent call last):\n File \"/home/dbaez/.ansible/tmp/ansible-tmp-1558641651.89-96864677547322/AnsiballZ_os_server_volume.py\", line 113, in <module>\n _ansiballz_main()\n File \"/home/dbaez/.ansible/tmp/ansible-tmp-1558641651.89-96864677547322/AnsiballZ_os_server_volume.py\", line 105, in _ansiballz_main\n invoke_module(zipped_mod, temp_path, ANSIBALLZ_PARAMS)\n File \"/home/dbaez/.ansible/tmp/ansible-tmp-1558641651.89-96864677547322/AnsiballZ_os_server_volume.py\", line 48, in invoke_module\n imp.load_module('__main__', mod, module, MOD_DESC)\n File \"/tmp/ansible_os_server_volume_payload_Waz3Ge/__main__.py\", line 147, in <module>\n File \"/tmp/ansible_os_server_volume_payload_Waz3Ge/__main__.py\", line 107, in main\nAttributeError: 'NoneType' object has no attribute 'id'\n", "module_stdout": "", "msg": "MODULE FAILURE\nSee stdout/stderr for the exact error", "rc": 1} ```
non_test
attaching additional volumes to an os server fails not able to find the system this is a contintuation of using linchpin where once i patched the schema locally i re ran the pinfile and this time it is able to provision the vm create the volume but fails to attach it to the system the error from the ansible task is below as well as the pinfile the problem seems to be that the server name is ci test client a but linchpin is appending a number based on count ci test client a which causes the os server volume module to fail because it can find the system it seems to stem from this filter plugin this conflicts with what the rest of the provision os server yml playbook is doing when it provisions the instance might make more sense to remove the system naming logic from the filter plugin and then just move the following tasks in provision os server yml before performing the rendering so that the filter plugin just creates the required server dict using the provided names from the two tasks below and volume info from the res def name generate res def names set fact res def names res def names with sequence start end res def int when res def loop control loop var inst c name generate res def names set fact res def names res def names when res def os server addl vols topology topology name os server addl vols resource groups resource group name os server addl vols resource group type openstack resource definitions name ci test client a role os server flavor small image rhel server released count keypair test additional volumes name test vol size device name dev vdb networks provider net cci credentials filename clouds yaml profile ci rhos layout inventory layout vars hostname ip hosts addl vols node count host groups ci test client a p u dbaez task p u dbaez an exception occurred during task execution to see the full traceback use vvv the error was attributeerror nonetype object has no attribute id p u dbaez failed item device name u dev vdb name u test vol ci test client server name u ci test client size add vol device name dev vdb name test vol ci test client server name ci test client size changed false module stderr traceback most recent call last n file home dbaez ansible tmp ansible tmp ansiballz os server volume py line in n ansiballz main n file home dbaez ansible tmp ansible tmp ansiballz os server volume py line in ansiballz main n invoke module zipped mod temp path ansiballz params n file home dbaez ansible tmp ansible tmp ansiballz os server volume py line in invoke module n imp load module main mod module mod desc n file tmp ansible os server volume payload main py line in n file tmp ansible os server volume payload main py line in main nattributeerror nonetype object has no attribute id n module stdout msg module failure nsee stdout stderr for the exact error rc
0
239,081
26,201,514,600
IssuesEvent
2023-01-03 17:55:56
snowdensb/nibrs
https://api.github.com/repos/snowdensb/nibrs
opened
CVE-2022-1471 (High) detected in multiple libraries
security vulnerability
## CVE-2022-1471 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>snakeyaml-1.17.jar</b>, <b>snakeyaml-1.23.jar</b>, <b>snakeyaml-1.19.jar</b></p></summary> <p> <details><summary><b>snakeyaml-1.17.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /tools/nibrs-fbi-service/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.17/snakeyaml-1.17.jar,/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/snakeyaml-1.17.jar</p> <p> Dependency Hierarchy: - :x: **snakeyaml-1.17.jar** (Vulnerable Library) </details> <details><summary><b>snakeyaml-1.23.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /tools/nibrs-summary-report-common/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.23/snakeyaml-1.23.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.5.RELEASE.jar (Root Library) - spring-boot-starter-2.1.5.RELEASE.jar - :x: **snakeyaml-1.23.jar** (Vulnerable Library) </details> <details><summary><b>snakeyaml-1.19.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /tools/nibrs-validation/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar,/web/nibrs-web/target/nibrs-web/WEB-INF/lib/snakeyaml-1.19.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar</p> <p> Dependency Hierarchy: - :x: **snakeyaml-1.19.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/snowdensb/nibrs/commit/e33ecd45d71662f63121c238ca1c416a6631a650">e33ecd45d71662f63121c238ca1c416a6631a650</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> SnakeYaml's Constructor() class does not restrict types which can be instantiated during deserialization. Deserializing yaml content provided by an attacker can lead to remote code execution. We recommend using SnakeYaml's SafeConsturctor when parsing untrusted content to restrict deserialization. <p>Publish Date: 2022-12-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1471>CVE-2022-1471</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
True
CVE-2022-1471 (High) detected in multiple libraries - ## CVE-2022-1471 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>snakeyaml-1.17.jar</b>, <b>snakeyaml-1.23.jar</b>, <b>snakeyaml-1.19.jar</b></p></summary> <p> <details><summary><b>snakeyaml-1.17.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /tools/nibrs-fbi-service/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.17/snakeyaml-1.17.jar,/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/snakeyaml-1.17.jar</p> <p> Dependency Hierarchy: - :x: **snakeyaml-1.17.jar** (Vulnerable Library) </details> <details><summary><b>snakeyaml-1.23.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /tools/nibrs-summary-report-common/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.23/snakeyaml-1.23.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.5.RELEASE.jar (Root Library) - spring-boot-starter-2.1.5.RELEASE.jar - :x: **snakeyaml-1.23.jar** (Vulnerable Library) </details> <details><summary><b>snakeyaml-1.19.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /tools/nibrs-validation/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar,/web/nibrs-web/target/nibrs-web/WEB-INF/lib/snakeyaml-1.19.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar</p> <p> Dependency Hierarchy: - :x: **snakeyaml-1.19.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/snowdensb/nibrs/commit/e33ecd45d71662f63121c238ca1c416a6631a650">e33ecd45d71662f63121c238ca1c416a6631a650</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> SnakeYaml's Constructor() class does not restrict types which can be instantiated during deserialization. Deserializing yaml content provided by an attacker can lead to remote code execution. We recommend using SnakeYaml's SafeConsturctor when parsing untrusted content to restrict deserialization. <p>Publish Date: 2022-12-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1471>CVE-2022-1471</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
non_test
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries snakeyaml jar snakeyaml jar snakeyaml jar snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file tools nibrs fbi service pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar tools nibrs fbi service target nibrs fbi service web inf lib snakeyaml jar dependency hierarchy x snakeyaml jar vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file tools nibrs summary report common pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy spring boot starter web release jar root library spring boot starter release jar x snakeyaml jar vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file tools nibrs validation pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar web nibrs web target nibrs web web inf lib snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy x snakeyaml jar vulnerable library found in head commit a href found in base branch master vulnerability details snakeyaml s constructor class does not restrict types which can be instantiated during deserialization deserializing yaml content provided by an attacker can lead to remote code execution we recommend using snakeyaml s safeconsturctor when parsing untrusted content to restrict deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href
0
11,762
3,223,331,847
IssuesEvent
2015-10-09 09:18:07
pyinstaller/pyinstaller
https://api.github.com/repos/pyinstaller/pyinstaller
opened
add tests for XML modules from the standard Python library
@low / cleanup test-suite
* PyInstaller contains several hook files `hook-xml.*.py`. * These files are old and some of them might be useless. * Create several tests for 'xml.*' module and remove not needed hook files.
1.0
add tests for XML modules from the standard Python library - * PyInstaller contains several hook files `hook-xml.*.py`. * These files are old and some of them might be useless. * Create several tests for 'xml.*' module and remove not needed hook files.
test
add tests for xml modules from the standard python library pyinstaller contains several hook files hook xml py these files are old and some of them might be useless create several tests for xml module and remove not needed hook files
1
138,423
11,200,865,707
IssuesEvent
2020-01-03 23:26:47
rancher/rancher
https://api.github.com/repos/rancher/rancher
opened
[forwardport]Cattle-cluster-agent and cattle-node-agent do not rollback to the previous version
[zube]: To Test
Forward port of https://github.com/rancher/rancher/issues/21386
1.0
[forwardport]Cattle-cluster-agent and cattle-node-agent do not rollback to the previous version - Forward port of https://github.com/rancher/rancher/issues/21386
test
cattle cluster agent and cattle node agent do not rollback to the previous version forward port of
1
506,918
14,675,898,885
IssuesEvent
2020-12-30 18:46:13
elementary/granite
https://api.github.com/repos/elementary/granite
closed
Support headerbar colour in new stylesheet
Bitesize Priority: Medium Status: Confirmed
Seems obvious but I'll open it since it hasn't got a tracker yet <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/92953310-support-headerbar-colour-in-new-stylesheet?utm_campaign=plugin&utm_content=tracker%2F60830613&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F60830613&utm_medium=issues&utm_source=github). </bountysource-plugin>
1.0
Support headerbar colour in new stylesheet - Seems obvious but I'll open it since it hasn't got a tracker yet <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/92953310-support-headerbar-colour-in-new-stylesheet?utm_campaign=plugin&utm_content=tracker%2F60830613&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F60830613&utm_medium=issues&utm_source=github). </bountysource-plugin>
non_test
support headerbar colour in new stylesheet seems obvious but i ll open it since it hasn t got a tracker yet want to back this issue we accept bounties via
0
261,156
22,700,566,327
IssuesEvent
2022-07-05 10:15:08
zkSNACKs/WalletWasabi
https://api.github.com/repos/zkSNACKs/WalletWasabi
closed
Log Out of Wallet
feature/enhancement UI ww2 testing
Would be nice if the user can _Log out_ of a wallet, without closing Wasabi. Quoting @danwalmsley: >I guess we could have "log out lite" where wallet remains loaded for all intents and purposes but the UI simply forces you to enter your password again
1.0
Log Out of Wallet - Would be nice if the user can _Log out_ of a wallet, without closing Wasabi. Quoting @danwalmsley: >I guess we could have "log out lite" where wallet remains loaded for all intents and purposes but the UI simply forces you to enter your password again
test
log out of wallet would be nice if the user can log out of a wallet without closing wasabi quoting danwalmsley i guess we could have log out lite where wallet remains loaded for all intents and purposes but the ui simply forces you to enter your password again
1
655,163
21,678,998,740
IssuesEvent
2022-05-09 03:04:38
StackExchange/dnscontrol
https://api.github.com/repos/StackExchange/dnscontrol
closed
Support for GeoIP features in BIND
Type: Enhancement Priority: p4 - Lowest
https://kb.isc.org/article/AA-01149/0/Using-the-GeoIP-Features-in-BIND-9.10.html Are there any plans to support that in BIND? Moreover, a lot of managed DNS providers out there support geolocation-based balancing. It may be a good idea to learn which providers do that via their API, and add support for that API as well.
1.0
Support for GeoIP features in BIND - https://kb.isc.org/article/AA-01149/0/Using-the-GeoIP-Features-in-BIND-9.10.html Are there any plans to support that in BIND? Moreover, a lot of managed DNS providers out there support geolocation-based balancing. It may be a good idea to learn which providers do that via their API, and add support for that API as well.
non_test
support for geoip features in bind are there any plans to support that in bind moreover a lot of managed dns providers out there support geolocation based balancing it may be a good idea to learn which providers do that via their api and add support for that api as well
0
550,069
16,104,339,364
IssuesEvent
2021-04-27 13:19:28
TEAM-SUITS/Suits
https://api.github.com/repos/TEAM-SUITS/Suits
closed
[BUG] 스토리북 오류
:bug: bug :orange_circle: Priority: High
**Describe the bug🦟** **이 버그를 발견하는 방법** 1. Go to 스토리북 실행 `npm run storybook` 2. See error **원래는 어떻게 동작하나요?** (버그가 없으면 원래 어떻게 작동해야 하는지 알려주세요.) 스토리북이 정상적으로 표시되어야 함. **Screenshots** <img width="1161" alt="Screen Shot 2021-04-26 at 11 44 42 PM" src="https://user-images.githubusercontent.com/72863748/116102484-b76e9300-a6e9-11eb-8744-e80fa3d7b70f.png"> **Additional context** theme.styled.js 파일에서 props로 받는 theme을 제대로 읽어들이지 못해서 발생하는 것 같습니다. 이런 저런 것들을 수정해봤는데 아직 해결하지 못하고 있습니다. 아래는 에러 메시지에 표시되는 theme.styled.js 파일의 문제 코드입니다. 해결해보려고 계속 노력해보겠지만 혹시 원인을 아실 것 같으면 조언 부탁드립니다. <img width="622" alt="Screen Shot 2021-04-26 at 11 45 15 PM" src="https://user-images.githubusercontent.com/72863748/116102709-e422aa80-a6e9-11eb-85cc-9cda0559b034.png">
1.0
[BUG] 스토리북 오류 - **Describe the bug🦟** **이 버그를 발견하는 방법** 1. Go to 스토리북 실행 `npm run storybook` 2. See error **원래는 어떻게 동작하나요?** (버그가 없으면 원래 어떻게 작동해야 하는지 알려주세요.) 스토리북이 정상적으로 표시되어야 함. **Screenshots** <img width="1161" alt="Screen Shot 2021-04-26 at 11 44 42 PM" src="https://user-images.githubusercontent.com/72863748/116102484-b76e9300-a6e9-11eb-8744-e80fa3d7b70f.png"> **Additional context** theme.styled.js 파일에서 props로 받는 theme을 제대로 읽어들이지 못해서 발생하는 것 같습니다. 이런 저런 것들을 수정해봤는데 아직 해결하지 못하고 있습니다. 아래는 에러 메시지에 표시되는 theme.styled.js 파일의 문제 코드입니다. 해결해보려고 계속 노력해보겠지만 혹시 원인을 아실 것 같으면 조언 부탁드립니다. <img width="622" alt="Screen Shot 2021-04-26 at 11 45 15 PM" src="https://user-images.githubusercontent.com/72863748/116102709-e422aa80-a6e9-11eb-85cc-9cda0559b034.png">
non_test
스토리북 오류 describe the bug🦟 이 버그를 발견하는 방법 go to 스토리북 실행 npm run storybook see error 원래는 어떻게 동작하나요 버그가 없으면 원래 어떻게 작동해야 하는지 알려주세요 스토리북이 정상적으로 표시되어야 함 screenshots img width alt screen shot at pm src additional context theme styled js 파일에서 props로 받는 theme을 제대로 읽어들이지 못해서 발생하는 것 같습니다 이런 저런 것들을 수정해봤는데 아직 해결하지 못하고 있습니다 아래는 에러 메시지에 표시되는 theme styled js 파일의 문제 코드입니다 해결해보려고 계속 노력해보겠지만 혹시 원인을 아실 것 같으면 조언 부탁드립니다 img width alt screen shot at pm src
0
81,395
7,780,380,343
IssuesEvent
2018-06-05 19:51:43
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
reopened
teamcity: failed tests on master: acceptance/TestVersionUpgrade
A-core-kv C-test-failure O-robot S-3-productivity
The following tests appear to have failed: [#680342](https://teamcity.cockroachdb.com/viewLog.html?buildId=680342): ``` --- FAIL: acceptance/TestVersionUpgrade (86.340s) test_log_scope.go:81: test logs captured to: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/logTestVersionUpgrade173016655 test_log_scope.go:62: use -show-logs to present logs inline --- FAIL: acceptance/TestVersionUpgrade: TestVersionUpgrade/runMode=local (86.340s) version_upgrade_test.go:331: binary=v1.1.0 version_upgrade_test.go:331: cluster=1.0 version_upgrade_test.go:331: cluster=1.0-3 version_upgrade_test.go:331: binary=v1.1.1 version_upgrade_test.go:331: cluster=1.1 version_upgrade_test.go:331: binary=v2.0.0 version_upgrade_test.go:331: binary=v2.0.0 version_upgrade_test.go:331: cluster=1.1-6 version_upgrade_test.go:331: cluster=2.0 version_upgrade_test.go:331: binary=source version_upgrade_test.go:331: cluster=2.0-5 version_upgrade_test.go:222: 0: pq: value differs between gossip (2.0-5) and KV (2.0); try again later (<nil> after 19.993038769s) ------- Stdout: ------- CockroachDB node starting at 2018-05-25 18:13:12.813229865 +0000 UTC build: CCL v1.0.6 @ 2017/09/14 15:15:48 (go1.8.3) admin: http://127.0.0.1:42959 sql: postgresql://root@127.0.0.1:38785?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: initialized new cluster clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:13:13.192344409 +0000 UTC build: CCL v1.0.6 @ 2017/09/14 15:15:48 (go1.8.3) admin: http://127.0.0.1:43053 sql: postgresql://root@127.0.0.1:40459?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/2 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2 status: initialized new node, joined pre-existing cluster clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 2 CockroachDB node starting at 2018-05-25 18:13:13.19361945 +0000 UTC build: CCL v1.0.6 @ 2017/09/14 15:15:48 (go1.8.3) admin: http://127.0.0.1:46589 sql: postgresql://root@127.0.0.1:32913?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/3 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3 status: initialized new node, joined pre-existing cluster clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 3 CockroachDB node starting at 2018-05-25 18:13:19.779502433 +0000 UTC (took 0.8s) build: CCL v1.1.0 @ 2017/10/12 14:50:18 (go1.8.3) admin: http://127.0.0.1:42461 sql: postgresql://root@127.0.0.1:45857?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/2 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 2 CockroachDB node starting at 2018-05-25 18:13:21.58739265 +0000 UTC (took 0.7s) build: CCL v1.1.0 @ 2017/10/12 14:50:18 (go1.8.3) admin: http://127.0.0.1:41479 sql: postgresql://root@127.0.0.1:40105?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/3 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 3 CockroachDB node starting at 2018-05-25 18:13:26.196230933 +0000 UTC (took 3.2s) build: CCL v1.1.0 @ 2017/10/12 14:50:18 (go1.8.3) admin: http://127.0.0.1:41489 sql: postgresql://root@127.0.0.1:45823?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:13:36.206082696 +0000 UTC (took 5.9s) build: CCL v1.1.1 @ 2017/10/19 15:31:46 (go1.8.3) admin: http://127.0.0.1:36077 sql: postgresql://root@127.0.0.1:40291?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:13:38.143422344 +0000 UTC (took 0.7s) build: CCL v1.1.1 @ 2017/10/19 15:31:46 (go1.8.3) admin: http://127.0.0.1:34061 sql: postgresql://root@127.0.0.1:33385?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/2 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 2 CockroachDB node starting at 2018-05-25 18:13:40.114076059 +0000 UTC (took 0.7s) build: CCL v1.1.1 @ 2017/10/19 15:31:46 (go1.8.3) admin: http://127.0.0.1:43419 sql: postgresql://root@127.0.0.1:44431?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/3 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 3 CockroachDB node starting at 2018-05-25 18:13:50.383953476 +0000 UTC (took 6.9s) build: CCL v2.0.0 @ 2018/04/03 20:56:09 (go1.10) admin: http://127.0.0.1:41605 sql: postgresql://root@127.0.0.1:38989?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/cockroach-temp234768692 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:13:52.471275507 +0000 UTC (took 0.7s) build: CCL v2.0.0 @ 2018/04/03 20:56:09 (go1.10) admin: http://127.0.0.1:34325 sql: postgresql://root@127.0.0.1:37533?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/cockroach-temp966885328 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:13:54.429337483 +0000 UTC (took 0.7s) build: CCL v2.0.0 @ 2018/04/03 20:56:09 (go1.10) admin: http://127.0.0.1:42565 sql: postgresql://root@127.0.0.1:35193?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/2 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2/cockroach-temp973020792 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 2 CockroachDB node starting at 2018-05-25 18:13:57.299563909 +0000 UTC (took 1.6s) build: CCL v2.0.0 @ 2018/04/03 20:56:09 (go1.10) admin: http://127.0.0.1:38897 sql: postgresql://root@127.0.0.1:45091?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/3 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3/cockroach-temp535152420 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 3 CockroachDB node starting at 2018-05-25 18:14:07.130215431 +0000 UTC (took 3.8s) build: CCL v2.1-alpha.20180507-673-g68b0299 @ 2018/05/25 18:03:36 (go1.10) admin: http://127.0.0.1:46101 sql: postgresql://root@127.0.0.1:46123?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/3 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3/cockroach-temp358177994 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 3 CockroachDB node starting at 2018-05-25 18:14:09.31867713 +0000 UTC (took 0.7s) build: CCL v2.1-alpha.20180507-673-g68b0299 @ 2018/05/25 18:03:36 (go1.10) admin: http://127.0.0.1:45579 sql: postgresql://root@127.0.0.1:39635?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/cockroach-temp354217346 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:14:16.329037657 +0000 UTC (took 5.7s) build: CCL v2.1-alpha.20180507-673-g68b0299 @ 2018/05/25 18:03:36 (go1.10) admin: http://127.0.0.1:35961 sql: postgresql://root@127.0.0.1:40361?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/2 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2/cockroach-temp910272817 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 2 test logs left over in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/logTestVersionUpgrade173016655 --- FAIL: acceptance/TestVersionUpgrade (86.340s) test_log_scope.go:81: test logs captured to: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/logTestVersionUpgrade173016655 test_log_scope.go:62: use -show-logs to present logs inline --- FAIL: acceptance/TestVersionUpgrade: TestVersionUpgrade/runMode=local (86.340s) version_upgrade_test.go:331: binary=v1.1.0 version_upgrade_test.go:331: cluster=1.0 version_upgrade_test.go:331: cluster=1.0-3 version_upgrade_test.go:331: binary=v1.1.1 version_upgrade_test.go:331: cluster=1.1 version_upgrade_test.go:331: binary=v2.0.0 version_upgrade_test.go:331: binary=v2.0.0 version_upgrade_test.go:331: cluster=1.1-6 version_upgrade_test.go:331: cluster=2.0 version_upgrade_test.go:331: binary=source version_upgrade_test.go:331: cluster=2.0-5 version_upgrade_test.go:222: 0: pq: value differs between gossip (2.0-5) and KV (2.0); try again later (<nil> after 19.993038769s) ------- Stdout: ------- CockroachDB node starting at 2018-05-25 18:13:12.813229865 +0000 UTC build: CCL v1.0.6 @ 2017/09/14 15:15:48 (go1.8.3) admin: http://127.0.0.1:42959 sql: postgresql://root@127.0.0.1:38785?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: initialized new cluster clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:13:13.192344409 +0000 UTC build: CCL v1.0.6 @ 2017/09/14 15:15:48 (go1.8.3) admin: http://127.0.0.1:43053 sql: postgresql://root@127.0.0.1:40459?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/2 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2 status: initialized new node, joined pre-existing cluster clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 2 CockroachDB node starting at 2018-05-25 18:13:13.19361945 +0000 UTC build: CCL v1.0.6 @ 2017/09/14 15:15:48 (go1.8.3) admin: http://127.0.0.1:46589 sql: postgresql://root@127.0.0.1:32913?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/3 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3 status: initialized new node, joined pre-existing cluster clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 3 CockroachDB node starting at 2018-05-25 18:13:19.779502433 +0000 UTC (took 0.8s) build: CCL v1.1.0 @ 2017/10/12 14:50:18 (go1.8.3) admin: http://127.0.0.1:42461 sql: postgresql://root@127.0.0.1:45857?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/2 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 2 CockroachDB node starting at 2018-05-25 18:13:21.58739265 +0000 UTC (took 0.7s) build: CCL v1.1.0 @ 2017/10/12 14:50:18 (go1.8.3) admin: http://127.0.0.1:41479 sql: postgresql://root@127.0.0.1:40105?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/3 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 3 CockroachDB node starting at 2018-05-25 18:13:26.196230933 +0000 UTC (took 3.2s) build: CCL v1.1.0 @ 2017/10/12 14:50:18 (go1.8.3) admin: http://127.0.0.1:41489 sql: postgresql://root@127.0.0.1:45823?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:13:36.206082696 +0000 UTC (took 5.9s) build: CCL v1.1.1 @ 2017/10/19 15:31:46 (go1.8.3) admin: http://127.0.0.1:36077 sql: postgresql://root@127.0.0.1:40291?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:13:38.143422344 +0000 UTC (took 0.7s) build: CCL v1.1.1 @ 2017/10/19 15:31:46 (go1.8.3) admin: http://127.0.0.1:34061 sql: postgresql://root@127.0.0.1:33385?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/2 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 2 CockroachDB node starting at 2018-05-25 18:13:40.114076059 +0000 UTC (took 0.7s) build: CCL v1.1.1 @ 2017/10/19 15:31:46 (go1.8.3) admin: http://127.0.0.1:43419 sql: postgresql://root@127.0.0.1:44431?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/3 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 3 CockroachDB node starting at 2018-05-25 18:13:50.383953476 +0000 UTC (took 6.9s) build: CCL v2.0.0 @ 2018/04/03 20:56:09 (go1.10) admin: http://127.0.0.1:41605 sql: postgresql://root@127.0.0.1:38989?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/cockroach-temp234768692 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:13:52.471275507 +0000 UTC (took 0.7s) build: CCL v2.0.0 @ 2018/04/03 20:56:09 (go1.10) admin: http://127.0.0.1:34325 sql: postgresql://root@127.0.0.1:37533?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/cockroach-temp966885328 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:13:54.429337483 +0000 UTC (took 0.7s) build: CCL v2.0.0 @ 2018/04/03 20:56:09 (go1.10) admin: http://127.0.0.1:42565 sql: postgresql://root@127.0.0.1:35193?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/2 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2/cockroach-temp973020792 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 2 CockroachDB node starting at 2018-05-25 18:13:57.299563909 +0000 UTC (took 1.6s) build: CCL v2.0.0 @ 2018/04/03 20:56:09 (go1.10) admin: http://127.0.0.1:38897 sql: postgresql://root@127.0.0.1:45091?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/3 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3/cockroach-temp535152420 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 3 CockroachDB node starting at 2018-05-25 18:14:07.130215431 +0000 UTC (took 3.8s) build: CCL v2.1-alpha.20180507-673-g68b0299 @ 2018/05/25 18:03:36 (go1.10) admin: http://127.0.0.1:46101 sql: postgresql://root@127.0.0.1:46123?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/3 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3/cockroach-temp358177994 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 3 CockroachDB node starting at 2018-05-25 18:14:09.31867713 +0000 UTC (took 0.7s) build: CCL v2.1-alpha.20180507-673-g68b0299 @ 2018/05/25 18:03:36 (go1.10) admin: http://127.0.0.1:45579 sql: postgresql://root@127.0.0.1:39635?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/cockroach-temp354217346 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:14:16.329037657 +0000 UTC (took 5.7s) build: CCL v2.1-alpha.20180507-673-g68b0299 @ 2018/05/25 18:03:36 (go1.10) admin: http://127.0.0.1:35961 sql: postgresql://root@127.0.0.1:40361?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/2 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2/cockroach-temp910272817 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 2 test logs left over in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/logTestVersionUpgrade173016655 ``` Please assign, take a look and update the issue accordingly.
1.0
teamcity: failed tests on master: acceptance/TestVersionUpgrade - The following tests appear to have failed: [#680342](https://teamcity.cockroachdb.com/viewLog.html?buildId=680342): ``` --- FAIL: acceptance/TestVersionUpgrade (86.340s) test_log_scope.go:81: test logs captured to: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/logTestVersionUpgrade173016655 test_log_scope.go:62: use -show-logs to present logs inline --- FAIL: acceptance/TestVersionUpgrade: TestVersionUpgrade/runMode=local (86.340s) version_upgrade_test.go:331: binary=v1.1.0 version_upgrade_test.go:331: cluster=1.0 version_upgrade_test.go:331: cluster=1.0-3 version_upgrade_test.go:331: binary=v1.1.1 version_upgrade_test.go:331: cluster=1.1 version_upgrade_test.go:331: binary=v2.0.0 version_upgrade_test.go:331: binary=v2.0.0 version_upgrade_test.go:331: cluster=1.1-6 version_upgrade_test.go:331: cluster=2.0 version_upgrade_test.go:331: binary=source version_upgrade_test.go:331: cluster=2.0-5 version_upgrade_test.go:222: 0: pq: value differs between gossip (2.0-5) and KV (2.0); try again later (<nil> after 19.993038769s) ------- Stdout: ------- CockroachDB node starting at 2018-05-25 18:13:12.813229865 +0000 UTC build: CCL v1.0.6 @ 2017/09/14 15:15:48 (go1.8.3) admin: http://127.0.0.1:42959 sql: postgresql://root@127.0.0.1:38785?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: initialized new cluster clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:13:13.192344409 +0000 UTC build: CCL v1.0.6 @ 2017/09/14 15:15:48 (go1.8.3) admin: http://127.0.0.1:43053 sql: postgresql://root@127.0.0.1:40459?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/2 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2 status: initialized new node, joined pre-existing cluster clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 2 CockroachDB node starting at 2018-05-25 18:13:13.19361945 +0000 UTC build: CCL v1.0.6 @ 2017/09/14 15:15:48 (go1.8.3) admin: http://127.0.0.1:46589 sql: postgresql://root@127.0.0.1:32913?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/3 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3 status: initialized new node, joined pre-existing cluster clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 3 CockroachDB node starting at 2018-05-25 18:13:19.779502433 +0000 UTC (took 0.8s) build: CCL v1.1.0 @ 2017/10/12 14:50:18 (go1.8.3) admin: http://127.0.0.1:42461 sql: postgresql://root@127.0.0.1:45857?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/2 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 2 CockroachDB node starting at 2018-05-25 18:13:21.58739265 +0000 UTC (took 0.7s) build: CCL v1.1.0 @ 2017/10/12 14:50:18 (go1.8.3) admin: http://127.0.0.1:41479 sql: postgresql://root@127.0.0.1:40105?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/3 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 3 CockroachDB node starting at 2018-05-25 18:13:26.196230933 +0000 UTC (took 3.2s) build: CCL v1.1.0 @ 2017/10/12 14:50:18 (go1.8.3) admin: http://127.0.0.1:41489 sql: postgresql://root@127.0.0.1:45823?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:13:36.206082696 +0000 UTC (took 5.9s) build: CCL v1.1.1 @ 2017/10/19 15:31:46 (go1.8.3) admin: http://127.0.0.1:36077 sql: postgresql://root@127.0.0.1:40291?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:13:38.143422344 +0000 UTC (took 0.7s) build: CCL v1.1.1 @ 2017/10/19 15:31:46 (go1.8.3) admin: http://127.0.0.1:34061 sql: postgresql://root@127.0.0.1:33385?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/2 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 2 CockroachDB node starting at 2018-05-25 18:13:40.114076059 +0000 UTC (took 0.7s) build: CCL v1.1.1 @ 2017/10/19 15:31:46 (go1.8.3) admin: http://127.0.0.1:43419 sql: postgresql://root@127.0.0.1:44431?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/3 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 3 CockroachDB node starting at 2018-05-25 18:13:50.383953476 +0000 UTC (took 6.9s) build: CCL v2.0.0 @ 2018/04/03 20:56:09 (go1.10) admin: http://127.0.0.1:41605 sql: postgresql://root@127.0.0.1:38989?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/cockroach-temp234768692 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:13:52.471275507 +0000 UTC (took 0.7s) build: CCL v2.0.0 @ 2018/04/03 20:56:09 (go1.10) admin: http://127.0.0.1:34325 sql: postgresql://root@127.0.0.1:37533?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/cockroach-temp966885328 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:13:54.429337483 +0000 UTC (took 0.7s) build: CCL v2.0.0 @ 2018/04/03 20:56:09 (go1.10) admin: http://127.0.0.1:42565 sql: postgresql://root@127.0.0.1:35193?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/2 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2/cockroach-temp973020792 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 2 CockroachDB node starting at 2018-05-25 18:13:57.299563909 +0000 UTC (took 1.6s) build: CCL v2.0.0 @ 2018/04/03 20:56:09 (go1.10) admin: http://127.0.0.1:38897 sql: postgresql://root@127.0.0.1:45091?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/3 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3/cockroach-temp535152420 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 3 CockroachDB node starting at 2018-05-25 18:14:07.130215431 +0000 UTC (took 3.8s) build: CCL v2.1-alpha.20180507-673-g68b0299 @ 2018/05/25 18:03:36 (go1.10) admin: http://127.0.0.1:46101 sql: postgresql://root@127.0.0.1:46123?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/3 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3/cockroach-temp358177994 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 3 CockroachDB node starting at 2018-05-25 18:14:09.31867713 +0000 UTC (took 0.7s) build: CCL v2.1-alpha.20180507-673-g68b0299 @ 2018/05/25 18:03:36 (go1.10) admin: http://127.0.0.1:45579 sql: postgresql://root@127.0.0.1:39635?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/cockroach-temp354217346 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:14:16.329037657 +0000 UTC (took 5.7s) build: CCL v2.1-alpha.20180507-673-g68b0299 @ 2018/05/25 18:03:36 (go1.10) admin: http://127.0.0.1:35961 sql: postgresql://root@127.0.0.1:40361?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/2 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2/cockroach-temp910272817 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 2 test logs left over in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/logTestVersionUpgrade173016655 --- FAIL: acceptance/TestVersionUpgrade (86.340s) test_log_scope.go:81: test logs captured to: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/logTestVersionUpgrade173016655 test_log_scope.go:62: use -show-logs to present logs inline --- FAIL: acceptance/TestVersionUpgrade: TestVersionUpgrade/runMode=local (86.340s) version_upgrade_test.go:331: binary=v1.1.0 version_upgrade_test.go:331: cluster=1.0 version_upgrade_test.go:331: cluster=1.0-3 version_upgrade_test.go:331: binary=v1.1.1 version_upgrade_test.go:331: cluster=1.1 version_upgrade_test.go:331: binary=v2.0.0 version_upgrade_test.go:331: binary=v2.0.0 version_upgrade_test.go:331: cluster=1.1-6 version_upgrade_test.go:331: cluster=2.0 version_upgrade_test.go:331: binary=source version_upgrade_test.go:331: cluster=2.0-5 version_upgrade_test.go:222: 0: pq: value differs between gossip (2.0-5) and KV (2.0); try again later (<nil> after 19.993038769s) ------- Stdout: ------- CockroachDB node starting at 2018-05-25 18:13:12.813229865 +0000 UTC build: CCL v1.0.6 @ 2017/09/14 15:15:48 (go1.8.3) admin: http://127.0.0.1:42959 sql: postgresql://root@127.0.0.1:38785?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: initialized new cluster clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:13:13.192344409 +0000 UTC build: CCL v1.0.6 @ 2017/09/14 15:15:48 (go1.8.3) admin: http://127.0.0.1:43053 sql: postgresql://root@127.0.0.1:40459?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/2 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2 status: initialized new node, joined pre-existing cluster clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 2 CockroachDB node starting at 2018-05-25 18:13:13.19361945 +0000 UTC build: CCL v1.0.6 @ 2017/09/14 15:15:48 (go1.8.3) admin: http://127.0.0.1:46589 sql: postgresql://root@127.0.0.1:32913?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/3 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3 status: initialized new node, joined pre-existing cluster clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 3 CockroachDB node starting at 2018-05-25 18:13:19.779502433 +0000 UTC (took 0.8s) build: CCL v1.1.0 @ 2017/10/12 14:50:18 (go1.8.3) admin: http://127.0.0.1:42461 sql: postgresql://root@127.0.0.1:45857?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/2 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 2 CockroachDB node starting at 2018-05-25 18:13:21.58739265 +0000 UTC (took 0.7s) build: CCL v1.1.0 @ 2017/10/12 14:50:18 (go1.8.3) admin: http://127.0.0.1:41479 sql: postgresql://root@127.0.0.1:40105?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/3 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 3 CockroachDB node starting at 2018-05-25 18:13:26.196230933 +0000 UTC (took 3.2s) build: CCL v1.1.0 @ 2017/10/12 14:50:18 (go1.8.3) admin: http://127.0.0.1:41489 sql: postgresql://root@127.0.0.1:45823?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:13:36.206082696 +0000 UTC (took 5.9s) build: CCL v1.1.1 @ 2017/10/19 15:31:46 (go1.8.3) admin: http://127.0.0.1:36077 sql: postgresql://root@127.0.0.1:40291?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:13:38.143422344 +0000 UTC (took 0.7s) build: CCL v1.1.1 @ 2017/10/19 15:31:46 (go1.8.3) admin: http://127.0.0.1:34061 sql: postgresql://root@127.0.0.1:33385?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/2 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 2 CockroachDB node starting at 2018-05-25 18:13:40.114076059 +0000 UTC (took 0.7s) build: CCL v1.1.1 @ 2017/10/19 15:31:46 (go1.8.3) admin: http://127.0.0.1:43419 sql: postgresql://root@127.0.0.1:44431?application_name=cockroach&sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/3 store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 3 CockroachDB node starting at 2018-05-25 18:13:50.383953476 +0000 UTC (took 6.9s) build: CCL v2.0.0 @ 2018/04/03 20:56:09 (go1.10) admin: http://127.0.0.1:41605 sql: postgresql://root@127.0.0.1:38989?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/cockroach-temp234768692 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:13:52.471275507 +0000 UTC (took 0.7s) build: CCL v2.0.0 @ 2018/04/03 20:56:09 (go1.10) admin: http://127.0.0.1:34325 sql: postgresql://root@127.0.0.1:37533?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/cockroach-temp966885328 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:13:54.429337483 +0000 UTC (took 0.7s) build: CCL v2.0.0 @ 2018/04/03 20:56:09 (go1.10) admin: http://127.0.0.1:42565 sql: postgresql://root@127.0.0.1:35193?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/2 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2/cockroach-temp973020792 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 2 CockroachDB node starting at 2018-05-25 18:13:57.299563909 +0000 UTC (took 1.6s) build: CCL v2.0.0 @ 2018/04/03 20:56:09 (go1.10) admin: http://127.0.0.1:38897 sql: postgresql://root@127.0.0.1:45091?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/3 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3/cockroach-temp535152420 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 3 CockroachDB node starting at 2018-05-25 18:14:07.130215431 +0000 UTC (took 3.8s) build: CCL v2.1-alpha.20180507-673-g68b0299 @ 2018/05/25 18:03:36 (go1.10) admin: http://127.0.0.1:46101 sql: postgresql://root@127.0.0.1:46123?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/3 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3/cockroach-temp358177994 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/3 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 3 CockroachDB node starting at 2018-05-25 18:14:09.31867713 +0000 UTC (took 0.7s) build: CCL v2.1-alpha.20180507-673-g68b0299 @ 2018/05/25 18:03:36 (go1.10) admin: http://127.0.0.1:45579 sql: postgresql://root@127.0.0.1:39635?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/1 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/cockroach-temp354217346 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/1 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 1 CockroachDB node starting at 2018-05-25 18:14:16.329037657 +0000 UTC (took 5.7s) build: CCL v2.1-alpha.20180507-673-g68b0299 @ 2018/05/25 18:03:36 (go1.10) admin: http://127.0.0.1:35961 sql: postgresql://root@127.0.0.1:40361?sslmode=disable logs: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/TestVersionUpgrade/runMode=local/2 temp dir: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2/cockroach-temp910272817 external I/O path: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2/extern store[0]: path=/home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/acceptance/.localcluster162504546/2 status: restarted pre-existing node clusterID: cc455c17-5953-4b4f-a05b-9696ea8182cf nodeID: 2 test logs left over in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/acceptance/logTestVersionUpgrade173016655 ``` Please assign, take a look and update the issue accordingly.
test
teamcity failed tests on master acceptance testversionupgrade the following tests appear to have failed fail acceptance testversionupgrade test log scope go test logs captured to home agent work go src github com cockroachdb cockroach artifacts acceptance test log scope go use show logs to present logs inline fail acceptance testversionupgrade testversionupgrade runmode local version upgrade test go binary version upgrade test go cluster version upgrade test go cluster version upgrade test go binary version upgrade test go cluster version upgrade test go binary version upgrade test go binary version upgrade test go cluster version upgrade test go cluster version upgrade test go binary source version upgrade test go cluster version upgrade test go pq value differs between gossip and kv try again later after stdout cockroachdb node starting at utc build ccl admin sql postgresql root sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local store path home agent work go src github com cockroachdb cockroach pkg acceptance status initialized new cluster clusterid nodeid cockroachdb node starting at utc build ccl admin sql postgresql root sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local store path home agent work go src github com cockroachdb cockroach pkg acceptance status initialized new node joined pre existing cluster clusterid nodeid cockroachdb node starting at utc build ccl admin sql postgresql root sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local store path home agent work go src github com cockroachdb cockroach pkg acceptance status initialized new node joined pre existing cluster clusterid nodeid cockroachdb node starting at utc took build ccl admin sql postgresql root application name cockroach sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl admin sql postgresql root application name cockroach sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl admin sql postgresql root application name cockroach sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl admin sql postgresql root application name cockroach sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl admin sql postgresql root application name cockroach sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl admin sql postgresql root application name cockroach sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl admin sql postgresql root sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local temp dir home agent work go src github com cockroachdb cockroach pkg acceptance cockroach external i o path home agent work go src github com cockroachdb cockroach pkg acceptance extern store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl admin sql postgresql root sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local temp dir home agent work go src github com cockroachdb cockroach pkg acceptance cockroach external i o path home agent work go src github com cockroachdb cockroach pkg acceptance extern store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl admin sql postgresql root sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local temp dir home agent work go src github com cockroachdb cockroach pkg acceptance cockroach external i o path home agent work go src github com cockroachdb cockroach pkg acceptance extern store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl admin sql postgresql root sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local temp dir home agent work go src github com cockroachdb cockroach pkg acceptance cockroach external i o path home agent work go src github com cockroachdb cockroach pkg acceptance extern store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl alpha admin sql postgresql root sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local temp dir home agent work go src github com cockroachdb cockroach pkg acceptance cockroach external i o path home agent work go src github com cockroachdb cockroach pkg acceptance extern store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl alpha admin sql postgresql root sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local temp dir home agent work go src github com cockroachdb cockroach pkg acceptance cockroach external i o path home agent work go src github com cockroachdb cockroach pkg acceptance extern store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl alpha admin sql postgresql root sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local temp dir home agent work go src github com cockroachdb cockroach pkg acceptance cockroach external i o path home agent work go src github com cockroachdb cockroach pkg acceptance extern store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid test logs left over in home agent work go src github com cockroachdb cockroach artifacts acceptance fail acceptance testversionupgrade test log scope go test logs captured to home agent work go src github com cockroachdb cockroach artifacts acceptance test log scope go use show logs to present logs inline fail acceptance testversionupgrade testversionupgrade runmode local version upgrade test go binary version upgrade test go cluster version upgrade test go cluster version upgrade test go binary version upgrade test go cluster version upgrade test go binary version upgrade test go binary version upgrade test go cluster version upgrade test go cluster version upgrade test go binary source version upgrade test go cluster version upgrade test go pq value differs between gossip and kv try again later after stdout cockroachdb node starting at utc build ccl admin sql postgresql root sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local store path home agent work go src github com cockroachdb cockroach pkg acceptance status initialized new cluster clusterid nodeid cockroachdb node starting at utc build ccl admin sql postgresql root sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local store path home agent work go src github com cockroachdb cockroach pkg acceptance status initialized new node joined pre existing cluster clusterid nodeid cockroachdb node starting at utc build ccl admin sql postgresql root sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local store path home agent work go src github com cockroachdb cockroach pkg acceptance status initialized new node joined pre existing cluster clusterid nodeid cockroachdb node starting at utc took build ccl admin sql postgresql root application name cockroach sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl admin sql postgresql root application name cockroach sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl admin sql postgresql root application name cockroach sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl admin sql postgresql root application name cockroach sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl admin sql postgresql root application name cockroach sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl admin sql postgresql root application name cockroach sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl admin sql postgresql root sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local temp dir home agent work go src github com cockroachdb cockroach pkg acceptance cockroach external i o path home agent work go src github com cockroachdb cockroach pkg acceptance extern store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl admin sql postgresql root sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local temp dir home agent work go src github com cockroachdb cockroach pkg acceptance cockroach external i o path home agent work go src github com cockroachdb cockroach pkg acceptance extern store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl admin sql postgresql root sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local temp dir home agent work go src github com cockroachdb cockroach pkg acceptance cockroach external i o path home agent work go src github com cockroachdb cockroach pkg acceptance extern store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl admin sql postgresql root sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local temp dir home agent work go src github com cockroachdb cockroach pkg acceptance cockroach external i o path home agent work go src github com cockroachdb cockroach pkg acceptance extern store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl alpha admin sql postgresql root sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local temp dir home agent work go src github com cockroachdb cockroach pkg acceptance cockroach external i o path home agent work go src github com cockroachdb cockroach pkg acceptance extern store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl alpha admin sql postgresql root sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local temp dir home agent work go src github com cockroachdb cockroach pkg acceptance cockroach external i o path home agent work go src github com cockroachdb cockroach pkg acceptance extern store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid cockroachdb node starting at utc took build ccl alpha admin sql postgresql root sslmode disable logs home agent work go src github com cockroachdb cockroach artifacts acceptance testversionupgrade runmode local temp dir home agent work go src github com cockroachdb cockroach pkg acceptance cockroach external i o path home agent work go src github com cockroachdb cockroach pkg acceptance extern store path home agent work go src github com cockroachdb cockroach pkg acceptance status restarted pre existing node clusterid nodeid test logs left over in home agent work go src github com cockroachdb cockroach artifacts acceptance please assign take a look and update the issue accordingly
1
326,978
28,035,969,254
IssuesEvent
2023-03-28 15:05:53
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
opened
Fix math.test_tensorflow_square
TensorFlow Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4543693013/jobs/8008709037" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4543693013/jobs/8008709037" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4543693013/jobs/8008709037" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4543693013/jobs/8008709037" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_math.py::test_tensorflow_square[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T13:40:43.6490511Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6493820Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6498999Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6499474Z E Falsifying example: test_tensorflow_square( 2023-03-28T13:40:43.6499884Z E dtype_and_x=(['float16'], [array(-1., dtype=float16)]), 2023-03-28T13:40:43.6500343Z E fn_tree='ivy.functional.frontends.tensorflow.math.square', 2023-03-28T13:40:43.6500718Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T13:40:43.6501003Z E num_positional_args=0, 2023-03-28T13:40:43.6501248Z E with_out=False, 2023-03-28T13:40:43.6501478Z E inplace=False, 2023-03-28T13:40:43.6501716Z E as_variable=[False], 2023-03-28T13:40:43.6501969Z E native_arrays=[False], 2023-03-28T13:40:43.6502237Z E generate_frontend_arrays=False, 2023-03-28T13:40:43.6502477Z E ), 2023-03-28T13:40:43.6502757Z E frontend='tensorflow', 2023-03-28T13:40:43.6503034Z E on_device='cpu', 2023-03-28T13:40:43.6503255Z E ) 2023-03-28T13:40:43.6503447Z E 2023-03-28T13:40:43.6503996Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2AAAkYGCIDQAAAnAAM=') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_math.py::test_tensorflow_square[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T13:40:43.6490511Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6493820Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6498999Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6499474Z E Falsifying example: test_tensorflow_square( 2023-03-28T13:40:43.6499884Z E dtype_and_x=(['float16'], [array(-1., dtype=float16)]), 2023-03-28T13:40:43.6500343Z E fn_tree='ivy.functional.frontends.tensorflow.math.square', 2023-03-28T13:40:43.6500718Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T13:40:43.6501003Z E num_positional_args=0, 2023-03-28T13:40:43.6501248Z E with_out=False, 2023-03-28T13:40:43.6501478Z E inplace=False, 2023-03-28T13:40:43.6501716Z E as_variable=[False], 2023-03-28T13:40:43.6501969Z E native_arrays=[False], 2023-03-28T13:40:43.6502237Z E generate_frontend_arrays=False, 2023-03-28T13:40:43.6502477Z E ), 2023-03-28T13:40:43.6502757Z E frontend='tensorflow', 2023-03-28T13:40:43.6503034Z E on_device='cpu', 2023-03-28T13:40:43.6503255Z E ) 2023-03-28T13:40:43.6503447Z E 2023-03-28T13:40:43.6503996Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2AAAkYGCIDQAAAnAAM=') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_math.py::test_tensorflow_square[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T13:40:43.6490511Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6493820Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6498999Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6499474Z E Falsifying example: test_tensorflow_square( 2023-03-28T13:40:43.6499884Z E dtype_and_x=(['float16'], [array(-1., dtype=float16)]), 2023-03-28T13:40:43.6500343Z E fn_tree='ivy.functional.frontends.tensorflow.math.square', 2023-03-28T13:40:43.6500718Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T13:40:43.6501003Z E num_positional_args=0, 2023-03-28T13:40:43.6501248Z E with_out=False, 2023-03-28T13:40:43.6501478Z E inplace=False, 2023-03-28T13:40:43.6501716Z E as_variable=[False], 2023-03-28T13:40:43.6501969Z E native_arrays=[False], 2023-03-28T13:40:43.6502237Z E generate_frontend_arrays=False, 2023-03-28T13:40:43.6502477Z E ), 2023-03-28T13:40:43.6502757Z E frontend='tensorflow', 2023-03-28T13:40:43.6503034Z E on_device='cpu', 2023-03-28T13:40:43.6503255Z E ) 2023-03-28T13:40:43.6503447Z E 2023-03-28T13:40:43.6503996Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2AAAkYGCIDQAAAnAAM=') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_math.py::test_tensorflow_square[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T13:40:43.6490511Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6493820Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6498999Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6499474Z E Falsifying example: test_tensorflow_square( 2023-03-28T13:40:43.6499884Z E dtype_and_x=(['float16'], [array(-1., dtype=float16)]), 2023-03-28T13:40:43.6500343Z E fn_tree='ivy.functional.frontends.tensorflow.math.square', 2023-03-28T13:40:43.6500718Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T13:40:43.6501003Z E num_positional_args=0, 2023-03-28T13:40:43.6501248Z E with_out=False, 2023-03-28T13:40:43.6501478Z E inplace=False, 2023-03-28T13:40:43.6501716Z E as_variable=[False], 2023-03-28T13:40:43.6501969Z E native_arrays=[False], 2023-03-28T13:40:43.6502237Z E generate_frontend_arrays=False, 2023-03-28T13:40:43.6502477Z E ), 2023-03-28T13:40:43.6502757Z E frontend='tensorflow', 2023-03-28T13:40:43.6503034Z E on_device='cpu', 2023-03-28T13:40:43.6503255Z E ) 2023-03-28T13:40:43.6503447Z E 2023-03-28T13:40:43.6503996Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2AAAkYGCIDQAAAnAAM=') as a decorator on your test case </details>
1.0
Fix math.test_tensorflow_square - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4543693013/jobs/8008709037" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4543693013/jobs/8008709037" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4543693013/jobs/8008709037" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4543693013/jobs/8008709037" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_math.py::test_tensorflow_square[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T13:40:43.6490511Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6493820Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6498999Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6499474Z E Falsifying example: test_tensorflow_square( 2023-03-28T13:40:43.6499884Z E dtype_and_x=(['float16'], [array(-1., dtype=float16)]), 2023-03-28T13:40:43.6500343Z E fn_tree='ivy.functional.frontends.tensorflow.math.square', 2023-03-28T13:40:43.6500718Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T13:40:43.6501003Z E num_positional_args=0, 2023-03-28T13:40:43.6501248Z E with_out=False, 2023-03-28T13:40:43.6501478Z E inplace=False, 2023-03-28T13:40:43.6501716Z E as_variable=[False], 2023-03-28T13:40:43.6501969Z E native_arrays=[False], 2023-03-28T13:40:43.6502237Z E generate_frontend_arrays=False, 2023-03-28T13:40:43.6502477Z E ), 2023-03-28T13:40:43.6502757Z E frontend='tensorflow', 2023-03-28T13:40:43.6503034Z E on_device='cpu', 2023-03-28T13:40:43.6503255Z E ) 2023-03-28T13:40:43.6503447Z E 2023-03-28T13:40:43.6503996Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2AAAkYGCIDQAAAnAAM=') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_math.py::test_tensorflow_square[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T13:40:43.6490511Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6493820Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6498999Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6499474Z E Falsifying example: test_tensorflow_square( 2023-03-28T13:40:43.6499884Z E dtype_and_x=(['float16'], [array(-1., dtype=float16)]), 2023-03-28T13:40:43.6500343Z E fn_tree='ivy.functional.frontends.tensorflow.math.square', 2023-03-28T13:40:43.6500718Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T13:40:43.6501003Z E num_positional_args=0, 2023-03-28T13:40:43.6501248Z E with_out=False, 2023-03-28T13:40:43.6501478Z E inplace=False, 2023-03-28T13:40:43.6501716Z E as_variable=[False], 2023-03-28T13:40:43.6501969Z E native_arrays=[False], 2023-03-28T13:40:43.6502237Z E generate_frontend_arrays=False, 2023-03-28T13:40:43.6502477Z E ), 2023-03-28T13:40:43.6502757Z E frontend='tensorflow', 2023-03-28T13:40:43.6503034Z E on_device='cpu', 2023-03-28T13:40:43.6503255Z E ) 2023-03-28T13:40:43.6503447Z E 2023-03-28T13:40:43.6503996Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2AAAkYGCIDQAAAnAAM=') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_math.py::test_tensorflow_square[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T13:40:43.6490511Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6493820Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6498999Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6499474Z E Falsifying example: test_tensorflow_square( 2023-03-28T13:40:43.6499884Z E dtype_and_x=(['float16'], [array(-1., dtype=float16)]), 2023-03-28T13:40:43.6500343Z E fn_tree='ivy.functional.frontends.tensorflow.math.square', 2023-03-28T13:40:43.6500718Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T13:40:43.6501003Z E num_positional_args=0, 2023-03-28T13:40:43.6501248Z E with_out=False, 2023-03-28T13:40:43.6501478Z E inplace=False, 2023-03-28T13:40:43.6501716Z E as_variable=[False], 2023-03-28T13:40:43.6501969Z E native_arrays=[False], 2023-03-28T13:40:43.6502237Z E generate_frontend_arrays=False, 2023-03-28T13:40:43.6502477Z E ), 2023-03-28T13:40:43.6502757Z E frontend='tensorflow', 2023-03-28T13:40:43.6503034Z E on_device='cpu', 2023-03-28T13:40:43.6503255Z E ) 2023-03-28T13:40:43.6503447Z E 2023-03-28T13:40:43.6503996Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2AAAkYGCIDQAAAnAAM=') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_math.py::test_tensorflow_square[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T13:40:43.6490511Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6493820Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6498999Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T13:40:43.6499474Z E Falsifying example: test_tensorflow_square( 2023-03-28T13:40:43.6499884Z E dtype_and_x=(['float16'], [array(-1., dtype=float16)]), 2023-03-28T13:40:43.6500343Z E fn_tree='ivy.functional.frontends.tensorflow.math.square', 2023-03-28T13:40:43.6500718Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T13:40:43.6501003Z E num_positional_args=0, 2023-03-28T13:40:43.6501248Z E with_out=False, 2023-03-28T13:40:43.6501478Z E inplace=False, 2023-03-28T13:40:43.6501716Z E as_variable=[False], 2023-03-28T13:40:43.6501969Z E native_arrays=[False], 2023-03-28T13:40:43.6502237Z E generate_frontend_arrays=False, 2023-03-28T13:40:43.6502477Z E ), 2023-03-28T13:40:43.6502757Z E frontend='tensorflow', 2023-03-28T13:40:43.6503034Z E on_device='cpu', 2023-03-28T13:40:43.6503255Z E ) 2023-03-28T13:40:43.6503447Z E 2023-03-28T13:40:43.6503996Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2AAAkYGCIDQAAAnAAM=') as a decorator on your test case </details>
test
fix math test tensorflow square tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test frontends test tensorflow test math py test tensorflow square e typeerror to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map tensorflow nested map to native got an unexpected keyword argument dtype e falsifying example test tensorflow square e dtype and x e fn tree ivy functional frontends tensorflow math square e test flags frontendfunctiontestflags e num positional args e with out false e inplace false e as variable e native arrays e generate frontend arrays false e e frontend tensorflow e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test tensorflow test math py test tensorflow square e typeerror to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map tensorflow nested map to native got an unexpected keyword argument dtype e falsifying example test tensorflow square e dtype and x e fn tree ivy functional frontends tensorflow math square e test flags frontendfunctiontestflags e num positional args e with out false e inplace false e as variable e native arrays e generate frontend arrays false e e frontend tensorflow e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test tensorflow test math py test tensorflow square e typeerror to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map tensorflow nested map to native got an unexpected keyword argument dtype e falsifying example test tensorflow square e dtype and x e fn tree ivy functional frontends tensorflow math square e test flags frontendfunctiontestflags e num positional args e with out false e inplace false e as variable e native arrays e generate frontend arrays false e e frontend tensorflow e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test tensorflow test math py test tensorflow square e typeerror to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map tensorflow nested map to native got an unexpected keyword argument dtype e falsifying example test tensorflow square e dtype and x e fn tree ivy functional frontends tensorflow math square e test flags frontendfunctiontestflags e num positional args e with out false e inplace false e as variable e native arrays e generate frontend arrays false e e frontend tensorflow e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case
1
113,533
9,652,519,355
IssuesEvent
2019-05-18 17:53:13
apache/incubator-mxnet
https://api.github.com/repos/apache/incubator-mxnet
opened
Flaky test: test_lstm_clip
Flaky Test
test_gluon_gpu.test_lstm_clip ``` ====================================================================== FAIL: test_gluon_gpu.test_lstm_clip ---------------------------------------------------------------------- Traceback (most recent call last): File "/usr/local/lib/python3.5/dist-packages/nose/case.py", line 198, in runTest self.test(*self.arg) File "/work/mxnet/tests/python/gpu/../unittest/common.py", line 177, in test_new orig_test(*args, **kwargs) File "/work/mxnet/tests/python/gpu/../unittest/common.py", line 110, in test_new orig_test(*args, **kwargs) File "/work/mxnet/tests/python/gpu/test_gluon_gpu.py", line 163, in test_lstm_clip assert (cell_states >= clip_min).all() and (cell_states <= clip_max).all() AssertionError: -------------------- >> begin captured logging << -------------------- common: INFO: Setting test np/mx/python random seeds, use MXNET_TEST_SEED=551104712 to reproduce. --------------------- >> end captured logging << --------------------- ```` http://jenkins.mxnet-ci.amazon-ml.com/blue/organizations/jenkins/mxnet-validation%2Funix-gpu/detail/PR-14935/6/pipeline https://github.com/apache/incubator-mxnet/pull/14935/files
1.0
Flaky test: test_lstm_clip - test_gluon_gpu.test_lstm_clip ``` ====================================================================== FAIL: test_gluon_gpu.test_lstm_clip ---------------------------------------------------------------------- Traceback (most recent call last): File "/usr/local/lib/python3.5/dist-packages/nose/case.py", line 198, in runTest self.test(*self.arg) File "/work/mxnet/tests/python/gpu/../unittest/common.py", line 177, in test_new orig_test(*args, **kwargs) File "/work/mxnet/tests/python/gpu/../unittest/common.py", line 110, in test_new orig_test(*args, **kwargs) File "/work/mxnet/tests/python/gpu/test_gluon_gpu.py", line 163, in test_lstm_clip assert (cell_states >= clip_min).all() and (cell_states <= clip_max).all() AssertionError: -------------------- >> begin captured logging << -------------------- common: INFO: Setting test np/mx/python random seeds, use MXNET_TEST_SEED=551104712 to reproduce. --------------------- >> end captured logging << --------------------- ```` http://jenkins.mxnet-ci.amazon-ml.com/blue/organizations/jenkins/mxnet-validation%2Funix-gpu/detail/PR-14935/6/pipeline https://github.com/apache/incubator-mxnet/pull/14935/files
test
flaky test test lstm clip test gluon gpu test lstm clip fail test gluon gpu test lstm clip traceback most recent call last file usr local lib dist packages nose case py line in runtest self test self arg file work mxnet tests python gpu unittest common py line in test new orig test args kwargs file work mxnet tests python gpu unittest common py line in test new orig test args kwargs file work mxnet tests python gpu test gluon gpu py line in test lstm clip assert cell states clip min all and cell states clip max all assertionerror begin captured logging common info setting test np mx python random seeds use mxnet test seed to reproduce end captured logging
1
3,358
4,350,171,228
IssuesEvent
2016-07-31 02:50:36
BOWiki/BOW
https://api.github.com/repos/BOWiki/BOW
closed
Perform Brakeman Audit
P1 security
[Brakeman](http://brakemanscanner.org/) is a scanner. We should perform an audit to determine what we need to fix, if anything.
True
Perform Brakeman Audit - [Brakeman](http://brakemanscanner.org/) is a scanner. We should perform an audit to determine what we need to fix, if anything.
non_test
perform brakeman audit is a scanner we should perform an audit to determine what we need to fix if anything
0
320,156
27,421,754,048
IssuesEvent
2023-03-01 17:10:45
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
closed
Chart workflow fails during GraphQL Helm test
bug test regression
### Description The GraphQL Helm test sporadically [fails](https://github.com/hashgraph/hedera-mirror-node/actions/runs/4290666403/jobs/7474919041) in CI during the charts workflow. ### Steps to reproduce Run charts install workflow ### Additional context ``` ------------------------------------------------------------------------------------------------------------------------ ==> Logs of container hedera-mirror-fbm9z10oue-graphql-test ------------------------------------------------------------------------------------------------------------------------ newman GraphQL API ❏ Negative Tests ↳ Get Account Non Existing field POST http://hedera-mirror-fbm9z10oue-graphql:80/graphql/alpha [200 OK, 288B, 282ms] ✓ Non Existing Field ↳ By Alias (Unimplemented) POST http://hedera-mirror-fbm9z10oue-graphql:80/graphql/alpha [200 OK, 243B, 99ms] ✓ By Id (Unimplemented) → Get Account POST http://hedera-mirror-fbm9z10oue-graphql:80/graphql/alpha [200 OK, 96B, 133ms] 1. Get By Entity Id (All Fields) ┌─────────────────────────┬────────────────────┬───────────────────┐ │ │ executed │ failed │ ├─────────────────────────┼────────────────────┼───────────────────┤ │ iterations │ 1 │ 0 │ ├─────────────────────────┼────────────────────┼───────────────────┤ │ requests │ 3 │ 0 │ ├─────────────────────────┼────────────────────┼───────────────────┤ │ test-scripts │ 3 │ 0 │ ├─────────────────────────┼────────────────────┼───────────────────┤ │ prerequest-scripts │ 0 │ 0 │ ├─────────────────────────┼────────────────────┼───────────────────┤ │ assertions │ 3 │ 1 │ ├─────────────────────────┴────────────────────┴───────────────────┤ │ total run duration: 850ms │ ├──────────────────────────────────────────────────────────────────┤ │ total data received: 412B (approx) │ ├──────────────────────────────────────────────────────────────────┤ │ average response time: 171ms [min: 99ms, max: 282ms, s.d.: 79ms] │ └──────────────────────────────────────────────────────────────────┘ # failure detail 1. TypeError Get By Entity Id (All Fields) Cannot convert undefined or null to object at assertion:0 in test-script inside "Get Account" ``` ### Hedera network other ### Version main ### Operating system None
1.0
Chart workflow fails during GraphQL Helm test - ### Description The GraphQL Helm test sporadically [fails](https://github.com/hashgraph/hedera-mirror-node/actions/runs/4290666403/jobs/7474919041) in CI during the charts workflow. ### Steps to reproduce Run charts install workflow ### Additional context ``` ------------------------------------------------------------------------------------------------------------------------ ==> Logs of container hedera-mirror-fbm9z10oue-graphql-test ------------------------------------------------------------------------------------------------------------------------ newman GraphQL API ❏ Negative Tests ↳ Get Account Non Existing field POST http://hedera-mirror-fbm9z10oue-graphql:80/graphql/alpha [200 OK, 288B, 282ms] ✓ Non Existing Field ↳ By Alias (Unimplemented) POST http://hedera-mirror-fbm9z10oue-graphql:80/graphql/alpha [200 OK, 243B, 99ms] ✓ By Id (Unimplemented) → Get Account POST http://hedera-mirror-fbm9z10oue-graphql:80/graphql/alpha [200 OK, 96B, 133ms] 1. Get By Entity Id (All Fields) ┌─────────────────────────┬────────────────────┬───────────────────┐ │ │ executed │ failed │ ├─────────────────────────┼────────────────────┼───────────────────┤ │ iterations │ 1 │ 0 │ ├─────────────────────────┼────────────────────┼───────────────────┤ │ requests │ 3 │ 0 │ ├─────────────────────────┼────────────────────┼───────────────────┤ │ test-scripts │ 3 │ 0 │ ├─────────────────────────┼────────────────────┼───────────────────┤ │ prerequest-scripts │ 0 │ 0 │ ├─────────────────────────┼────────────────────┼───────────────────┤ │ assertions │ 3 │ 1 │ ├─────────────────────────┴────────────────────┴───────────────────┤ │ total run duration: 850ms │ ├──────────────────────────────────────────────────────────────────┤ │ total data received: 412B (approx) │ ├──────────────────────────────────────────────────────────────────┤ │ average response time: 171ms [min: 99ms, max: 282ms, s.d.: 79ms] │ └──────────────────────────────────────────────────────────────────┘ # failure detail 1. TypeError Get By Entity Id (All Fields) Cannot convert undefined or null to object at assertion:0 in test-script inside "Get Account" ``` ### Hedera network other ### Version main ### Operating system None
test
chart workflow fails during graphql helm test description the graphql helm test sporadically in ci during the charts workflow steps to reproduce run charts install workflow additional context logs of container hedera mirror graphql test newman graphql api ❏ negative tests ↳ get account non existing field post ✓ non existing field ↳ by alias unimplemented post ✓ by id unimplemented → get account post get by entity id all fields ┌─────────────────────────┬────────────────────┬───────────────────┐ │ │ executed │ failed │ ├─────────────────────────┼────────────────────┼───────────────────┤ │ iterations │ │ │ ├─────────────────────────┼────────────────────┼───────────────────┤ │ requests │ │ │ ├─────────────────────────┼────────────────────┼───────────────────┤ │ test scripts │ │ │ ├─────────────────────────┼────────────────────┼───────────────────┤ │ prerequest scripts │ │ │ ├─────────────────────────┼────────────────────┼───────────────────┤ │ assertions │ │ │ ├─────────────────────────┴────────────────────┴───────────────────┤ │ total run duration │ ├──────────────────────────────────────────────────────────────────┤ │ total data received approx │ ├──────────────────────────────────────────────────────────────────┤ │ average response time │ └──────────────────────────────────────────────────────────────────┘ failure detail typeerror get by entity id all fields cannot convert undefined or null to object at assertion in test script inside get account hedera network other version main operating system none
1
222,543
17,460,355,016
IssuesEvent
2021-08-06 09:33:17
Tencent/bk-job
https://api.github.com/repos/Tencent/bk-job
closed
bugfix: 文件任务包含大量的源和目标主机,导致执行引擎阻塞
kind/bug stage/test stage/uat
<!-- Please use this template while reporting a bug and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks! --> **What happened**: **What you expected to happen**: **How to reproduce it (as minimally and precisely as possible)**: **Anything else we need to know?**: **Environment**: **Required**: - bk-job version (use `cat VERSION` in installed dir): **Optional**: - Cloud provider or hardware configuration: - OS (e.g: `cat /etc/os-release`): - Kernel (e.g. `uname -a`): - Install tools: - Others:
1.0
bugfix: 文件任务包含大量的源和目标主机,导致执行引擎阻塞 - <!-- Please use this template while reporting a bug and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks! --> **What happened**: **What you expected to happen**: **How to reproduce it (as minimally and precisely as possible)**: **Anything else we need to know?**: **Environment**: **Required**: - bk-job version (use `cat VERSION` in installed dir): **Optional**: - Cloud provider or hardware configuration: - OS (e.g: `cat /etc/os-release`): - Kernel (e.g. `uname -a`): - Install tools: - Others:
test
bugfix 文件任务包含大量的源和目标主机,导致执行引擎阻塞 please use this template while reporting a bug and provide as much info as possible not doing so may result in your bug not being addressed in a timely manner thanks what happened what you expected to happen how to reproduce it as minimally and precisely as possible anything else we need to know environment required bk job version use cat version in installed dir optional cloud provider or hardware configuration os e g cat etc os release kernel e g uname a install tools others
1
185,338
6,721,119,551
IssuesEvent
2017-10-16 10:23:36
CS2103AUG2017-T17-B1/main
https://api.github.com/repos/CS2103AUG2017-T17-B1/main
opened
As a efficient debt collector I want to have an automated journey scheduler
priority.medium
so that I can maximise my efficiency
1.0
As a efficient debt collector I want to have an automated journey scheduler - so that I can maximise my efficiency
non_test
as a efficient debt collector i want to have an automated journey scheduler so that i can maximise my efficiency
0
770,945
27,062,455,423
IssuesEvent
2023-02-13 20:57:36
calcom/cal.com
https://api.github.com/repos/calcom/cal.com
closed
[CAL-1030] mobile dropdown menu layout issues
✅ good first issue 🙋🏻‍♂️help wanted 🧹 Improvements ⚡ Quick Wins Medium priority
found another issue on mobile - [ ] flickering border - [ ] double border for delete button https://user-images.githubusercontent.com/8019099/217485440-22d1d56d-4c65-4475-ac3a-a8008bb5073a.mp4 _Originally posted by @PeerRich in https://github.com/calcom/cal.com/pull/6854#pullrequestreview-1288750773_ <sub>[CAL-1030](https://linear.app/calcom/issue/CAL-1030/mobile-dropdown-menu-layout-issues)</sub>
1.0
[CAL-1030] mobile dropdown menu layout issues - found another issue on mobile - [ ] flickering border - [ ] double border for delete button https://user-images.githubusercontent.com/8019099/217485440-22d1d56d-4c65-4475-ac3a-a8008bb5073a.mp4 _Originally posted by @PeerRich in https://github.com/calcom/cal.com/pull/6854#pullrequestreview-1288750773_ <sub>[CAL-1030](https://linear.app/calcom/issue/CAL-1030/mobile-dropdown-menu-layout-issues)</sub>
non_test
mobile dropdown menu layout issues found another issue on mobile flickering border double border for delete button originally posted by peerrich in
0
163,510
12,733,284,737
IssuesEvent
2020-06-25 12:01:32
DiSSCo/ELViS
https://api.github.com/repos/DiSSCo/ELViS
closed
Possibility for requesters and VA Coordinators to filter own/other requests
enhancement resolved to test
#### Description Since requesters and VA Coordinators now have the possibility to see all requests from others (either other requesters or requests related to other institutions) in the main menu option "Requests", it is necessary to offer them a filter via which they can choose to either see their own requests and requests related to their own institution, or other requests, from other requesters (except the ones still in the draft status ofc) or only related to other institution than their own. Sprint target for sprint 13
1.0
Possibility for requesters and VA Coordinators to filter own/other requests - #### Description Since requesters and VA Coordinators now have the possibility to see all requests from others (either other requesters or requests related to other institutions) in the main menu option "Requests", it is necessary to offer them a filter via which they can choose to either see their own requests and requests related to their own institution, or other requests, from other requesters (except the ones still in the draft status ofc) or only related to other institution than their own. Sprint target for sprint 13
test
possibility for requesters and va coordinators to filter own other requests description since requesters and va coordinators now have the possibility to see all requests from others either other requesters or requests related to other institutions in the main menu option requests it is necessary to offer them a filter via which they can choose to either see their own requests and requests related to their own institution or other requests from other requesters except the ones still in the draft status ofc or only related to other institution than their own sprint target for sprint
1
42,238
5,429,494,756
IssuesEvent
2017-03-03 18:37:38
webcompat/webcompat.com
https://api.github.com/repos/webcompat/webcompat.com
closed
Only print expected fixture file path when we can't find it
scope: testing status: good first bug
Currently we output a message "Expected fixture file: ..." every time a fixture file is used when running tests. It's not very useful when things are working. <img width="1199" alt="screen shot 2016-10-04 at 4 28 27 pm" src="https://cloud.githubusercontent.com/assets/67283/19093213/d9b57494-8a4f-11e6-8349-bc2afe94efe3.png"> To fix this bug, we should only show that message when a fixture file is _not_ found, so we can have a better idea of what the problem is. The relevant code is in the `mockable_response` decorator: https://github.com/webcompat/webcompat.com/blob/master/webcompat/helpers.py#L364
1.0
Only print expected fixture file path when we can't find it - Currently we output a message "Expected fixture file: ..." every time a fixture file is used when running tests. It's not very useful when things are working. <img width="1199" alt="screen shot 2016-10-04 at 4 28 27 pm" src="https://cloud.githubusercontent.com/assets/67283/19093213/d9b57494-8a4f-11e6-8349-bc2afe94efe3.png"> To fix this bug, we should only show that message when a fixture file is _not_ found, so we can have a better idea of what the problem is. The relevant code is in the `mockable_response` decorator: https://github.com/webcompat/webcompat.com/blob/master/webcompat/helpers.py#L364
test
only print expected fixture file path when we can t find it currently we output a message expected fixture file every time a fixture file is used when running tests it s not very useful when things are working img width alt screen shot at pm src to fix this bug we should only show that message when a fixture file is not found so we can have a better idea of what the problem is the relevant code is in the mockable response decorator
1
284,103
21,389,515,148
IssuesEvent
2022-04-21 05:04:48
cengage/react-magma
https://api.github.com/repos/cengage/react-magma
closed
Documentation> Dropzone> "View Design Guidelines" button is not appearing to navigate back from API page to Guidelines page.
documentation
**Describe the bug** Button "View Design Guidelines" is not apearing when clicks on button "View Component API". **To Reproduce** Steps to reproduce the behavior: 1. Go to https://react-magma.cengage.com/version/2.5.7/design/dropzone/ 2. Click on "View Component API" button 3. Verify that "View Design Guidelines" button is not appearing to navigate back from API page to Guidelines page **Expected behavior** Button "View Design Guidelines" should appear when clicks on button "View Component API" **Screenshots** https://somup.com/c3nQXVZN4k **Desktop :** - OS: [Win 10] - Browser [chrome] - Version [Chrome version: Version 98.0.4758.81 (Official Build) (64-bit)] **Additional context** We have observed that "Dropzone" is appearing under Design guidelines and API both. When we clicks on "View Component API" button, the API page appears. However, user is unable to navigate back to Guidelines page as "View Design Guidelines" button is missing.
1.0
Documentation> Dropzone> "View Design Guidelines" button is not appearing to navigate back from API page to Guidelines page. - **Describe the bug** Button "View Design Guidelines" is not apearing when clicks on button "View Component API". **To Reproduce** Steps to reproduce the behavior: 1. Go to https://react-magma.cengage.com/version/2.5.7/design/dropzone/ 2. Click on "View Component API" button 3. Verify that "View Design Guidelines" button is not appearing to navigate back from API page to Guidelines page **Expected behavior** Button "View Design Guidelines" should appear when clicks on button "View Component API" **Screenshots** https://somup.com/c3nQXVZN4k **Desktop :** - OS: [Win 10] - Browser [chrome] - Version [Chrome version: Version 98.0.4758.81 (Official Build) (64-bit)] **Additional context** We have observed that "Dropzone" is appearing under Design guidelines and API both. When we clicks on "View Component API" button, the API page appears. However, user is unable to navigate back to Guidelines page as "View Design Guidelines" button is missing.
non_test
documentation dropzone view design guidelines button is not appearing to navigate back from api page to guidelines page describe the bug button view design guidelines is not apearing when clicks on button view component api to reproduce steps to reproduce the behavior go to click on view component api button verify that view design guidelines button is not appearing to navigate back from api page to guidelines page expected behavior button view design guidelines should appear when clicks on button view component api screenshots desktop os browser version additional context we have observed that dropzone is appearing under design guidelines and api both when we clicks on view component api button the api page appears however user is unable to navigate back to guidelines page as view design guidelines button is missing
0
137,742
11,161,041,777
IssuesEvent
2019-12-26 12:07:45
Rexbas/Teletubbies-Mod
https://api.github.com/repos/Rexbas/Teletubbies-Mod
closed
Test 1.14 release
1.14 test
- [x] Test machine sounds on server - [x] Export all textures - [x] Recipes - [x] Sounds
1.0
Test 1.14 release - - [x] Test machine sounds on server - [x] Export all textures - [x] Recipes - [x] Sounds
test
test release test machine sounds on server export all textures recipes sounds
1
213,781
16,536,979,804
IssuesEvent
2021-05-27 12:56:12
GIScience/ohsome-quality-analyst
https://api.github.com/repos/GIScience/ohsome-quality-analyst
opened
Tests should use a fixed timerange
comments welcome tests
Tests should be reproducible. But OQT currently allows the tests to produce different results depending on external resources (e.g. ohsome API). The results of the queries should not be changing only because the ohsome API covers a larger timerange. E.g. https://github.com/GIScience/ohsome-quality-analyst/commit/a2628e234d60c64bc537c67d8d6324cc3b33d5e5#diff-63f2f7efc0e30d4dd22cc038f27d3793beddbde215d954a961cd45dd2858f570L80-R80
1.0
Tests should use a fixed timerange - Tests should be reproducible. But OQT currently allows the tests to produce different results depending on external resources (e.g. ohsome API). The results of the queries should not be changing only because the ohsome API covers a larger timerange. E.g. https://github.com/GIScience/ohsome-quality-analyst/commit/a2628e234d60c64bc537c67d8d6324cc3b33d5e5#diff-63f2f7efc0e30d4dd22cc038f27d3793beddbde215d954a961cd45dd2858f570L80-R80
test
tests should use a fixed timerange tests should be reproducible but oqt currently allows the tests to produce different results depending on external resources e g ohsome api the results of the queries should not be changing only because the ohsome api covers a larger timerange e g
1
201,498
22,972,728,729
IssuesEvent
2022-07-20 05:46:15
MetaMask/snaps-skunkworks
https://api.github.com/repos/MetaMask/snaps-skunkworks
opened
Add `wallet_getBip32Entropy_*` restricted method
enhancement snaps:capabilities snaps:security
Some forms of key derivation necessitate deviations from the [BIP-44](https://github.com/bitcoin/bips/blob/master/bip-0044.mediawiki) standard, for example SegWit ([BIP-49](https://github.com/bitcoin/bips/blob/master/bip-0049.mediawiki)) and native SegWit ([BIP-84](https://github.com/bitcoin/bips/blob/master/bip-0084.mediawiki)) Bitcoin accounts. Since we shipped [`@metamask/key-tree@4.0.0`](https://github.com/MetaMask/key-tree/releases/tag/v4.0.0), we are able to provide entropy for arbitrary [BIP-32](https://github.com/bitcoin/bips/blob/master/bip-0032.mediawiki] paths. We should leverage the full capabilities of this package and ship one or more restricted methods supporting these use cases. The method could be named `wallet_getBip32Entropy` and modeled on the existing `getSlip44Entropy` method, with the following differences: - The user should be able to specify one or more arbitrary BIP-32 paths using this permission, of arbitrary depths. - We could go with a wildcard implementation and do something like `wallet_getBip32Entropy_m/84'/0'` - We could also go with some kind of caveat format where the permission is parameterized in the snap manifest, like we plan to do with snap keyrings. - We should make sure that the path `m/44'` cannot be requested, and we should in general maximally attenuate the permission while ensuring that it's still flexible enough for all/the vast majority of BIP-32 use cases.
True
Add `wallet_getBip32Entropy_*` restricted method - Some forms of key derivation necessitate deviations from the [BIP-44](https://github.com/bitcoin/bips/blob/master/bip-0044.mediawiki) standard, for example SegWit ([BIP-49](https://github.com/bitcoin/bips/blob/master/bip-0049.mediawiki)) and native SegWit ([BIP-84](https://github.com/bitcoin/bips/blob/master/bip-0084.mediawiki)) Bitcoin accounts. Since we shipped [`@metamask/key-tree@4.0.0`](https://github.com/MetaMask/key-tree/releases/tag/v4.0.0), we are able to provide entropy for arbitrary [BIP-32](https://github.com/bitcoin/bips/blob/master/bip-0032.mediawiki] paths. We should leverage the full capabilities of this package and ship one or more restricted methods supporting these use cases. The method could be named `wallet_getBip32Entropy` and modeled on the existing `getSlip44Entropy` method, with the following differences: - The user should be able to specify one or more arbitrary BIP-32 paths using this permission, of arbitrary depths. - We could go with a wildcard implementation and do something like `wallet_getBip32Entropy_m/84'/0'` - We could also go with some kind of caveat format where the permission is parameterized in the snap manifest, like we plan to do with snap keyrings. - We should make sure that the path `m/44'` cannot be requested, and we should in general maximally attenuate the permission while ensuring that it's still flexible enough for all/the vast majority of BIP-32 use cases.
non_test
add wallet restricted method some forms of key derivation necessitate deviations from the standard for example segwit and native segwit bitcoin accounts since we shipped we are able to provide entropy for arbitrary paths we should leverage the full capabilities of this package and ship one or more restricted methods supporting these use cases the method could be named wallet and modeled on the existing method with the following differences the user should be able to specify one or more arbitrary bip paths using this permission of arbitrary depths we could go with a wildcard implementation and do something like wallet m we could also go with some kind of caveat format where the permission is parameterized in the snap manifest like we plan to do with snap keyrings we should make sure that the path m cannot be requested and we should in general maximally attenuate the permission while ensuring that it s still flexible enough for all the vast majority of bip use cases
0
277,007
30,590,467,263
IssuesEvent
2023-07-21 16:34:47
divviup/prio-server
https://api.github.com/repos/divviup/prio-server
closed
Enforce minimum number of contributions when releasing sum parts
schema-protocol security
If there are too few contributions (i.e., individual data share packets) to a sum part being emitted at the multi-batch reduce step, then it's possible the final aggregation would reveal information about individual participants. If that's the case, then data share processors should refuse to emit sum parts to the portal server if the total number of contributions is insufficient.
True
Enforce minimum number of contributions when releasing sum parts - If there are too few contributions (i.e., individual data share packets) to a sum part being emitted at the multi-batch reduce step, then it's possible the final aggregation would reveal information about individual participants. If that's the case, then data share processors should refuse to emit sum parts to the portal server if the total number of contributions is insufficient.
non_test
enforce minimum number of contributions when releasing sum parts if there are too few contributions i e individual data share packets to a sum part being emitted at the multi batch reduce step then it s possible the final aggregation would reveal information about individual participants if that s the case then data share processors should refuse to emit sum parts to the portal server if the total number of contributions is insufficient
0
166,018
12,888,499,566
IssuesEvent
2020-07-13 13:08:04
pingcap/tidb
https://api.github.com/repos/pingcap/tidb
closed
unit test TestKillTableReader fails
component/test type/bug
FAIL: executor_test.go:5920: testSuite.TestKillTableReader ## Bug Report Please answer these questions before submitting your issue. Thanks! ### 1. Minimal reproduce step (Required) <!-- a step by step guide for reproducing the bug. --> ``` [2020-07-10T02:34:55.037Z] ---------------------------------------------------------------------- [2020-07-10T02:34:55.037Z] FAIL: executor_test.go:5920: testSuite.TestKillTableReader [2020-07-10T02:34:55.037Z] [2020-07-10T02:34:55.038Z] executor_test.go:5932: [2020-07-10T02:34:55.038Z] c.Assert(tk.QueryToErr("select * from t"), Equals, executor.ErrQueryInterrupted) [2020-07-10T02:34:55.038Z] ... obtained *errors.withStack = [tikv:1317]Query execution was interrupted ("[tikv:1317]Query execution was interrupted") [2020-07-10T02:34:55.038Z] ... expected *terror.Error = &terror.Error{class:5, code:1317, message:"Query execution was interrupted", args:[]interface {}(nil), file:"", line:0} ("[executor:1317]Query execution was interrupted") [2020-07-10T02:34:55.038Z] [2020-07-10T02:34:55.038Z] [2020-07-10T02:34:55.038Z] ---------------------------------------------------------------------- ``` ### 2. What did you expect to see? (Required) ### 3. What did you see instead (Required) ### 4. Affected version (Required) <!-- v3.0.0, v4.0.0, etc --> ### 5. Root Cause Analysis <!-- should be filled by the investigator before it's closed -->
1.0
unit test TestKillTableReader fails - FAIL: executor_test.go:5920: testSuite.TestKillTableReader ## Bug Report Please answer these questions before submitting your issue. Thanks! ### 1. Minimal reproduce step (Required) <!-- a step by step guide for reproducing the bug. --> ``` [2020-07-10T02:34:55.037Z] ---------------------------------------------------------------------- [2020-07-10T02:34:55.037Z] FAIL: executor_test.go:5920: testSuite.TestKillTableReader [2020-07-10T02:34:55.037Z] [2020-07-10T02:34:55.038Z] executor_test.go:5932: [2020-07-10T02:34:55.038Z] c.Assert(tk.QueryToErr("select * from t"), Equals, executor.ErrQueryInterrupted) [2020-07-10T02:34:55.038Z] ... obtained *errors.withStack = [tikv:1317]Query execution was interrupted ("[tikv:1317]Query execution was interrupted") [2020-07-10T02:34:55.038Z] ... expected *terror.Error = &terror.Error{class:5, code:1317, message:"Query execution was interrupted", args:[]interface {}(nil), file:"", line:0} ("[executor:1317]Query execution was interrupted") [2020-07-10T02:34:55.038Z] [2020-07-10T02:34:55.038Z] [2020-07-10T02:34:55.038Z] ---------------------------------------------------------------------- ``` ### 2. What did you expect to see? (Required) ### 3. What did you see instead (Required) ### 4. Affected version (Required) <!-- v3.0.0, v4.0.0, etc --> ### 5. Root Cause Analysis <!-- should be filled by the investigator before it's closed -->
test
unit test testkilltablereader fails fail executor test go testsuite testkilltablereader bug report please answer these questions before submitting your issue thanks minimal reproduce step required fail executor test go testsuite testkilltablereader executor test go c assert tk querytoerr select from t equals executor errqueryinterrupted obtained errors withstack query execution was interrupted query execution was interrupted expected terror error terror error class code message query execution was interrupted args interface nil file line query execution was interrupted what did you expect to see required what did you see instead required affected version required root cause analysis
1
255,695
21,946,051,533
IssuesEvent
2022-05-24 00:48:01
FuelLabs/sway
https://api.github.com/repos/FuelLabs/sway
opened
Expand the scope of the IR unit tests to allow for use of the core library and standard library
bug lib: std testing ir lib: core P: high
Currently, the IR unit tests are unable to use the core and standard libraries, due to a limitation in how the unit test infra is set up. I propose that we change this so that we can expand the scope of the IR unit tests. Unit tests are here: https://github.com/FuelLabs/sway/blob/53361ee726dacdb5d1ffa2f4f5354b1476b02f05/sway-core/src/optimize.rs#L2546
1.0
Expand the scope of the IR unit tests to allow for use of the core library and standard library - Currently, the IR unit tests are unable to use the core and standard libraries, due to a limitation in how the unit test infra is set up. I propose that we change this so that we can expand the scope of the IR unit tests. Unit tests are here: https://github.com/FuelLabs/sway/blob/53361ee726dacdb5d1ffa2f4f5354b1476b02f05/sway-core/src/optimize.rs#L2546
test
expand the scope of the ir unit tests to allow for use of the core library and standard library currently the ir unit tests are unable to use the core and standard libraries due to a limitation in how the unit test infra is set up i propose that we change this so that we can expand the scope of the ir unit tests unit tests are here
1
319,802
23,790,891,460
IssuesEvent
2022-09-02 14:26:51
zio/zio
https://api.github.com/repos/zio/zio
closed
Add Documentation Regarding Creating Custom Top Level Runtimes
documentation
Sometimes users want the ability to customize a `Runtime` at the top of their application to create a `Runtime` with their configuration settings . In ZIO 2 layers provide a consistent way to do all runtime customization, including setting `RuntimeConfigFlags`, setting `FiberRef` values, and performing other configuration that is potentially effectual and resourceful such as loading configuration information or starting and stopping third party services. However, it is often not clear how to go from layers, which are applied to `ZIO` workflows, to a top level runtime. We actually have an operator for this, `Runtime.unsafe.fromLayer`. With this a user can create a top level runtime that they can use throughout their application. For example: ```scala val layer: ZLayer[Any, Nothing, Unit] = Runtime.removeDefaultLoggers ++ Runtime.addLogger(sl4jlogger) val runtime: Runtime[Any] = Unsafe.unsafe { implicit unsafe => Runtime.unsafe.fromLayer(layer) } ``` This allows performing top level configuration in a principled way for all types of configuration changes, regardless of whether they are pure, effectual, or resourceful. It also ensures that these layers are only built once instead of each time the runtime is used to run a workflow. We should add documentation regarding this feature so users know how to do this.
1.0
Add Documentation Regarding Creating Custom Top Level Runtimes - Sometimes users want the ability to customize a `Runtime` at the top of their application to create a `Runtime` with their configuration settings . In ZIO 2 layers provide a consistent way to do all runtime customization, including setting `RuntimeConfigFlags`, setting `FiberRef` values, and performing other configuration that is potentially effectual and resourceful such as loading configuration information or starting and stopping third party services. However, it is often not clear how to go from layers, which are applied to `ZIO` workflows, to a top level runtime. We actually have an operator for this, `Runtime.unsafe.fromLayer`. With this a user can create a top level runtime that they can use throughout their application. For example: ```scala val layer: ZLayer[Any, Nothing, Unit] = Runtime.removeDefaultLoggers ++ Runtime.addLogger(sl4jlogger) val runtime: Runtime[Any] = Unsafe.unsafe { implicit unsafe => Runtime.unsafe.fromLayer(layer) } ``` This allows performing top level configuration in a principled way for all types of configuration changes, regardless of whether they are pure, effectual, or resourceful. It also ensures that these layers are only built once instead of each time the runtime is used to run a workflow. We should add documentation regarding this feature so users know how to do this.
non_test
add documentation regarding creating custom top level runtimes sometimes users want the ability to customize a runtime at the top of their application to create a runtime with their configuration settings in zio layers provide a consistent way to do all runtime customization including setting runtimeconfigflags setting fiberref values and performing other configuration that is potentially effectual and resourceful such as loading configuration information or starting and stopping third party services however it is often not clear how to go from layers which are applied to zio workflows to a top level runtime we actually have an operator for this runtime unsafe fromlayer with this a user can create a top level runtime that they can use throughout their application for example scala val layer zlayer runtime removedefaultloggers runtime addlogger val runtime runtime unsafe unsafe implicit unsafe runtime unsafe fromlayer layer this allows performing top level configuration in a principled way for all types of configuration changes regardless of whether they are pure effectual or resourceful it also ensures that these layers are only built once instead of each time the runtime is used to run a workflow we should add documentation regarding this feature so users know how to do this
0
223,931
17,647,222,889
IssuesEvent
2021-08-20 08:07:26
navid-farjad/yollochat
https://api.github.com/repos/navid-farjad/yollochat
closed
Call Duration recording
approved done test required
After Call is done, as is expected, the call conversation need to be recorded and added as a note for the agent
1.0
Call Duration recording - After Call is done, as is expected, the call conversation need to be recorded and added as a note for the agent
test
call duration recording after call is done as is expected the call conversation need to be recorded and added as a note for the agent
1
159,142
20,036,653,775
IssuesEvent
2022-02-02 12:38:48
kapseliboi/webuild
https://api.github.com/repos/kapseliboi/webuild
opened
CVE-2021-21353 (High) detected in pug-2.0.3.tgz, pug-code-gen-2.0.1.tgz
security vulnerability
## CVE-2021-21353 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>pug-2.0.3.tgz</b>, <b>pug-code-gen-2.0.1.tgz</b></p></summary> <p> <details><summary><b>pug-2.0.3.tgz</b></p></summary> <p>A clean, whitespace-sensitive template language for writing HTML</p> <p>Library home page: <a href="https://registry.npmjs.org/pug/-/pug-2.0.3.tgz">https://registry.npmjs.org/pug/-/pug-2.0.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/pug/package.json</p> <p> Dependency Hierarchy: - :x: **pug-2.0.3.tgz** (Vulnerable Library) </details> <details><summary><b>pug-code-gen-2.0.1.tgz</b></p></summary> <p>Default code-generator for pug. It generates HTML via a JavaScript template function.</p> <p>Library home page: <a href="https://registry.npmjs.org/pug-code-gen/-/pug-code-gen-2.0.1.tgz">https://registry.npmjs.org/pug-code-gen/-/pug-code-gen-2.0.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/pug-code-gen/package.json</p> <p> Dependency Hierarchy: - pug-2.0.3.tgz (Root Library) - :x: **pug-code-gen-2.0.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/webuild/commit/cfe337e0888ede79847ee4201c96d43ab06dc89e">cfe337e0888ede79847ee4201c96d43ab06dc89e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Pug is an npm package which is a high-performance template engine. In pug before version 3.0.1, if a remote attacker was able to control the `pretty` option of the pug compiler, e.g. if you spread a user provided object such as the query parameters of a request into the pug template inputs, it was possible for them to achieve remote code execution on the node.js backend. This is fixed in version 3.0.1. This advisory applies to multiple pug packages including "pug", "pug-code-gen". pug-code-gen has a backported fix at version 2.0.3. This advisory is not exploitable if there is no way for un-trusted input to be passed to pug as the `pretty` option, e.g. if you compile templates in advance before applying user input to them, you do not need to upgrade. <p>Publish Date: 2021-03-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21353>CVE-2021-21353</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-p493-635q-r6gr">https://github.com/advisories/GHSA-p493-635q-r6gr</a></p> <p>Release Date: 2021-03-03</p> <p>Fix Resolution (pug-code-gen): 2.0.3</p> <p>Direct dependency fix Resolution (pug): 2.0.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-21353 (High) detected in pug-2.0.3.tgz, pug-code-gen-2.0.1.tgz - ## CVE-2021-21353 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>pug-2.0.3.tgz</b>, <b>pug-code-gen-2.0.1.tgz</b></p></summary> <p> <details><summary><b>pug-2.0.3.tgz</b></p></summary> <p>A clean, whitespace-sensitive template language for writing HTML</p> <p>Library home page: <a href="https://registry.npmjs.org/pug/-/pug-2.0.3.tgz">https://registry.npmjs.org/pug/-/pug-2.0.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/pug/package.json</p> <p> Dependency Hierarchy: - :x: **pug-2.0.3.tgz** (Vulnerable Library) </details> <details><summary><b>pug-code-gen-2.0.1.tgz</b></p></summary> <p>Default code-generator for pug. It generates HTML via a JavaScript template function.</p> <p>Library home page: <a href="https://registry.npmjs.org/pug-code-gen/-/pug-code-gen-2.0.1.tgz">https://registry.npmjs.org/pug-code-gen/-/pug-code-gen-2.0.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/pug-code-gen/package.json</p> <p> Dependency Hierarchy: - pug-2.0.3.tgz (Root Library) - :x: **pug-code-gen-2.0.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/webuild/commit/cfe337e0888ede79847ee4201c96d43ab06dc89e">cfe337e0888ede79847ee4201c96d43ab06dc89e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Pug is an npm package which is a high-performance template engine. In pug before version 3.0.1, if a remote attacker was able to control the `pretty` option of the pug compiler, e.g. if you spread a user provided object such as the query parameters of a request into the pug template inputs, it was possible for them to achieve remote code execution on the node.js backend. This is fixed in version 3.0.1. This advisory applies to multiple pug packages including "pug", "pug-code-gen". pug-code-gen has a backported fix at version 2.0.3. This advisory is not exploitable if there is no way for un-trusted input to be passed to pug as the `pretty` option, e.g. if you compile templates in advance before applying user input to them, you do not need to upgrade. <p>Publish Date: 2021-03-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21353>CVE-2021-21353</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-p493-635q-r6gr">https://github.com/advisories/GHSA-p493-635q-r6gr</a></p> <p>Release Date: 2021-03-03</p> <p>Fix Resolution (pug-code-gen): 2.0.3</p> <p>Direct dependency fix Resolution (pug): 2.0.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in pug tgz pug code gen tgz cve high severity vulnerability vulnerable libraries pug tgz pug code gen tgz pug tgz a clean whitespace sensitive template language for writing html library home page a href path to dependency file package json path to vulnerable library node modules pug package json dependency hierarchy x pug tgz vulnerable library pug code gen tgz default code generator for pug it generates html via a javascript template function library home page a href path to dependency file package json path to vulnerable library node modules pug code gen package json dependency hierarchy pug tgz root library x pug code gen tgz vulnerable library found in head commit a href found in base branch master vulnerability details pug is an npm package which is a high performance template engine in pug before version if a remote attacker was able to control the pretty option of the pug compiler e g if you spread a user provided object such as the query parameters of a request into the pug template inputs it was possible for them to achieve remote code execution on the node js backend this is fixed in version this advisory applies to multiple pug packages including pug pug code gen pug code gen has a backported fix at version this advisory is not exploitable if there is no way for un trusted input to be passed to pug as the pretty option e g if you compile templates in advance before applying user input to them you do not need to upgrade publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope changed impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution pug code gen direct dependency fix resolution pug step up your open source security game with whitesource
0
488,469
14,077,996,306
IssuesEvent
2020-11-04 12:55:29
gnosis/conditional-tokens-explorer
https://api.github.com/repos/gnosis/conditional-tokens-explorer
closed
Cannot merge 3 or more positions of a deeper level
Low priority bug
The issue might be fixed after #512 is implemented **Steps:** 1. create a condition with 3 or more outcomes 2. split positions for a condition (p.1, p.2. p.3) 3. then split any position (as an example, p.3) for the same condition (result = p.31, p.32, p.33) 4. open Merge position page 5. select p.31, p.32, p.33 positions **Actual Result:** Loader is eternally displayed in the positions section (positions can't be loaded) ![loaded.jpg](https://images.zenhubusercontent.com/5f55ec970a2e890ec10f7778/c03853ca-66bd-4eef-b344-2db6c4fd2f0c) **Expected Result:** positions are loaded, user is able to merge position
1.0
Cannot merge 3 or more positions of a deeper level - The issue might be fixed after #512 is implemented **Steps:** 1. create a condition with 3 or more outcomes 2. split positions for a condition (p.1, p.2. p.3) 3. then split any position (as an example, p.3) for the same condition (result = p.31, p.32, p.33) 4. open Merge position page 5. select p.31, p.32, p.33 positions **Actual Result:** Loader is eternally displayed in the positions section (positions can't be loaded) ![loaded.jpg](https://images.zenhubusercontent.com/5f55ec970a2e890ec10f7778/c03853ca-66bd-4eef-b344-2db6c4fd2f0c) **Expected Result:** positions are loaded, user is able to merge position
non_test
cannot merge or more positions of a deeper level the issue might be fixed after is implemented steps create a condition with or more outcomes split positions for a condition p p p then split any position as an example p for the same condition result p p p open merge position page select p p p positions actual result loader is eternally displayed in the positions section positions can t be loaded expected result positions are loaded user is able to merge position
0
497,214
14,365,643,006
IssuesEvent
2020-12-01 02:19:47
vmware/singleton
https://api.github.com/repos/vmware/singleton
opened
[REQUIREMENT] Downloadable offline resource bundle zip + sample app
area/java-client kind/feature priority/high
**Is your feature request related to a problem? Please describe.** Offline resource bundle zip is not available for download. **Describe the solution you'd like** 1. For every new product, downloadable offline resource bundle zip with folder hierarchy: <component> <component>/messages_<locale>.json 2. Update the sample app to use a sample downloadable zip from VIP test/dev.
1.0
[REQUIREMENT] Downloadable offline resource bundle zip + sample app - **Is your feature request related to a problem? Please describe.** Offline resource bundle zip is not available for download. **Describe the solution you'd like** 1. For every new product, downloadable offline resource bundle zip with folder hierarchy: <component> <component>/messages_<locale>.json 2. Update the sample app to use a sample downloadable zip from VIP test/dev.
non_test
downloadable offline resource bundle zip sample app is your feature request related to a problem please describe offline resource bundle zip is not available for download describe the solution you d like for every new product downloadable offline resource bundle zip with folder hierarchy messages json update the sample app to use a sample downloadable zip from vip test dev
0
187,701
14,429,240,561
IssuesEvent
2020-12-06 13:28:16
kalexmills/github-vet-tests-dec2020
https://api.github.com/repos/kalexmills/github-vet-tests-dec2020
closed
hashicorp/nomad: command/agent/agent_test.go; 12 LoC
fresh small test
Found a possible issue in [hashicorp/nomad](https://www.github.com/hashicorp/nomad) at [command/agent/agent_test.go](https://github.com/hashicorp/nomad/blob/d76340f0f7e7b44ad8ecb2b10cfc9bc710c940de/command/agent/agent_test.go#L462-L473) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > reference to tc is reassigned at line 467 [Click here to see the code in its original context.](https://github.com/hashicorp/nomad/blob/d76340f0f7e7b44ad8ecb2b10cfc9bc710c940de/command/agent/agent_test.go#L462-L473) <details> <summary>Click here to show the 12 line(s) of Go which triggered the analyzer.</summary> ```go for _, tc := range cases { t.Run(fmt.Sprintf("%v", tc), func(t *testing.T) { conf := DevConfig(nil) require.NoError(t, conf.normalizeAddrs()) conf.Server.RaftMultiplier = &tc _, err := convertServerConfig(conf) require.Error(t, err) require.Contains(t, err.Error(), "raft_multiplier cannot be") }) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: d76340f0f7e7b44ad8ecb2b10cfc9bc710c940de
1.0
hashicorp/nomad: command/agent/agent_test.go; 12 LoC - Found a possible issue in [hashicorp/nomad](https://www.github.com/hashicorp/nomad) at [command/agent/agent_test.go](https://github.com/hashicorp/nomad/blob/d76340f0f7e7b44ad8ecb2b10cfc9bc710c940de/command/agent/agent_test.go#L462-L473) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > reference to tc is reassigned at line 467 [Click here to see the code in its original context.](https://github.com/hashicorp/nomad/blob/d76340f0f7e7b44ad8ecb2b10cfc9bc710c940de/command/agent/agent_test.go#L462-L473) <details> <summary>Click here to show the 12 line(s) of Go which triggered the analyzer.</summary> ```go for _, tc := range cases { t.Run(fmt.Sprintf("%v", tc), func(t *testing.T) { conf := DevConfig(nil) require.NoError(t, conf.normalizeAddrs()) conf.Server.RaftMultiplier = &tc _, err := convertServerConfig(conf) require.Error(t, err) require.Contains(t, err.Error(), "raft_multiplier cannot be") }) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: d76340f0f7e7b44ad8ecb2b10cfc9bc710c940de
test
hashicorp nomad command agent agent test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message reference to tc is reassigned at line click here to show the line s of go which triggered the analyzer go for tc range cases t run fmt sprintf v tc func t testing t conf devconfig nil require noerror t conf normalizeaddrs conf server raftmultiplier tc err convertserverconfig conf require error t err require contains t err error raft multiplier cannot be leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
106,032
16,664,121,915
IssuesEvent
2021-06-06 21:31:48
uniquelyparticular/micro-cors-http-proxy
https://api.github.com/repos/uniquelyparticular/micro-cors-http-proxy
opened
CVE-2021-33587 (High) detected in css-what-2.1.3.tgz
security vulnerability
## CVE-2021-33587 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>css-what-2.1.3.tgz</b></p></summary> <p>a CSS selector parser</p> <p>Library home page: <a href="https://registry.npmjs.org/css-what/-/css-what-2.1.3.tgz">https://registry.npmjs.org/css-what/-/css-what-2.1.3.tgz</a></p> <p>Path to dependency file: micro-cors-http-proxy/package.json</p> <p>Path to vulnerable library: micro-cors-http-proxy/node_modules/css-what/package.json</p> <p> Dependency Hierarchy: - micro-dev-3.0.0.tgz (Root Library) - pretty-error-2.1.1.tgz - renderkid-2.0.3.tgz - css-select-1.2.0.tgz - :x: **css-what-2.1.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/uniquelyparticular/micro-cors-http-proxy/commit/abf77a0c3270a91ef5bbbf9a3dccf29ac0aef673">abf77a0c3270a91ef5bbbf9a3dccf29ac0aef673</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The css-what package before 5.0.1 for Node.js does not ensure that attribute parsing has Linear Time Complexity relative to the size of the input. <p>Publish Date: 2021-05-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33587>CVE-2021-33587</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33587">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33587</a></p> <p>Release Date: 2021-05-28</p> <p>Fix Resolution: css-what - 5.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-33587 (High) detected in css-what-2.1.3.tgz - ## CVE-2021-33587 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>css-what-2.1.3.tgz</b></p></summary> <p>a CSS selector parser</p> <p>Library home page: <a href="https://registry.npmjs.org/css-what/-/css-what-2.1.3.tgz">https://registry.npmjs.org/css-what/-/css-what-2.1.3.tgz</a></p> <p>Path to dependency file: micro-cors-http-proxy/package.json</p> <p>Path to vulnerable library: micro-cors-http-proxy/node_modules/css-what/package.json</p> <p> Dependency Hierarchy: - micro-dev-3.0.0.tgz (Root Library) - pretty-error-2.1.1.tgz - renderkid-2.0.3.tgz - css-select-1.2.0.tgz - :x: **css-what-2.1.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/uniquelyparticular/micro-cors-http-proxy/commit/abf77a0c3270a91ef5bbbf9a3dccf29ac0aef673">abf77a0c3270a91ef5bbbf9a3dccf29ac0aef673</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The css-what package before 5.0.1 for Node.js does not ensure that attribute parsing has Linear Time Complexity relative to the size of the input. <p>Publish Date: 2021-05-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33587>CVE-2021-33587</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33587">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33587</a></p> <p>Release Date: 2021-05-28</p> <p>Fix Resolution: css-what - 5.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in css what tgz cve high severity vulnerability vulnerable library css what tgz a css selector parser library home page a href path to dependency file micro cors http proxy package json path to vulnerable library micro cors http proxy node modules css what package json dependency hierarchy micro dev tgz root library pretty error tgz renderkid tgz css select tgz x css what tgz vulnerable library found in head commit a href vulnerability details the css what package before for node js does not ensure that attribute parsing has linear time complexity relative to the size of the input publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution css what step up your open source security game with whitesource
0
176,022
28,013,842,513
IssuesEvent
2023-03-27 20:47:18
Esri/calcite-components
https://api.github.com/repos/Esri/calcite-components
opened
[Action Menu] Responsive layout
enhancement design p - high 0 - new estimate - 8 needs milestone
### Description Mobile responsiveness across Calcite Design System, focused on web (traditionally desktop sized web browser), tablet, phone screen sizes, and large monitors. ### Acceptance Criteria - [ ] 1. Consider breaking changes for folks. Consultation should occur on both the design and development teams. - [ ] 2. Components need to be designed for the following breakpoints: - [ ] Small (s), `476px` - [ ] Medium (m), `768px` - [ ] Large (l), `1152px` - [ ] Extra larger (xl), `1440px` ### Relevant Info Part of epic https://github.com/Esri/calcite-components/issues/6670 ### Which Component `action-menu` ### Example Use Case _No response_ ### Priority impact p4 - not time sensitive ### Esri team Calcite (design)
1.0
[Action Menu] Responsive layout - ### Description Mobile responsiveness across Calcite Design System, focused on web (traditionally desktop sized web browser), tablet, phone screen sizes, and large monitors. ### Acceptance Criteria - [ ] 1. Consider breaking changes for folks. Consultation should occur on both the design and development teams. - [ ] 2. Components need to be designed for the following breakpoints: - [ ] Small (s), `476px` - [ ] Medium (m), `768px` - [ ] Large (l), `1152px` - [ ] Extra larger (xl), `1440px` ### Relevant Info Part of epic https://github.com/Esri/calcite-components/issues/6670 ### Which Component `action-menu` ### Example Use Case _No response_ ### Priority impact p4 - not time sensitive ### Esri team Calcite (design)
non_test
responsive layout description mobile responsiveness across calcite design system focused on web traditionally desktop sized web browser tablet phone screen sizes and large monitors acceptance criteria consider breaking changes for folks consultation should occur on both the design and development teams components need to be designed for the following breakpoints small s medium m large l extra larger xl relevant info part of epic which component action menu example use case no response priority impact not time sensitive esri team calcite design
0
38,166
5,168,353,474
IssuesEvent
2017-01-17 21:21:24
easydigitaldownloads/easy-digital-downloads
https://api.github.com/repos/easydigitaldownloads/easy-digital-downloads
closed
PayPal: disable redirect on Buy Now buttons if PayPal gateway is disabled
Bug Frontend Has PR Needs Testing Payments Priority: Medium
We've seen it happen numerous times where a site is set up with Stripe, 2Checkout or any other gateway that supports Buy Now buttons and when a purchase button is clicked, the customer is redirected to PayPal. Typically this means there is a JS error or something blocking the Stripe Checkout JS but regardless the redirect to PayPal should not occur. If the PayPal gateway is disabled, let's bail before redirecting to PayPal.
1.0
PayPal: disable redirect on Buy Now buttons if PayPal gateway is disabled - We've seen it happen numerous times where a site is set up with Stripe, 2Checkout or any other gateway that supports Buy Now buttons and when a purchase button is clicked, the customer is redirected to PayPal. Typically this means there is a JS error or something blocking the Stripe Checkout JS but regardless the redirect to PayPal should not occur. If the PayPal gateway is disabled, let's bail before redirecting to PayPal.
test
paypal disable redirect on buy now buttons if paypal gateway is disabled we ve seen it happen numerous times where a site is set up with stripe or any other gateway that supports buy now buttons and when a purchase button is clicked the customer is redirected to paypal typically this means there is a js error or something blocking the stripe checkout js but regardless the redirect to paypal should not occur if the paypal gateway is disabled let s bail before redirecting to paypal
1
6,440
3,022,684,698
IssuesEvent
2015-07-31 21:57:31
mk23/webhdfs
https://api.github.com/repos/mk23/webhdfs
closed
Add one-shot command support to webhdfs shell.
documentation enhancement
Allow additional positional parameters to specify a one-shot command that executes and exits.
1.0
Add one-shot command support to webhdfs shell. - Allow additional positional parameters to specify a one-shot command that executes and exits.
non_test
add one shot command support to webhdfs shell allow additional positional parameters to specify a one shot command that executes and exits
0
163,768
12,744,112,552
IssuesEvent
2020-06-26 11:50:40
UBitSandBox/HYBRID-REST
https://api.github.com/repos/UBitSandBox/HYBRID-REST
opened
Add Unit Test for Sync Part
enhancement security test
We should talk about this part... It's not dead easy to develop some unit Tests there
1.0
Add Unit Test for Sync Part - We should talk about this part... It's not dead easy to develop some unit Tests there
test
add unit test for sync part we should talk about this part it s not dead easy to develop some unit tests there
1
173,398
13,399,964,645
IssuesEvent
2020-09-03 15:10:30
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
CI: VariableWidthHistogramAggregatorTests#testMultipleSegments
:Analytics/Aggregations >test-failure Team:Analytics
Error Message java.lang.AssertionError: expected:<false> but was:<true> Stacktrace ``` java.lang.AssertionError: expected:<false> but was:<true> at __randomizedtesting.SeedInfo.seed([7B50D97E959CDE64:C7276C27523B7221]:0) at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:834) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.elasticsearch.search.aggregations.bucket.histogram.VariableWidthHistogramAggregatorTests.lambda$testMultipleSegments$31(VariableWidthHistogramAggregatorTests.java:442) at org.elasticsearch.search.aggregations.bucket.histogram.VariableWidthHistogramAggregatorTests.executeTestCase(VariableWidthHistogramAggregatorTests.java:638) at org.elasticsearch.search.aggregations.bucket.histogram.VariableWidthHistogramAggregatorTests.testSearchAndReduceCase(VariableWidthHistogramAggregatorTests.java:576) at org.elasticsearch.search.aggregations.bucket.histogram.VariableWidthHistogramAggregatorTests.testMultipleSegments(VariableWidthHistogramAggregatorTests.java:429) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1750) ```
1.0
CI: VariableWidthHistogramAggregatorTests#testMultipleSegments - Error Message java.lang.AssertionError: expected:<false> but was:<true> Stacktrace ``` java.lang.AssertionError: expected:<false> but was:<true> at __randomizedtesting.SeedInfo.seed([7B50D97E959CDE64:C7276C27523B7221]:0) at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:834) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.elasticsearch.search.aggregations.bucket.histogram.VariableWidthHistogramAggregatorTests.lambda$testMultipleSegments$31(VariableWidthHistogramAggregatorTests.java:442) at org.elasticsearch.search.aggregations.bucket.histogram.VariableWidthHistogramAggregatorTests.executeTestCase(VariableWidthHistogramAggregatorTests.java:638) at org.elasticsearch.search.aggregations.bucket.histogram.VariableWidthHistogramAggregatorTests.testSearchAndReduceCase(VariableWidthHistogramAggregatorTests.java:576) at org.elasticsearch.search.aggregations.bucket.histogram.VariableWidthHistogramAggregatorTests.testMultipleSegments(VariableWidthHistogramAggregatorTests.java:429) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1750) ```
test
ci variablewidthhistogramaggregatortests testmultiplesegments error message java lang assertionerror expected but was stacktrace java lang assertionerror expected but was at randomizedtesting seedinfo seed at org junit assert fail assert java at org junit assert failnotequals assert java at org junit assert assertequals assert java at org junit assert assertequals assert java at org elasticsearch search aggregations bucket histogram variablewidthhistogramaggregatortests lambda testmultiplesegments variablewidthhistogramaggregatortests java at org elasticsearch search aggregations bucket histogram variablewidthhistogramaggregatortests executetestcase variablewidthhistogramaggregatortests java at org elasticsearch search aggregations bucket histogram variablewidthhistogramaggregatortests testsearchandreducecase variablewidthhistogramaggregatortests java at org elasticsearch search aggregations bucket histogram variablewidthhistogramaggregatortests testmultiplesegments variablewidthhistogramaggregatortests java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java
1
182,261
21,664,494,763
IssuesEvent
2022-05-07 01:32:36
emilwareus/NodeGoat
https://api.github.com/repos/emilwareus/NodeGoat
closed
CVE-2021-3918 (High) detected in json-schema-0.2.3.tgz - autoclosed
security vulnerability
## CVE-2021-3918 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p></summary> <p>JSON Schema validation and specifications</p> <p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/json-schema/package.json,/node_modules/npm/node_modules/request/node_modules/http-signature/node_modules/jsprim/node_modules/json-schema/package.json</p> <p> Dependency Hierarchy: - grunt-retire-0.3.12.tgz (Root Library) - request-2.67.0.tgz - http-signature-1.1.1.tgz - jsprim-1.4.1.tgz - :x: **json-schema-0.2.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/emilwareus/NodeGoat/git/commits/8e0ff86cf8b947654265bbb4cd5fd2e727f2bdac">8e0ff86cf8b947654265bbb4cd5fd2e727f2bdac</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution') <p>Publish Date: 2021-11-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918>CVE-2021-3918</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-3918">https://nvd.nist.gov/vuln/detail/CVE-2021-3918</a></p> <p>Release Date: 2021-11-13</p> <p>Fix Resolution: json-schema - 0.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-3918 (High) detected in json-schema-0.2.3.tgz - autoclosed - ## CVE-2021-3918 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p></summary> <p>JSON Schema validation and specifications</p> <p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/json-schema/package.json,/node_modules/npm/node_modules/request/node_modules/http-signature/node_modules/jsprim/node_modules/json-schema/package.json</p> <p> Dependency Hierarchy: - grunt-retire-0.3.12.tgz (Root Library) - request-2.67.0.tgz - http-signature-1.1.1.tgz - jsprim-1.4.1.tgz - :x: **json-schema-0.2.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/emilwareus/NodeGoat/git/commits/8e0ff86cf8b947654265bbb4cd5fd2e727f2bdac">8e0ff86cf8b947654265bbb4cd5fd2e727f2bdac</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution') <p>Publish Date: 2021-11-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918>CVE-2021-3918</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-3918">https://nvd.nist.gov/vuln/detail/CVE-2021-3918</a></p> <p>Release Date: 2021-11-13</p> <p>Fix Resolution: json-schema - 0.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in json schema tgz autoclosed cve high severity vulnerability vulnerable library json schema tgz json schema validation and specifications library home page a href path to dependency file package json path to vulnerable library node modules json schema package json node modules npm node modules request node modules http signature node modules jsprim node modules json schema package json dependency hierarchy grunt retire tgz root library request tgz http signature tgz jsprim tgz x json schema tgz vulnerable library found in head commit a href vulnerability details json schema is vulnerable to improperly controlled modification of object prototype attributes prototype pollution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution json schema step up your open source security game with whitesource
0
26,534
4,750,475,202
IssuesEvent
2016-10-22 10:56:57
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
ValueError on DSBRDB called by scipy.linalg.eigvalsh
defect scipy.linalg
Hello, I am getting an error with large arrays when using scipy.linalg.eigvalsh: `>>> import numpy as np` `>>> from scipy.linalg import eigvalsh` `>>> eigvalsh(np.zeros((6143,6143)))` `array([ 0., 0., 0., ..., 0., 0., 0.])` `>>> eigvalsh(np.zeros((6144,6144)))` `Traceback (most recent call last):` ` File "<stdin>", line 1, in <module>` ` File "/users/gauthe/Library/anaconda2/lib/python2.7/site-packages/scipy/linalg/decomp.py", line 684, in eigvalsh` ` check_finite=check_finite)` ` File "/users/gauthe/Library/anaconda2/lib/python2.7/site-packages/scipy/linalg/decomp.py", line 345, in eigh` ` iu=a1.shape[0], overwrite_a=overwrite_a)` `ValueError: On entry to DSBRDB parameter number 12 had an illegal value` The issue seems close to #5401, but refers to another LAPACK function. It is a pure effect of size, since 6143 * 6143 is ok, but 6144 * 6144 is not. 6144 = 3 * (2**11) is big, but not that big. I have no problem on my macbook, with an older scipy (0.14.0), I think it's a LAPACK bug. Version I'm using; scipy version 0.18.1 numpy version 1.11.1 Python 2.7.12 Anaconda 4.2.0 (64-bit) openSUSE 13.2 (Harlequin) `>>> scipy.__config__.show()` lapack_opt_info: libraries = ['mkl_lapack95_lp64', 'mkl_intel_lp64', 'mkl_intel_thread', 'mkl_core', 'iomp5', 'pthread'] library_dirs = ['/users/gauthe/anaconda2/lib'] define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)] include_dirs = ['/users/gauthe/anaconda2/include'] blas_opt_info: libraries = ['mkl_intel_lp64', 'mkl_intel_thread', 'mkl_core', 'iomp5', 'pthread'] library_dirs = ['/users/gauthe/anaconda2/lib'] define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)] include_dirs = ['/users/gauthe/anaconda2/include'] openblas_lapack_info: NOT AVAILABLE lapack_mkl_info: libraries = ['mkl_lapack95_lp64', 'mkl_intel_lp64', 'mkl_intel_thread', 'mkl_core', 'iomp5', 'pthread'] library_dirs = ['/users/gauthe/anaconda2/lib'] define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)] include_dirs = ['/users/gauthe/anaconda2/include'] blas_mkl_info: libraries = ['mkl_intel_lp64', 'mkl_intel_thread', 'mkl_core', 'iomp5', 'pthread'] library_dirs = ['/users/gauthe/anaconda2/lib'] define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)] include_dirs = ['/users/gauthe/anaconda2/include'] mkl_info: libraries = ['mkl_intel_lp64', 'mkl_intel_thread', 'mkl_core', 'iomp5', 'pthread'] library_dirs = ['/users/gauthe/anaconda2/lib'] define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)] include_dirs = ['/users/gauthe/anaconda2/include']
1.0
ValueError on DSBRDB called by scipy.linalg.eigvalsh - Hello, I am getting an error with large arrays when using scipy.linalg.eigvalsh: `>>> import numpy as np` `>>> from scipy.linalg import eigvalsh` `>>> eigvalsh(np.zeros((6143,6143)))` `array([ 0., 0., 0., ..., 0., 0., 0.])` `>>> eigvalsh(np.zeros((6144,6144)))` `Traceback (most recent call last):` ` File "<stdin>", line 1, in <module>` ` File "/users/gauthe/Library/anaconda2/lib/python2.7/site-packages/scipy/linalg/decomp.py", line 684, in eigvalsh` ` check_finite=check_finite)` ` File "/users/gauthe/Library/anaconda2/lib/python2.7/site-packages/scipy/linalg/decomp.py", line 345, in eigh` ` iu=a1.shape[0], overwrite_a=overwrite_a)` `ValueError: On entry to DSBRDB parameter number 12 had an illegal value` The issue seems close to #5401, but refers to another LAPACK function. It is a pure effect of size, since 6143 * 6143 is ok, but 6144 * 6144 is not. 6144 = 3 * (2**11) is big, but not that big. I have no problem on my macbook, with an older scipy (0.14.0), I think it's a LAPACK bug. Version I'm using; scipy version 0.18.1 numpy version 1.11.1 Python 2.7.12 Anaconda 4.2.0 (64-bit) openSUSE 13.2 (Harlequin) `>>> scipy.__config__.show()` lapack_opt_info: libraries = ['mkl_lapack95_lp64', 'mkl_intel_lp64', 'mkl_intel_thread', 'mkl_core', 'iomp5', 'pthread'] library_dirs = ['/users/gauthe/anaconda2/lib'] define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)] include_dirs = ['/users/gauthe/anaconda2/include'] blas_opt_info: libraries = ['mkl_intel_lp64', 'mkl_intel_thread', 'mkl_core', 'iomp5', 'pthread'] library_dirs = ['/users/gauthe/anaconda2/lib'] define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)] include_dirs = ['/users/gauthe/anaconda2/include'] openblas_lapack_info: NOT AVAILABLE lapack_mkl_info: libraries = ['mkl_lapack95_lp64', 'mkl_intel_lp64', 'mkl_intel_thread', 'mkl_core', 'iomp5', 'pthread'] library_dirs = ['/users/gauthe/anaconda2/lib'] define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)] include_dirs = ['/users/gauthe/anaconda2/include'] blas_mkl_info: libraries = ['mkl_intel_lp64', 'mkl_intel_thread', 'mkl_core', 'iomp5', 'pthread'] library_dirs = ['/users/gauthe/anaconda2/lib'] define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)] include_dirs = ['/users/gauthe/anaconda2/include'] mkl_info: libraries = ['mkl_intel_lp64', 'mkl_intel_thread', 'mkl_core', 'iomp5', 'pthread'] library_dirs = ['/users/gauthe/anaconda2/lib'] define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)] include_dirs = ['/users/gauthe/anaconda2/include']
non_test
valueerror on dsbrdb called by scipy linalg eigvalsh hello i am getting an error with large arrays when using scipy linalg eigvalsh import numpy as np from scipy linalg import eigvalsh eigvalsh np zeros array eigvalsh np zeros traceback most recent call last file line in file users gauthe library lib site packages scipy linalg decomp py line in eigvalsh check finite check finite file users gauthe library lib site packages scipy linalg decomp py line in eigh iu shape overwrite a overwrite a valueerror on entry to dsbrdb parameter number had an illegal value the issue seems close to but refers to another lapack function it is a pure effect of size since is ok but is not is big but not that big i have no problem on my macbook with an older scipy i think it s a lapack bug version i m using scipy version numpy version python anaconda bit opensuse harlequin scipy config show lapack opt info libraries library dirs define macros include dirs blas opt info libraries library dirs define macros include dirs openblas lapack info not available lapack mkl info libraries library dirs define macros include dirs blas mkl info libraries library dirs define macros include dirs mkl info libraries library dirs define macros include dirs
0
110,510
9,458,776,195
IssuesEvent
2019-04-17 06:39:52
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Failing test: X-Pack API Integration Tests.x-pack/test/api_integration/apis/management/rollup/rollup·js - apis management rollup jobs "after all" hook
failed-test
Latest failure: https://kibana-ci.elastic.co/job/elastic+kibana+master/906/JOB=x-pack-ciGroup6,node=immutable/testReport/junit/X-Pack%20API%20Integration%20Tests/x-pack_test_api_integration_apis_management_rollup_rollup%C2%B7js/apis_management_rollup_jobs__after_all__hook/ Stacktrace ``` { [index_not_found_exception] no such index [statusCode], with { resource.type="index_or_alias" & resource.id="statusCode" & index_uuid="_na_" & index="statusCode" } :: {"path":"/statusCode%2Cerror%2Cmessage%2Ccause","query":{},"statusCode":404,"response":"{\"error\":{\"root_cause\":[{\"type\":\"index_not_found_exception\",\"reason\":\"no such index [statusCode]\",\"resource.type\":\"index_or_alias\",\"resource.id\":\"statusCode\",\"index_uuid\":\"_na_\",\"index\":\"statusCode\"}],\"type\":\"index_not_found_exception\",\"reason\":\"no such index [statusCode]\",\"resource.type\":\"index_or_alias\",\"resource.id\":\"statusCode\",\"index_uuid\":\"_na_\",\"index\":\"statusCode\"},\"status\":404}"} at respond (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-ciGroup6/node/immutable/kibana/node_modules/elasticsearch/src/lib/transport.js:308:15) at checkRespForFailure (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-ciGroup6/node/immutable/kibana/node_modules/elasticsearch/src/lib/transport.js:267:7) at HttpConnector.<anonymous> (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-ciGroup6/node/immutable/kibana/node_modules/elasticsearch/src/lib/connectors/http.js:166:7) at IncomingMessage.wrapper (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-ciGroup6/node/immutable/kibana/node_modules/elasticsearch/node_modules/lodash/lodash.js:4935:19) at endReadableNT (_stream_readable.js:1103:12) at process._tickCallback (internal/process/next_tick.js:63:19) status: 404, displayName: 'NotFound', message: '[index_not_found_exception] no such index [statusCode], with { resource.type="index_or_alias" & resource.id="statusCode" & index_uuid="_na_" & index="statusCode" }', path: '/statusCode%2Cerror%2Cmessage%2Ccause', query: {}, body: { error: { root_cause: [Array], type: 'index_not_found_exception', reason: 'no such index [statusCode]', 'resource.type': 'index_or_alias', 'resource.id': 'statusCode', index_uuid: '_na_', index: 'statusCode' }, status: 404 }, statusCode: 404, response: '{"error":{"root_cause":[{"type":"index_not_found_exception","reason":"no such index [statusCode]","resource.type":"index_or_alias","resource.id":"statusCode","index_uuid":"_na_","index":"statusCode"}],"type":"index_not_found_exception","reason":"no such index [statusCode]","resource.type":"index_or_alias","resource.id":"statusCode","index_uuid":"_na_","index":"statusCode"},"status":404}', toString: [Function], toJSON: [Function] } ``` Standard Output ``` [00:00:00] │ [00:00:00] └-: apis [00:00:00] └-> "before all" hook [00:02:41] └-: management [00:02:41] └-> "before all" hook [00:03:14] └-: rollup [00:03:14] └-> "before all" hook [00:03:14] └-: jobs [00:03:14] └-> "before all" hook [00:03:16] └-> "after all" hook [00:03:16] │ info [o.e.c.m.MetaDataDeleteIndexService] [kibana-ci-immutable-ubuntu-1553798051849348932] [qzzxalv-1553799729545/t4apKqyVS72K7p-AERzZSA] deleting index [00:03:16] │ info [o.e.c.m.MetaDataDeleteIndexService] [kibana-ci-immutable-ubuntu-1553798051849348932] [igujhehntqg-1553799729687/sDwIJWVlQnytpygaM2VX7Q] deleting index [00:03:16] └- ✖ fail: "apis management rollup jobs "after all" hook" [00:03:16] │ [index_not_found_exception] no such index [statusCode], with { resource.type="index_or_alias" & resource.id="statusCode" & index_uuid="_na_" & index="statusCode" } [00:03:16] │ :: {"path":"/statusCode%2Cerror%2Cmessage%2Ccause","query":{},"statusCode":404,"response":"{\"error\":{\"root_cause\":[{\"type\":\"index_not_found_exception\",\"reason\":\"no such index [statusCode]\",\"resource.type\":\"index_or_alias\",\"resource.id\":\"statusCode\",\"index_uuid\":\"_na_\",\"index\":\"statusCode\"}],\"type\":\"index_not_found_exception\",\"reason\":\"no such index [statusCode]\",\"resource.type\":\"index_or_alias\",\"resource.id\":\"statusCode\",\"index_uuid\":\"_na_\",\"index\":\"statusCode\"},\"status\":404}"} [00:03:16] │ at respond (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-ciGroup6/node/immutable/kibana/node_modules/elasticsearch/src/lib/transport.js:308:15) [00:03:16] │ at checkRespForFailure (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-ciGroup6/node/immutable/kibana/node_modules/elasticsearch/src/lib/transport.js:267:7) [00:03:16] │ at HttpConnector.<anonymous> (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-ciGroup6/node/immutable/kibana/node_modules/elasticsearch/src/lib/connectors/http.js:166:7) [00:03:16] │ at IncomingMessage.wrapper (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-ciGroup6/node/immutable/kibana/node_modules/elasticsearch/node_modules/lodash/lodash.js:4935:19) [00:03:16] │ at endReadableNT (_stream_readable.js:1103:12) [00:03:16] │ at process._tickCallback (internal/process/next_tick.js:63:19) [00:03:16] │ [00:03:16] │ ``` <details> <summary>Previous failure</summary> A test failed on a tracked branch ``` [Error: the object { "_events": {} "_eventsCount": 0 "_maxListeners": [undefined] "accepted": false "badRequest": true "body": { "error": "Bad Request" "message": "Could not delete job [nbsxrhzvoq-1552629059154] because indexer state is [ABORTING]. Job must be [STOPPED] before deletion." "statusCode": 400 } "buffered": true "charset": "utf-8" "clientError": true "error": "Error: cannot POST /api/rollup/delete (400)" "files": [undefined] "forbidden": false "header": { "cache-control": "no-cache" "connection": "close" "content-length": "177" "content-type": "application/json; charset=utf-8" "date": "Fri, 15 Mar 2019 05:51:00 GMT" "kbn-name": "kibana" "kbn-xpack-sig": "a422b35703abc517c92812429d155891" } "headers": { "cache-control": "no-cache" "connection": "close" "content-length": "177" "content-type": "application/json; charset=utf-8" "date": "Fri, 15 Mar 2019 05:51:00 GMT" "kbn-name": "kibana" "kbn-xpack-sig": "a422b35703abc517c92812429d155891" } "info": false "links": {} "noContent": false "notAcceptable": false "notFound": false "ok": false "redirect": false "redirects": [] "req": { "_contentLength": 301 "_ended": true "_events": { "drain": { "listener": [Function] } "error": { "listener": [Function] } "prefinish": [Function] } "_eventsCount": 3 "_hasBody": true "_header": "POST /api/rollup/delete HTTP/1.1\r\nHost: localhost:5620\r\nAccept-Encoding: gzip, deflate\r\nUser-Agent: node-superagent/3.8.2\r\nkbn-xsrf: xxx\r\nContent-Type: application/json\r\nAuthorization: Basic ZWxhc3RpYzpjaGFuZ2VtZQ==\r\nContent-Length: 301\r\nConnection: close\r\n\r\n" "_headerSent": true "_last": true "_maxListeners": [undefined] "_onPendingData": [Function] "_removedConnection": false "_removedContLen": false "_removedTE": false "_trailer": "" "aborted": [undefined] "agent": { "_events": { "free": [Function] } "_eventsCount": 1 "_maxListeners": [undefined] "defaultPort": 80 "freeSockets": {} "keepAlive": false "keepAliveMsecs": 1000 "maxFreeSockets": 256 "maxSockets": Infinity "options": { "path": [null] } "protocol": "http:" "requests": {} "sockets": { "localhost:5620:": [ { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } ] } } "chunkedEncoding": false "connection": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "finished": true "maxHeadersCount": [null] "method": "POST" "output": [] "outputCallbacks": [] "outputEncodings": [] "outputSize": 0 "parser": [null] "path": "/api/rollup/delete" "res": { "_consuming": false "_dumped": false "_events": { "close": [Function] "data": [ [Function] [Function] [Function] ] "end": [ [Function] [Function] [Function] ] "error": [ { "listener": [Function] } [Function] ] } "_eventsCount": 4 "_maxListeners": [undefined] "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": { "encoding": "utf8" } "defaultEncoding": "utf8" "destroyed": false "emitClose": true "emittedReadable": false "encoding": "utf8" "endEmitted": true "ended": true "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": false "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": false "readingMore": true "resumeScheduled": false "sync": true } "aborted": false "client": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "complete": true "connection": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "headers": { "cache-control": "no-cache" "connection": "close" "content-length": "177" "content-type": "application/json; charset=utf-8" "date": "Fri, 15 Mar 2019 05:51:00 GMT" "kbn-name": "kibana" "kbn-xpack-sig": "a422b35703abc517c92812429d155891" } "httpVersion": "1.1" "httpVersionMajor": 1 "httpVersionMinor": 1 "method": [null] "rawHeaders": [ "kbn-name" "kibana" "kbn-xpack-sig" "a422b35703abc517c92812429d155891" "content-type" "application/json; charset=utf-8" "cache-control" "no-cache" "content-length" "177" "connection" "close" "Date" "Fri, 15 Mar 2019 05:51:00 GMT" ] "rawTrailers": [] "readable": false "req": [Circular] "socket": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "statusCode": 400 "statusMessage": "Bad Request" "text": "{\"statusCode\":400,\"error\":\"Bad Request\",\"message\":\"Could not delete job [nbsxrhzvoq-1552629059154] because indexer state is [ABORTING]. Job must be [STOPPED] before deletion.\"}" "trailers": {} "upgrade": false "url": "" } "sendDate": false "shouldKeepAlive": false "socket": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "socketPath": [undefined] "timeout": [undefined] "timeoutCb": [null] "upgradeOrConnect": false "useChunkedEncodingByDefault": true "writable": true } "request": { "_agent": false "_asserts": [] "_buffer": true "_callback": [Function] "_data": { "jobIds": [ "wwcnompptsjrbm-1552629058926" "nbsxrhzvoq-1552629059154" "nbsxrhzvoq-1552629059154" "abc" "ovcwg-1552629059483" "xppabncglokc-1552629059650" "dwumwtgaf-1552629059812" "oxtnysgbuh-1552629060001" "fcctx-1552629060204" "zunqhjjamgdoieylchvt-1552629060397" "arbhsuiiggqwytxmhu-1552629060620" ] } "_endCalled": true "_events": {} "_eventsCount": 0 "_formData": [null] "_header": { "authorization": "Basic ZWxhc3RpYzpjaGFuZ2VtZQ==" "content-type": "application/json" "kbn-xsrf": "xxx" "user-agent": "node-superagent/3.8.2" } "_maxListeners": [undefined] "_maxRedirects": 0 "_query": [] "_redirectList": [] "_redirects": 0 "_streamRequest": false "app": "http://elastic:changeme@localhost:5620" "called": true "catch": [Function] "cookies": "" "header": { "Authorization": "Basic ZWxhc3RpYzpjaGFuZ2VtZQ==" "Content-Type": "application/json" "User-Agent": "node-superagent/3.8.2" "kbn-xsrf": "xxx" } "host": "localhost:5620" "method": "POST" "protocol": "http:" "qs": {} "qsRaw": [] "req": { "_contentLength": 301 "_ended": true "_events": { "drain": { "listener": [Function] } "error": { "listener": [Function] } "prefinish": [Function] } "_eventsCount": 3 "_hasBody": true "_header": "POST /api/rollup/delete HTTP/1.1\r\nHost: localhost:5620\r\nAccept-Encoding: gzip, deflate\r\nUser-Agent: node-superagent/3.8.2\r\nkbn-xsrf: xxx\r\nContent-Type: application/json\r\nAuthorization: Basic ZWxhc3RpYzpjaGFuZ2VtZQ==\r\nContent-Length: 301\r\nConnection: close\r\n\r\n" "_headerSent": true "_last": true "_maxListeners": [undefined] "_onPendingData": [Function] "_removedConnection": false "_removedContLen": false "_removedTE": false "_trailer": "" "aborted": [undefined] "agent": { "_events": { "free": [Function] } "_eventsCount": 1 "_maxListeners": [undefined] "defaultPort": 80 "freeSockets": {} "keepAlive": false "keepAliveMsecs": 1000 "maxFreeSockets": 256 "maxSockets": Infinity "options": { "path": [null] } "protocol": "http:" "requests": {} "sockets": { "localhost:5620:": [ { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } ] } } "chunkedEncoding": false "connection": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "finished": true "maxHeadersCount": [null] "method": "POST" "output": [] "outputCallbacks": [] "outputEncodings": [] "outputSize": 0 "parser": [null] "path": "/api/rollup/delete" "res": { "_consuming": false "_dumped": false "_events": { "close": [Function] "data": [ [Function] [Function] [Function] ] "end": [ [Function] [Function] [Function] ] "error": [ { "listener": [Function] } [Function] ] } "_eventsCount": 4 "_maxListeners": [undefined] "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": { "encoding": "utf8" } "defaultEncoding": "utf8" "destroyed": false "emitClose": true "emittedReadable": false "encoding": "utf8" "endEmitted": true "ended": true "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": false "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": false "readingMore": true "resumeScheduled": false "sync": true } "aborted": false "client": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "complete": true "connection": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "headers": { "cache-control": "no-cache" "connection": "close" "content-length": "177" "content-type": "application/json; charset=utf-8" "date": "Fri, 15 Mar 2019 05:51:00 GMT" "kbn-name": "kibana" "kbn-xpack-sig": "a422b35703abc517c92812429d155891" } "httpVersion": "1.1" "httpVersionMajor": 1 "httpVersionMinor": 1 "method": [null] "rawHeaders": [ "kbn-name" "kibana" "kbn-xpack-sig" "a422b35703abc517c92812429d155891" "content-type" "application/json; charset=utf-8" "cache-control" "no-cache" "content-length" "177" "connection" "close" "Date" "Fri, 15 Mar 2019 05:51:00 GMT" ] "rawTrailers": [] "readable": false "req": [Circular] "socket": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "statusCode": 400 "statusMessage": "Bad Request" "text": "{\"statusCode\":400,\"error\":\"Bad Request\",\"message\":\"Could not delete job [nbsxrhzvoq-1552629059154] because indexer state is [ABORTING]. Job must be [STOPPED] before deletion.\"}" "trailers": {} "upgrade": false "url": "" } "sendDate": false "shouldKeepAlive": false "socket": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "socketPath": [undefined] "timeout": [undefined] "timeoutCb": [null] "upgradeOrConnect": false "useChunkedEncodingByDefault": true "writable": true } "res": { "_consuming": false "_dumped": false "_events": { "close": [Function] "data": [ [Function] [Function] [Function] ] "end": [ [Function] [Function] [Function] ] "error": [ { "listener": [Function] } [Function] ] } "_eventsCount": 4 "_maxListeners": [undefined] "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": { "encoding": "utf8" } "defaultEncoding": "utf8" "destroyed": false "emitClose": true "emittedReadable": false "encoding": "utf8" "endEmitted": true "ended": true "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": false "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": false "readingMore": true "resumeScheduled": false "sync": true } "aborted": false "client": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": { "_contentLength": 301 "_ended": true "_events": { "drain": { "listener": [Function] } "error": { "listener": [Function] } "prefinish": [Function] } "_eventsCount": 3 "_hasBody": true "_header": "POST /api/rollup/delete HTTP/1.1\r\nHost: localhost:5620\r\nAccept-Encoding: gzip, deflate\r\nUser-Agent: node-superagent/3.8.2\r\nkbn-xsrf: xxx\r\nContent-Type: application/json\r\nAuthorization: Basic ZWxhc3RpYzpjaGFuZ2VtZQ==\r\nContent-Length: 301\r\nConnection: close\r\n\r\n" "_headerSent": true "_last": true "_maxListeners": [undefined] "_onPendingData": [Function] "_removedConnection": false "_removedContLen": false "_removedTE": false "_trailer": "" "aborted": [undefined] "agent": { "_events": { "free": [Function] } "_eventsCount": 1 "_maxListeners": [undefined] "defaultPort": 80 "freeSockets": {} "keepAlive": false "keepAliveMsecs": 1000 "maxFreeSockets": 256 "maxSockets": Infinity "options": { "path": [null] } "protocol": "http:" "requests": {} "sockets": { "localhost:5620:": [ [Circular] ] } } "chunkedEncoding": false "connection": [Circular] "finished": true "maxHeadersCount": [null] "method": "POST" "output": [] "outputCallbacks": [] "outputEncodings": [] "outputSize": 0 "parser": [null] "path": "/api/rollup/delete" "res": [Circular] "sendDate": false "shouldKeepAlive": false "socket": [Circular] "socketPath": [undefined] "timeout": [undefined] "timeoutCb": [null] "upgradeOrConnect": false "useChunkedEncodingByDefault": true "writable": true } "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "complete": true "connection": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": { "_contentLength": 301 "_ended": true "_events": { "drain": { "listener": [Function] } "error": { "listener": [Function] } "prefinish": [Function] } "_eventsCount": 3 "_hasBody": true "_header": "POST /api/rollup/delete HTTP/1.1\r\nHost: localhost:5620\r\nAccept-Encoding: gzip, deflate\r\nUser-Agent: node-superagent/3.8.2\r\nkbn-xsrf: xxx\r\nContent-Type: application/json\r\nAuthorization: Basic ZWxhc3RpYzpjaGFuZ2VtZQ==\r\nContent-Length: 301\r\nConnection: close\r\n\r\n" "_headerSent": true "_last": true "_maxListeners": [undefined] "_onPendingData": [Function] "_removedConnection": false "_removedContLen": false "_removedTE": false "_trailer": "" "aborted": [undefined] "agent": { "_events": { "free": [Function] } "_eventsCount": 1 "_maxListeners": [undefined] "defaultPort": 80 "freeSockets": {} "keepAlive": false "keepAliveMsecs": 1000 "maxFreeSockets": 256 "maxSockets": Infinity "options": { "path": [null] } "protocol": "http:" "requests": {} "sockets": { "localhost:5620:": [ [Circular] ] } } "chunkedEncoding": false "connection": [Circular] "finished": true "maxHeadersCount": [null] "method": "POST" "output": [] "outputCallbacks": [] "outputEncodings": [] "outputSize": 0 "parser": [null] "path": "/api/rollup/delete" "res": [Circular] "sendDate": false "shouldKeepAlive": false "socket": [Circular] "socketPath": [undefined] "timeout": [undefined] "timeoutCb": [null] "upgradeOrConnect": false "useChunkedEncodingByDefault": true "writable": true } "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "headers": { "cache-control": "no-cache" "connection": "close" "content-length": "177" "content-type": "application/json; charset=utf-8" "date": "Fri, 15 Mar 2019 05:51:00 GMT" "kbn-name": "kibana" "kbn-xpack-sig": "a422b35703abc517c92812429d155891" } "httpVersion": "1.1" "httpVersionMajor": 1 "httpVersionMinor": 1 "method": [null] "rawHeaders": [ "kbn-name" "kibana" "kbn-xpack-sig" "a422b35703abc517c92812429d155891" "content-type" "application/json; charset=utf-8" "cache-control" "no-cache" "content-length" "177" "connection" "close" "Date" "Fri, 15 Mar 2019 05:51:00 GMT" ] "rawTrailers": [] "readable": false "req": { "_contentLength": 301 "_ended": true "_events": { "drain": { "listener": [Function] } "error": { "listener": [Function] } "prefinish": [Function] } "_eventsCount": 3 "_hasBody": true "_header": "POST /api/rollup/delete HTTP/1.1\r\nHost: localhost:5620\r\nAccept-Encoding: gzip, deflate\r\nUser-Agent: node-superagent/3.8.2\r\nkbn-xsrf: xxx\r\nContent-Type: application/json\r\nAuthorization: Basic ZWxhc3RpYzpjaGFuZ2VtZQ==\r\nContent-Length: 301\r\nConnection: close\r\n\r\n" "_headerSent": true "_last": true "_maxListeners": [undefined] "_onPendingData": [Function] "_removedConnection": false "_removedContLen": false "_removedTE": false "_trailer": "" "aborted": [undefined] "agent": { "_events": { "free": [Function] } "_eventsCount": 1 "_maxListeners": [undefined] "defaultPort": 80 "freeSockets": {} "keepAlive": false "keepAliveMsecs": 1000 "maxFreeSockets": 256 "maxSockets": Infinity "options": { "path": [null] } "protocol": "http:" "requests": {} "sockets": { "localhost:5620:": [ { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } ] } } "chunkedEncoding": false "connection": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "finished": true "maxHeadersCount": [null] "method": "POST" "output": [] "outputCallbacks": [] "outputEncodings": [] "outputSize": 0 "parser": [null] "path": "/api/rollup/delete" "res": [Circular] "sendDate": false "shouldKeepAlive": false "socket": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "socketPath": [undefined] "timeout": [undefined] "timeoutCb": [null] "upgradeOrConnect": false "useChunkedEncodingByDefault": true "writable": true } "socket": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": { "_contentLength": 301 "_ended": true "_events": { "drain": { "listener": [Function] } "error": { "listener": [Function] } "prefinish": [Function] } "_eventsCount": 3 "_hasBody": true "_header": "POST /api/rollup/delete HTTP/1.1\r\nHost: localhost:5620\r\nAccept-Encoding: gzip, deflate\r\nUser-Agent: node-superagent/3.8.2\r\nkbn-xsrf: xxx\r\nContent-Type: application/json\r\nAuthorization: Basic ZWxhc3RpYzpjaGFuZ2VtZQ==\r\nContent-Length: 301\r\nConnection: close\r\n\r\n" "_headerSent": true "_last": true "_maxListeners": [undefined] "_onPendingData": [Function] "_removedConnection": false "_removedContLen": false "_removedTE": false "_trailer": "" "aborted": [undefined] "agent": { "_events": { "free": [Function] } "_eventsCount": 1 "_maxListeners": [undefined] "defaultPort": 80 "freeSockets": {} "keepAlive": false "keepAliveMsecs": 1000 "maxFreeSockets": 256 "maxSockets": Infinity "options": { "path": [null] } "protocol": "http:" "requests": {} "sockets": { "localhost:5620:": [ [Circular] ] } } "chunkedEncoding": false "connection": [Circular] "finished": true "maxHeadersCount": [null] "method": "POST" "output": [] "outputCallbacks": [] "outputEncodings": [] "outputSize": 0 "parser": [null] "path": "/api/rollup/delete" "res": [Circular] "sendDate": false "shouldKeepAlive": false "socket": [Circular] "socketPath": [undefined] "timeout": [undefined] "timeoutCb": [null] "upgradeOrConnect": false "useChunkedEncodingByDefault": true "writable": true } "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "statusCode": 400 "statusMessage": "Bad Request" "text": "{\"statusCode\":400,\"error\":\"Bad Request\",\"message\":\"Could not delete job [nbsxrhzvoq-1552629059154] because indexer state is [ABORTING]. Job must be [STOPPED] before deletion.\"}" "trailers": {} "upgrade": false "url": "" } "response": [Circular] "then": [Function] "toPromise": [Function] "url": "http://elastic:changeme@localhost:5620/api/rollup/delete" "writable": true } "res": { "_consuming": false "_dumped": false "_events": { "close": [Function] "data": [ [Function] [Function] [Function] ] "end": [ [Function] [Function] [Function] ] "error": [ { "listener": [Function] } [Function] ] } "_eventsCount": 4 "_maxListeners": [undefined] "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": { "encoding": "utf8" } "defaultEncoding": "utf8" "destroyed": false "emitClose": true "emittedReadable": false "encoding": "utf8" "endEmitted": true "ended": true "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": false "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": false "readingMore": true "resumeScheduled": false "sync": true } "aborted": false "client": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": { "_contentLength": 301 "_ended": true "_events": { "drain": { "listener": [Function] } "error": { "listener": [Function] } "prefinish": [Function] } "_eventsCount": 3 "_hasBody": true "_header": "POST /api/rollup/delete HTTP/1.1\r\nHost: localhost:5620\r\nAccept-Encoding: gzip, deflate\r\nUser-Agent: node-superagent/3.8.2\r\nkbn-xsrf: xxx\r\nContent-Type: application/json\r\nAuthorization: Basic ZWxhc3RpYzpjaGFuZ2VtZQ==\r\nContent-Length: 301\r\nConnection: close\r\n\r\n" "_headerSent": true "_last": true "_maxListeners": [undefined] "_onPendingData": [Function] "_removedConnection": false "_removedContLen": false "_removedTE": false "_trailer": "" "aborted": [undefined] "agent": { "_events": { "free": [Function] } "_eventsCount": 1 "_maxListeners": [undefined] "defaultPort": 80 "freeSockets": {} "keepAlive": false "keepAliveMsecs": 1000 "maxFreeSockets": 256 "maxSockets": Infinity "options": { "path": [null] } "protocol": "http:" "requests": {} "sockets": { "localhost:5620:": [ [Circular] ] } } "chunkedEncoding": false "connection": [Circular] "finished": true "maxHeadersCount": [null] "method": "POST" "output": [] "outputCallbacks": [] "outputEncodings": [] "outputSize": 0 "parser": [null] "path": "/api/rollup/delete" "res": [Circular] "sendDate": false "shouldKeepAlive": false "socket": [Circular] "socketPath": [undefined] "timeout": [undefined] "timeoutCb": [null] "upgradeOrConnect": false "useChunkedEncodingByDefault": true "writable": true } "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "complete": true "connection": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": { "_contentLength": 301 "_ended": true "_events": { "drain": { "listener": [Function] } "error": { "listener": [Function] } "prefinish": [Function] } "_eventsCount": 3 "_hasBody": true "_header": "POST /api/rollup/delete HTTP/1.1\r\nHost: localhost:5620\r\nAccept-Encoding: gzip, deflate\r\nUser-Agent: node-superagent/3.8.2\r\nkbn-xsrf: xxx\r\nContent-Type: application/json\r\nAuthorization: Basic ZWxhc3RpYzpjaGFuZ2VtZQ==\r\nContent-Length: 301\r\nConnection: close\r\n\r\n" "_headerSent": true "_last": true "_maxListeners": [undefined] "_onPendingData": [Function] "_removedConnection": false "_removedContLen": false "_removedTE": false "_trailer": "" "aborted": [undefined] "agent": { "_events": { "free": [Function] } "_eventsCount": 1 "_maxListeners": [undefined] "defaultPort": 80 "freeSockets": {} "keepAlive": false "keepAliveMsecs": 1000 "maxFreeSockets": 256 "maxSockets": Infinity "options": { "path": [null] } "protocol": "http:" "requests": {} "sockets": { "localhost:5620:": [ [Circular] ] } } "chunkedEncoding": false "connection": [Circular] "finished": true "maxHeadersCount": [null] "method": "POST" "output": [] "outputCallbacks": [] "outputEncodings": [] "outputSize": 0 "parser": [null] "path": "/api/rollup/delete" "res": [Circular] "sendDate": false "shouldKeepAlive": false "socket": [Circular] "socketPath": [undefined] "timeout": [undefined] "timeoutCb": [null] "upgradeOrConnect": false "useChunkedEncodingByDefault": true "writable": true } "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "headers": { "cache-control": "no-cache" "connection": "close" "content-length": "177" "content-type": "application/json; charset=utf-8" "date": "Fri, 15 Mar 2019 05:51:00 GMT" "kbn-name": "kibana" "kbn-xpack-sig": "a422b35703abc517c92812429d155891" } "httpVersion": "1.1" "httpVersionMajor": 1 "httpVersionMinor": 1 "method": [null] "rawHeaders": [ "kbn-name" "kibana" "kbn-xpack-sig" "a422b35703abc517c92812429d155891" "content-type" "application/json; charset=utf-8" "cache-control" "no-cache" "content-length" "177" "connection" "close" "Date" "Fri, 15 Mar 2019 05:51:00 GMT" ] "rawTrailers": [] "readable": false "req": { "_contentLength": 301 "_ended": true "_events": { "drain": { "listener": [Function] } "error": { "listener": [Function] } "prefinish": [Function] } "_eventsCount": 3 "_hasBody": true "_header": "POST /api/rollup/delete HTTP/1.1\r\nHost: localhost:5620\r\nAccept-Encoding: gzip, deflate\r\nUser-Agent: node-superagent/3.8.2\r\nkbn-xsrf: xxx\r\nContent-Type: application/json\r\nAuthorization: Basic ZWxhc3RpYzpjaGFuZ2VtZQ==\r\nContent-Length: 301\r\nConnection: close\r\n\r\n" "_headerSent": true "_last": true "_maxListeners": [undefined] "_onPendingData": [Function] "_removedConnection": false "_removedContLen": false "_removedTE": false "_trailer": "" "aborted": [undefined] "agent": { "_events": { "free": [Function] } "_eventsCount": 1 "_maxListeners": [undefined] "defaultPort": 80 "freeSockets": {} "keepAlive": false "keepAliveMsecs": 1000 "maxFreeSockets": 256 "maxSockets": Infinity "options": { "path": [null] } "protocol": "http:" "requests": {} "sockets": { "localhost:5620:": [ { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } ] } } "chunkedEncoding": false "connection": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "finished": true "maxHeadersCount": [null] "method": "POST" "output": [] "outputCallbacks": [] "outputEncodings": [] "outputSize": 0 "parser": [null] "path": "/api/rollup/delete" "res": [Circular] "sendDate": false "shouldKeepAlive": false "socket": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "socketPath": [undefined] "timeout": [undefined] "timeoutCb": [null] "upgradeOrConnect": false "useChunkedEncodingByDefault": true "writable": true } "socket": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": { "_contentLength": 301 "_ended": true "_events": { "drain": { "listener": [Function] } "error": { "listener": [Function] } "prefinish": [Function] } "_eventsCount": 3 "_hasBody": true "_header": "POST /api/rollup/delete HTTP/1.1\r\nHost: localhost:5620\r\nAccept-Encoding: gzip, deflate\r\nUser-Agent: node-superagent/3.8.2\r\nkbn-xsrf: xxx\r\nContent-Type: application/json\r\nAuthorization: Basic ZWxhc3RpYzpjaGFuZ2VtZQ==\r\nContent-Length: 301\r\nConnection: close\r\n\r\n" "_headerSent": true "_last": true "_maxListeners": [undefined] "_onPendingData": [Function] "_removedConnection": false "_removedContLen": false "_removedTE": false "_trailer": "" "aborted": [undefined] "agent": { "_events": { "free": [Function] } "_eventsCount": 1 "_maxListeners": [undefined] "defaultPort": 80 "freeSockets": {} "keepAlive": false "keepAliveMsecs": 1000 "maxFreeSockets": 256 "maxSockets": Infinity "options": { "path": [null] } "protocol": "http:" "requests": {} "sockets": { "localhost:5620:": [ [Circular] ] } } "chunkedEncoding": false "connection": [Circular] "finished": true "maxHeadersCount": [null] "method": "POST" "output": [] "outputCallbacks": [] "outputEncodings": [] "outputSize": 0 "parser": [null] "path": "/api/rollup/delete" "res": [Circular] "sendDate": false "shouldKeepAlive": false "socket": [Circular] "socketPath": [undefined] "timeout": [undefined] "timeoutCb": [null] "upgradeOrConnect": false "useChunkedEncodingByDefault": true "writable": true } "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "statusCode": 400 "statusMessage": "Bad Request" "text": "{\"statusCode\":400,\"error\":\"Bad Request\",\"message\":\"Could not delete job [nbsxrhzvoq-1552629059154] because indexer state is [ABORTING]. Job must be [STOPPED] before deletion.\"}" "trailers": {} "upgrade": false "url": "" } "serverError": false "setEncoding": [Function] "status": 400 "statusCode": 400 "statusType": 4 "text": "{\"statusCode\":400,\"error\":\"Bad Request\",\"message\":\"Could not delete job [nbsxrhzvoq-1552629059154] because indexer state is [ABORTING]. Job must be [STOPPED] before deletion.\"}" "type": "application/json" "unauthorized": false } was thrown, throw an Error :)] ``` </details> First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+7.x/JOB=x-pack-ciGroup5,node=immutable/466/) <!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack API Integration Tests.x-pack/test/api_integration/apis/management/rollup/rollup·js","test.name":"apis management rollup jobs \"after all\" hook","test.failCount":3}} -->
1.0
Failing test: X-Pack API Integration Tests.x-pack/test/api_integration/apis/management/rollup/rollup·js - apis management rollup jobs "after all" hook - Latest failure: https://kibana-ci.elastic.co/job/elastic+kibana+master/906/JOB=x-pack-ciGroup6,node=immutable/testReport/junit/X-Pack%20API%20Integration%20Tests/x-pack_test_api_integration_apis_management_rollup_rollup%C2%B7js/apis_management_rollup_jobs__after_all__hook/ Stacktrace ``` { [index_not_found_exception] no such index [statusCode], with { resource.type="index_or_alias" & resource.id="statusCode" & index_uuid="_na_" & index="statusCode" } :: {"path":"/statusCode%2Cerror%2Cmessage%2Ccause","query":{},"statusCode":404,"response":"{\"error\":{\"root_cause\":[{\"type\":\"index_not_found_exception\",\"reason\":\"no such index [statusCode]\",\"resource.type\":\"index_or_alias\",\"resource.id\":\"statusCode\",\"index_uuid\":\"_na_\",\"index\":\"statusCode\"}],\"type\":\"index_not_found_exception\",\"reason\":\"no such index [statusCode]\",\"resource.type\":\"index_or_alias\",\"resource.id\":\"statusCode\",\"index_uuid\":\"_na_\",\"index\":\"statusCode\"},\"status\":404}"} at respond (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-ciGroup6/node/immutable/kibana/node_modules/elasticsearch/src/lib/transport.js:308:15) at checkRespForFailure (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-ciGroup6/node/immutable/kibana/node_modules/elasticsearch/src/lib/transport.js:267:7) at HttpConnector.<anonymous> (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-ciGroup6/node/immutable/kibana/node_modules/elasticsearch/src/lib/connectors/http.js:166:7) at IncomingMessage.wrapper (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-ciGroup6/node/immutable/kibana/node_modules/elasticsearch/node_modules/lodash/lodash.js:4935:19) at endReadableNT (_stream_readable.js:1103:12) at process._tickCallback (internal/process/next_tick.js:63:19) status: 404, displayName: 'NotFound', message: '[index_not_found_exception] no such index [statusCode], with { resource.type="index_or_alias" & resource.id="statusCode" & index_uuid="_na_" & index="statusCode" }', path: '/statusCode%2Cerror%2Cmessage%2Ccause', query: {}, body: { error: { root_cause: [Array], type: 'index_not_found_exception', reason: 'no such index [statusCode]', 'resource.type': 'index_or_alias', 'resource.id': 'statusCode', index_uuid: '_na_', index: 'statusCode' }, status: 404 }, statusCode: 404, response: '{"error":{"root_cause":[{"type":"index_not_found_exception","reason":"no such index [statusCode]","resource.type":"index_or_alias","resource.id":"statusCode","index_uuid":"_na_","index":"statusCode"}],"type":"index_not_found_exception","reason":"no such index [statusCode]","resource.type":"index_or_alias","resource.id":"statusCode","index_uuid":"_na_","index":"statusCode"},"status":404}', toString: [Function], toJSON: [Function] } ``` Standard Output ``` [00:00:00] │ [00:00:00] └-: apis [00:00:00] └-> "before all" hook [00:02:41] └-: management [00:02:41] └-> "before all" hook [00:03:14] └-: rollup [00:03:14] └-> "before all" hook [00:03:14] └-: jobs [00:03:14] └-> "before all" hook [00:03:16] └-> "after all" hook [00:03:16] │ info [o.e.c.m.MetaDataDeleteIndexService] [kibana-ci-immutable-ubuntu-1553798051849348932] [qzzxalv-1553799729545/t4apKqyVS72K7p-AERzZSA] deleting index [00:03:16] │ info [o.e.c.m.MetaDataDeleteIndexService] [kibana-ci-immutable-ubuntu-1553798051849348932] [igujhehntqg-1553799729687/sDwIJWVlQnytpygaM2VX7Q] deleting index [00:03:16] └- ✖ fail: "apis management rollup jobs "after all" hook" [00:03:16] │ [index_not_found_exception] no such index [statusCode], with { resource.type="index_or_alias" & resource.id="statusCode" & index_uuid="_na_" & index="statusCode" } [00:03:16] │ :: {"path":"/statusCode%2Cerror%2Cmessage%2Ccause","query":{},"statusCode":404,"response":"{\"error\":{\"root_cause\":[{\"type\":\"index_not_found_exception\",\"reason\":\"no such index [statusCode]\",\"resource.type\":\"index_or_alias\",\"resource.id\":\"statusCode\",\"index_uuid\":\"_na_\",\"index\":\"statusCode\"}],\"type\":\"index_not_found_exception\",\"reason\":\"no such index [statusCode]\",\"resource.type\":\"index_or_alias\",\"resource.id\":\"statusCode\",\"index_uuid\":\"_na_\",\"index\":\"statusCode\"},\"status\":404}"} [00:03:16] │ at respond (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-ciGroup6/node/immutable/kibana/node_modules/elasticsearch/src/lib/transport.js:308:15) [00:03:16] │ at checkRespForFailure (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-ciGroup6/node/immutable/kibana/node_modules/elasticsearch/src/lib/transport.js:267:7) [00:03:16] │ at HttpConnector.<anonymous> (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-ciGroup6/node/immutable/kibana/node_modules/elasticsearch/src/lib/connectors/http.js:166:7) [00:03:16] │ at IncomingMessage.wrapper (/var/lib/jenkins/workspace/elastic+kibana+master/JOB/x-pack-ciGroup6/node/immutable/kibana/node_modules/elasticsearch/node_modules/lodash/lodash.js:4935:19) [00:03:16] │ at endReadableNT (_stream_readable.js:1103:12) [00:03:16] │ at process._tickCallback (internal/process/next_tick.js:63:19) [00:03:16] │ [00:03:16] │ ``` <details> <summary>Previous failure</summary> A test failed on a tracked branch ``` [Error: the object { "_events": {} "_eventsCount": 0 "_maxListeners": [undefined] "accepted": false "badRequest": true "body": { "error": "Bad Request" "message": "Could not delete job [nbsxrhzvoq-1552629059154] because indexer state is [ABORTING]. Job must be [STOPPED] before deletion." "statusCode": 400 } "buffered": true "charset": "utf-8" "clientError": true "error": "Error: cannot POST /api/rollup/delete (400)" "files": [undefined] "forbidden": false "header": { "cache-control": "no-cache" "connection": "close" "content-length": "177" "content-type": "application/json; charset=utf-8" "date": "Fri, 15 Mar 2019 05:51:00 GMT" "kbn-name": "kibana" "kbn-xpack-sig": "a422b35703abc517c92812429d155891" } "headers": { "cache-control": "no-cache" "connection": "close" "content-length": "177" "content-type": "application/json; charset=utf-8" "date": "Fri, 15 Mar 2019 05:51:00 GMT" "kbn-name": "kibana" "kbn-xpack-sig": "a422b35703abc517c92812429d155891" } "info": false "links": {} "noContent": false "notAcceptable": false "notFound": false "ok": false "redirect": false "redirects": [] "req": { "_contentLength": 301 "_ended": true "_events": { "drain": { "listener": [Function] } "error": { "listener": [Function] } "prefinish": [Function] } "_eventsCount": 3 "_hasBody": true "_header": "POST /api/rollup/delete HTTP/1.1\r\nHost: localhost:5620\r\nAccept-Encoding: gzip, deflate\r\nUser-Agent: node-superagent/3.8.2\r\nkbn-xsrf: xxx\r\nContent-Type: application/json\r\nAuthorization: Basic ZWxhc3RpYzpjaGFuZ2VtZQ==\r\nContent-Length: 301\r\nConnection: close\r\n\r\n" "_headerSent": true "_last": true "_maxListeners": [undefined] "_onPendingData": [Function] "_removedConnection": false "_removedContLen": false "_removedTE": false "_trailer": "" "aborted": [undefined] "agent": { "_events": { "free": [Function] } "_eventsCount": 1 "_maxListeners": [undefined] "defaultPort": 80 "freeSockets": {} "keepAlive": false "keepAliveMsecs": 1000 "maxFreeSockets": 256 "maxSockets": Infinity "options": { "path": [null] } "protocol": "http:" "requests": {} "sockets": { "localhost:5620:": [ { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } ] } } "chunkedEncoding": false "connection": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "finished": true "maxHeadersCount": [null] "method": "POST" "output": [] "outputCallbacks": [] "outputEncodings": [] "outputSize": 0 "parser": [null] "path": "/api/rollup/delete" "res": { "_consuming": false "_dumped": false "_events": { "close": [Function] "data": [ [Function] [Function] [Function] ] "end": [ [Function] [Function] [Function] ] "error": [ { "listener": [Function] } [Function] ] } "_eventsCount": 4 "_maxListeners": [undefined] "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": { "encoding": "utf8" } "defaultEncoding": "utf8" "destroyed": false "emitClose": true "emittedReadable": false "encoding": "utf8" "endEmitted": true "ended": true "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": false "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": false "readingMore": true "resumeScheduled": false "sync": true } "aborted": false "client": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "complete": true "connection": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "headers": { "cache-control": "no-cache" "connection": "close" "content-length": "177" "content-type": "application/json; charset=utf-8" "date": "Fri, 15 Mar 2019 05:51:00 GMT" "kbn-name": "kibana" "kbn-xpack-sig": "a422b35703abc517c92812429d155891" } "httpVersion": "1.1" "httpVersionMajor": 1 "httpVersionMinor": 1 "method": [null] "rawHeaders": [ "kbn-name" "kibana" "kbn-xpack-sig" "a422b35703abc517c92812429d155891" "content-type" "application/json; charset=utf-8" "cache-control" "no-cache" "content-length" "177" "connection" "close" "Date" "Fri, 15 Mar 2019 05:51:00 GMT" ] "rawTrailers": [] "readable": false "req": [Circular] "socket": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "statusCode": 400 "statusMessage": "Bad Request" "text": "{\"statusCode\":400,\"error\":\"Bad Request\",\"message\":\"Could not delete job [nbsxrhzvoq-1552629059154] because indexer state is [ABORTING]. Job must be [STOPPED] before deletion.\"}" "trailers": {} "upgrade": false "url": "" } "sendDate": false "shouldKeepAlive": false "socket": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "socketPath": [undefined] "timeout": [undefined] "timeoutCb": [null] "upgradeOrConnect": false "useChunkedEncodingByDefault": true "writable": true } "request": { "_agent": false "_asserts": [] "_buffer": true "_callback": [Function] "_data": { "jobIds": [ "wwcnompptsjrbm-1552629058926" "nbsxrhzvoq-1552629059154" "nbsxrhzvoq-1552629059154" "abc" "ovcwg-1552629059483" "xppabncglokc-1552629059650" "dwumwtgaf-1552629059812" "oxtnysgbuh-1552629060001" "fcctx-1552629060204" "zunqhjjamgdoieylchvt-1552629060397" "arbhsuiiggqwytxmhu-1552629060620" ] } "_endCalled": true "_events": {} "_eventsCount": 0 "_formData": [null] "_header": { "authorization": "Basic ZWxhc3RpYzpjaGFuZ2VtZQ==" "content-type": "application/json" "kbn-xsrf": "xxx" "user-agent": "node-superagent/3.8.2" } "_maxListeners": [undefined] "_maxRedirects": 0 "_query": [] "_redirectList": [] "_redirects": 0 "_streamRequest": false "app": "http://elastic:changeme@localhost:5620" "called": true "catch": [Function] "cookies": "" "header": { "Authorization": "Basic ZWxhc3RpYzpjaGFuZ2VtZQ==" "Content-Type": "application/json" "User-Agent": "node-superagent/3.8.2" "kbn-xsrf": "xxx" } "host": "localhost:5620" "method": "POST" "protocol": "http:" "qs": {} "qsRaw": [] "req": { "_contentLength": 301 "_ended": true "_events": { "drain": { "listener": [Function] } "error": { "listener": [Function] } "prefinish": [Function] } "_eventsCount": 3 "_hasBody": true "_header": "POST /api/rollup/delete HTTP/1.1\r\nHost: localhost:5620\r\nAccept-Encoding: gzip, deflate\r\nUser-Agent: node-superagent/3.8.2\r\nkbn-xsrf: xxx\r\nContent-Type: application/json\r\nAuthorization: Basic ZWxhc3RpYzpjaGFuZ2VtZQ==\r\nContent-Length: 301\r\nConnection: close\r\n\r\n" "_headerSent": true "_last": true "_maxListeners": [undefined] "_onPendingData": [Function] "_removedConnection": false "_removedContLen": false "_removedTE": false "_trailer": "" "aborted": [undefined] "agent": { "_events": { "free": [Function] } "_eventsCount": 1 "_maxListeners": [undefined] "defaultPort": 80 "freeSockets": {} "keepAlive": false "keepAliveMsecs": 1000 "maxFreeSockets": 256 "maxSockets": Infinity "options": { "path": [null] } "protocol": "http:" "requests": {} "sockets": { "localhost:5620:": [ { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } ] } } "chunkedEncoding": false "connection": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "finished": true "maxHeadersCount": [null] "method": "POST" "output": [] "outputCallbacks": [] "outputEncodings": [] "outputSize": 0 "parser": [null] "path": "/api/rollup/delete" "res": { "_consuming": false "_dumped": false "_events": { "close": [Function] "data": [ [Function] [Function] [Function] ] "end": [ [Function] [Function] [Function] ] "error": [ { "listener": [Function] } [Function] ] } "_eventsCount": 4 "_maxListeners": [undefined] "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": { "encoding": "utf8" } "defaultEncoding": "utf8" "destroyed": false "emitClose": true "emittedReadable": false "encoding": "utf8" "endEmitted": true "ended": true "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": false "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": false "readingMore": true "resumeScheduled": false "sync": true } "aborted": false "client": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "complete": true "connection": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "headers": { "cache-control": "no-cache" "connection": "close" "content-length": "177" "content-type": "application/json; charset=utf-8" "date": "Fri, 15 Mar 2019 05:51:00 GMT" "kbn-name": "kibana" "kbn-xpack-sig": "a422b35703abc517c92812429d155891" } "httpVersion": "1.1" "httpVersionMajor": 1 "httpVersionMinor": 1 "method": [null] "rawHeaders": [ "kbn-name" "kibana" "kbn-xpack-sig" "a422b35703abc517c92812429d155891" "content-type" "application/json; charset=utf-8" "cache-control" "no-cache" "content-length" "177" "connection" "close" "Date" "Fri, 15 Mar 2019 05:51:00 GMT" ] "rawTrailers": [] "readable": false "req": [Circular] "socket": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "statusCode": 400 "statusMessage": "Bad Request" "text": "{\"statusCode\":400,\"error\":\"Bad Request\",\"message\":\"Could not delete job [nbsxrhzvoq-1552629059154] because indexer state is [ABORTING]. Job must be [STOPPED] before deletion.\"}" "trailers": {} "upgrade": false "url": "" } "sendDate": false "shouldKeepAlive": false "socket": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "socketPath": [undefined] "timeout": [undefined] "timeoutCb": [null] "upgradeOrConnect": false "useChunkedEncodingByDefault": true "writable": true } "res": { "_consuming": false "_dumped": false "_events": { "close": [Function] "data": [ [Function] [Function] [Function] ] "end": [ [Function] [Function] [Function] ] "error": [ { "listener": [Function] } [Function] ] } "_eventsCount": 4 "_maxListeners": [undefined] "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": { "encoding": "utf8" } "defaultEncoding": "utf8" "destroyed": false "emitClose": true "emittedReadable": false "encoding": "utf8" "endEmitted": true "ended": true "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": false "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": false "readingMore": true "resumeScheduled": false "sync": true } "aborted": false "client": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": { "_contentLength": 301 "_ended": true "_events": { "drain": { "listener": [Function] } "error": { "listener": [Function] } "prefinish": [Function] } "_eventsCount": 3 "_hasBody": true "_header": "POST /api/rollup/delete HTTP/1.1\r\nHost: localhost:5620\r\nAccept-Encoding: gzip, deflate\r\nUser-Agent: node-superagent/3.8.2\r\nkbn-xsrf: xxx\r\nContent-Type: application/json\r\nAuthorization: Basic ZWxhc3RpYzpjaGFuZ2VtZQ==\r\nContent-Length: 301\r\nConnection: close\r\n\r\n" "_headerSent": true "_last": true "_maxListeners": [undefined] "_onPendingData": [Function] "_removedConnection": false "_removedContLen": false "_removedTE": false "_trailer": "" "aborted": [undefined] "agent": { "_events": { "free": [Function] } "_eventsCount": 1 "_maxListeners": [undefined] "defaultPort": 80 "freeSockets": {} "keepAlive": false "keepAliveMsecs": 1000 "maxFreeSockets": 256 "maxSockets": Infinity "options": { "path": [null] } "protocol": "http:" "requests": {} "sockets": { "localhost:5620:": [ [Circular] ] } } "chunkedEncoding": false "connection": [Circular] "finished": true "maxHeadersCount": [null] "method": "POST" "output": [] "outputCallbacks": [] "outputEncodings": [] "outputSize": 0 "parser": [null] "path": "/api/rollup/delete" "res": [Circular] "sendDate": false "shouldKeepAlive": false "socket": [Circular] "socketPath": [undefined] "timeout": [undefined] "timeoutCb": [null] "upgradeOrConnect": false "useChunkedEncodingByDefault": true "writable": true } "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "complete": true "connection": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": { "_contentLength": 301 "_ended": true "_events": { "drain": { "listener": [Function] } "error": { "listener": [Function] } "prefinish": [Function] } "_eventsCount": 3 "_hasBody": true "_header": "POST /api/rollup/delete HTTP/1.1\r\nHost: localhost:5620\r\nAccept-Encoding: gzip, deflate\r\nUser-Agent: node-superagent/3.8.2\r\nkbn-xsrf: xxx\r\nContent-Type: application/json\r\nAuthorization: Basic ZWxhc3RpYzpjaGFuZ2VtZQ==\r\nContent-Length: 301\r\nConnection: close\r\n\r\n" "_headerSent": true "_last": true "_maxListeners": [undefined] "_onPendingData": [Function] "_removedConnection": false "_removedContLen": false "_removedTE": false "_trailer": "" "aborted": [undefined] "agent": { "_events": { "free": [Function] } "_eventsCount": 1 "_maxListeners": [undefined] "defaultPort": 80 "freeSockets": {} "keepAlive": false "keepAliveMsecs": 1000 "maxFreeSockets": 256 "maxSockets": Infinity "options": { "path": [null] } "protocol": "http:" "requests": {} "sockets": { "localhost:5620:": [ [Circular] ] } } "chunkedEncoding": false "connection": [Circular] "finished": true "maxHeadersCount": [null] "method": "POST" "output": [] "outputCallbacks": [] "outputEncodings": [] "outputSize": 0 "parser": [null] "path": "/api/rollup/delete" "res": [Circular] "sendDate": false "shouldKeepAlive": false "socket": [Circular] "socketPath": [undefined] "timeout": [undefined] "timeoutCb": [null] "upgradeOrConnect": false "useChunkedEncodingByDefault": true "writable": true } "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "headers": { "cache-control": "no-cache" "connection": "close" "content-length": "177" "content-type": "application/json; charset=utf-8" "date": "Fri, 15 Mar 2019 05:51:00 GMT" "kbn-name": "kibana" "kbn-xpack-sig": "a422b35703abc517c92812429d155891" } "httpVersion": "1.1" "httpVersionMajor": 1 "httpVersionMinor": 1 "method": [null] "rawHeaders": [ "kbn-name" "kibana" "kbn-xpack-sig" "a422b35703abc517c92812429d155891" "content-type" "application/json; charset=utf-8" "cache-control" "no-cache" "content-length" "177" "connection" "close" "Date" "Fri, 15 Mar 2019 05:51:00 GMT" ] "rawTrailers": [] "readable": false "req": { "_contentLength": 301 "_ended": true "_events": { "drain": { "listener": [Function] } "error": { "listener": [Function] } "prefinish": [Function] } "_eventsCount": 3 "_hasBody": true "_header": "POST /api/rollup/delete HTTP/1.1\r\nHost: localhost:5620\r\nAccept-Encoding: gzip, deflate\r\nUser-Agent: node-superagent/3.8.2\r\nkbn-xsrf: xxx\r\nContent-Type: application/json\r\nAuthorization: Basic ZWxhc3RpYzpjaGFuZ2VtZQ==\r\nContent-Length: 301\r\nConnection: close\r\n\r\n" "_headerSent": true "_last": true "_maxListeners": [undefined] "_onPendingData": [Function] "_removedConnection": false "_removedContLen": false "_removedTE": false "_trailer": "" "aborted": [undefined] "agent": { "_events": { "free": [Function] } "_eventsCount": 1 "_maxListeners": [undefined] "defaultPort": 80 "freeSockets": {} "keepAlive": false "keepAliveMsecs": 1000 "maxFreeSockets": 256 "maxSockets": Infinity "options": { "path": [null] } "protocol": "http:" "requests": {} "sockets": { "localhost:5620:": [ { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } ] } } "chunkedEncoding": false "connection": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "finished": true "maxHeadersCount": [null] "method": "POST" "output": [] "outputCallbacks": [] "outputEncodings": [] "outputSize": 0 "parser": [null] "path": "/api/rollup/delete" "res": [Circular] "sendDate": false "shouldKeepAlive": false "socket": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "socketPath": [undefined] "timeout": [undefined] "timeoutCb": [null] "upgradeOrConnect": false "useChunkedEncodingByDefault": true "writable": true } "socket": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": { "_contentLength": 301 "_ended": true "_events": { "drain": { "listener": [Function] } "error": { "listener": [Function] } "prefinish": [Function] } "_eventsCount": 3 "_hasBody": true "_header": "POST /api/rollup/delete HTTP/1.1\r\nHost: localhost:5620\r\nAccept-Encoding: gzip, deflate\r\nUser-Agent: node-superagent/3.8.2\r\nkbn-xsrf: xxx\r\nContent-Type: application/json\r\nAuthorization: Basic ZWxhc3RpYzpjaGFuZ2VtZQ==\r\nContent-Length: 301\r\nConnection: close\r\n\r\n" "_headerSent": true "_last": true "_maxListeners": [undefined] "_onPendingData": [Function] "_removedConnection": false "_removedContLen": false "_removedTE": false "_trailer": "" "aborted": [undefined] "agent": { "_events": { "free": [Function] } "_eventsCount": 1 "_maxListeners": [undefined] "defaultPort": 80 "freeSockets": {} "keepAlive": false "keepAliveMsecs": 1000 "maxFreeSockets": 256 "maxSockets": Infinity "options": { "path": [null] } "protocol": "http:" "requests": {} "sockets": { "localhost:5620:": [ [Circular] ] } } "chunkedEncoding": false "connection": [Circular] "finished": true "maxHeadersCount": [null] "method": "POST" "output": [] "outputCallbacks": [] "outputEncodings": [] "outputSize": 0 "parser": [null] "path": "/api/rollup/delete" "res": [Circular] "sendDate": false "shouldKeepAlive": false "socket": [Circular] "socketPath": [undefined] "timeout": [undefined] "timeoutCb": [null] "upgradeOrConnect": false "useChunkedEncodingByDefault": true "writable": true } "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "statusCode": 400 "statusMessage": "Bad Request" "text": "{\"statusCode\":400,\"error\":\"Bad Request\",\"message\":\"Could not delete job [nbsxrhzvoq-1552629059154] because indexer state is [ABORTING]. Job must be [STOPPED] before deletion.\"}" "trailers": {} "upgrade": false "url": "" } "response": [Circular] "then": [Function] "toPromise": [Function] "url": "http://elastic:changeme@localhost:5620/api/rollup/delete" "writable": true } "res": { "_consuming": false "_dumped": false "_events": { "close": [Function] "data": [ [Function] [Function] [Function] ] "end": [ [Function] [Function] [Function] ] "error": [ { "listener": [Function] } [Function] ] } "_eventsCount": 4 "_maxListeners": [undefined] "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": { "encoding": "utf8" } "defaultEncoding": "utf8" "destroyed": false "emitClose": true "emittedReadable": false "encoding": "utf8" "endEmitted": true "ended": true "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": false "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": false "readingMore": true "resumeScheduled": false "sync": true } "aborted": false "client": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": { "_contentLength": 301 "_ended": true "_events": { "drain": { "listener": [Function] } "error": { "listener": [Function] } "prefinish": [Function] } "_eventsCount": 3 "_hasBody": true "_header": "POST /api/rollup/delete HTTP/1.1\r\nHost: localhost:5620\r\nAccept-Encoding: gzip, deflate\r\nUser-Agent: node-superagent/3.8.2\r\nkbn-xsrf: xxx\r\nContent-Type: application/json\r\nAuthorization: Basic ZWxhc3RpYzpjaGFuZ2VtZQ==\r\nContent-Length: 301\r\nConnection: close\r\n\r\n" "_headerSent": true "_last": true "_maxListeners": [undefined] "_onPendingData": [Function] "_removedConnection": false "_removedContLen": false "_removedTE": false "_trailer": "" "aborted": [undefined] "agent": { "_events": { "free": [Function] } "_eventsCount": 1 "_maxListeners": [undefined] "defaultPort": 80 "freeSockets": {} "keepAlive": false "keepAliveMsecs": 1000 "maxFreeSockets": 256 "maxSockets": Infinity "options": { "path": [null] } "protocol": "http:" "requests": {} "sockets": { "localhost:5620:": [ [Circular] ] } } "chunkedEncoding": false "connection": [Circular] "finished": true "maxHeadersCount": [null] "method": "POST" "output": [] "outputCallbacks": [] "outputEncodings": [] "outputSize": 0 "parser": [null] "path": "/api/rollup/delete" "res": [Circular] "sendDate": false "shouldKeepAlive": false "socket": [Circular] "socketPath": [undefined] "timeout": [undefined] "timeoutCb": [null] "upgradeOrConnect": false "useChunkedEncodingByDefault": true "writable": true } "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "complete": true "connection": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": { "_contentLength": 301 "_ended": true "_events": { "drain": { "listener": [Function] } "error": { "listener": [Function] } "prefinish": [Function] } "_eventsCount": 3 "_hasBody": true "_header": "POST /api/rollup/delete HTTP/1.1\r\nHost: localhost:5620\r\nAccept-Encoding: gzip, deflate\r\nUser-Agent: node-superagent/3.8.2\r\nkbn-xsrf: xxx\r\nContent-Type: application/json\r\nAuthorization: Basic ZWxhc3RpYzpjaGFuZ2VtZQ==\r\nContent-Length: 301\r\nConnection: close\r\n\r\n" "_headerSent": true "_last": true "_maxListeners": [undefined] "_onPendingData": [Function] "_removedConnection": false "_removedContLen": false "_removedTE": false "_trailer": "" "aborted": [undefined] "agent": { "_events": { "free": [Function] } "_eventsCount": 1 "_maxListeners": [undefined] "defaultPort": 80 "freeSockets": {} "keepAlive": false "keepAliveMsecs": 1000 "maxFreeSockets": 256 "maxSockets": Infinity "options": { "path": [null] } "protocol": "http:" "requests": {} "sockets": { "localhost:5620:": [ [Circular] ] } } "chunkedEncoding": false "connection": [Circular] "finished": true "maxHeadersCount": [null] "method": "POST" "output": [] "outputCallbacks": [] "outputEncodings": [] "outputSize": 0 "parser": [null] "path": "/api/rollup/delete" "res": [Circular] "sendDate": false "shouldKeepAlive": false "socket": [Circular] "socketPath": [undefined] "timeout": [undefined] "timeoutCb": [null] "upgradeOrConnect": false "useChunkedEncodingByDefault": true "writable": true } "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "headers": { "cache-control": "no-cache" "connection": "close" "content-length": "177" "content-type": "application/json; charset=utf-8" "date": "Fri, 15 Mar 2019 05:51:00 GMT" "kbn-name": "kibana" "kbn-xpack-sig": "a422b35703abc517c92812429d155891" } "httpVersion": "1.1" "httpVersionMajor": 1 "httpVersionMinor": 1 "method": [null] "rawHeaders": [ "kbn-name" "kibana" "kbn-xpack-sig" "a422b35703abc517c92812429d155891" "content-type" "application/json; charset=utf-8" "cache-control" "no-cache" "content-length" "177" "connection" "close" "Date" "Fri, 15 Mar 2019 05:51:00 GMT" ] "rawTrailers": [] "readable": false "req": { "_contentLength": 301 "_ended": true "_events": { "drain": { "listener": [Function] } "error": { "listener": [Function] } "prefinish": [Function] } "_eventsCount": 3 "_hasBody": true "_header": "POST /api/rollup/delete HTTP/1.1\r\nHost: localhost:5620\r\nAccept-Encoding: gzip, deflate\r\nUser-Agent: node-superagent/3.8.2\r\nkbn-xsrf: xxx\r\nContent-Type: application/json\r\nAuthorization: Basic ZWxhc3RpYzpjaGFuZ2VtZQ==\r\nContent-Length: 301\r\nConnection: close\r\n\r\n" "_headerSent": true "_last": true "_maxListeners": [undefined] "_onPendingData": [Function] "_removedConnection": false "_removedContLen": false "_removedTE": false "_trailer": "" "aborted": [undefined] "agent": { "_events": { "free": [Function] } "_eventsCount": 1 "_maxListeners": [undefined] "defaultPort": 80 "freeSockets": {} "keepAlive": false "keepAliveMsecs": 1000 "maxFreeSockets": 256 "maxSockets": Infinity "options": { "path": [null] } "protocol": "http:" "requests": {} "sockets": { "localhost:5620:": [ { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } ] } } "chunkedEncoding": false "connection": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "finished": true "maxHeadersCount": [null] "method": "POST" "output": [] "outputCallbacks": [] "outputEncodings": [] "outputSize": 0 "parser": [null] "path": "/api/rollup/delete" "res": [Circular] "sendDate": false "shouldKeepAlive": false "socket": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": [Circular] "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "socketPath": [undefined] "timeout": [undefined] "timeoutCb": [null] "upgradeOrConnect": false "useChunkedEncodingByDefault": true "writable": true } "socket": { "_events": { "agentRemove": [Function] "close": [ [Function] [Function] ] "drain": [Function] "end": [Function] "error": [Function] "finish": { "listener": [Function] } "free": [Function] } "_eventsCount": 7 "_hadError": false "_handle": "[object TCP]" "_host": "localhost" "_httpMessage": { "_contentLength": 301 "_ended": true "_events": { "drain": { "listener": [Function] } "error": { "listener": [Function] } "prefinish": [Function] } "_eventsCount": 3 "_hasBody": true "_header": "POST /api/rollup/delete HTTP/1.1\r\nHost: localhost:5620\r\nAccept-Encoding: gzip, deflate\r\nUser-Agent: node-superagent/3.8.2\r\nkbn-xsrf: xxx\r\nContent-Type: application/json\r\nAuthorization: Basic ZWxhc3RpYzpjaGFuZ2VtZQ==\r\nContent-Length: 301\r\nConnection: close\r\n\r\n" "_headerSent": true "_last": true "_maxListeners": [undefined] "_onPendingData": [Function] "_removedConnection": false "_removedContLen": false "_removedTE": false "_trailer": "" "aborted": [undefined] "agent": { "_events": { "free": [Function] } "_eventsCount": 1 "_maxListeners": [undefined] "defaultPort": 80 "freeSockets": {} "keepAlive": false "keepAliveMsecs": 1000 "maxFreeSockets": 256 "maxSockets": Infinity "options": { "path": [null] } "protocol": "http:" "requests": {} "sockets": { "localhost:5620:": [ [Circular] ] } } "chunkedEncoding": false "connection": [Circular] "finished": true "maxHeadersCount": [null] "method": "POST" "output": [] "outputCallbacks": [] "outputEncodings": [] "outputSize": 0 "parser": [null] "path": "/api/rollup/delete" "res": [Circular] "sendDate": false "shouldKeepAlive": false "socket": [Circular] "socketPath": [undefined] "timeout": [undefined] "timeoutCb": [null] "upgradeOrConnect": false "useChunkedEncodingByDefault": true "writable": true } "_maxListeners": [undefined] "_parent": [null] "_pendingData": [null] "_pendingEncoding": "" "_readableState": { "awaitDrain": 0 "buffer": { "head": [null] "length": 0 "tail": [null] } "decoder": [null] "defaultEncoding": "utf8" "destroyed": false "emitClose": false "emittedReadable": false "encoding": [null] "endEmitted": false "ended": false "flowing": true "highWaterMark": 16384 "length": 0 "needReadable": true "objectMode": false "paused": false "pipes": [null] "pipesCount": 0 "readableListening": false "reading": true "readingMore": false "resumeScheduled": false "sync": false } "_server": [null] "_sockname": [null] "_writableState": { "bufferProcessing": false "bufferedRequest": [null] "bufferedRequestCount": 0 "corked": 0 "corkedRequestsFree": { "entry": [null] "finish": [Function] "next": { "entry": [null] "finish": [Function] "next": [null] } } "decodeStrings": false "defaultEncoding": "utf8" "destroyed": false "emitClose": false "ended": true "ending": true "errorEmitted": false "finalCalled": true "finished": false "highWaterMark": 16384 "lastBufferedRequest": [null] "length": 0 "needDrain": false "objectMode": false "onwrite": [Function] "pendingcb": 1 "prefinished": false "sync": false "writecb": [null] "writelen": 0 "writing": false } "allowHalfOpen": false "connecting": false "parser": [null] "readable": true "server": [null] "writable": false } "statusCode": 400 "statusMessage": "Bad Request" "text": "{\"statusCode\":400,\"error\":\"Bad Request\",\"message\":\"Could not delete job [nbsxrhzvoq-1552629059154] because indexer state is [ABORTING]. Job must be [STOPPED] before deletion.\"}" "trailers": {} "upgrade": false "url": "" } "serverError": false "setEncoding": [Function] "status": 400 "statusCode": 400 "statusType": 4 "text": "{\"statusCode\":400,\"error\":\"Bad Request\",\"message\":\"Could not delete job [nbsxrhzvoq-1552629059154] because indexer state is [ABORTING]. Job must be [STOPPED] before deletion.\"}" "type": "application/json" "unauthorized": false } was thrown, throw an Error :)] ``` </details> First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+7.x/JOB=x-pack-ciGroup5,node=immutable/466/) <!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack API Integration Tests.x-pack/test/api_integration/apis/management/rollup/rollup·js","test.name":"apis management rollup jobs \"after all\" hook","test.failCount":3}} -->
test
failing test x pack api integration tests x pack test api integration apis management rollup rollup·js apis management rollup jobs after all hook latest failure stacktrace no such index with resource type index or alias resource id statuscode index uuid na index statuscode path statuscode query statuscode response error root cause resource type index or alias resource id statuscode index uuid na index statuscode type index not found exception reason no such index resource type index or alias resource id statuscode index uuid na index statuscode status at respond var lib jenkins workspace elastic kibana master job x pack node immutable kibana node modules elasticsearch src lib transport js at checkrespforfailure var lib jenkins workspace elastic kibana master job x pack node immutable kibana node modules elasticsearch src lib transport js at httpconnector var lib jenkins workspace elastic kibana master job x pack node immutable kibana node modules elasticsearch src lib connectors http js at incomingmessage wrapper var lib jenkins workspace elastic kibana master job x pack node immutable kibana node modules elasticsearch node modules lodash lodash js at endreadablent stream readable js at process tickcallback internal process next tick js status displayname notfound message no such index with resource type index or alias resource id statuscode index uuid na index statuscode path statuscode query body error root cause type index not found exception reason no such index resource type index or alias resource id statuscode index uuid na index statuscode status statuscode response error root cause resource type index or alias resource id statuscode index uuid na index statuscode type index not found exception reason no such index resource type index or alias resource id statuscode index uuid na index statuscode status tostring tojson standard output │ └ apis └ before all hook └ management └ before all hook └ rollup └ before all hook └ jobs └ before all hook └ after all hook │ info deleting index │ info deleting index └ ✖ fail apis management rollup jobs after all hook │ no such index with resource type index or alias resource id statuscode index uuid na index statuscode │ path statuscode query statuscode response error root cause resource type index or alias resource id statuscode index uuid na index statuscode type index not found exception reason no such index resource type index or alias resource id statuscode index uuid na index statuscode status │ at respond var lib jenkins workspace elastic kibana master job x pack node immutable kibana node modules elasticsearch src lib transport js │ at checkrespforfailure var lib jenkins workspace elastic kibana master job x pack node immutable kibana node modules elasticsearch src lib transport js │ at httpconnector var lib jenkins workspace elastic kibana master job x pack node immutable kibana node modules elasticsearch src lib connectors http js │ at incomingmessage wrapper var lib jenkins workspace elastic kibana master job x pack node immutable kibana node modules elasticsearch node modules lodash lodash js │ at endreadablent stream readable js │ at process tickcallback internal process next tick js │ │ previous failure a test failed on a tracked branch error the object events eventscount maxlisteners accepted false badrequest true body error bad request message could not delete job because indexer state is job must be before deletion statuscode buffered true charset utf clienterror true error error cannot post api rollup delete files forbidden false header cache control no cache connection close content length content type application json charset utf date fri mar gmt kbn name kibana kbn xpack sig headers cache control no cache connection close content length content type application json charset utf date fri mar gmt kbn name kibana kbn xpack sig info false links nocontent false notacceptable false notfound false ok false redirect false redirects req contentlength ended true events drain listener error listener prefinish eventscount hasbody true header post api rollup delete http r nhost localhost r naccept encoding gzip deflate r nuser agent node superagent r nkbn xsrf xxx r ncontent type application json r nauthorization basic r ncontent length r nconnection close r n r n headersent true last true maxlisteners onpendingdata removedconnection false removedcontlen false removedte false trailer aborted agent events free eventscount maxlisteners defaultport freesockets keepalive false keepalivemsecs maxfreesockets maxsockets infinity options path protocol http requests sockets localhost events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false chunkedencoding false connection events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false finished true maxheaderscount method post output outputcallbacks outputencodings outputsize parser path api rollup delete res consuming false dumped false events close data end error listener eventscount maxlisteners readablestate awaitdrain buffer head length tail decoder encoding defaultencoding destroyed false emitclose true emittedreadable false encoding endemitted true ended true flowing true highwatermark length needreadable false objectmode false paused false pipes pipescount readablelistening false reading false readingmore true resumescheduled false sync true aborted false client events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false complete true connection events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false headers cache control no cache connection close content length content type application json charset utf date fri mar gmt kbn name kibana kbn xpack sig httpversion httpversionmajor httpversionminor method rawheaders kbn name kibana kbn xpack sig content type application json charset utf cache control no cache content length connection close date fri mar gmt rawtrailers readable false req socket events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false statuscode statusmessage bad request text statuscode error bad request message could not delete job because indexer state is job must be before deletion trailers upgrade false url senddate false shouldkeepalive false socket events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false socketpath timeout timeoutcb upgradeorconnect false usechunkedencodingbydefault true writable true request agent false asserts buffer true callback data jobids wwcnompptsjrbm nbsxrhzvoq nbsxrhzvoq abc ovcwg xppabncglokc dwumwtgaf oxtnysgbuh fcctx zunqhjjamgdoieylchvt arbhsuiiggqwytxmhu endcalled true events eventscount formdata header authorization basic content type application json kbn xsrf xxx user agent node superagent maxlisteners maxredirects query redirectlist redirects streamrequest false app called true catch cookies header authorization basic content type application json user agent node superagent kbn xsrf xxx host localhost method post protocol http qs qsraw req contentlength ended true events drain listener error listener prefinish eventscount hasbody true header post api rollup delete http r nhost localhost r naccept encoding gzip deflate r nuser agent node superagent r nkbn xsrf xxx r ncontent type application json r nauthorization basic r ncontent length r nconnection close r n r n headersent true last true maxlisteners onpendingdata removedconnection false removedcontlen false removedte false trailer aborted agent events free eventscount maxlisteners defaultport freesockets keepalive false keepalivemsecs maxfreesockets maxsockets infinity options path protocol http requests sockets localhost events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false chunkedencoding false connection events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false finished true maxheaderscount method post output outputcallbacks outputencodings outputsize parser path api rollup delete res consuming false dumped false events close data end error listener eventscount maxlisteners readablestate awaitdrain buffer head length tail decoder encoding defaultencoding destroyed false emitclose true emittedreadable false encoding endemitted true ended true flowing true highwatermark length needreadable false objectmode false paused false pipes pipescount readablelistening false reading false readingmore true resumescheduled false sync true aborted false client events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false complete true connection events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false headers cache control no cache connection close content length content type application json charset utf date fri mar gmt kbn name kibana kbn xpack sig httpversion httpversionmajor httpversionminor method rawheaders kbn name kibana kbn xpack sig content type application json charset utf cache control no cache content length connection close date fri mar gmt rawtrailers readable false req socket events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false statuscode statusmessage bad request text statuscode error bad request message could not delete job because indexer state is job must be before deletion trailers upgrade false url senddate false shouldkeepalive false socket events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false socketpath timeout timeoutcb upgradeorconnect false usechunkedencodingbydefault true writable true res consuming false dumped false events close data end error listener eventscount maxlisteners readablestate awaitdrain buffer head length tail decoder encoding defaultencoding destroyed false emitclose true emittedreadable false encoding endemitted true ended true flowing true highwatermark length needreadable false objectmode false paused false pipes pipescount readablelistening false reading false readingmore true resumescheduled false sync true aborted false client events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage contentlength ended true events drain listener error listener prefinish eventscount hasbody true header post api rollup delete http r nhost localhost r naccept encoding gzip deflate r nuser agent node superagent r nkbn xsrf xxx r ncontent type application json r nauthorization basic r ncontent length r nconnection close r n r n headersent true last true maxlisteners onpendingdata removedconnection false removedcontlen false removedte false trailer aborted agent events free eventscount maxlisteners defaultport freesockets keepalive false keepalivemsecs maxfreesockets maxsockets infinity options path protocol http requests sockets localhost chunkedencoding false connection finished true maxheaderscount method post output outputcallbacks outputencodings outputsize parser path api rollup delete res senddate false shouldkeepalive false socket socketpath timeout timeoutcb upgradeorconnect false usechunkedencodingbydefault true writable true maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false complete true connection events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage contentlength ended true events drain listener error listener prefinish eventscount hasbody true header post api rollup delete http r nhost localhost r naccept encoding gzip deflate r nuser agent node superagent r nkbn xsrf xxx r ncontent type application json r nauthorization basic r ncontent length r nconnection close r n r n headersent true last true maxlisteners onpendingdata removedconnection false removedcontlen false removedte false trailer aborted agent events free eventscount maxlisteners defaultport freesockets keepalive false keepalivemsecs maxfreesockets maxsockets infinity options path protocol http requests sockets localhost chunkedencoding false connection finished true maxheaderscount method post output outputcallbacks outputencodings outputsize parser path api rollup delete res senddate false shouldkeepalive false socket socketpath timeout timeoutcb upgradeorconnect false usechunkedencodingbydefault true writable true maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false headers cache control no cache connection close content length content type application json charset utf date fri mar gmt kbn name kibana kbn xpack sig httpversion httpversionmajor httpversionminor method rawheaders kbn name kibana kbn xpack sig content type application json charset utf cache control no cache content length connection close date fri mar gmt rawtrailers readable false req contentlength ended true events drain listener error listener prefinish eventscount hasbody true header post api rollup delete http r nhost localhost r naccept encoding gzip deflate r nuser agent node superagent r nkbn xsrf xxx r ncontent type application json r nauthorization basic r ncontent length r nconnection close r n r n headersent true last true maxlisteners onpendingdata removedconnection false removedcontlen false removedte false trailer aborted agent events free eventscount maxlisteners defaultport freesockets keepalive false keepalivemsecs maxfreesockets maxsockets infinity options path protocol http requests sockets localhost events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false chunkedencoding false connection events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false finished true maxheaderscount method post output outputcallbacks outputencodings outputsize parser path api rollup delete res senddate false shouldkeepalive false socket events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false socketpath timeout timeoutcb upgradeorconnect false usechunkedencodingbydefault true writable true socket events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage contentlength ended true events drain listener error listener prefinish eventscount hasbody true header post api rollup delete http r nhost localhost r naccept encoding gzip deflate r nuser agent node superagent r nkbn xsrf xxx r ncontent type application json r nauthorization basic r ncontent length r nconnection close r n r n headersent true last true maxlisteners onpendingdata removedconnection false removedcontlen false removedte false trailer aborted agent events free eventscount maxlisteners defaultport freesockets keepalive false keepalivemsecs maxfreesockets maxsockets infinity options path protocol http requests sockets localhost chunkedencoding false connection finished true maxheaderscount method post output outputcallbacks outputencodings outputsize parser path api rollup delete res senddate false shouldkeepalive false socket socketpath timeout timeoutcb upgradeorconnect false usechunkedencodingbydefault true writable true maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false statuscode statusmessage bad request text statuscode error bad request message could not delete job because indexer state is job must be before deletion trailers upgrade false url response then topromise url writable true res consuming false dumped false events close data end error listener eventscount maxlisteners readablestate awaitdrain buffer head length tail decoder encoding defaultencoding destroyed false emitclose true emittedreadable false encoding endemitted true ended true flowing true highwatermark length needreadable false objectmode false paused false pipes pipescount readablelistening false reading false readingmore true resumescheduled false sync true aborted false client events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage contentlength ended true events drain listener error listener prefinish eventscount hasbody true header post api rollup delete http r nhost localhost r naccept encoding gzip deflate r nuser agent node superagent r nkbn xsrf xxx r ncontent type application json r nauthorization basic r ncontent length r nconnection close r n r n headersent true last true maxlisteners onpendingdata removedconnection false removedcontlen false removedte false trailer aborted agent events free eventscount maxlisteners defaultport freesockets keepalive false keepalivemsecs maxfreesockets maxsockets infinity options path protocol http requests sockets localhost chunkedencoding false connection finished true maxheaderscount method post output outputcallbacks outputencodings outputsize parser path api rollup delete res senddate false shouldkeepalive false socket socketpath timeout timeoutcb upgradeorconnect false usechunkedencodingbydefault true writable true maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false complete true connection events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage contentlength ended true events drain listener error listener prefinish eventscount hasbody true header post api rollup delete http r nhost localhost r naccept encoding gzip deflate r nuser agent node superagent r nkbn xsrf xxx r ncontent type application json r nauthorization basic r ncontent length r nconnection close r n r n headersent true last true maxlisteners onpendingdata removedconnection false removedcontlen false removedte false trailer aborted agent events free eventscount maxlisteners defaultport freesockets keepalive false keepalivemsecs maxfreesockets maxsockets infinity options path protocol http requests sockets localhost chunkedencoding false connection finished true maxheaderscount method post output outputcallbacks outputencodings outputsize parser path api rollup delete res senddate false shouldkeepalive false socket socketpath timeout timeoutcb upgradeorconnect false usechunkedencodingbydefault true writable true maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false headers cache control no cache connection close content length content type application json charset utf date fri mar gmt kbn name kibana kbn xpack sig httpversion httpversionmajor httpversionminor method rawheaders kbn name kibana kbn xpack sig content type application json charset utf cache control no cache content length connection close date fri mar gmt rawtrailers readable false req contentlength ended true events drain listener error listener prefinish eventscount hasbody true header post api rollup delete http r nhost localhost r naccept encoding gzip deflate r nuser agent node superagent r nkbn xsrf xxx r ncontent type application json r nauthorization basic r ncontent length r nconnection close r n r n headersent true last true maxlisteners onpendingdata removedconnection false removedcontlen false removedte false trailer aborted agent events free eventscount maxlisteners defaultport freesockets keepalive false keepalivemsecs maxfreesockets maxsockets infinity options path protocol http requests sockets localhost events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false chunkedencoding false connection events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false finished true maxheaderscount method post output outputcallbacks outputencodings outputsize parser path api rollup delete res senddate false shouldkeepalive false socket events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false socketpath timeout timeoutcb upgradeorconnect false usechunkedencodingbydefault true writable true socket events agentremove close drain end error finish listener free eventscount haderror false handle host localhost httpmessage contentlength ended true events drain listener error listener prefinish eventscount hasbody true header post api rollup delete http r nhost localhost r naccept encoding gzip deflate r nuser agent node superagent r nkbn xsrf xxx r ncontent type application json r nauthorization basic r ncontent length r nconnection close r n r n headersent true last true maxlisteners onpendingdata removedconnection false removedcontlen false removedte false trailer aborted agent events free eventscount maxlisteners defaultport freesockets keepalive false keepalivemsecs maxfreesockets maxsockets infinity options path protocol http requests sockets localhost chunkedencoding false connection finished true maxheaderscount method post output outputcallbacks outputencodings outputsize parser path api rollup delete res senddate false shouldkeepalive false socket socketpath timeout timeoutcb upgradeorconnect false usechunkedencodingbydefault true writable true maxlisteners parent pendingdata pendingencoding readablestate awaitdrain buffer head length tail decoder defaultencoding destroyed false emitclose false emittedreadable false encoding endemitted false ended false flowing true highwatermark length needreadable true objectmode false paused false pipes pipescount readablelistening false reading true readingmore false resumescheduled false sync false server sockname writablestate bufferprocessing false bufferedrequest bufferedrequestcount corked corkedrequestsfree entry finish next entry finish next decodestrings false defaultencoding destroyed false emitclose false ended true ending true erroremitted false finalcalled true finished false highwatermark lastbufferedrequest length needdrain false objectmode false onwrite pendingcb prefinished false sync false writecb writelen writing false allowhalfopen false connecting false parser readable true server writable false statuscode statusmessage bad request text statuscode error bad request message could not delete job because indexer state is job must be before deletion trailers upgrade false url servererror false setencoding status statuscode statustype text statuscode error bad request message could not delete job because indexer state is job must be before deletion type application json unauthorized false was thrown throw an error first failure
1
193,693
14,660,034,633
IssuesEvent
2020-12-28 22:14:30
github-vet/rangeloop-pointer-findings
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
closed
praparn/kubernetes_201904: WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go; 5 LoC
fresh test tiny
Found a possible issue in [praparn/kubernetes_201904](https://www.github.com/praparn/kubernetes_201904) at [WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go](https://github.com/praparn/kubernetes_201904/blob/92c6c826b9a75a27925c13dcf921100d54c3e015/WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go#L56-L60) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to m at line 58 may start a goroutine [Click here to see the code in its original context.](https://github.com/praparn/kubernetes_201904/blob/92c6c826b9a75a27925c13dcf921100d54c3e015/WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go#L56-L60) <details> <summary>Click here to show the 5 line(s) of Go which triggered the analyzer.</summary> ```go for _, m := range metas[5:] { var buf bytes.Buffer testutil.Ok(t, json.NewEncoder(&buf).Encode(&m)) testutil.Ok(t, bkt.Upload(ctx, path.Join(m.ULID.String(), block.MetaFilename), &buf)) } ``` </details> <details> <summary>Click here to show extra information the analyzer produced.</summary> ``` The following graphviz dot graph describes paths through the callgraph that could lead to a function calling a goroutine: digraph G { "(Encode, 1)" -> {"(Write, 1)";"(panicToErr, 1)";"(Encode, 2)";} "(newClient, 1)" -> {"(dial, 2)";"(NewLease, 1)";} "(apply, 2)" -> {"(applyRequest, 1)";} "(Get, 2)" -> {"(handle, 2)";"(Get, 3)";} "(copy, 2)" -> {"(get, 1)";"(findObject, 2)";} "(tryUpgrade, 2)" -> {} "(NewServerConn, 2)" -> {"(serverHandshake, 1)";} "(NewClientConn, 3)" -> {"(clientHandshake, 2)";} "(clientHandshake, 2)" -> {"(newClientTransport, 6)";} "(New, 1)" -> {"(newClient, 1)";"(Start, 1)";"(parse, 1)";"(New, 2)";"(apply, 2)";"(get, 1)";} "(Start, 1)" -> {} "(applyRequest, 1)" -> {"(Copy, 2)";} "(Do, 3)" -> {} "(panicValToErr, 2)" -> {"(New, 1)";} "(serverHandshake, 1)" -> {"(newServerTransport, 4)";} "(newClientTransport, 6)" -> {} "(Write, 1)" -> {"(Add, 1)";"(Call, 5)";"(copy, 2)";"(Set, 2)";"(New, 1)";"(Get, 1)";} "(Set, 2)" -> {"(Do, 3)";"(New, 1)";} "(Get, 1)" -> {"(get, 1)";"(Do, 2)";"(Add, 2)";"(Get, 2)";} "(Do, 2)" -> {"(RoundTrip, 1)";} "(Add, 2)" -> {"(Do, 2)";} "(Copy, 2)" -> {"(ServeHTTP, 2)";} "(RoundTrip, 1)" -> {"(awaitOpenSlotForRequest, 1)";} "(ConfigureTransport, 1)" -> {"(configureTransport, 1)";} "(Decode, 2)" -> {"(Decode, 1)";} "(newServerTransport, 4)" -> {} "(Encode, 2)" -> {"(Write, 1)";"(Decode, 3)";"(WriteTo, 1)";"(copy, 2)";} "(dial, 2)" -> {"(NewServerConn, 2)";"(NewClientConn, 3)";} "(NewLease, 1)" -> {} "(handle, 2)" -> {"(ServeHTTP, 2)";} "(ServeHTTP, 2)" -> {"(tryUpgrade, 2)";"(tryUpgrade, 7)";} "(WriteTo, 1)" -> {"(Get, 1)";} "(parse, 1)" -> {"(append, 1)";} "(append, 1)" -> {} "(New, 2)" -> {"(Start, 1)";} "(Call, 5)" -> {"(Get, 1)";} "(Get, 3)" -> {"(Copy, 2)";"(RoundTrip, 1)";} "(awaitOpenSlotForRequest, 1)" -> {} "(configureTransport, 1)" -> {"(addConnIfNeeded, 3)";} "(addConnIfNeeded, 3)" -> {} "(panicToErr, 1)" -> {"(panicValToErr, 2)";} "(Add, 1)" -> {"(New, 1)";"(Add, 2)";} "(Decode, 1)" -> {"(unmarshal, 2)";"(DecodeElement, 2)";} "(unmarshal, 2)" -> {"(Copy, 2)";} "(DecodeElement, 2)" -> {"(unmarshal, 2)";} "(tryUpgrade, 7)" -> {} "(Decode, 3)" -> {"(New, 1)";"(Decode, 1)";"(Decode, 2)";} "(get, 1)" -> {"(SetTransportDefaults, 1)";} "(SetTransportDefaults, 1)" -> {"(ConfigureTransport, 1)";} "(findObject, 2)" -> {"(get, 1)";} } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 92c6c826b9a75a27925c13dcf921100d54c3e015
1.0
praparn/kubernetes_201904: WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go; 5 LoC - Found a possible issue in [praparn/kubernetes_201904](https://www.github.com/praparn/kubernetes_201904) at [WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go](https://github.com/praparn/kubernetes_201904/blob/92c6c826b9a75a27925c13dcf921100d54c3e015/WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go#L56-L60) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to m at line 58 may start a goroutine [Click here to see the code in its original context.](https://github.com/praparn/kubernetes_201904/blob/92c6c826b9a75a27925c13dcf921100d54c3e015/WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go#L56-L60) <details> <summary>Click here to show the 5 line(s) of Go which triggered the analyzer.</summary> ```go for _, m := range metas[5:] { var buf bytes.Buffer testutil.Ok(t, json.NewEncoder(&buf).Encode(&m)) testutil.Ok(t, bkt.Upload(ctx, path.Join(m.ULID.String(), block.MetaFilename), &buf)) } ``` </details> <details> <summary>Click here to show extra information the analyzer produced.</summary> ``` The following graphviz dot graph describes paths through the callgraph that could lead to a function calling a goroutine: digraph G { "(Encode, 1)" -> {"(Write, 1)";"(panicToErr, 1)";"(Encode, 2)";} "(newClient, 1)" -> {"(dial, 2)";"(NewLease, 1)";} "(apply, 2)" -> {"(applyRequest, 1)";} "(Get, 2)" -> {"(handle, 2)";"(Get, 3)";} "(copy, 2)" -> {"(get, 1)";"(findObject, 2)";} "(tryUpgrade, 2)" -> {} "(NewServerConn, 2)" -> {"(serverHandshake, 1)";} "(NewClientConn, 3)" -> {"(clientHandshake, 2)";} "(clientHandshake, 2)" -> {"(newClientTransport, 6)";} "(New, 1)" -> {"(newClient, 1)";"(Start, 1)";"(parse, 1)";"(New, 2)";"(apply, 2)";"(get, 1)";} "(Start, 1)" -> {} "(applyRequest, 1)" -> {"(Copy, 2)";} "(Do, 3)" -> {} "(panicValToErr, 2)" -> {"(New, 1)";} "(serverHandshake, 1)" -> {"(newServerTransport, 4)";} "(newClientTransport, 6)" -> {} "(Write, 1)" -> {"(Add, 1)";"(Call, 5)";"(copy, 2)";"(Set, 2)";"(New, 1)";"(Get, 1)";} "(Set, 2)" -> {"(Do, 3)";"(New, 1)";} "(Get, 1)" -> {"(get, 1)";"(Do, 2)";"(Add, 2)";"(Get, 2)";} "(Do, 2)" -> {"(RoundTrip, 1)";} "(Add, 2)" -> {"(Do, 2)";} "(Copy, 2)" -> {"(ServeHTTP, 2)";} "(RoundTrip, 1)" -> {"(awaitOpenSlotForRequest, 1)";} "(ConfigureTransport, 1)" -> {"(configureTransport, 1)";} "(Decode, 2)" -> {"(Decode, 1)";} "(newServerTransport, 4)" -> {} "(Encode, 2)" -> {"(Write, 1)";"(Decode, 3)";"(WriteTo, 1)";"(copy, 2)";} "(dial, 2)" -> {"(NewServerConn, 2)";"(NewClientConn, 3)";} "(NewLease, 1)" -> {} "(handle, 2)" -> {"(ServeHTTP, 2)";} "(ServeHTTP, 2)" -> {"(tryUpgrade, 2)";"(tryUpgrade, 7)";} "(WriteTo, 1)" -> {"(Get, 1)";} "(parse, 1)" -> {"(append, 1)";} "(append, 1)" -> {} "(New, 2)" -> {"(Start, 1)";} "(Call, 5)" -> {"(Get, 1)";} "(Get, 3)" -> {"(Copy, 2)";"(RoundTrip, 1)";} "(awaitOpenSlotForRequest, 1)" -> {} "(configureTransport, 1)" -> {"(addConnIfNeeded, 3)";} "(addConnIfNeeded, 3)" -> {} "(panicToErr, 1)" -> {"(panicValToErr, 2)";} "(Add, 1)" -> {"(New, 1)";"(Add, 2)";} "(Decode, 1)" -> {"(unmarshal, 2)";"(DecodeElement, 2)";} "(unmarshal, 2)" -> {"(Copy, 2)";} "(DecodeElement, 2)" -> {"(unmarshal, 2)";} "(tryUpgrade, 7)" -> {} "(Decode, 3)" -> {"(New, 1)";"(Decode, 1)";"(Decode, 2)";} "(get, 1)" -> {"(SetTransportDefaults, 1)";} "(SetTransportDefaults, 1)" -> {"(ConfigureTransport, 1)";} "(findObject, 2)" -> {"(get, 1)";} } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 92c6c826b9a75a27925c13dcf921100d54c3e015
test
praparn kubernetes workshop monitor prometheus operator vendor github com improbable eng thanos pkg compact compact test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to m at line may start a goroutine click here to show the line s of go which triggered the analyzer go for m range metas var buf bytes buffer testutil ok t json newencoder buf encode m testutil ok t bkt upload ctx path join m ulid string block metafilename buf click here to show extra information the analyzer produced the following graphviz dot graph describes paths through the callgraph that could lead to a function calling a goroutine digraph g encode write panictoerr encode newclient dial newlease apply applyrequest get handle get copy get findobject tryupgrade newserverconn serverhandshake newclientconn clienthandshake clienthandshake newclienttransport new newclient start parse new apply get start applyrequest copy do panicvaltoerr new serverhandshake newservertransport newclienttransport write add call copy set new get set do new get get do add get do roundtrip add do copy servehttp roundtrip awaitopenslotforrequest configuretransport configuretransport decode decode newservertransport encode write decode writeto copy dial newserverconn newclientconn newlease handle servehttp servehttp tryupgrade tryupgrade writeto get parse append append new start call get get copy roundtrip awaitopenslotforrequest configuretransport addconnifneeded addconnifneeded panictoerr panicvaltoerr add new add decode unmarshal decodeelement unmarshal copy decodeelement unmarshal tryupgrade decode new decode decode get settransportdefaults settransportdefaults configuretransport findobject get leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
65,904
6,978,162,358
IssuesEvent
2017-12-12 16:41:06
geosolutions-it/MapStore2
https://api.github.com/repos/geosolutions-it/MapStore2
reopened
Always-on GetFeatureInfo
In Test Internal Priority: High Project: C040 Tested user feedback user story
We should evaluate the default design of having the Info tool always active. User behaviour shows that info-on-click is expected. Related to #1125
2.0
Always-on GetFeatureInfo - We should evaluate the default design of having the Info tool always active. User behaviour shows that info-on-click is expected. Related to #1125
test
always on getfeatureinfo we should evaluate the default design of having the info tool always active user behaviour shows that info on click is expected related to
1
116,675
11,939,557,780
IssuesEvent
2020-04-02 15:22:44
ESMValGroup/ESMValTool
https://api.github.com/repos/ESMValGroup/ESMValTool
closed
Add new user mailing list to the documentation
documentation help wanted
# New user mailing list Subscribe by sending an empty mail to: `sympa@listserv.dfn.de` with the subject: `subscribe esmvaltool` or `subscribe esmvaltool FIRSTNAME LASTNAME` # Caveats + confirmation mail after subscription is in German + confirmation mail is send with a delay of some hours # Alternatively The admin of the list can add people manually (single or bulk). # Archive A public archive is available at [https://www.listserv.dfn.de/sympa/arc/esmvaltool](https://www.listserv.dfn.de/sympa/arc/esmvaltool) # Background The old mailing list was broken and the archive was not public. So we removed it from the docs #1556 .
1.0
Add new user mailing list to the documentation - # New user mailing list Subscribe by sending an empty mail to: `sympa@listserv.dfn.de` with the subject: `subscribe esmvaltool` or `subscribe esmvaltool FIRSTNAME LASTNAME` # Caveats + confirmation mail after subscription is in German + confirmation mail is send with a delay of some hours # Alternatively The admin of the list can add people manually (single or bulk). # Archive A public archive is available at [https://www.listserv.dfn.de/sympa/arc/esmvaltool](https://www.listserv.dfn.de/sympa/arc/esmvaltool) # Background The old mailing list was broken and the archive was not public. So we removed it from the docs #1556 .
non_test
add new user mailing list to the documentation new user mailing list subscribe by sending an empty mail to sympa listserv dfn de with the subject subscribe esmvaltool or subscribe esmvaltool firstname lastname caveats confirmation mail after subscription is in german confirmation mail is send with a delay of some hours alternatively the admin of the list can add people manually single or bulk archive a public archive is available at background the old mailing list was broken and the archive was not public so we removed it from the docs
0
197,211
14,912,401,714
IssuesEvent
2021-01-22 12:37:05
jetstack/cert-manager
https://api.github.com/repos/jetstack/cert-manager
closed
Adds external issuer e2e tests
area/testing kind/feature priority/important-soon
As of 0.16.0 external issuers are broken, as the Issuer group was not set on Secret resource annotations managed by the issuing controller. This causes CertificateRequests that belong to an external issuer to be constantly renewed. This has not work around, apart from downgrading to 0.15. We should create an e2e external issuer to catch these cases in future. This external issuer can then provide as an example to others on how to build their own external issuers. /kind feature
1.0
Adds external issuer e2e tests - As of 0.16.0 external issuers are broken, as the Issuer group was not set on Secret resource annotations managed by the issuing controller. This causes CertificateRequests that belong to an external issuer to be constantly renewed. This has not work around, apart from downgrading to 0.15. We should create an e2e external issuer to catch these cases in future. This external issuer can then provide as an example to others on how to build their own external issuers. /kind feature
test
adds external issuer tests as of external issuers are broken as the issuer group was not set on secret resource annotations managed by the issuing controller this causes certificaterequests that belong to an external issuer to be constantly renewed this has not work around apart from downgrading to we should create an external issuer to catch these cases in future this external issuer can then provide as an example to others on how to build their own external issuers kind feature
1
765,697
26,858,211,895
IssuesEvent
2023-02-03 16:10:52
Poobslag/turbofat
https://api.github.com/repos/Poobslag/turbofat
closed
Intro/boss level filler creatures should appear during puzzles
priority-4
During intro/boss levels, three characters are always shown. Some of these are secondary filler characters, but they don't show up during the puzzles. These secondary filler characters should appear during the puzzles.
1.0
Intro/boss level filler creatures should appear during puzzles - During intro/boss levels, three characters are always shown. Some of these are secondary filler characters, but they don't show up during the puzzles. These secondary filler characters should appear during the puzzles.
non_test
intro boss level filler creatures should appear during puzzles during intro boss levels three characters are always shown some of these are secondary filler characters but they don t show up during the puzzles these secondary filler characters should appear during the puzzles
0
706,768
24,283,473,116
IssuesEvent
2022-09-28 19:37:05
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
opened
LE Audio: Audio shell: Unicast server cannot execute commands for the default_stream
bug priority: low area: Bluetooth area: Bluetooth Audio
**Describe the bug** The `default_stream` is not properly for the unicast server, and thus remains NULL. The means that none of the `bt_audio_stream_*` commands can be executed. **To Reproduce** N/A **Expected behavior** The `default_stream` should be set and all the commands callable as the unicast server. **Impact** Minor; just missing features for the unicast server. **Logs and console output** N/A **Environment (please complete the following information):** - Commit SHA or Version used: a5ce3da6ab425d104d73d34f47352ae721c5d216 **Additional context** N/A
1.0
LE Audio: Audio shell: Unicast server cannot execute commands for the default_stream - **Describe the bug** The `default_stream` is not properly for the unicast server, and thus remains NULL. The means that none of the `bt_audio_stream_*` commands can be executed. **To Reproduce** N/A **Expected behavior** The `default_stream` should be set and all the commands callable as the unicast server. **Impact** Minor; just missing features for the unicast server. **Logs and console output** N/A **Environment (please complete the following information):** - Commit SHA or Version used: a5ce3da6ab425d104d73d34f47352ae721c5d216 **Additional context** N/A
non_test
le audio audio shell unicast server cannot execute commands for the default stream describe the bug the default stream is not properly for the unicast server and thus remains null the means that none of the bt audio stream commands can be executed to reproduce n a expected behavior the default stream should be set and all the commands callable as the unicast server impact minor just missing features for the unicast server logs and console output n a environment please complete the following information commit sha or version used additional context n a
0
131,225
12,478,816,834
IssuesEvent
2020-05-29 17:03:40
getting-things-gnome/gtg
https://api.github.com/repos/getting-things-gnome/gtg
closed
Thoroughly document the multilingual dates and keyword parsing syntax
documentation
An initial commit I've made in my own userdocs branch in my personal repo: https://github.com/nekohayo/GTG/commit/3b23e5b8 If @vansia43 likes it I could merge it to master. Afterwards, potentially, subsequent work could include splitting it into some modular page that can be reused/included/cited/linked to from other pages, because that syntax is valid anywhere in the app, so both the Quick Add entry and the Task Editor window, the "Pick a date..." dialogs, and maybe elsewhere (?). A starting point about the multilingual aspect was mentioned in https://github.com/getting-things-gnome/gtg/issues/243#issuecomment-623752476 too.
1.0
Thoroughly document the multilingual dates and keyword parsing syntax - An initial commit I've made in my own userdocs branch in my personal repo: https://github.com/nekohayo/GTG/commit/3b23e5b8 If @vansia43 likes it I could merge it to master. Afterwards, potentially, subsequent work could include splitting it into some modular page that can be reused/included/cited/linked to from other pages, because that syntax is valid anywhere in the app, so both the Quick Add entry and the Task Editor window, the "Pick a date..." dialogs, and maybe elsewhere (?). A starting point about the multilingual aspect was mentioned in https://github.com/getting-things-gnome/gtg/issues/243#issuecomment-623752476 too.
non_test
thoroughly document the multilingual dates and keyword parsing syntax an initial commit i ve made in my own userdocs branch in my personal repo if likes it i could merge it to master afterwards potentially subsequent work could include splitting it into some modular page that can be reused included cited linked to from other pages because that syntax is valid anywhere in the app so both the quick add entry and the task editor window the pick a date dialogs and maybe elsewhere a starting point about the multilingual aspect was mentioned in too
0