Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
308,778
26,631,368,497
IssuesEvent
2023-01-24 18:06:46
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/maps/group2/index·js - maps app "before all" hook in "maps app"
failed-test
A test failed on a tracked branch ``` Error: 400 resp: '' req: { transitional: { silentJSONParsing: true, forcedJSONParsing: true, clarifyTimeoutError: false }, adapter: [Function: httpAdapter], transformRequest: [ [Function: transformRequest] ], transformResponse: [ [Function: transformResponse] ], timeout: 0, xsrfCookieName: 'XSRF-TOKEN', xsrfHeaderName: 'X-XSRF-TOKEN', maxContentLength: 30000000, maxBodyLength: 30000000, env: { FormData: [Function: FormData] { LINE_BREAK: '\r\n', DEFAULT_CONTENT_TYPE: 'application/octet-stream' } }, validateStatus: [Function: validateStatus], headers: { Accept: 'application/json, text/plain, */*', 'Content-Type': 'multipart/form-data; boundary=--------------------------756925804713543151415896', 'kbn-xsrf': 'kbn-client', 'User-Agent': 'axios/0.27.2', 'content-type': 'multipart/form-data; boundary=--------------------------756925804713543151415896' }, method: 'post', url: 'http://elastic:changeme@localhost:5620/api/saved_objects/_import', data: FormData { _overheadLength: 129, _valueLength: 73008, _valuesToMeasure: [], writable: false, readable: true, dataSize: 0, maxDataSize: 2097152, pauseStreams: true, _released: true, _streams: [], _currentStream: null, _insideLoop: false, _pendingNext: false, _boundary: '--------------------------756925804713543151415896', _events: [Object: null prototype] { error: [Function: handleStreamError] }, _eventsCount: 1 }, params: { overwrite: true }, httpsAgent: null, paramsSerializer: [Function: paramsSerializer] } at createFailError (dev_cli_errors.ts:27:24) at KbnClientImportExport.req (kbn_client_import_export.ts:162:28) at processTicksAndRejections (node:internal/process/task_queues:95:5) at KbnClientImportExport.load (kbn_client_import_export.ts:67:18) at Context.<anonymous> (index.js:20:7) at Object.apply (wrap_function.js:73:16) { exitCode: 1, showHelp: false, [Symbol(fail error)]: true } ``` First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/25959#0185e4db-a23b-4b2c-89cf-a6a19a47024b) <!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/maps/group2/index·js","test.name":"maps app \"before all\" hook in \"maps app\"","test.failCount":1}} -->
1.0
Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/maps/group2/index·js - maps app "before all" hook in "maps app" - A test failed on a tracked branch ``` Error: 400 resp: '' req: { transitional: { silentJSONParsing: true, forcedJSONParsing: true, clarifyTimeoutError: false }, adapter: [Function: httpAdapter], transformRequest: [ [Function: transformRequest] ], transformResponse: [ [Function: transformResponse] ], timeout: 0, xsrfCookieName: 'XSRF-TOKEN', xsrfHeaderName: 'X-XSRF-TOKEN', maxContentLength: 30000000, maxBodyLength: 30000000, env: { FormData: [Function: FormData] { LINE_BREAK: '\r\n', DEFAULT_CONTENT_TYPE: 'application/octet-stream' } }, validateStatus: [Function: validateStatus], headers: { Accept: 'application/json, text/plain, */*', 'Content-Type': 'multipart/form-data; boundary=--------------------------756925804713543151415896', 'kbn-xsrf': 'kbn-client', 'User-Agent': 'axios/0.27.2', 'content-type': 'multipart/form-data; boundary=--------------------------756925804713543151415896' }, method: 'post', url: 'http://elastic:changeme@localhost:5620/api/saved_objects/_import', data: FormData { _overheadLength: 129, _valueLength: 73008, _valuesToMeasure: [], writable: false, readable: true, dataSize: 0, maxDataSize: 2097152, pauseStreams: true, _released: true, _streams: [], _currentStream: null, _insideLoop: false, _pendingNext: false, _boundary: '--------------------------756925804713543151415896', _events: [Object: null prototype] { error: [Function: handleStreamError] }, _eventsCount: 1 }, params: { overwrite: true }, httpsAgent: null, paramsSerializer: [Function: paramsSerializer] } at createFailError (dev_cli_errors.ts:27:24) at KbnClientImportExport.req (kbn_client_import_export.ts:162:28) at processTicksAndRejections (node:internal/process/task_queues:95:5) at KbnClientImportExport.load (kbn_client_import_export.ts:67:18) at Context.<anonymous> (index.js:20:7) at Object.apply (wrap_function.js:73:16) { exitCode: 1, showHelp: false, [Symbol(fail error)]: true } ``` First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/25959#0185e4db-a23b-4b2c-89cf-a6a19a47024b) <!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/maps/group2/index·js","test.name":"maps app \"before all\" hook in \"maps app\"","test.failCount":1}} -->
test
failing test chrome x pack ui functional tests x pack test functional apps maps index·js maps app before all hook in maps app a test failed on a tracked branch error resp req transitional silentjsonparsing true forcedjsonparsing true clarifytimeouterror false adapter transformrequest transformresponse timeout xsrfcookiename xsrf token xsrfheadername x xsrf token maxcontentlength maxbodylength env formdata line break r n default content type application octet stream validatestatus headers accept application json text plain content type multipart form data boundary kbn xsrf kbn client user agent axios content type multipart form data boundary method post url data formdata overheadlength valuelength valuestomeasure writable false readable true datasize maxdatasize pausestreams true released true streams currentstream null insideloop false pendingnext false boundary events error eventscount params overwrite true httpsagent null paramsserializer at createfailerror dev cli errors ts at kbnclientimportexport req kbn client import export ts at processticksandrejections node internal process task queues at kbnclientimportexport load kbn client import export ts at context index js at object apply wrap function js exitcode showhelp false true first failure
1
440,508
30,748,258,562
IssuesEvent
2023-07-28 16:47:15
scikit-learn/scikit-learn
https://api.github.com/repos/scikit-learn/scikit-learn
closed
[DOCS] Missing values are now supported in Decision Trees
Documentation module:tree
v1.3 of scikit-learn introduced some missing value support as evident in the same documentation [file](https://scikit-learn.org/stable/modules/tree.html#tree-missing-value-support) later on but it still states in the beginning of the dos that missing values are ["not supported in this module"](https://github.com/scikit-learn/scikit-learn/blob/1090121815f983a96a3a903987aea92c804bb740/doc/modules/tree.rst?plain=1#L30-L31). I think it would be fine to just remove that sentence or reference the section explaining the compatibility with missing values.
1.0
[DOCS] Missing values are now supported in Decision Trees - v1.3 of scikit-learn introduced some missing value support as evident in the same documentation [file](https://scikit-learn.org/stable/modules/tree.html#tree-missing-value-support) later on but it still states in the beginning of the dos that missing values are ["not supported in this module"](https://github.com/scikit-learn/scikit-learn/blob/1090121815f983a96a3a903987aea92c804bb740/doc/modules/tree.rst?plain=1#L30-L31). I think it would be fine to just remove that sentence or reference the section explaining the compatibility with missing values.
non_test
missing values are now supported in decision trees of scikit learn introduced some missing value support as evident in the same documentation later on but it still states in the beginning of the dos that missing values are i think it would be fine to just remove that sentence or reference the section explaining the compatibility with missing values
0
241,230
20,110,765,039
IssuesEvent
2022-02-07 14:54:04
bithyve/hexa
https://api.github.com/repos/bithyve/hexa
closed
Post restore:Test account issue.
fixed Test Account 2.0.70
![Screenshot_20220204-152752](https://user-images.githubusercontent.com/62699947/152509934-b3cc93b7-cbcf-4699-b798-51a1be793006.jpg) ![Screenshot_20220204-152810](https://user-images.githubusercontent.com/62699947/152509988-4fd52ed9-30da-42f0-bec0-9f46fc48916b.jpg) - Post recovery When I open the test account, the UI was different and confirmation was undefined.on refresh the UI was different.
1.0
Post restore:Test account issue. - ![Screenshot_20220204-152752](https://user-images.githubusercontent.com/62699947/152509934-b3cc93b7-cbcf-4699-b798-51a1be793006.jpg) ![Screenshot_20220204-152810](https://user-images.githubusercontent.com/62699947/152509988-4fd52ed9-30da-42f0-bec0-9f46fc48916b.jpg) - Post recovery When I open the test account, the UI was different and confirmation was undefined.on refresh the UI was different.
test
post restore test account issue post recovery when i open the test account the ui was different and confirmation was undefined on refresh the ui was different
1
115,195
9,783,620,131
IssuesEvent
2019-06-08 12:01:06
imixs/imixs-workflow
https://api.github.com/repos/imixs/imixs-workflow
closed
SchedulerService must not throw application exceptions
bug testing
The bean: SchedulerService has timeout method onTimeout which must not throw application exceptions
1.0
SchedulerService must not throw application exceptions - The bean: SchedulerService has timeout method onTimeout which must not throw application exceptions
test
schedulerservice must not throw application exceptions the bean schedulerservice has timeout method ontimeout which must not throw application exceptions
1
248,020
20,989,772,430
IssuesEvent
2022-03-29 08:15:32
OpenSID/OpenSID
https://api.github.com/repos/OpenSID/OpenSID
closed
[Premium V22.03-Rev02] Setelah Impor Database melalui PhpMyadmin, Database berubah bukan menjadi Collation utf8_general_ci
bug tester-22.03
### Jelaskan error yg dialami 1. Impor Data_contoh_awal ataupun dari data base backup melalui PHPMYADMIN, Setelah di Impor Database berubah bukan menjadi Collation utf8_general_ci. 2. Hal ini terjadi baik impor di locolhost maupun impor pada hosting 3. Pada hosting setelah impor database, ujicoba loading->website menjadi error 4. Pilih Perintah "Periksa" --> Pilih Tombol "Perbaiki" -> Database juga tidak terperbaiki ### Cara untuk mereplikasi errornya 1. Lakukan import data melalui phpmyadmin baik di localhost maupun hosting. 2. Jalankan System->dan Login ke Database 3. Pilih Perintah "Periksa" di url _>http://localhost/Premium/periksa ### Hasil yg diharapkan Agar bisa normal dan database tidak merubah Collation dari yang diinginkan dengan Collation type utf8_general_ci, baik impor di localhost maupun di hosting. ### Tangkapan layar dan log error ![image](https://user-images.githubusercontent.com/64481641/158544249-5214cc4b-7bac-4ef2-a01e-99fd1b6532b6.png) Log Error Loading awal setelah mengimpor database dari contoh_data_awal pada hosting <?php defined('BASEPATH') OR exit('No direct script access allowed'); ?> ERROR - 2022-03-16 03:28:21 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 03:28:21 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 05:04:31 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 05:04:31 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 05:04:53 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 05:04:53 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 05:06:36 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 05:06:36 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 11:05:51 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 11:05:51 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 11:06:57 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 11:06:57 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 13:56:47 --> 404 Page Not Found: ERROR - 2022-03-16 13:57:55 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 13:57:56 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 07:18:03 --> 404 Page Not Found: Wp-loginphp/index ERROR - 2022-03-16 14:32:14 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 14:32:15 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 14:33:11 --> 404 Page Not Found: ERROR - 2022-03-16 07:49:36 --> Unable to connect to the database ERROR - 2022-03-16 07:49:36 --> Severity: error --> Exception: Call to a member function error() on string /home/teupinbe/public_html/donjo-app/core/MY_Security.php(226) : eval()'d code(1) : eval()'d code 70 ERROR - 2022-03-16 07:49:38 --> Unable to connect to the database ERROR - 2022-03-16 07:49:38 --> Severity: error --> Exception: Call to a member function error() on string /home/teupinbe/public_html/donjo-app/core/MY_Security.php(226) : eval()'d code(1) : eval()'d code 70 ERROR - 2022-03-16 07:49:40 --> Unable to connect to the database ERROR - 2022-03-16 07:49:40 --> Severity: error --> Exception: Call to a member function error() on string /home/teupinbe/public_html/donjo-app/core/MY_Security.php(226) : eval()'d code(1) : eval()'d code 70 ERROR - 2022-03-16 07:49:56 --> Unable to connect to the database ERROR - 2022-03-16 07:49:56 --> Severity: error --> Exception: Call to a member function error() on string /home/teupinbe/public_html/donjo-app/core/MY_Security.php(226) : eval()'d code(1) : eval()'d code 70 ERROR - 2022-03-16 07:50:42 --> Unable to connect to the database ERROR - 2022-03-16 07:50:42 --> Severity: error --> Exception: Call to a member function error() on string /home/teupinbe/public_html/donjo-app/core/MY_Security.php(226) : eval()'d code(1) : eval()'d code 70 ERROR - 2022-03-16 14:51:18 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 14:51:19 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com**** ### Rilis Versi OpenSID Rilis Premium ### Versi OpenSID Premium (v22.03-Rev01) ### Tema Yang Digunakan default ### Informasi tambahan .
1.0
[Premium V22.03-Rev02] Setelah Impor Database melalui PhpMyadmin, Database berubah bukan menjadi Collation utf8_general_ci - ### Jelaskan error yg dialami 1. Impor Data_contoh_awal ataupun dari data base backup melalui PHPMYADMIN, Setelah di Impor Database berubah bukan menjadi Collation utf8_general_ci. 2. Hal ini terjadi baik impor di locolhost maupun impor pada hosting 3. Pada hosting setelah impor database, ujicoba loading->website menjadi error 4. Pilih Perintah "Periksa" --> Pilih Tombol "Perbaiki" -> Database juga tidak terperbaiki ### Cara untuk mereplikasi errornya 1. Lakukan import data melalui phpmyadmin baik di localhost maupun hosting. 2. Jalankan System->dan Login ke Database 3. Pilih Perintah "Periksa" di url _>http://localhost/Premium/periksa ### Hasil yg diharapkan Agar bisa normal dan database tidak merubah Collation dari yang diinginkan dengan Collation type utf8_general_ci, baik impor di localhost maupun di hosting. ### Tangkapan layar dan log error ![image](https://user-images.githubusercontent.com/64481641/158544249-5214cc4b-7bac-4ef2-a01e-99fd1b6532b6.png) Log Error Loading awal setelah mengimpor database dari contoh_data_awal pada hosting <?php defined('BASEPATH') OR exit('No direct script access allowed'); ?> ERROR - 2022-03-16 03:28:21 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 03:28:21 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 05:04:31 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 05:04:31 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 05:04:53 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 05:04:53 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 05:06:36 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 05:06:36 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 11:05:51 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 11:05:51 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 11:06:57 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 11:06:57 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 13:56:47 --> 404 Page Not Found: ERROR - 2022-03-16 13:57:55 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 13:57:56 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 07:18:03 --> 404 Page Not Found: Wp-loginphp/index ERROR - 2022-03-16 14:32:14 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 14:32:15 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 14:33:11 --> 404 Page Not Found: ERROR - 2022-03-16 07:49:36 --> Unable to connect to the database ERROR - 2022-03-16 07:49:36 --> Severity: error --> Exception: Call to a member function error() on string /home/teupinbe/public_html/donjo-app/core/MY_Security.php(226) : eval()'d code(1) : eval()'d code 70 ERROR - 2022-03-16 07:49:38 --> Unable to connect to the database ERROR - 2022-03-16 07:49:38 --> Severity: error --> Exception: Call to a member function error() on string /home/teupinbe/public_html/donjo-app/core/MY_Security.php(226) : eval()'d code(1) : eval()'d code 70 ERROR - 2022-03-16 07:49:40 --> Unable to connect to the database ERROR - 2022-03-16 07:49:40 --> Severity: error --> Exception: Call to a member function error() on string /home/teupinbe/public_html/donjo-app/core/MY_Security.php(226) : eval()'d code(1) : eval()'d code 70 ERROR - 2022-03-16 07:49:56 --> Unable to connect to the database ERROR - 2022-03-16 07:49:56 --> Severity: error --> Exception: Call to a member function error() on string /home/teupinbe/public_html/donjo-app/core/MY_Security.php(226) : eval()'d code(1) : eval()'d code 70 ERROR - 2022-03-16 07:50:42 --> Unable to connect to the database ERROR - 2022-03-16 07:50:42 --> Severity: error --> Exception: Call to a member function error() on string /home/teupinbe/public_html/donjo-app/core/MY_Security.php(226) : eval()'d code(1) : eval()'d code 70 ERROR - 2022-03-16 14:51:18 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com ERROR - 2022-03-16 14:51:19 --> Error occured while loading url by cURL. <br /> Could not resolve host: api.kawalcorona.com**** ### Rilis Versi OpenSID Rilis Premium ### Versi OpenSID Premium (v22.03-Rev01) ### Tema Yang Digunakan default ### Informasi tambahan .
test
setelah impor database melalui phpmyadmin database berubah bukan menjadi collation general ci jelaskan error yg dialami impor data contoh awal ataupun dari data base backup melalui phpmyadmin setelah di impor database berubah bukan menjadi collation general ci hal ini terjadi baik impor di locolhost maupun impor pada hosting pada hosting setelah impor database ujicoba loading website menjadi error pilih perintah periksa pilih tombol perbaiki database juga tidak terperbaiki cara untuk mereplikasi errornya lakukan import data melalui phpmyadmin baik di localhost maupun hosting jalankan system dan login ke database pilih perintah periksa di url hasil yg diharapkan agar bisa normal dan database tidak merubah collation dari yang diinginkan dengan collation type general ci baik impor di localhost maupun di hosting tangkapan layar dan log error log error loading awal setelah mengimpor database dari contoh data awal pada hosting error error occured while loading url by curl could not resolve host api kawalcorona com error error occured while loading url by curl could not resolve host api kawalcorona com error error occured while loading url by curl could not resolve host api kawalcorona com error error occured while loading url by curl could not resolve host api kawalcorona com error error occured while loading url by curl could not resolve host api kawalcorona com error error occured while loading url by curl could not resolve host api kawalcorona com error error occured while loading url by curl could not resolve host api kawalcorona com error error occured while loading url by curl could not resolve host api kawalcorona com error error occured while loading url by curl could not resolve host api kawalcorona com error error occured while loading url by curl could not resolve host api kawalcorona com error error occured while loading url by curl could not resolve host api kawalcorona com error error occured while loading url by curl could not resolve host api kawalcorona com error page not found error error occured while loading url by curl could not resolve host api kawalcorona com error error occured while loading url by curl could not resolve host api kawalcorona com error page not found wp loginphp index error error occured while loading url by curl could not resolve host api kawalcorona com error error occured while loading url by curl could not resolve host api kawalcorona com error page not found error unable to connect to the database error severity error exception call to a member function error on string home teupinbe public html donjo app core my security php eval d code eval d code error unable to connect to the database error severity error exception call to a member function error on string home teupinbe public html donjo app core my security php eval d code eval d code error unable to connect to the database error severity error exception call to a member function error on string home teupinbe public html donjo app core my security php eval d code eval d code error unable to connect to the database error severity error exception call to a member function error on string home teupinbe public html donjo app core my security php eval d code eval d code error unable to connect to the database error severity error exception call to a member function error on string home teupinbe public html donjo app core my security php eval d code eval d code error error occured while loading url by curl could not resolve host api kawalcorona com error error occured while loading url by curl could not resolve host api kawalcorona com rilis versi opensid rilis premium versi opensid premium tema yang digunakan default informasi tambahan
1
185,963
14,394,532,873
IssuesEvent
2020-12-03 01:31:00
github-vet/rangeclosure-findings
https://api.github.com/repos/github-vet/rangeclosure-findings
closed
deferpanic/gorump: 1.7/go/src/time/format_test.go; 8 LoC
fresh test tiny
Found a possible issue in [deferpanic/gorump](https://www.github.com/deferpanic/gorump) at [1.7/go/src/time/format_test.go](https://github.com/deferpanic/gorump/blob/313ecc2ef408fbfd85123cdfcf448042787b53ea/1.7/go/src/time/format_test.go#L186-L193) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to test at line 191 may start a goroutine [Click here to see the code in its original context.](https://github.com/deferpanic/gorump/blob/313ecc2ef408fbfd85123cdfcf448042787b53ea/1.7/go/src/time/format_test.go#L186-L193) <details> <summary>Click here to show the 8 line(s) of Go which triggered the analyzer.</summary> ```go for _, test := range parseTests { time, err := Parse(test.format, test.value) if err != nil { t.Errorf("%s error: %v", test.name, err) } else { checkTime(time, &test, t) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 313ecc2ef408fbfd85123cdfcf448042787b53ea
1.0
deferpanic/gorump: 1.7/go/src/time/format_test.go; 8 LoC - Found a possible issue in [deferpanic/gorump](https://www.github.com/deferpanic/gorump) at [1.7/go/src/time/format_test.go](https://github.com/deferpanic/gorump/blob/313ecc2ef408fbfd85123cdfcf448042787b53ea/1.7/go/src/time/format_test.go#L186-L193) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to test at line 191 may start a goroutine [Click here to see the code in its original context.](https://github.com/deferpanic/gorump/blob/313ecc2ef408fbfd85123cdfcf448042787b53ea/1.7/go/src/time/format_test.go#L186-L193) <details> <summary>Click here to show the 8 line(s) of Go which triggered the analyzer.</summary> ```go for _, test := range parseTests { time, err := Parse(test.format, test.value) if err != nil { t.Errorf("%s error: %v", test.name, err) } else { checkTime(time, &test, t) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 313ecc2ef408fbfd85123cdfcf448042787b53ea
test
deferpanic gorump go src time format test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to test at line may start a goroutine click here to show the line s of go which triggered the analyzer go for test range parsetests time err parse test format test value if err nil t errorf s error v test name err else checktime time test t leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
289,817
25,015,920,752
IssuesEvent
2022-11-03 18:48:00
lowRISC/opentitan
https://api.github.com/repos/lowRISC/opentitan
closed
//sw/device/tests:entropy_src_ast_rng_req_test_sim_verilator Times out waiting for entropy
Type:Bug Component:TestTriage
Timed out after 1000 usec üI00000 test_rom.c:133] Version: earlgrey_silver_release_v5-8368-g62844d571, Build Date: 2022-11-02 10:54:46^M^M I00001 test_rom.c:235] Test ROM complete, jumping to flash (addr: 20000480)!^M^M I00000 ottf_main.c:126] Running sw/device/tests/entropy_src_ast_rng_req_test.c^M^M E00001 entropy_src_ast_rng_req_test.c:60] CHECK-fail: Timed out after 1000 usec (500 CPU cycles) waiting for read_fifo_depth(&entropy_src) > 0^M^M I00002 status.c:34] FAIL!^M^M
1.0
//sw/device/tests:entropy_src_ast_rng_req_test_sim_verilator Times out waiting for entropy - Timed out after 1000 usec üI00000 test_rom.c:133] Version: earlgrey_silver_release_v5-8368-g62844d571, Build Date: 2022-11-02 10:54:46^M^M I00001 test_rom.c:235] Test ROM complete, jumping to flash (addr: 20000480)!^M^M I00000 ottf_main.c:126] Running sw/device/tests/entropy_src_ast_rng_req_test.c^M^M E00001 entropy_src_ast_rng_req_test.c:60] CHECK-fail: Timed out after 1000 usec (500 CPU cycles) waiting for read_fifo_depth(&entropy_src) > 0^M^M I00002 status.c:34] FAIL!^M^M
test
sw device tests entropy src ast rng req test sim verilator times out waiting for entropy timed out after usec test rom c version earlgrey silver release build date m m test rom c test rom complete jumping to flash addr m m ottf main c running sw device tests entropy src ast rng req test c m m entropy src ast rng req test c check fail timed out after usec cpu cycles waiting for read fifo depth entropy src m m status c fail m m
1
206,931
7,123,013,195
IssuesEvent
2018-01-19 14:04:02
fxi/map-x-mgl
https://api.github.com/repos/fxi/map-x-mgl
closed
Layout tweak: adding mapx logo to the app
Priority 1
Can the new mapx logo be added to the mapx app engine. Can the logo actually have a pop up tag called training and can it link to the training pages on the main landing site ?
1.0
Layout tweak: adding mapx logo to the app - Can the new mapx logo be added to the mapx app engine. Can the logo actually have a pop up tag called training and can it link to the training pages on the main landing site ?
non_test
layout tweak adding mapx logo to the app can the new mapx logo be added to the mapx app engine can the logo actually have a pop up tag called training and can it link to the training pages on the main landing site
0
67,681
7,057,564,497
IssuesEvent
2018-01-04 16:53:28
JuliaGraphs/LightGraphs.jl
https://api.github.com/repos/JuliaGraphs/LightGraphs.jl
closed
Nonbacktracking matrix eigenvalues are sometime NaN
CI / tests wontfix
Sometimes when you compute the eigenvalues of the Nonbacktracking matrix the leading eigenvalue comes out as NaN. This is happening on the explicitly computed matrix and happens on both 0.5 and 0.6. I need to investigate for changes to eigs that make it more common on 0.6.
1.0
Nonbacktracking matrix eigenvalues are sometime NaN - Sometimes when you compute the eigenvalues of the Nonbacktracking matrix the leading eigenvalue comes out as NaN. This is happening on the explicitly computed matrix and happens on both 0.5 and 0.6. I need to investigate for changes to eigs that make it more common on 0.6.
test
nonbacktracking matrix eigenvalues are sometime nan sometimes when you compute the eigenvalues of the nonbacktracking matrix the leading eigenvalue comes out as nan this is happening on the explicitly computed matrix and happens on both and i need to investigate for changes to eigs that make it more common on
1
427,315
12,393,982,397
IssuesEvent
2020-05-20 16:13:30
googleapis/elixir-google-api
https://api.github.com/repos/googleapis/elixir-google-api
closed
Synthesis failed for SQLAdmin
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate SQLAdmin. :broken_heart: Here's the output from running `synth.py`: ``` led to remove deps/parse_trans/ebin/parse_trans.app: Permission denied warning: failed to remove deps/parse_trans/ebin/parse_trans_mod.beam: Permission denied warning: failed to remove deps/parse_trans/ebin/parse_trans_codegen.beam: Permission denied warning: failed to remove deps/parse_trans/ebin/ct_expand.beam: Permission denied warning: failed to remove deps/parse_trans/ebin/parse_trans.beam: Permission denied warning: failed to remove deps/parse_trans/ebin/exprecs.beam: Permission denied warning: failed to remove deps/parse_trans/ebin/parse_trans_pp.beam: Permission denied warning: failed to remove deps/parse_trans/.rebar3/erlcinfo: Permission denied warning: failed to remove deps/parse_trans/hex_metadata.config: Permission denied warning: failed to remove deps/parse_trans/README.md: Permission denied warning: failed to remove deps/parse_trans/rebar.config: Permission denied warning: failed to remove deps/parse_trans/include/codegen.hrl: Permission denied warning: failed to remove deps/parse_trans/include/exprecs.hrl: Permission denied warning: failed to remove deps/parse_trans/.fetch: Permission denied warning: failed to remove deps/parse_trans/.hex: Permission denied warning: failed to remove deps/idna/LICENSE: Permission denied warning: failed to remove deps/idna/rebar.lock: Permission denied warning: failed to remove deps/idna/src/idna.erl: Permission denied warning: failed to remove deps/idna/src/idna_logger.hrl: Permission denied warning: failed to remove deps/idna/src/idna_ucs.erl: Permission denied warning: failed to remove deps/idna/src/punycode.erl: Permission denied warning: failed to remove deps/idna/src/idna_table.erl: Permission denied warning: failed to remove deps/idna/src/idna_context.erl: Permission denied warning: failed to remove deps/idna/src/idna.app.src: Permission denied warning: failed to remove deps/idna/src/idna_mapping.erl: Permission denied warning: failed to remove deps/idna/src/idna_data.erl: Permission denied warning: failed to remove deps/idna/src/idna_bidi.erl: Permission denied warning: failed to remove deps/idna/ebin/idna_mapping.beam: Permission denied warning: failed to remove deps/idna/ebin/idna_context.beam: Permission denied warning: failed to remove deps/idna/ebin/idna_bidi.beam: Permission denied warning: failed to remove deps/idna/ebin/punycode.beam: Permission denied warning: failed to remove deps/idna/ebin/idna_table.beam: Permission denied warning: failed to remove deps/idna/ebin/idna_data.beam: Permission denied warning: failed to remove deps/idna/ebin/idna_ucs.beam: Permission denied warning: failed to remove deps/idna/ebin/idna.app: Permission denied warning: failed to remove deps/idna/ebin/idna.beam: Permission denied warning: failed to remove deps/idna/.rebar3/erlcinfo: Permission denied warning: failed to remove deps/idna/hex_metadata.config: Permission denied warning: failed to remove deps/idna/README.md: Permission denied warning: failed to remove deps/idna/rebar.config: Permission denied warning: failed to remove deps/idna/.fetch: Permission denied warning: failed to remove deps/idna/rebar.config.script: Permission denied warning: failed to remove deps/idna/.hex: Permission denied warning: failed to remove deps/hackney/MAINTAINERS: Permission denied warning: failed to remove deps/hackney/LICENSE: Permission denied warning: failed to remove deps/hackney/rebar.lock: Permission denied warning: failed to remove deps/hackney/src/hackney_ssl.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_response.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_tcp.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_http.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_cookie.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_url.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_headers.erl: Permission denied warning: failed to remove deps/hackney/src/hackney.app.src: Permission denied warning: failed to remove deps/hackney/src/hackney_pool_handler.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_trace.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_multipart.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_headers_new.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_http_connect.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_util.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_socks5.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_request.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_app.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_internal.hrl: Permission denied warning: failed to remove deps/hackney/src/hackney_date.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_manager.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_connect.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_bstr.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_sup.erl: Permission denied warning: failed to remove deps/hackney/src/hackney.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_local_tcp.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_stream.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_pool.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_metrics.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_methods.hrl: Permission denied warning: failed to remove deps/hackney/NOTICE: Permission denied warning: failed to remove deps/hackney/ebin/hackney_pool.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_trace.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_pool_handler.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_headers.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_url.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_manager.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_metrics.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_stream.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_sup.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_multipart.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_http.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_socks5.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_app.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_http_connect.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_response.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney.app: Permission denied warning: failed to remove deps/hackney/ebin/hackney_headers_new.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_cookie.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_request.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_util.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_connect.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_date.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_ssl.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_bstr.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_tcp.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_local_tcp.beam: Permission denied warning: failed to remove deps/hackney/.rebar3/erlcinfo: Permission denied warning: failed to remove deps/hackney/hex_metadata.config: Permission denied warning: failed to remove deps/hackney/README.md: Permission denied warning: failed to remove deps/hackney/rebar.config: Permission denied warning: failed to remove deps/hackney/include/hackney.hrl: Permission denied warning: failed to remove deps/hackney/include/hackney_lib.hrl: Permission denied warning: failed to remove deps/hackney/.fetch: Permission denied warning: failed to remove deps/hackney/.hex: Permission denied warning: failed to remove deps/hackney/NEWS.md: Permission denied Removing __pycache__/ Removing specifications/gdd/SQLAdmin-v1beta4.json Traceback (most recent call last): File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 559, in _inner_main sys.exit(EXIT_CODE_SKIPPED) SystemExit: 28 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 615, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 476, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 611, in _inner_main executor.check_call(["git", "clean", "-fdx"], cwd=working_repo_path) File "/tmpfs/src/github/synthtool/autosynth/executor.py", line 29, in check_call subprocess.check_call(command, **args) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 311, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command '['git', 'clean', '-fdx']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge/c19bc74d-4d42-4305-94c3-323bf90f3ebc).
1.0
Synthesis failed for SQLAdmin - Hello! Autosynth couldn't regenerate SQLAdmin. :broken_heart: Here's the output from running `synth.py`: ``` led to remove deps/parse_trans/ebin/parse_trans.app: Permission denied warning: failed to remove deps/parse_trans/ebin/parse_trans_mod.beam: Permission denied warning: failed to remove deps/parse_trans/ebin/parse_trans_codegen.beam: Permission denied warning: failed to remove deps/parse_trans/ebin/ct_expand.beam: Permission denied warning: failed to remove deps/parse_trans/ebin/parse_trans.beam: Permission denied warning: failed to remove deps/parse_trans/ebin/exprecs.beam: Permission denied warning: failed to remove deps/parse_trans/ebin/parse_trans_pp.beam: Permission denied warning: failed to remove deps/parse_trans/.rebar3/erlcinfo: Permission denied warning: failed to remove deps/parse_trans/hex_metadata.config: Permission denied warning: failed to remove deps/parse_trans/README.md: Permission denied warning: failed to remove deps/parse_trans/rebar.config: Permission denied warning: failed to remove deps/parse_trans/include/codegen.hrl: Permission denied warning: failed to remove deps/parse_trans/include/exprecs.hrl: Permission denied warning: failed to remove deps/parse_trans/.fetch: Permission denied warning: failed to remove deps/parse_trans/.hex: Permission denied warning: failed to remove deps/idna/LICENSE: Permission denied warning: failed to remove deps/idna/rebar.lock: Permission denied warning: failed to remove deps/idna/src/idna.erl: Permission denied warning: failed to remove deps/idna/src/idna_logger.hrl: Permission denied warning: failed to remove deps/idna/src/idna_ucs.erl: Permission denied warning: failed to remove deps/idna/src/punycode.erl: Permission denied warning: failed to remove deps/idna/src/idna_table.erl: Permission denied warning: failed to remove deps/idna/src/idna_context.erl: Permission denied warning: failed to remove deps/idna/src/idna.app.src: Permission denied warning: failed to remove deps/idna/src/idna_mapping.erl: Permission denied warning: failed to remove deps/idna/src/idna_data.erl: Permission denied warning: failed to remove deps/idna/src/idna_bidi.erl: Permission denied warning: failed to remove deps/idna/ebin/idna_mapping.beam: Permission denied warning: failed to remove deps/idna/ebin/idna_context.beam: Permission denied warning: failed to remove deps/idna/ebin/idna_bidi.beam: Permission denied warning: failed to remove deps/idna/ebin/punycode.beam: Permission denied warning: failed to remove deps/idna/ebin/idna_table.beam: Permission denied warning: failed to remove deps/idna/ebin/idna_data.beam: Permission denied warning: failed to remove deps/idna/ebin/idna_ucs.beam: Permission denied warning: failed to remove deps/idna/ebin/idna.app: Permission denied warning: failed to remove deps/idna/ebin/idna.beam: Permission denied warning: failed to remove deps/idna/.rebar3/erlcinfo: Permission denied warning: failed to remove deps/idna/hex_metadata.config: Permission denied warning: failed to remove deps/idna/README.md: Permission denied warning: failed to remove deps/idna/rebar.config: Permission denied warning: failed to remove deps/idna/.fetch: Permission denied warning: failed to remove deps/idna/rebar.config.script: Permission denied warning: failed to remove deps/idna/.hex: Permission denied warning: failed to remove deps/hackney/MAINTAINERS: Permission denied warning: failed to remove deps/hackney/LICENSE: Permission denied warning: failed to remove deps/hackney/rebar.lock: Permission denied warning: failed to remove deps/hackney/src/hackney_ssl.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_response.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_tcp.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_http.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_cookie.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_url.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_headers.erl: Permission denied warning: failed to remove deps/hackney/src/hackney.app.src: Permission denied warning: failed to remove deps/hackney/src/hackney_pool_handler.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_trace.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_multipart.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_headers_new.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_http_connect.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_util.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_socks5.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_request.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_app.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_internal.hrl: Permission denied warning: failed to remove deps/hackney/src/hackney_date.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_manager.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_connect.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_bstr.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_sup.erl: Permission denied warning: failed to remove deps/hackney/src/hackney.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_local_tcp.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_stream.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_pool.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_metrics.erl: Permission denied warning: failed to remove deps/hackney/src/hackney_methods.hrl: Permission denied warning: failed to remove deps/hackney/NOTICE: Permission denied warning: failed to remove deps/hackney/ebin/hackney_pool.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_trace.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_pool_handler.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_headers.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_url.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_manager.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_metrics.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_stream.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_sup.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_multipart.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_http.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_socks5.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_app.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_http_connect.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_response.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney.app: Permission denied warning: failed to remove deps/hackney/ebin/hackney_headers_new.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_cookie.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_request.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_util.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_connect.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_date.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_ssl.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_bstr.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_tcp.beam: Permission denied warning: failed to remove deps/hackney/ebin/hackney_local_tcp.beam: Permission denied warning: failed to remove deps/hackney/.rebar3/erlcinfo: Permission denied warning: failed to remove deps/hackney/hex_metadata.config: Permission denied warning: failed to remove deps/hackney/README.md: Permission denied warning: failed to remove deps/hackney/rebar.config: Permission denied warning: failed to remove deps/hackney/include/hackney.hrl: Permission denied warning: failed to remove deps/hackney/include/hackney_lib.hrl: Permission denied warning: failed to remove deps/hackney/.fetch: Permission denied warning: failed to remove deps/hackney/.hex: Permission denied warning: failed to remove deps/hackney/NEWS.md: Permission denied Removing __pycache__/ Removing specifications/gdd/SQLAdmin-v1beta4.json Traceback (most recent call last): File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 559, in _inner_main sys.exit(EXIT_CODE_SKIPPED) SystemExit: 28 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 615, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 476, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 611, in _inner_main executor.check_call(["git", "clean", "-fdx"], cwd=working_repo_path) File "/tmpfs/src/github/synthtool/autosynth/executor.py", line 29, in check_call subprocess.check_call(command, **args) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 311, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command '['git', 'clean', '-fdx']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge/c19bc74d-4d42-4305-94c3-323bf90f3ebc).
non_test
synthesis failed for sqladmin hello autosynth couldn t regenerate sqladmin broken heart here s the output from running synth py led to remove deps parse trans ebin parse trans app permission denied warning failed to remove deps parse trans ebin parse trans mod beam permission denied warning failed to remove deps parse trans ebin parse trans codegen beam permission denied warning failed to remove deps parse trans ebin ct expand beam permission denied warning failed to remove deps parse trans ebin parse trans beam permission denied warning failed to remove deps parse trans ebin exprecs beam permission denied warning failed to remove deps parse trans ebin parse trans pp beam permission denied warning failed to remove deps parse trans erlcinfo permission denied warning failed to remove deps parse trans hex metadata config permission denied warning failed to remove deps parse trans readme md permission denied warning failed to remove deps parse trans rebar config permission denied warning failed to remove deps parse trans include codegen hrl permission denied warning failed to remove deps parse trans include exprecs hrl permission denied warning failed to remove deps parse trans fetch permission denied warning failed to remove deps parse trans hex permission denied warning failed to remove deps idna license permission denied warning failed to remove deps idna rebar lock permission denied warning failed to remove deps idna src idna erl permission denied warning failed to remove deps idna src idna logger hrl permission denied warning failed to remove deps idna src idna ucs erl permission denied warning failed to remove deps idna src punycode erl permission denied warning failed to remove deps idna src idna table erl permission denied warning failed to remove deps idna src idna context erl permission denied warning failed to remove deps idna src idna app src permission denied warning failed to remove deps idna src idna mapping erl permission denied warning failed to remove deps idna src idna data erl permission denied warning failed to remove deps idna src idna bidi erl permission denied warning failed to remove deps idna ebin idna mapping beam permission denied warning failed to remove deps idna ebin idna context beam permission denied warning failed to remove deps idna ebin idna bidi beam permission denied warning failed to remove deps idna ebin punycode beam permission denied warning failed to remove deps idna ebin idna table beam permission denied warning failed to remove deps idna ebin idna data beam permission denied warning failed to remove deps idna ebin idna ucs beam permission denied warning failed to remove deps idna ebin idna app permission denied warning failed to remove deps idna ebin idna beam permission denied warning failed to remove deps idna erlcinfo permission denied warning failed to remove deps idna hex metadata config permission denied warning failed to remove deps idna readme md permission denied warning failed to remove deps idna rebar config permission denied warning failed to remove deps idna fetch permission denied warning failed to remove deps idna rebar config script permission denied warning failed to remove deps idna hex permission denied warning failed to remove deps hackney maintainers permission denied warning failed to remove deps hackney license permission denied warning failed to remove deps hackney rebar lock permission denied warning failed to remove deps hackney src hackney ssl erl permission denied warning failed to remove deps hackney src hackney response erl permission denied warning failed to remove deps hackney src hackney tcp erl permission denied warning failed to remove deps hackney src hackney http erl permission denied warning failed to remove deps hackney src hackney cookie erl permission denied warning failed to remove deps hackney src hackney url erl permission denied warning failed to remove deps hackney src hackney headers erl permission denied warning failed to remove deps hackney src hackney app src permission denied warning failed to remove deps hackney src hackney pool handler erl permission denied warning failed to remove deps hackney src hackney trace erl permission denied warning failed to remove deps hackney src hackney multipart erl permission denied warning failed to remove deps hackney src hackney headers new erl permission denied warning failed to remove deps hackney src hackney http connect erl permission denied warning failed to remove deps hackney src hackney util erl permission denied warning failed to remove deps hackney src hackney erl permission denied warning failed to remove deps hackney src hackney request erl permission denied warning failed to remove deps hackney src hackney app erl permission denied warning failed to remove deps hackney src hackney internal hrl permission denied warning failed to remove deps hackney src hackney date erl permission denied warning failed to remove deps hackney src hackney manager erl permission denied warning failed to remove deps hackney src hackney connect erl permission denied warning failed to remove deps hackney src hackney bstr erl permission denied warning failed to remove deps hackney src hackney sup erl permission denied warning failed to remove deps hackney src hackney erl permission denied warning failed to remove deps hackney src hackney local tcp erl permission denied warning failed to remove deps hackney src hackney stream erl permission denied warning failed to remove deps hackney src hackney pool erl permission denied warning failed to remove deps hackney src hackney metrics erl permission denied warning failed to remove deps hackney src hackney methods hrl permission denied warning failed to remove deps hackney notice permission denied warning failed to remove deps hackney ebin hackney pool beam permission denied warning failed to remove deps hackney ebin hackney trace beam permission denied warning failed to remove deps hackney ebin hackney pool handler beam permission denied warning failed to remove deps hackney ebin hackney beam permission denied warning failed to remove deps hackney ebin hackney headers beam permission denied warning failed to remove deps hackney ebin hackney url beam permission denied warning failed to remove deps hackney ebin hackney manager beam permission denied warning failed to remove deps hackney ebin hackney metrics beam permission denied warning failed to remove deps hackney ebin hackney stream beam permission denied warning failed to remove deps hackney ebin hackney sup beam permission denied warning failed to remove deps hackney ebin hackney multipart beam permission denied warning failed to remove deps hackney ebin hackney http beam permission denied warning failed to remove deps hackney ebin hackney beam permission denied warning failed to remove deps hackney ebin hackney app beam permission denied warning failed to remove deps hackney ebin hackney http connect beam permission denied warning failed to remove deps hackney ebin hackney response beam permission denied warning failed to remove deps hackney ebin hackney app permission denied warning failed to remove deps hackney ebin hackney headers new beam permission denied warning failed to remove deps hackney ebin hackney cookie beam permission denied warning failed to remove deps hackney ebin hackney request beam permission denied warning failed to remove deps hackney ebin hackney util beam permission denied warning failed to remove deps hackney ebin hackney connect beam permission denied warning failed to remove deps hackney ebin hackney date beam permission denied warning failed to remove deps hackney ebin hackney ssl beam permission denied warning failed to remove deps hackney ebin hackney bstr beam permission denied warning failed to remove deps hackney ebin hackney tcp beam permission denied warning failed to remove deps hackney ebin hackney local tcp beam permission denied warning failed to remove deps hackney erlcinfo permission denied warning failed to remove deps hackney hex metadata config permission denied warning failed to remove deps hackney readme md permission denied warning failed to remove deps hackney rebar config permission denied warning failed to remove deps hackney include hackney hrl permission denied warning failed to remove deps hackney include hackney lib hrl permission denied warning failed to remove deps hackney fetch permission denied warning failed to remove deps hackney hex permission denied warning failed to remove deps hackney news md permission denied removing pycache removing specifications gdd sqladmin json traceback most recent call last file tmpfs src github synthtool autosynth synth py line in inner main sys exit exit code skipped systemexit during handling of the above exception another exception occurred traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main executor check call cwd working repo path file tmpfs src github synthtool autosynth executor py line in check call subprocess check call command args file home kbuilder pyenv versions lib subprocess py line in check call raise calledprocesserror retcode cmd subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
0
33,847
9,206,177,855
IssuesEvent
2019-03-08 12:58:20
qissue-bot/QGIS
https://api.github.com/repos/qissue-bot/QGIS
closed
Build error in 1.0.0 (linguist problem in French translation), Qt 4.5.0 beta1
Category: Build/Install Component: Affected QGIS version Component: Crashes QGIS or corrupts data Component: Easy fix? Component: Operating System Component: Pull Request or Patch supplied Component: Regression? Component: Resolution Priority: Low Project: QGIS Application Status: Closed Tracker: Bug report
--- Author Name: **vince -** (vince -) Original Redmine Issue: 1470, https://issues.qgis.org/issues/1470 Original Assignee: nobody - --- I am currently trying to port Qgis to Macports. I have fetched the 1.0.0 source code as found on download.osgeo.org/qgis/src, and I use the prerelease of Qt 4.5.0 (not yet posted to Macports) in order to get 4 way universal builds on [[MacOS]] X (i386, x86_64, ppc, ppc64) (Qt 4.4 and less are 32-bit only). While buildling, I find this error: [ 56%] Generating qgis_fr.qm lrelease error: duplicate messages found in '/usr/pkg/var/macports/build/_Users_vincent_Developpement_ports_gis_qgis/work/qgis_1.0.0/i18n/qgis_fr.ts': * Context: [[QgsLabelDialogBase]] * Source: ° Well, I have dug a bit into the file with the new Linguist (which happens to signal the same error when loading the file), but I must admit I haven't found what could possibly have gone astray. The problem is that this is a fatal error and the build aborts. Any idea on this very basic problem? Merci, Vincent
1.0
Build error in 1.0.0 (linguist problem in French translation), Qt 4.5.0 beta1 - --- Author Name: **vince -** (vince -) Original Redmine Issue: 1470, https://issues.qgis.org/issues/1470 Original Assignee: nobody - --- I am currently trying to port Qgis to Macports. I have fetched the 1.0.0 source code as found on download.osgeo.org/qgis/src, and I use the prerelease of Qt 4.5.0 (not yet posted to Macports) in order to get 4 way universal builds on [[MacOS]] X (i386, x86_64, ppc, ppc64) (Qt 4.4 and less are 32-bit only). While buildling, I find this error: [ 56%] Generating qgis_fr.qm lrelease error: duplicate messages found in '/usr/pkg/var/macports/build/_Users_vincent_Developpement_ports_gis_qgis/work/qgis_1.0.0/i18n/qgis_fr.ts': * Context: [[QgsLabelDialogBase]] * Source: ° Well, I have dug a bit into the file with the new Linguist (which happens to signal the same error when loading the file), but I must admit I haven't found what could possibly have gone astray. The problem is that this is a fatal error and the build aborts. Any idea on this very basic problem? Merci, Vincent
non_test
build error in linguist problem in french translation qt author name vince vince original redmine issue original assignee nobody i am currently trying to port qgis to macports i have fetched the source code as found on download osgeo org qgis src and i use the prerelease of qt not yet posted to macports in order to get way universal builds on x ppc qt and less are bit only while buildling i find this error generating qgis fr qm lrelease error duplicate messages found in usr pkg var macports build users vincent developpement ports gis qgis work qgis qgis fr ts context source ° well i have dug a bit into the file with the new linguist which happens to signal the same error when loading the file but i must admit i haven t found what could possibly have gone astray the problem is that this is a fatal error and the build aborts any idea on this very basic problem merci vincent
0
5,460
2,576,490,341
IssuesEvent
2015-02-12 10:29:12
KIZI/EasyMiner-EasyMinerCenter
https://api.github.com/repos/KIZI/EasyMiner-EasyMinerCenter
closed
Chyba importu csv s "id"
priority: high state: commited type: bug
Pokud csv obsahuje sloupec ID, dojde k pádu importu (ani se nevytvoří příslušná tabulka)... Zároveň kontrolovat unikátnost názvu sloupců
1.0
Chyba importu csv s "id" - Pokud csv obsahuje sloupec ID, dojde k pádu importu (ani se nevytvoří příslušná tabulka)... Zároveň kontrolovat unikátnost názvu sloupců
non_test
chyba importu csv s id pokud csv obsahuje sloupec id dojde k pádu importu ani se nevytvoří příslušná tabulka zároveň kontrolovat unikátnost názvu sloupců
0
218,267
16,981,410,793
IssuesEvent
2021-06-30 09:19:33
moby/moby
https://api.github.com/repos/moby/moby
opened
Flaky test: libnetwork TestCreateParallel (arm64)
area/networking area/testing kind/bug
logs: https://ci-next.docker.com/public/blue/rest/organizations/jenkins/pipelines/moby/branches/PR-42576/runs/1/nodes/295/log/?start=0 Seen failing on https://github.com/moby/moby/pull/42576 ``` === RUN TestCreateParallel time="2021-06-29T04:56:20Z" level=warning msg="bridge store not initialized. kv object docker/network/v1.0/bridge/net26/ is not added to the store" time="2021-06-29T04:56:20Z" level=warning msg="bridge store not initialized. kv object docker/network/v1.0/bridge/net70/ is not added to the store" bridge_test.go:1133: Success should be 1 instead: 2 --- FAIL: TestCreateParallel (0.09s) ``` Bundles: [arm64-integration-bundles.tar.gz](https://github.com/moby/moby/files/6739623/arm64-integration-bundles.tar.gz)
1.0
Flaky test: libnetwork TestCreateParallel (arm64) - logs: https://ci-next.docker.com/public/blue/rest/organizations/jenkins/pipelines/moby/branches/PR-42576/runs/1/nodes/295/log/?start=0 Seen failing on https://github.com/moby/moby/pull/42576 ``` === RUN TestCreateParallel time="2021-06-29T04:56:20Z" level=warning msg="bridge store not initialized. kv object docker/network/v1.0/bridge/net26/ is not added to the store" time="2021-06-29T04:56:20Z" level=warning msg="bridge store not initialized. kv object docker/network/v1.0/bridge/net70/ is not added to the store" bridge_test.go:1133: Success should be 1 instead: 2 --- FAIL: TestCreateParallel (0.09s) ``` Bundles: [arm64-integration-bundles.tar.gz](https://github.com/moby/moby/files/6739623/arm64-integration-bundles.tar.gz)
test
flaky test libnetwork testcreateparallel logs seen failing on run testcreateparallel time level warning msg bridge store not initialized kv object docker network bridge is not added to the store time level warning msg bridge store not initialized kv object docker network bridge is not added to the store bridge test go success should be instead fail testcreateparallel bundles
1
6,849
10,040,321,111
IssuesEvent
2019-07-18 19:39:26
westmary48/le-voyage-
https://api.github.com/repos/westmary48/le-voyage-
closed
Single Trip
CRUD MVP requirement
## User Story As a user, when view my home page, I should see a link called view memory and when I click the link, I should be taken to a page that has my single memory information. ## Acceptance Criteria **WHEN** I look at the home page **THEN** I should see a link on each of my cards that says view memory **AND** when I click on the link **THEN** I should be taken to the a page **THAT** shows the memory card in an individual view. ## Developer Notes - Create an axios get request called `getSingleMemory` - See comments for axios call - In your SingleTrip Function - In your TripCard Component: - Import {Link} from react-router-dom - Create a variable called singleLink add the correct path for routing - Add the Link in Trip Card in the return and pass in the route singleLink. - In your SingleTrip Component: - Import scatData - Set the state to trip and set it to an empty object - Create the life cycle function componentDidMount - Create a variable called trip id and set it equal to const scatId = this.props.match.params.id; - Call your axios get request for getSingleMemory - create a promise and set the tripPromise to set state - Create a deleteTrip function - Create a variable called tripId in this function - Call the axios delete function `deleteTrip` - In the promise, define the path that the page will go to after the card has been deleted - In the render, set trip equal to state - Create a domString displaying all the keys - Add a delete button in the domString - In the button tag create an onClick and then pass in this.deleteTrip
1.0
Single Trip - ## User Story As a user, when view my home page, I should see a link called view memory and when I click the link, I should be taken to a page that has my single memory information. ## Acceptance Criteria **WHEN** I look at the home page **THEN** I should see a link on each of my cards that says view memory **AND** when I click on the link **THEN** I should be taken to the a page **THAT** shows the memory card in an individual view. ## Developer Notes - Create an axios get request called `getSingleMemory` - See comments for axios call - In your SingleTrip Function - In your TripCard Component: - Import {Link} from react-router-dom - Create a variable called singleLink add the correct path for routing - Add the Link in Trip Card in the return and pass in the route singleLink. - In your SingleTrip Component: - Import scatData - Set the state to trip and set it to an empty object - Create the life cycle function componentDidMount - Create a variable called trip id and set it equal to const scatId = this.props.match.params.id; - Call your axios get request for getSingleMemory - create a promise and set the tripPromise to set state - Create a deleteTrip function - Create a variable called tripId in this function - Call the axios delete function `deleteTrip` - In the promise, define the path that the page will go to after the card has been deleted - In the render, set trip equal to state - Create a domString displaying all the keys - Add a delete button in the domString - In the button tag create an onClick and then pass in this.deleteTrip
non_test
single trip user story as a user when view my home page i should see a link called view memory and when i click the link i should be taken to a page that has my single memory information acceptance criteria when i look at the home page then i should see a link on each of my cards that says view memory and when i click on the link then i should be taken to the a page that shows the memory card in an individual view developer notes create an axios get request called getsinglememory see comments for axios call in your singletrip function in your tripcard component import link from react router dom create a variable called singlelink add the correct path for routing add the link in trip card in the return and pass in the route singlelink in your singletrip component import scatdata set the state to trip and set it to an empty object create the life cycle function componentdidmount create a variable called trip id and set it equal to const scatid this props match params id call your axios get request for getsinglememory create a promise and set the trippromise to set state create a deletetrip function create a variable called tripid in this function call the axios delete function deletetrip in the promise define the path that the page will go to after the card has been deleted in the render set trip equal to state create a domstring displaying all the keys add a delete button in the domstring in the button tag create an onclick and then pass in this deletetrip
0
135,714
30,351,332,914
IssuesEvent
2023-07-11 19:13:10
creativecommons/cc-resource-archive
https://api.github.com/repos/creativecommons/cc-resource-archive
closed
Unnecessary <a> tags in resource page
🟧 priority: high 🏁 status: ready for work 🛠 goal: fix 💻 aspect: code
## Description Unnecessary <a> tags are used in all the resource pages. ## Reproduction 1. open docs folder 2. then navigate to _site folder 3. all the index.html files in each folder contains this error ( specifically while rendering the title for the page ) 4. See error. 5. This is also seen in the resource.html in the _layout file. ## Expectation remove this redundant a tag as it not only confuses the reader but disrupts the intended working of the website ## Screenshots ![image](https://user-images.githubusercontent.com/100520932/227621108-a26cc37f-9a0d-4dcd-ab43-bf62486ace24.png) ![Screenshot from 2023-03-25 00-51-30](https://user-images.githubusercontent.com/100520932/227621165-0d95c52c-b842-4ea4-be98-24af9b2b99c2.png) ## Resolution - [x] I would be interested in resolving this bug.
1.0
Unnecessary <a> tags in resource page - ## Description Unnecessary <a> tags are used in all the resource pages. ## Reproduction 1. open docs folder 2. then navigate to _site folder 3. all the index.html files in each folder contains this error ( specifically while rendering the title for the page ) 4. See error. 5. This is also seen in the resource.html in the _layout file. ## Expectation remove this redundant a tag as it not only confuses the reader but disrupts the intended working of the website ## Screenshots ![image](https://user-images.githubusercontent.com/100520932/227621108-a26cc37f-9a0d-4dcd-ab43-bf62486ace24.png) ![Screenshot from 2023-03-25 00-51-30](https://user-images.githubusercontent.com/100520932/227621165-0d95c52c-b842-4ea4-be98-24af9b2b99c2.png) ## Resolution - [x] I would be interested in resolving this bug.
non_test
unnecessary tags in resource page description unnecessary tags are used in all the resource pages reproduction open docs folder then navigate to site folder all the index html files in each folder contains this error specifically while rendering the title for the page see error this is also seen in the resource html in the layout file expectation remove this redundant a tag as it not only confuses the reader but disrupts the intended working of the website screenshots resolution i would be interested in resolving this bug
0
60,987
6,720,332,610
IssuesEvent
2017-10-16 07:21:19
zalando/zalenium
https://api.github.com/repos/zalando/zalenium
closed
Kubernetes GKE support
waiting-retest
Running kubernetes version 1.6.6 on GKE and I'm having trouble getting started with zalenium. Manifest: ``` --- apiVersion: v1 kind: ServiceAccount metadata: name: zalenium --- apiVersion: rbac.authorization.k8s.io/v1beta1 kind: RoleBinding metadata: name: zalenium subjects: - kind: ServiceAccount name: zalenium roleRef: kind: ClusterRole name: admin apiGroup: rbac.authorization.k8s.io --- apiVersion: v1 kind: Service metadata: name: zalenium-grid spec: selector: app: zalenium role: grid ports: - name: zalenium-grid protocol: TCP port: 4444 targetPort: 4444 type: NodePort --- apiVersion: v1 kind: Service metadata: name: zalenium-videos spec: selector: app: zalenium role: grid ports: - name: zalenium-videos protocol: TCP port: 5555 targetPort: 5555 type: NodePort --- apiVersion: v1 kind: PersistentVolume metadata: name: zalenium-shared spec: capacity: storage: 10Gi accessModes: - ReadWriteMany gcePersistentDisk: pdName: "zalenium-shared" fsType: "ext4" --- kind: PersistentVolumeClaim apiVersion: v1 metadata: name: zalenium-shared spec: accessModes: - ReadWriteMany resources: requests: storage: 10Gi --- apiVersion: v1 kind: PersistentVolume metadata: name: zalenium-videos spec: capacity: storage: 10Gi accessModes: - ReadWriteMany gcePersistentDisk: pdName: "zalenium-videos" fsType: "ext4" --- kind: PersistentVolumeClaim apiVersion: v1 metadata: name: zalenium-videos spec: accessModes: - ReadWriteMany resources: requests: storage: 10Gi --- apiVersion: extensions/v1beta1 kind: Deployment metadata: name: zalenium labels: app: zalenium role: grid spec: replicas: 1 selector: matchLabels: app: zalenium role: grid strategy: type: RollingUpdate rollingUpdate: maxSurge: 1 maxUnavailable: 1 minReadySeconds: 5 template: metadata: creationTimestamp: null labels: app: zalenium role: grid spec: serviceAccountName: zalenium serviceAccount: zalenium dnsPolicy: ClusterFirst imagePullSecrets: - name: gcr-json-key volumes: - name: zalenium-videos persistentVolumeClaim: claimName: zalenium-videos - name: zalenium-shared persistentVolumeClaim: claimName: zalenium-shared - name: dshm emptyDir: medium: Memory - name: dockersocket hostPath: path: /var/run/docker.sock restartPolicy: Always terminationGracePeriodSeconds: 120 containers: - name: zalenium image: dosel/zalenium imagePullPolicy: Always terminationMessagePath: /dev/termination-log args: - start - --chromeContainers - '2' - --firefoxContainers - '2' - --screenWidth - '1440' - --screenHeight - '810' - --timeZone - Sweden/Stockholm ports: - name: zalenium-grid containerPort: 4444 - name: zalenium-videos containerPort: 5555 volumeMounts: - name: zalenium-videos mountPath: /home/seluser/videos - name: zalenium-shared mountPath: /tmp/mounted - name: dshm mountPath: /dev/shm - name: dockersocket mountPath: /var/run/docker.sock ``` Gives: ``` Kubernetes service account found. -- LOG 17:29:09:606904905 Running additional checks... Copying files for Dashboard... Starting Nginx reverse proxy... Starting Selenium Hub... 17:29:09.906 INFO - Selenium build info: version: '3.3.1', revision: '5234b32' 17:29:09.907 INFO - Launching Selenium Grid hub 17:29:10.474 INFO - binding de.zalando.ep.zalenium.servlet.LivePreviewServlet to /grid/admin/LivePreviewServlet/* 17:29:10.475 INFO - binding de.zalando.ep.zalenium.servlet.ZaleniumConsoleServlet to /grid/admin/ZaleniumConsoleServlet/* 17:29:10.476 INFO - binding de.zalando.ep.zalenium.servlet.ZaleniumResourceServlet to /grid/admin/ZaleniumResourceServlet/* 17:29:10.476 INFO - binding de.zalando.ep.zalenium.dashboard.DashboardCleanupServlet to /grid/admin/DashboardCleanupServlet/* 17:29:10.546 INFO - Will listen on 4445 17:29:10.710 INFO - Nodes should register to http://10.0.17.54:4445/grid/register/ 17:29:10.710 INFO - Selenium Grid hub is up and running .....Selenium Hub started! Starting DockerSeleniumStarter node... 17:29:11.136 INFO - Selenium build info: version: '3.3.1', revision: '5234b32' 17:29:11.137 INFO - Launching a Selenium Grid node 2017-07-13 17:29:11.544:INFO::main: Logging initialized @676ms to org.seleniumhq.jetty9.util.log.StdErrLog 17:29:11.581 INFO - Driver provider org.openqa.selenium.ie.InternetExplorerDriver registration is skipped: registration capabilities Capabilities [{ensureCleanSession=true, browserName=internet explorer, version=, platform=WINDOWS}] does not match the current platform LINUX 17:29:11.581 INFO - Driver provider org.openqa.selenium.edge.EdgeDriver registration is skipped: registration capabilities Capabilities [{browserName=MicrosoftEdge, version=, platform=WINDOWS}] does not match the current platform LINUX 17:29:11.582 INFO - Driver class not found: com.opera.core.systems.OperaDriver 17:29:11.582 INFO - Driver provider com.opera.core.systems.OperaDriver registration is skipped: Unable to create new instances on this machine. 17:29:11.582 INFO - Driver class not found: com.opera.core.systems.OperaDriver 17:29:11.582 INFO - Driver provider com.opera.core.systems.OperaDriver is not registered 17:29:11.583 INFO - Driver provider org.openqa.selenium.safari.SafariDriver registration is skipped: registration capabilities Capabilities [{browserName=safari, version=, platform=MAC}] does not match the current platform LINUX 2017-07-13 17:29:11.627:INFO:osjs.Server:main: jetty-9.2.20.v20161216 2017-07-13 17:29:11.668:INFO:osjsh.ContextHandler:main: Started o.s.j.s.ServletContextHandler@cd3fee8{/,null,AVAILABLE} 2017-07-13 17:29:11.683:INFO:osjs.AbstractConnector:main: Started ServerConnector@25306456{HTTP/1.1,[http/1.1]}{0.0.0.0:30000} 2017-07-13 17:29:11.684:INFO:osjs.Server:main: Started @816ms 17:29:11.684 INFO - Selenium Grid node is up and ready to register to the hub 17:29:11.699 INFO - Registering the node to the hub: http://localhost:4444/grid/register 17:29:11.942 INFO - Initialising Kubernetes support 17:29:11.975 INFO - SessionCleaner initialized with insideBrowserTimeout 0 and clientGoneTimeout 1800000 polling every 180000 org.openqa.grid.common.exception.GridException: Error sending the registration request: The hub responded with 500:Server Error at org.openqa.grid.internal.utils.SelfRegisteringRemote.registerToHub(SelfRegisteringRemote.java:277) at org.openqa.grid.internal.utils.SelfRegisteringRemote.startRegistrationProcess(SelfRegisteringRemote.java:218) at org.openqa.grid.selenium.GridLauncherV3$3.launch(GridLauncherV3.java:290) at org.openqa.grid.selenium.GridLauncherV3.main(GridLauncherV3.java:99) ..............................................................................................................................................StarterRemoteProxy failed to register to the hub after 30 seconds, failing... ```
1.0
Kubernetes GKE support - Running kubernetes version 1.6.6 on GKE and I'm having trouble getting started with zalenium. Manifest: ``` --- apiVersion: v1 kind: ServiceAccount metadata: name: zalenium --- apiVersion: rbac.authorization.k8s.io/v1beta1 kind: RoleBinding metadata: name: zalenium subjects: - kind: ServiceAccount name: zalenium roleRef: kind: ClusterRole name: admin apiGroup: rbac.authorization.k8s.io --- apiVersion: v1 kind: Service metadata: name: zalenium-grid spec: selector: app: zalenium role: grid ports: - name: zalenium-grid protocol: TCP port: 4444 targetPort: 4444 type: NodePort --- apiVersion: v1 kind: Service metadata: name: zalenium-videos spec: selector: app: zalenium role: grid ports: - name: zalenium-videos protocol: TCP port: 5555 targetPort: 5555 type: NodePort --- apiVersion: v1 kind: PersistentVolume metadata: name: zalenium-shared spec: capacity: storage: 10Gi accessModes: - ReadWriteMany gcePersistentDisk: pdName: "zalenium-shared" fsType: "ext4" --- kind: PersistentVolumeClaim apiVersion: v1 metadata: name: zalenium-shared spec: accessModes: - ReadWriteMany resources: requests: storage: 10Gi --- apiVersion: v1 kind: PersistentVolume metadata: name: zalenium-videos spec: capacity: storage: 10Gi accessModes: - ReadWriteMany gcePersistentDisk: pdName: "zalenium-videos" fsType: "ext4" --- kind: PersistentVolumeClaim apiVersion: v1 metadata: name: zalenium-videos spec: accessModes: - ReadWriteMany resources: requests: storage: 10Gi --- apiVersion: extensions/v1beta1 kind: Deployment metadata: name: zalenium labels: app: zalenium role: grid spec: replicas: 1 selector: matchLabels: app: zalenium role: grid strategy: type: RollingUpdate rollingUpdate: maxSurge: 1 maxUnavailable: 1 minReadySeconds: 5 template: metadata: creationTimestamp: null labels: app: zalenium role: grid spec: serviceAccountName: zalenium serviceAccount: zalenium dnsPolicy: ClusterFirst imagePullSecrets: - name: gcr-json-key volumes: - name: zalenium-videos persistentVolumeClaim: claimName: zalenium-videos - name: zalenium-shared persistentVolumeClaim: claimName: zalenium-shared - name: dshm emptyDir: medium: Memory - name: dockersocket hostPath: path: /var/run/docker.sock restartPolicy: Always terminationGracePeriodSeconds: 120 containers: - name: zalenium image: dosel/zalenium imagePullPolicy: Always terminationMessagePath: /dev/termination-log args: - start - --chromeContainers - '2' - --firefoxContainers - '2' - --screenWidth - '1440' - --screenHeight - '810' - --timeZone - Sweden/Stockholm ports: - name: zalenium-grid containerPort: 4444 - name: zalenium-videos containerPort: 5555 volumeMounts: - name: zalenium-videos mountPath: /home/seluser/videos - name: zalenium-shared mountPath: /tmp/mounted - name: dshm mountPath: /dev/shm - name: dockersocket mountPath: /var/run/docker.sock ``` Gives: ``` Kubernetes service account found. -- LOG 17:29:09:606904905 Running additional checks... Copying files for Dashboard... Starting Nginx reverse proxy... Starting Selenium Hub... 17:29:09.906 INFO - Selenium build info: version: '3.3.1', revision: '5234b32' 17:29:09.907 INFO - Launching Selenium Grid hub 17:29:10.474 INFO - binding de.zalando.ep.zalenium.servlet.LivePreviewServlet to /grid/admin/LivePreviewServlet/* 17:29:10.475 INFO - binding de.zalando.ep.zalenium.servlet.ZaleniumConsoleServlet to /grid/admin/ZaleniumConsoleServlet/* 17:29:10.476 INFO - binding de.zalando.ep.zalenium.servlet.ZaleniumResourceServlet to /grid/admin/ZaleniumResourceServlet/* 17:29:10.476 INFO - binding de.zalando.ep.zalenium.dashboard.DashboardCleanupServlet to /grid/admin/DashboardCleanupServlet/* 17:29:10.546 INFO - Will listen on 4445 17:29:10.710 INFO - Nodes should register to http://10.0.17.54:4445/grid/register/ 17:29:10.710 INFO - Selenium Grid hub is up and running .....Selenium Hub started! Starting DockerSeleniumStarter node... 17:29:11.136 INFO - Selenium build info: version: '3.3.1', revision: '5234b32' 17:29:11.137 INFO - Launching a Selenium Grid node 2017-07-13 17:29:11.544:INFO::main: Logging initialized @676ms to org.seleniumhq.jetty9.util.log.StdErrLog 17:29:11.581 INFO - Driver provider org.openqa.selenium.ie.InternetExplorerDriver registration is skipped: registration capabilities Capabilities [{ensureCleanSession=true, browserName=internet explorer, version=, platform=WINDOWS}] does not match the current platform LINUX 17:29:11.581 INFO - Driver provider org.openqa.selenium.edge.EdgeDriver registration is skipped: registration capabilities Capabilities [{browserName=MicrosoftEdge, version=, platform=WINDOWS}] does not match the current platform LINUX 17:29:11.582 INFO - Driver class not found: com.opera.core.systems.OperaDriver 17:29:11.582 INFO - Driver provider com.opera.core.systems.OperaDriver registration is skipped: Unable to create new instances on this machine. 17:29:11.582 INFO - Driver class not found: com.opera.core.systems.OperaDriver 17:29:11.582 INFO - Driver provider com.opera.core.systems.OperaDriver is not registered 17:29:11.583 INFO - Driver provider org.openqa.selenium.safari.SafariDriver registration is skipped: registration capabilities Capabilities [{browserName=safari, version=, platform=MAC}] does not match the current platform LINUX 2017-07-13 17:29:11.627:INFO:osjs.Server:main: jetty-9.2.20.v20161216 2017-07-13 17:29:11.668:INFO:osjsh.ContextHandler:main: Started o.s.j.s.ServletContextHandler@cd3fee8{/,null,AVAILABLE} 2017-07-13 17:29:11.683:INFO:osjs.AbstractConnector:main: Started ServerConnector@25306456{HTTP/1.1,[http/1.1]}{0.0.0.0:30000} 2017-07-13 17:29:11.684:INFO:osjs.Server:main: Started @816ms 17:29:11.684 INFO - Selenium Grid node is up and ready to register to the hub 17:29:11.699 INFO - Registering the node to the hub: http://localhost:4444/grid/register 17:29:11.942 INFO - Initialising Kubernetes support 17:29:11.975 INFO - SessionCleaner initialized with insideBrowserTimeout 0 and clientGoneTimeout 1800000 polling every 180000 org.openqa.grid.common.exception.GridException: Error sending the registration request: The hub responded with 500:Server Error at org.openqa.grid.internal.utils.SelfRegisteringRemote.registerToHub(SelfRegisteringRemote.java:277) at org.openqa.grid.internal.utils.SelfRegisteringRemote.startRegistrationProcess(SelfRegisteringRemote.java:218) at org.openqa.grid.selenium.GridLauncherV3$3.launch(GridLauncherV3.java:290) at org.openqa.grid.selenium.GridLauncherV3.main(GridLauncherV3.java:99) ..............................................................................................................................................StarterRemoteProxy failed to register to the hub after 30 seconds, failing... ```
test
kubernetes gke support running kubernetes version on gke and i m having trouble getting started with zalenium manifest apiversion kind serviceaccount metadata name zalenium apiversion rbac authorization io kind rolebinding metadata name zalenium subjects kind serviceaccount name zalenium roleref kind clusterrole name admin apigroup rbac authorization io apiversion kind service metadata name zalenium grid spec selector app zalenium role grid ports name zalenium grid protocol tcp port targetport type nodeport apiversion kind service metadata name zalenium videos spec selector app zalenium role grid ports name zalenium videos protocol tcp port targetport type nodeport apiversion kind persistentvolume metadata name zalenium shared spec capacity storage accessmodes readwritemany gcepersistentdisk pdname zalenium shared fstype kind persistentvolumeclaim apiversion metadata name zalenium shared spec accessmodes readwritemany resources requests storage apiversion kind persistentvolume metadata name zalenium videos spec capacity storage accessmodes readwritemany gcepersistentdisk pdname zalenium videos fstype kind persistentvolumeclaim apiversion metadata name zalenium videos spec accessmodes readwritemany resources requests storage apiversion extensions kind deployment metadata name zalenium labels app zalenium role grid spec replicas selector matchlabels app zalenium role grid strategy type rollingupdate rollingupdate maxsurge maxunavailable minreadyseconds template metadata creationtimestamp null labels app zalenium role grid spec serviceaccountname zalenium serviceaccount zalenium dnspolicy clusterfirst imagepullsecrets name gcr json key volumes name zalenium videos persistentvolumeclaim claimname zalenium videos name zalenium shared persistentvolumeclaim claimname zalenium shared name dshm emptydir medium memory name dockersocket hostpath path var run docker sock restartpolicy always terminationgraceperiodseconds containers name zalenium image dosel zalenium imagepullpolicy always terminationmessagepath dev termination log args start chromecontainers firefoxcontainers screenwidth screenheight timezone sweden stockholm ports name zalenium grid containerport name zalenium videos containerport volumemounts name zalenium videos mountpath home seluser videos name zalenium shared mountpath tmp mounted name dshm mountpath dev shm name dockersocket mountpath var run docker sock gives kubernetes service account found log running additional checks copying files for dashboard starting nginx reverse proxy starting selenium hub info selenium build info version revision info launching selenium grid hub info binding de zalando ep zalenium servlet livepreviewservlet to grid admin livepreviewservlet info binding de zalando ep zalenium servlet zaleniumconsoleservlet to grid admin zaleniumconsoleservlet info binding de zalando ep zalenium servlet zaleniumresourceservlet to grid admin zaleniumresourceservlet info binding de zalando ep zalenium dashboard dashboardcleanupservlet to grid admin dashboardcleanupservlet info will listen on info nodes should register to info selenium grid hub is up and running selenium hub started starting dockerseleniumstarter node info selenium build info version revision info launching a selenium grid node info main logging initialized to org seleniumhq util log stderrlog info driver provider org openqa selenium ie internetexplorerdriver registration is skipped registration capabilities capabilities does not match the current platform linux info driver provider org openqa selenium edge edgedriver registration is skipped registration capabilities capabilities does not match the current platform linux info driver class not found com opera core systems operadriver info driver provider com opera core systems operadriver registration is skipped unable to create new instances on this machine info driver class not found com opera core systems operadriver info driver provider com opera core systems operadriver is not registered info driver provider org openqa selenium safari safaridriver registration is skipped registration capabilities capabilities does not match the current platform linux info osjs server main jetty info osjsh contexthandler main started o s j s servletcontexthandler null available info osjs abstractconnector main started serverconnector http info osjs server main started info selenium grid node is up and ready to register to the hub info registering the node to the hub info initialising kubernetes support info sessioncleaner initialized with insidebrowsertimeout and clientgonetimeout polling every org openqa grid common exception gridexception error sending the registration request the hub responded with server error at org openqa grid internal utils selfregisteringremote registertohub selfregisteringremote java at org openqa grid internal utils selfregisteringremote startregistrationprocess selfregisteringremote java at org openqa grid selenium launch java at org openqa grid selenium main java starterremoteproxy failed to register to the hub after seconds failing
1
25,764
5,198,035,376
IssuesEvent
2017-01-23 17:03:15
arquillian/arquillian-cube
https://api.github.com/repos/arquillian/arquillian-cube
closed
Refactor ftest-containerobject project
bug documentation enhancement
##### Issue Overview Refactor `ftest-containerobject` so instead of using Ping Pong container which is not a real example of when to use Container Objects pattern and change it to use an FTP server (https://github.com/m-creations/docker-openwrt-ftp) with Apache Commons (https://commons.apache.org/proper/commons-net/apidocs/org/apache/commons/net/ftp/FTPClient.html) ##### Expected Behaviour Use FTP which is a more real example. ##### Current Behaviour Ping Pong simple example used
1.0
Refactor ftest-containerobject project - ##### Issue Overview Refactor `ftest-containerobject` so instead of using Ping Pong container which is not a real example of when to use Container Objects pattern and change it to use an FTP server (https://github.com/m-creations/docker-openwrt-ftp) with Apache Commons (https://commons.apache.org/proper/commons-net/apidocs/org/apache/commons/net/ftp/FTPClient.html) ##### Expected Behaviour Use FTP which is a more real example. ##### Current Behaviour Ping Pong simple example used
non_test
refactor ftest containerobject project issue overview refactor ftest containerobject so instead of using ping pong container which is not a real example of when to use container objects pattern and change it to use an ftp server with apache commons expected behaviour use ftp which is a more real example current behaviour ping pong simple example used
0
101,946
31,771,595,658
IssuesEvent
2023-09-12 12:09:47
llvm/llvm-project
https://api.github.com/repos/llvm/llvm-project
closed
Failed Profile-i386 tests on new Windows 32bit MSVC buildbot
clang bugzilla build-problem obsolete platform:windows
| | | | --- | --- | | Bugzilla Link | [47759](https://llvm.org/bz47759) | | Version | trunk | | OS | Windows 2000 | | CC | @zygoloid | ## Extended Description When setting up a new 32bit WIndows buildbot with MSVC, I noticed a few of the instrprof tests are failing. The buildbot is currently in staging as I want to resolve the failing tests before moving it to production. http://lab.llvm.org:8014/#/builders/27
1.0
Failed Profile-i386 tests on new Windows 32bit MSVC buildbot - | | | | --- | --- | | Bugzilla Link | [47759](https://llvm.org/bz47759) | | Version | trunk | | OS | Windows 2000 | | CC | @zygoloid | ## Extended Description When setting up a new 32bit WIndows buildbot with MSVC, I noticed a few of the instrprof tests are failing. The buildbot is currently in staging as I want to resolve the failing tests before moving it to production. http://lab.llvm.org:8014/#/builders/27
non_test
failed profile tests on new windows msvc buildbot bugzilla link version trunk os windows cc zygoloid extended description when setting up a new windows buildbot with msvc i noticed a few of the instrprof tests are failing the buildbot is currently in staging as i want to resolve the failing tests before moving it to production
0
216,947
16,675,268,121
IssuesEvent
2021-06-07 15:27:28
bounswe/2021SpringGroup10
https://api.github.com/repos/bounswe/2021SpringGroup10
closed
Creation of RAM table template for the Milestone
Everybody Priority: High Status: Needs Review Status: Pending Type: Communication Type: Documentation Type: Wiki
We should show the task explicitly. The main headings are Wiki Documentation, Communication, Requirements, User Scenarios and Mockups, UML Diagrams, Project Plan, and RAM.
1.0
Creation of RAM table template for the Milestone - We should show the task explicitly. The main headings are Wiki Documentation, Communication, Requirements, User Scenarios and Mockups, UML Diagrams, Project Plan, and RAM.
non_test
creation of ram table template for the milestone we should show the task explicitly the main headings are wiki documentation communication requirements user scenarios and mockups uml diagrams project plan and ram
0
93,620
8,439,255,578
IssuesEvent
2018-10-18 00:47:22
kubeflow/testing
https://api.github.com/repos/kubeflow/testing
closed
NFS share is out of space.
area/testing priority/p1
``` W + SRC_DIR=/mnt/test-data-volume/kubeflow-presubmit-kfctl-1776-268c7f3-3833-549e/src + mkdir -p /src/kubeflow + git clone https://github.com/kubeflow/kubeflow.git /mnt/test-data-volume/kubeflow-presubmit-kfctl-1776-268c7f3-3833-549e/src/kubeflow/kubeflow Cloning into '/mnt/test-data-volume/kubeflow-presubmit-kfctl-1776-268c7f3-3833-549e/src/kubeflow/kubeflow'... fatal: fsync error on '/mnt/test-data-volume/kubeflow-presubmit-kfctl-1776-268c7f3-3833-549e/src/kubeflow/kubeflow/.git/objects/pack/tmp_pack_gTR7N5': No space left on device fatal: index-pack failed ```
1.0
NFS share is out of space. - ``` W + SRC_DIR=/mnt/test-data-volume/kubeflow-presubmit-kfctl-1776-268c7f3-3833-549e/src + mkdir -p /src/kubeflow + git clone https://github.com/kubeflow/kubeflow.git /mnt/test-data-volume/kubeflow-presubmit-kfctl-1776-268c7f3-3833-549e/src/kubeflow/kubeflow Cloning into '/mnt/test-data-volume/kubeflow-presubmit-kfctl-1776-268c7f3-3833-549e/src/kubeflow/kubeflow'... fatal: fsync error on '/mnt/test-data-volume/kubeflow-presubmit-kfctl-1776-268c7f3-3833-549e/src/kubeflow/kubeflow/.git/objects/pack/tmp_pack_gTR7N5': No space left on device fatal: index-pack failed ```
test
nfs share is out of space w src dir mnt test data volume kubeflow presubmit kfctl src mkdir p src kubeflow git clone mnt test data volume kubeflow presubmit kfctl src kubeflow kubeflow cloning into mnt test data volume kubeflow presubmit kfctl src kubeflow kubeflow fatal fsync error on mnt test data volume kubeflow presubmit kfctl src kubeflow kubeflow git objects pack tmp pack no space left on device fatal index pack failed
1
20,002
11,355,895,127
IssuesEvent
2020-01-24 21:11:39
edgexfoundry/edgex-go
https://api.github.com/repos/edgexfoundry/edgex-go
opened
Update OpenAPI Docs for V2 API -- Generic Errors
core-services f2f-geneva support-services
Remove ErrorEnvelope (batch endpoint DTO -- wraps use-case DTO). Add ErrorResponse (use-case DTO). Add ErrorResponse to all relevant use-case-request paths. Related conversation: ![image](https://user-images.githubusercontent.com/464327/73100384-0f00cc00-3eb3-11ea-9f4b-0cb1238875a0.png)
2.0
Update OpenAPI Docs for V2 API -- Generic Errors - Remove ErrorEnvelope (batch endpoint DTO -- wraps use-case DTO). Add ErrorResponse (use-case DTO). Add ErrorResponse to all relevant use-case-request paths. Related conversation: ![image](https://user-images.githubusercontent.com/464327/73100384-0f00cc00-3eb3-11ea-9f4b-0cb1238875a0.png)
non_test
update openapi docs for api generic errors remove errorenvelope batch endpoint dto wraps use case dto add errorresponse use case dto add errorresponse to all relevant use case request paths related conversation
0
24,378
4,076,375,653
IssuesEvent
2016-05-29 21:21:50
GTNewHorizons/NewHorizons
https://api.github.com/repos/GTNewHorizons/NewHorizons
opened
Remote Accessor (id:7048) vs. ingame IRC and /say Command
bugMinor need to be tested
(Ingame) IRC chat msg's and /say msg's show up additional times equal to the number of times that item was used. Relogging resets the "use count" to 0. Needs to be tested if can be fixed by IRC relay change. Bug does not affect normal ingame player msg's. So chatting woth other players ingame is flawless.
1.0
Remote Accessor (id:7048) vs. ingame IRC and /say Command - (Ingame) IRC chat msg's and /say msg's show up additional times equal to the number of times that item was used. Relogging resets the "use count" to 0. Needs to be tested if can be fixed by IRC relay change. Bug does not affect normal ingame player msg's. So chatting woth other players ingame is flawless.
test
remote accessor id vs ingame irc and say command ingame irc chat msg s and say msg s show up additional times equal to the number of times that item was used relogging resets the use count to needs to be tested if can be fixed by irc relay change bug does not affect normal ingame player msg s so chatting woth other players ingame is flawless
1
745,963
26,008,164,419
IssuesEvent
2022-12-20 21:39:19
robotframework/robotframework
https://api.github.com/repos/robotframework/robotframework
closed
Bug in `--reportbackgroundcolor` documentation in the User Guide
bug priority: medium
_Observation_: User guide and robot mismatching in #setting-background-colors section _User Guide_: If you specify three colors, the first one will be used when all the tests pass, the second when all tests have been skipped, and the last when there are any failures. (pass:skip:fail) _Robot Command Line Help_: '--reportbackground': Expected format 'pass:fail:skip'
1.0
Bug in `--reportbackgroundcolor` documentation in the User Guide - _Observation_: User guide and robot mismatching in #setting-background-colors section _User Guide_: If you specify three colors, the first one will be used when all the tests pass, the second when all tests have been skipped, and the last when there are any failures. (pass:skip:fail) _Robot Command Line Help_: '--reportbackground': Expected format 'pass:fail:skip'
non_test
bug in reportbackgroundcolor documentation in the user guide observation user guide and robot mismatching in setting background colors section user guide if you specify three colors the first one will be used when all the tests pass the second when all tests have been skipped and the last when there are any failures pass skip fail robot command line help reportbackground expected format pass fail skip
0
177,829
29,170,712,408
IssuesEvent
2023-05-19 01:20:08
wpumacay/renderer
https://api.github.com/repos/wpumacay/renderer
closed
Shader Manager Implementation
component: python component: shaders component: assets priority: high type: design type: documentation type: feature
# Description This tracks the implementation of the `shader_manager` singleton, used to both `create` and `store` shaders in a easier way. The rationale is that `shaders` will be considered as assets, and this manager will be in charge of easily create them and share the ownership of this shaders with user code that might request them. ## Tasks - [ ] Implementatio of the **`sbader_manager`** - [ ] Create an **`example`** of its usage - [ ] Implement some **`unittests`** - [ ] Implement **`python`** bindings - [ ] Make **`documentation`** explaining its usage ## Notes * There's the legacy implementation of the old [**`shader_manager`**][0], which we could use as a starting point. * There's a legacy [**`example`**] of its usage; however, the API should change in order to have a cleaner and easier to use API. Notice in this example that the renderer is requesting the shaders that it might use. However, the role will likely be swapped, as the materials are going to be the ones that have a specific shader according to its shading model. [0]: <https://github.com/wpumacay/loco_renderer/blob/legacy/legacy/include/shaders/CShaderManager.h> (reference-impl-1) [1]: <https://github.com/wpumacay/loco_renderer/blob/legacy/legacy/src/renderers/CMeshRenderer.cpp#L9> (reference-sample-usage-1)
1.0
Shader Manager Implementation - # Description This tracks the implementation of the `shader_manager` singleton, used to both `create` and `store` shaders in a easier way. The rationale is that `shaders` will be considered as assets, and this manager will be in charge of easily create them and share the ownership of this shaders with user code that might request them. ## Tasks - [ ] Implementatio of the **`sbader_manager`** - [ ] Create an **`example`** of its usage - [ ] Implement some **`unittests`** - [ ] Implement **`python`** bindings - [ ] Make **`documentation`** explaining its usage ## Notes * There's the legacy implementation of the old [**`shader_manager`**][0], which we could use as a starting point. * There's a legacy [**`example`**] of its usage; however, the API should change in order to have a cleaner and easier to use API. Notice in this example that the renderer is requesting the shaders that it might use. However, the role will likely be swapped, as the materials are going to be the ones that have a specific shader according to its shading model. [0]: <https://github.com/wpumacay/loco_renderer/blob/legacy/legacy/include/shaders/CShaderManager.h> (reference-impl-1) [1]: <https://github.com/wpumacay/loco_renderer/blob/legacy/legacy/src/renderers/CMeshRenderer.cpp#L9> (reference-sample-usage-1)
non_test
shader manager implementation description this tracks the implementation of the shader manager singleton used to both create and store shaders in a easier way the rationale is that shaders will be considered as assets and this manager will be in charge of easily create them and share the ownership of this shaders with user code that might request them tasks implementatio of the sbader manager create an example of its usage implement some unittests implement python bindings make documentation explaining its usage notes there s the legacy implementation of the old which we could use as a starting point there s a legacy of its usage however the api should change in order to have a cleaner and easier to use api notice in this example that the renderer is requesting the shaders that it might use however the role will likely be swapped as the materials are going to be the ones that have a specific shader according to its shading model reference impl reference sample usage
0
236,717
19,569,640,259
IssuesEvent
2022-01-04 08:14:59
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
No completion support for Qualified name reference
Type/Bug Priority/High Team/LanguageServer Area/Completion GA-Test-Hackathon
**Description:** Consider the following capture. ![qnameref](https://user-images.githubusercontent.com/35211477/146872841-0cc230f0-cd95-4169-afba-6fe86c7222eb.png) Visible symbols in the float module is not suggested as depicted above. **Steps to reproduce:** ```ballerina class Circle { float radius; function init(float radius) { self.radius = radius; } function getCircumference() returns float { return 2 * float:PI * self.radius; } function getArea() returns float { return float:P<cursor> * float:pow(self.radius, 2); } } ``` **Affected Versions:** slbeta6
1.0
No completion support for Qualified name reference - **Description:** Consider the following capture. ![qnameref](https://user-images.githubusercontent.com/35211477/146872841-0cc230f0-cd95-4169-afba-6fe86c7222eb.png) Visible symbols in the float module is not suggested as depicted above. **Steps to reproduce:** ```ballerina class Circle { float radius; function init(float radius) { self.radius = radius; } function getCircumference() returns float { return 2 * float:PI * self.radius; } function getArea() returns float { return float:P<cursor> * float:pow(self.radius, 2); } } ``` **Affected Versions:** slbeta6
test
no completion support for qualified name reference description consider the following capture visible symbols in the float module is not suggested as depicted above steps to reproduce ballerina class circle float radius function init float radius self radius radius function getcircumference returns float return float pi self radius function getarea returns float return float p float pow self radius affected versions
1
93,084
10,764,499,396
IssuesEvent
2019-11-01 08:28:39
Kzrthikz/ped
https://api.github.com/repos/Kzrthikz/ped
opened
Optional inputs are confusing.
severity.Medium type.DocumentationBug
From the user guide, optional user inputs are not clear. Could consider using [CAPS] to show its an optional user input. This is especially confusing for filter feature as to what is defined by "FIELD", "QUANTIFIER" and "VALUE".
1.0
Optional inputs are confusing. - From the user guide, optional user inputs are not clear. Could consider using [CAPS] to show its an optional user input. This is especially confusing for filter feature as to what is defined by "FIELD", "QUANTIFIER" and "VALUE".
non_test
optional inputs are confusing from the user guide optional user inputs are not clear could consider using to show its an optional user input this is especially confusing for filter feature as to what is defined by field quantifier and value
0
50,690
6,107,374,474
IssuesEvent
2017-06-21 07:56:12
brave/browser-laptop
https://api.github.com/repos/brave/browser-laptop
closed
Disable fetching background images on test runs
automated-tests suggestion
**Describe the issue you encountered:** Create an option to disable fetching background images on test runs. https://github.com/brave/browser-laptop/issues/6503#issuecomment-304769664 > I notice `npm run test` often fails due to timeout because images are not fetched in time. Per @bsclifton; > another option: we could default that setting to false if `NODE_ENV` is test
1.0
Disable fetching background images on test runs - **Describe the issue you encountered:** Create an option to disable fetching background images on test runs. https://github.com/brave/browser-laptop/issues/6503#issuecomment-304769664 > I notice `npm run test` often fails due to timeout because images are not fetched in time. Per @bsclifton; > another option: we could default that setting to false if `NODE_ENV` is test
test
disable fetching background images on test runs describe the issue you encountered create an option to disable fetching background images on test runs i notice npm run test often fails due to timeout because images are not fetched in time per bsclifton another option we could default that setting to false if node env is test
1
589,217
17,692,463,131
IssuesEvent
2021-08-24 11:46:11
kirbydesign/designsystem
https://api.github.com/repos/kirbydesign/designsystem
closed
[Housekeeping] Keep a CHANGELOG.md
NOT Tech refined housekeeping priority 2
**Short description of housekeeping request** Maintain a changelog next to the code instead of only in the cookbook site named CHANGELOG.md This is a pretty standard location for changelogs, thus making it easier for people unfamiliar with Kirby Design System to see changes between versions at a glance. The format could be https://keepachangelog.com/en/1.0.0/ **Alternatives** The changelog is already present in the cookbook site, so a link in the README.md file could also be a solution. **Additional context** I am an administrator in a project that uses Kirby Design System. Bumping package versions in our project requires administrator approval, but I am not well-versed enough in Kirby Design System to know what impact a version bump has. ## Tasks ### Kick Off * [ ] Ensure this issue has been `Tech refined` with @kirbydesign/kirby-guild and is updated with a clear implementation description _This issue should be in the [Ready to do](https://github.com/kirbydesign/designsystem/projects/1#column-4590936) column of the [Kirby kan-ban board](https://github.com/kirbydesign/designsystem/projects/1) before starting implementation_ * [ ] Assign yourself to this issue and move it to the [In progress](https://github.com/kirbydesign/designsystem/projects/1#column-4590937) column of the [Kirby kan-ban board](https://github.com/kirbydesign/designsystem/projects/1) ### Code * [ ] Create Feature Branch from [master branch](https://github.com/kirbydesign/designsystem/tree/master) * [ ] Implement/update unit tests * [ ] Create a draft implementation and push to Github * [ ] Ask a member of @kirbydesign/kirby-guild for a WIP review by creating a draft Pull Request ### Code Review * [ ] Open a pull request (or mark the existing draft PR as `Ready for review`) and ask @kirbydesign/kirby-guild for a review _Remember to add `closes #issueno` to the description of the PR._ * [ ] Once approved, merge feature branch/PR to master :tada: Celebrate
1.0
[Housekeeping] Keep a CHANGELOG.md - **Short description of housekeeping request** Maintain a changelog next to the code instead of only in the cookbook site named CHANGELOG.md This is a pretty standard location for changelogs, thus making it easier for people unfamiliar with Kirby Design System to see changes between versions at a glance. The format could be https://keepachangelog.com/en/1.0.0/ **Alternatives** The changelog is already present in the cookbook site, so a link in the README.md file could also be a solution. **Additional context** I am an administrator in a project that uses Kirby Design System. Bumping package versions in our project requires administrator approval, but I am not well-versed enough in Kirby Design System to know what impact a version bump has. ## Tasks ### Kick Off * [ ] Ensure this issue has been `Tech refined` with @kirbydesign/kirby-guild and is updated with a clear implementation description _This issue should be in the [Ready to do](https://github.com/kirbydesign/designsystem/projects/1#column-4590936) column of the [Kirby kan-ban board](https://github.com/kirbydesign/designsystem/projects/1) before starting implementation_ * [ ] Assign yourself to this issue and move it to the [In progress](https://github.com/kirbydesign/designsystem/projects/1#column-4590937) column of the [Kirby kan-ban board](https://github.com/kirbydesign/designsystem/projects/1) ### Code * [ ] Create Feature Branch from [master branch](https://github.com/kirbydesign/designsystem/tree/master) * [ ] Implement/update unit tests * [ ] Create a draft implementation and push to Github * [ ] Ask a member of @kirbydesign/kirby-guild for a WIP review by creating a draft Pull Request ### Code Review * [ ] Open a pull request (or mark the existing draft PR as `Ready for review`) and ask @kirbydesign/kirby-guild for a review _Remember to add `closes #issueno` to the description of the PR._ * [ ] Once approved, merge feature branch/PR to master :tada: Celebrate
non_test
keep a changelog md short description of housekeeping request maintain a changelog next to the code instead of only in the cookbook site named changelog md this is a pretty standard location for changelogs thus making it easier for people unfamiliar with kirby design system to see changes between versions at a glance the format could be alternatives the changelog is already present in the cookbook site so a link in the readme md file could also be a solution additional context i am an administrator in a project that uses kirby design system bumping package versions in our project requires administrator approval but i am not well versed enough in kirby design system to know what impact a version bump has tasks kick off ensure this issue has been tech refined with kirbydesign kirby guild and is updated with a clear implementation description this issue should be in the column of the before starting implementation assign yourself to this issue and move it to the column of the code create feature branch from implement update unit tests create a draft implementation and push to github ask a member of kirbydesign kirby guild for a wip review by creating a draft pull request code review open a pull request or mark the existing draft pr as ready for review and ask kirbydesign kirby guild for a review remember to add closes issueno to the description of the pr once approved merge feature branch pr to master tada celebrate
0
85,998
15,755,314,366
IssuesEvent
2021-03-31 01:33:29
heltondoria/event-system
https://api.github.com/repos/heltondoria/event-system
opened
CVE-2020-7751 (High) detected in pathval-1.1.0.tgz
security vulnerability
## CVE-2020-7751 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>pathval-1.1.0.tgz</b></p></summary> <p>Object value retrieval given a string path</p> <p>Library home page: <a href="https://registry.npmjs.org/pathval/-/pathval-1.1.0.tgz">https://registry.npmjs.org/pathval/-/pathval-1.1.0.tgz</a></p> <p>Path to dependency file: /event-system/package.json</p> <p>Path to vulnerable library: event-system/node_modules/pathval/package.json</p> <p> Dependency Hierarchy: - chai-4.2.0.tgz (Root Library) - :x: **pathval-1.1.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects all versions of package pathval. <p>Publish Date: 2020-10-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7751>CVE-2020-7751</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7751 (High) detected in pathval-1.1.0.tgz - ## CVE-2020-7751 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>pathval-1.1.0.tgz</b></p></summary> <p>Object value retrieval given a string path</p> <p>Library home page: <a href="https://registry.npmjs.org/pathval/-/pathval-1.1.0.tgz">https://registry.npmjs.org/pathval/-/pathval-1.1.0.tgz</a></p> <p>Path to dependency file: /event-system/package.json</p> <p>Path to vulnerable library: event-system/node_modules/pathval/package.json</p> <p> Dependency Hierarchy: - chai-4.2.0.tgz (Root Library) - :x: **pathval-1.1.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects all versions of package pathval. <p>Publish Date: 2020-10-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7751>CVE-2020-7751</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in pathval tgz cve high severity vulnerability vulnerable library pathval tgz object value retrieval given a string path library home page a href path to dependency file event system package json path to vulnerable library event system node modules pathval package json dependency hierarchy chai tgz root library x pathval tgz vulnerable library vulnerability details this affects all versions of package pathval publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href step up your open source security game with whitesource
0
37,122
5,103,110,167
IssuesEvent
2017-01-04 20:21:52
dhermes/bezier
https://api.github.com/repos/dhermes/bezier
opened
Implemented Speedups in C/C++/Fortran
hygiene perf testing
e.g. for `Surface.evaluate_cartesian`. To provide test coverage for these extensions, can [use `gcov`][1] [1]: http://stackoverflow.com/questions/5144603/shared-library-coverage-test-with-gcov-linux-fortran
1.0
Implemented Speedups in C/C++/Fortran - e.g. for `Surface.evaluate_cartesian`. To provide test coverage for these extensions, can [use `gcov`][1] [1]: http://stackoverflow.com/questions/5144603/shared-library-coverage-test-with-gcov-linux-fortran
test
implemented speedups in c c fortran e g for surface evaluate cartesian to provide test coverage for these extensions can
1
256,861
22,107,069,652
IssuesEvent
2022-06-01 17:50:29
owid/covid-19-data
https://api.github.com/repos/owid/covid-19-data
opened
Ending our COVID-19 testing data updates
dom:testing announcement
As of 23 June 2022, we will no longer add new data points to our COVID-19 testing dataset. We will continue updates of all other metrics in our COVID-19 dataset. You can read more [in this discussion](https://github.com/owid/covid-19-data/discussions/2667)
1.0
Ending our COVID-19 testing data updates - As of 23 June 2022, we will no longer add new data points to our COVID-19 testing dataset. We will continue updates of all other metrics in our COVID-19 dataset. You can read more [in this discussion](https://github.com/owid/covid-19-data/discussions/2667)
test
ending our covid testing data updates as of june we will no longer add new data points to our covid testing dataset we will continue updates of all other metrics in our covid dataset you can read more
1
211,634
16,330,998,344
IssuesEvent
2021-05-12 09:12:22
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
opened
No error info displays in the 'Error Details' dialog after clicking the 'Details...' next to a promote version failed activity log
🧪 testing
**Storage Explorer Version**: 1.20.0-dev **Build Number**: 20210512.2 **Branch**: main **Platform/OS**: Windows 10/ Linux Ubuntu 18.04/ MacOS Big Sur 11.3 **Architecture**: ia32/x64 **How found**: exploratory testing **Regression From**: Not a regression ## Steps to Reproduce ## 1. Expand the storage account -> Blob Containers.(Make sure blob versioning is enabled) 2. Create one blob container -> Upload one blob to it. 3. Change the blob to get the blob version -> Right click the blob -> Click 'Manage History -> Manage Versions'. 4. Acquire lease for the blob -> Select the previous version -> Click 'Promote Version'. 5. Click the 'Details...' next to the activity log -> Check whether an error info displays in the 'Error Details' dialog.. ## Expected Experience ## An error info displays. ## Actual Experience ## No error info displays, just display {} in the 'Error Details' dialog. ![image](https://user-images.githubusercontent.com/41351993/117948558-d1dc7980-b343-11eb-89dc-94503e7c3444.png)
1.0
No error info displays in the 'Error Details' dialog after clicking the 'Details...' next to a promote version failed activity log - **Storage Explorer Version**: 1.20.0-dev **Build Number**: 20210512.2 **Branch**: main **Platform/OS**: Windows 10/ Linux Ubuntu 18.04/ MacOS Big Sur 11.3 **Architecture**: ia32/x64 **How found**: exploratory testing **Regression From**: Not a regression ## Steps to Reproduce ## 1. Expand the storage account -> Blob Containers.(Make sure blob versioning is enabled) 2. Create one blob container -> Upload one blob to it. 3. Change the blob to get the blob version -> Right click the blob -> Click 'Manage History -> Manage Versions'. 4. Acquire lease for the blob -> Select the previous version -> Click 'Promote Version'. 5. Click the 'Details...' next to the activity log -> Check whether an error info displays in the 'Error Details' dialog.. ## Expected Experience ## An error info displays. ## Actual Experience ## No error info displays, just display {} in the 'Error Details' dialog. ![image](https://user-images.githubusercontent.com/41351993/117948558-d1dc7980-b343-11eb-89dc-94503e7c3444.png)
test
no error info displays in the error details dialog after clicking the details next to a promote version failed activity log storage explorer version dev build number branch main platform os windows linux ubuntu macos big sur architecture how found exploratory testing regression from not a regression steps to reproduce expand the storage account blob containers make sure blob versioning is enabled create one blob container upload one blob to it change the blob to get the blob version right click the blob click manage history manage versions acquire lease for the blob select the previous version click promote version click the details next to the activity log check whether an error info displays in the error details dialog expected experience an error info displays actual experience no error info displays just display in the error details dialog
1
30,922
2,729,656,404
IssuesEvent
2015-04-16 09:59:29
jkall/qgis-midvatten-plugin
https://api.github.com/repos/jkall/qgis-midvatten-plugin
opened
allow obsid dubplicates by introducing another ID as primary key
enhancement Priority-High
Introduce a unique observation id that is independent of obsid and name. This will allow the existance of duplicates among obsid. Major code revisions are needed. Several security checks are needed during imports and also, when obsid duplicates are found, user interaction to distinguish between observations of same obsid. Probably time to introduce a main table with id for each observation as the primary key. This table will hold all observations, no matter if they are points or lines.
1.0
allow obsid dubplicates by introducing another ID as primary key - Introduce a unique observation id that is independent of obsid and name. This will allow the existance of duplicates among obsid. Major code revisions are needed. Several security checks are needed during imports and also, when obsid duplicates are found, user interaction to distinguish between observations of same obsid. Probably time to introduce a main table with id for each observation as the primary key. This table will hold all observations, no matter if they are points or lines.
non_test
allow obsid dubplicates by introducing another id as primary key introduce a unique observation id that is independent of obsid and name this will allow the existance of duplicates among obsid major code revisions are needed several security checks are needed during imports and also when obsid duplicates are found user interaction to distinguish between observations of same obsid probably time to introduce a main table with id for each observation as the primary key this table will hold all observations no matter if they are points or lines
0
179,704
13,895,971,515
IssuesEvent
2020-10-19 16:31:49
WuriGuinea/mosip-guinea-ref-impl
https://api.github.com/repos/WuriGuinea/mosip-guinea-ref-impl
closed
Additional address details-i.e. "complément d'adresse" not populated in notification email and/or INU card
In progress Priority 1 Test
See Row 1131 in Master templates, $!additionalAddressDetails_fra
1.0
Additional address details-i.e. "complément d'adresse" not populated in notification email and/or INU card - See Row 1131 in Master templates, $!additionalAddressDetails_fra
test
additional address details i e complément d adresse not populated in notification email and or inu card see row in master templates additionaladdressdetails fra
1
339,046
10,241,019,185
IssuesEvent
2019-08-19 22:33:33
kubernetes-sigs/cluster-api-provider-aws
https://api.github.com/repos/kubernetes-sigs/cluster-api-provider-aws
closed
service of type LoadBalancer stuck in pending, missing tag on subnets
kind/bug lifecycle/active priority/important-soon
/kind bug **What steps did you take and what happened:** deployed CAPA v0.3.7 succesfully `kubectl run test1 --image=nginx` `kubectl expose deploy test1 --port=80 --type=LoadBalancer` ``` E0813 18:59:18.479595 1 service_controller.go:219] error processing service default/test1 (will retry): failed to ensure load balancer for service default/test1: could not find any suitable subnets for creating the ELB I0813 18:59:18.479816 1 event.go:209] Event(v1.ObjectReference{Kind:"Service", Namespace:"default", Name:"test1", UID:"469f0ef1-bdfc-11e9-85ce-025ee73143fe", APIVersion:"v1", ResourceVersion:"2057", FieldPath:""}): type: 'Warning' reason: 'CreatingLoadBalancerFailed' Error creating load balancer (will retry): failed to ensure load balancer for service default/test1: could not find any suitable subnets for creating the ELB ``` **What did you expect to happen:** The ELB to be created by the `LoadBalancer` service type. **Anything else you would like to add:** Manually adding `kubernetes.io/cluster/<name>` tag with value `owned` to the subnets resolves the issue. The controller manager re-tries and creates the ELB. **Environment:** - Cluster-api-provider-aws version: v0.3.7 - Kubernetes version: (use `kubectl version`): 1.14.4 - OS (e.g. from `/etc/os-release`): N/A
1.0
service of type LoadBalancer stuck in pending, missing tag on subnets - /kind bug **What steps did you take and what happened:** deployed CAPA v0.3.7 succesfully `kubectl run test1 --image=nginx` `kubectl expose deploy test1 --port=80 --type=LoadBalancer` ``` E0813 18:59:18.479595 1 service_controller.go:219] error processing service default/test1 (will retry): failed to ensure load balancer for service default/test1: could not find any suitable subnets for creating the ELB I0813 18:59:18.479816 1 event.go:209] Event(v1.ObjectReference{Kind:"Service", Namespace:"default", Name:"test1", UID:"469f0ef1-bdfc-11e9-85ce-025ee73143fe", APIVersion:"v1", ResourceVersion:"2057", FieldPath:""}): type: 'Warning' reason: 'CreatingLoadBalancerFailed' Error creating load balancer (will retry): failed to ensure load balancer for service default/test1: could not find any suitable subnets for creating the ELB ``` **What did you expect to happen:** The ELB to be created by the `LoadBalancer` service type. **Anything else you would like to add:** Manually adding `kubernetes.io/cluster/<name>` tag with value `owned` to the subnets resolves the issue. The controller manager re-tries and creates the ELB. **Environment:** - Cluster-api-provider-aws version: v0.3.7 - Kubernetes version: (use `kubectl version`): 1.14.4 - OS (e.g. from `/etc/os-release`): N/A
non_test
service of type loadbalancer stuck in pending missing tag on subnets kind bug what steps did you take and what happened deployed capa succesfully kubectl run image nginx kubectl expose deploy port type loadbalancer service controller go error processing service default will retry failed to ensure load balancer for service default could not find any suitable subnets for creating the elb event go event objectreference kind service namespace default name uid bdfc apiversion resourceversion fieldpath type warning reason creatingloadbalancerfailed error creating load balancer will retry failed to ensure load balancer for service default could not find any suitable subnets for creating the elb what did you expect to happen the elb to be created by the loadbalancer service type anything else you would like to add manually adding kubernetes io cluster tag with value owned to the subnets resolves the issue the controller manager re tries and creates the elb environment cluster api provider aws version kubernetes version use kubectl version os e g from etc os release n a
0
227,232
18,054,175,771
IssuesEvent
2021-09-20 05:09:48
logicmoo/logicmoo_workspace
https://api.github.com/repos/logicmoo/logicmoo_workspace
opened
logicmoo.pfc.test.sanity_base.ATTVAR_04 JUnit
Test_9999 logicmoo.pfc.test.sanity_base unit_test ATTVAR_04 Passing
(cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s swipl -x /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-clif attvar_04.pl) % ISSUE: https://github.com/logicmoo/logicmoo_workspace/issues/ % EDIT: https://github.com/logicmoo/logicmoo_workspace/edit/master/packs_sys/pfc/t/sanity_base/attvar_04.pl % JENKINS: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/ATTVAR_04/logicmoo_pfc_test_sanity_base_ATTVAR_04_JUnit/ % ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3AATTVAR_04 ``` %~ init_phase(after_load) %~ init_phase(restore_state) % running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/attvar_04.pl'), %~ this_test_might_need( :-( expects_dialect(pfc))) %~ this_test_might_need( :-( use_module( library(logicmoo_plarkc)))) :- if(\+ current_module(attvar_reader)). :- use_module(library(logicmoo/attvar_reader)). :- endif. % Tests assertable attributed variables :- debug_logicmoo(_). :- nodebug_logicmoo(http(_)). :- debug_logicmoo(logicmoo(_)). % :- mpred_trace_exec. % :- mpred_trace_exec. :- dynamic(sk_out/1). :- dynamic(sk_in/1). % :- ain(baseKB:rtArgsVerbatum(my_sk)). % :- ain(baseKB:rtArgsVerbatum(my_sk)). :- read_attvars(true). % :- expects_dialect(swi). %~ install_attvar_expander(baseKB) %~ term_xform( my_sk( aVar( [ vn='Ex', %~ sk='SKF-666'])) --> %~ %~ my_sk(Sk) :- %~ %~ attr_bind([put_attr(Sk,sk,'SKF-666')],true)) % :- expects_dialect(swi). my_sk(aVar([vn='Ex',sk='SKF-666'])). No source location!? :- mpred_test((my_sk(Ex),get_attr(Ex,sk,What),What=='SKF-666')). % ISSUE: https://github.com/logicmoo/logicmoo_workspace/issues/ % EDIT: https://github.com/logicmoo/logicmoo_workspace/edit/master/packs_sys/pfc/t/sanity_base/attvar_04.pl % JENKINS: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/ATTVAR_04/logicmoo_pfc_test_sanity_base_ATTVAR_04_JUnit/ % ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3AATTVAR_04 %~ mpred_test( "Test_0001_Line_0000__my_sk_1", %~ baseKB : ( my_sk(Ex) , %~ get_attr(Ex,sk,What) , %~ What=='SKF-666')) /*~ %~ mpred_test("Test_0001_Line_0000__my_sk_1",baseKB:(my_sk(_3830),get_attr(_3830,sk,_3852),_3852=='SKF-666')) passed=info(why_was_true(baseKB:(my_sk(_19832),get_attr(_19832,sk,'SKF-666'),'SKF-666'=='SKF-666'))) no_proof_for((my_sk(Ex),get_attr(Ex,sk,'SKF-666'),'SKF-666'=='SKF-666')). no_proof_for((my_sk(Ex),get_attr(Ex,sk,'SKF-666'),'SKF-666'=='SKF-666')). no_proof_for((my_sk(Ex),get_attr(Ex,sk,'SKF-666'),'SKF-666'=='SKF-666')). name = 'logicmoo.pfc.test.sanity_base.ATTVAR_04-Test_0001_Line_0000__my_sk_1'. JUNIT_CLASSNAME = 'logicmoo.pfc.test.sanity_base.ATTVAR_04'. JUNIT_CMD = 'timeout --foreground --preserve-status -s SIGKILL -k 10s 10s swipl -x /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-clif attvar_04.pl'. % saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-pfc-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.ATTVAR_04-Test_0001_Line_0000__my_sk_1-junit.xml ~*/ %~ unused(no_junit_results) Test_0001_Line_0000__my_sk_1 result = passed. %~ test_completed_exit(64) ``` totalTime=1.000 SUCCESS: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k attvar_04.pl (returned 64) Add_LABELS='' Rem_LABELS='Skipped,Errors,Warnings,Overtime,Skipped,Skipped'
3.0
logicmoo.pfc.test.sanity_base.ATTVAR_04 JUnit - (cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s swipl -x /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-clif attvar_04.pl) % ISSUE: https://github.com/logicmoo/logicmoo_workspace/issues/ % EDIT: https://github.com/logicmoo/logicmoo_workspace/edit/master/packs_sys/pfc/t/sanity_base/attvar_04.pl % JENKINS: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/ATTVAR_04/logicmoo_pfc_test_sanity_base_ATTVAR_04_JUnit/ % ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3AATTVAR_04 ``` %~ init_phase(after_load) %~ init_phase(restore_state) % running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/attvar_04.pl'), %~ this_test_might_need( :-( expects_dialect(pfc))) %~ this_test_might_need( :-( use_module( library(logicmoo_plarkc)))) :- if(\+ current_module(attvar_reader)). :- use_module(library(logicmoo/attvar_reader)). :- endif. % Tests assertable attributed variables :- debug_logicmoo(_). :- nodebug_logicmoo(http(_)). :- debug_logicmoo(logicmoo(_)). % :- mpred_trace_exec. % :- mpred_trace_exec. :- dynamic(sk_out/1). :- dynamic(sk_in/1). % :- ain(baseKB:rtArgsVerbatum(my_sk)). % :- ain(baseKB:rtArgsVerbatum(my_sk)). :- read_attvars(true). % :- expects_dialect(swi). %~ install_attvar_expander(baseKB) %~ term_xform( my_sk( aVar( [ vn='Ex', %~ sk='SKF-666'])) --> %~ %~ my_sk(Sk) :- %~ %~ attr_bind([put_attr(Sk,sk,'SKF-666')],true)) % :- expects_dialect(swi). my_sk(aVar([vn='Ex',sk='SKF-666'])). No source location!? :- mpred_test((my_sk(Ex),get_attr(Ex,sk,What),What=='SKF-666')). % ISSUE: https://github.com/logicmoo/logicmoo_workspace/issues/ % EDIT: https://github.com/logicmoo/logicmoo_workspace/edit/master/packs_sys/pfc/t/sanity_base/attvar_04.pl % JENKINS: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/ATTVAR_04/logicmoo_pfc_test_sanity_base_ATTVAR_04_JUnit/ % ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3AATTVAR_04 %~ mpred_test( "Test_0001_Line_0000__my_sk_1", %~ baseKB : ( my_sk(Ex) , %~ get_attr(Ex,sk,What) , %~ What=='SKF-666')) /*~ %~ mpred_test("Test_0001_Line_0000__my_sk_1",baseKB:(my_sk(_3830),get_attr(_3830,sk,_3852),_3852=='SKF-666')) passed=info(why_was_true(baseKB:(my_sk(_19832),get_attr(_19832,sk,'SKF-666'),'SKF-666'=='SKF-666'))) no_proof_for((my_sk(Ex),get_attr(Ex,sk,'SKF-666'),'SKF-666'=='SKF-666')). no_proof_for((my_sk(Ex),get_attr(Ex,sk,'SKF-666'),'SKF-666'=='SKF-666')). no_proof_for((my_sk(Ex),get_attr(Ex,sk,'SKF-666'),'SKF-666'=='SKF-666')). name = 'logicmoo.pfc.test.sanity_base.ATTVAR_04-Test_0001_Line_0000__my_sk_1'. JUNIT_CLASSNAME = 'logicmoo.pfc.test.sanity_base.ATTVAR_04'. JUNIT_CMD = 'timeout --foreground --preserve-status -s SIGKILL -k 10s 10s swipl -x /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-clif attvar_04.pl'. % saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-pfc-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.ATTVAR_04-Test_0001_Line_0000__my_sk_1-junit.xml ~*/ %~ unused(no_junit_results) Test_0001_Line_0000__my_sk_1 result = passed. %~ test_completed_exit(64) ``` totalTime=1.000 SUCCESS: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k attvar_04.pl (returned 64) Add_LABELS='' Rem_LABELS='Skipped,Errors,Warnings,Overtime,Skipped,Skipped'
test
logicmoo pfc test sanity base attvar junit cd var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base timeout foreground preserve status s sigkill k swipl x var lib jenkins workspace logicmoo workspace bin lmoo clif attvar pl issue edit jenkins issue search init phase after load init phase restore state running var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base attvar pl this test might need expects dialect pfc this test might need use module library logicmoo plarkc if current module attvar reader use module library logicmoo attvar reader endif tests assertable attributed variables debug logicmoo nodebug logicmoo http debug logicmoo logicmoo mpred trace exec mpred trace exec dynamic sk out dynamic sk in ain basekb rtargsverbatum my sk ain basekb rtargsverbatum my sk read attvars true expects dialect swi install attvar expander basekb term xform my sk avar vn ex sk skf my sk sk attr bind true expects dialect swi my sk avar no source location mpred test my sk ex get attr ex sk what what skf issue edit jenkins issue search mpred test test line my sk basekb my sk ex get attr ex sk what what skf mpred test test line my sk basekb my sk get attr sk skf passed info why was true basekb my sk get attr sk skf skf skf no proof for my sk ex get attr ex sk skf skf skf no proof for my sk ex get attr ex sk skf skf skf no proof for my sk ex get attr ex sk skf skf skf name logicmoo pfc test sanity base attvar test line my sk junit classname logicmoo pfc test sanity base attvar junit cmd timeout foreground preserve status s sigkill k swipl x var lib jenkins workspace logicmoo workspace bin lmoo clif attvar pl saving junit var lib jenkins workspace logicmoo workspace test results jenkins report logicmoo pfc test sanity base units logicmoo pfc test sanity base attvar test line my sk junit xml unused no junit results test line my sk result passed test completed exit totaltime success var lib jenkins workspace logicmoo workspace bin lmoo junit minor k attvar pl returned add labels rem labels skipped errors warnings overtime skipped skipped
1
40,911
2,868,949,721
IssuesEvent
2015-06-05 22:08:45
dart-lang/pub
https://api.github.com/repos/dart-lang/pub
closed
Make the pub shell script work from within the repo
enhancement Fixed Priority-Medium
<a href="https://github.com/munificent"><img src="https://avatars.githubusercontent.com/u/46275?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [munificent](https://github.com/munificent)** _Originally opened as dart-lang/sdk#10928_ ---- The shell script under sdk/bin/ for running pub only works in the built SDK, not in the source repo. It should support both. For reference the dart2js script handles this.
1.0
Make the pub shell script work from within the repo - <a href="https://github.com/munificent"><img src="https://avatars.githubusercontent.com/u/46275?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [munificent](https://github.com/munificent)** _Originally opened as dart-lang/sdk#10928_ ---- The shell script under sdk/bin/ for running pub only works in the built SDK, not in the source repo. It should support both. For reference the dart2js script handles this.
non_test
make the pub shell script work from within the repo issue by originally opened as dart lang sdk the shell script under sdk bin for running pub only works in the built sdk not in the source repo it should support both for reference the script handles this
0
170,788
27,015,556,558
IssuesEvent
2023-02-10 19:01:22
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
closed
Tuple Declaration Hiding
Area-IDE Feature Request Developer Community Need Design Review
VSF_TYPE_MARKDOWNNew code editor suggestion. I love Tuples but they elongate the method declaration. The suggestion is to allow a toggle to collapse them. For example... ``` private static (bool showHelp, bool deleteExistingFiles, string inputFolder, string outputFolder, bool pauseCommandWindow, List&lt;string&gt; Files) ProcessCommandLineArguments(string[] args) ``` ... to something like this: ``` private static (tuple... ») ProcessCommandLineArguments(string[] args) ``` _This issue has been moved from https://developercommunity.visualstudio.com/content/idea/582406/tuple-declaration-hiding.html VSTS ticketId: 899239_ _These are the original issue comments:_ Jane Wu [MSFT] on 5/27/2019, 01:08 AM (10 days ago): <p>Thank you for taking the time to provide your suggestion. We will do some preliminary checks to make sure we can proceed further.&nbsp;We'll provide an update once the issue has been triaged by the product team.</p>
1.0
Tuple Declaration Hiding - VSF_TYPE_MARKDOWNNew code editor suggestion. I love Tuples but they elongate the method declaration. The suggestion is to allow a toggle to collapse them. For example... ``` private static (bool showHelp, bool deleteExistingFiles, string inputFolder, string outputFolder, bool pauseCommandWindow, List&lt;string&gt; Files) ProcessCommandLineArguments(string[] args) ``` ... to something like this: ``` private static (tuple... ») ProcessCommandLineArguments(string[] args) ``` _This issue has been moved from https://developercommunity.visualstudio.com/content/idea/582406/tuple-declaration-hiding.html VSTS ticketId: 899239_ _These are the original issue comments:_ Jane Wu [MSFT] on 5/27/2019, 01:08 AM (10 days ago): <p>Thank you for taking the time to provide your suggestion. We will do some preliminary checks to make sure we can proceed further.&nbsp;We'll provide an update once the issue has been triaged by the product team.</p>
non_test
tuple declaration hiding vsf type markdownnew code editor suggestion i love tuples but they elongate the method declaration the suggestion is to allow a toggle to collapse them for example private static bool showhelp bool deleteexistingfiles string inputfolder string outputfolder bool pausecommandwindow list lt string gt files processcommandlinearguments string args to something like this private static tuple » processcommandlinearguments string args this issue has been moved from vsts ticketid these are the original issue comments jane wu on am days ago thank you for taking the time to provide your suggestion we will do some preliminary checks to make sure we can proceed further nbsp we ll provide an update once the issue has been triaged by the product team
0
273,208
23,738,083,294
IssuesEvent
2022-08-31 09:52:26
pingcap/tidb
https://api.github.com/repos/pingcap/tidb
closed
unstable test in the TestNowAndUTCTimestamp
type/bug component/test component/expression severity/minor
## Bug Report Please answer these questions before submitting your issue. Thanks! ### 1. Minimal reproduce step (Required) ``` === RUN TestNowAndUTCTimestamp builtin_time_test.go:849: Error Trace: /home/jenkins/.tidb/tmp/04446c229c5a73c16deb3edddcb4db34/sandbox/processwrapper-sandbox/5550/execroot/__main__/bazel-out/k8-fastbuild/bin/expression/expression_test_/expression_test.runfiles/__main__/expression/builtin_time_test.go:849 Error: "3.099621454s" is not less than or equal to "3s" Test: TestNowAndUTCTimestamp --- FAIL: TestNowAndUTCTimestamp (3.05s) ``` https://prow.pingcap.net/view/gs/pingcapprow/logs/bazel_test_tidb/1564103182567608320#1:build-log.txt%3A2832 <!-- a step by step guide for reproducing the bug. --> ### 2. What did you expect to see? (Required) ### 3. What did you see instead (Required) ### 4. What is your TiDB version? (Required) <!-- Paste the output of SELECT tidb_version() -->
1.0
unstable test in the TestNowAndUTCTimestamp - ## Bug Report Please answer these questions before submitting your issue. Thanks! ### 1. Minimal reproduce step (Required) ``` === RUN TestNowAndUTCTimestamp builtin_time_test.go:849: Error Trace: /home/jenkins/.tidb/tmp/04446c229c5a73c16deb3edddcb4db34/sandbox/processwrapper-sandbox/5550/execroot/__main__/bazel-out/k8-fastbuild/bin/expression/expression_test_/expression_test.runfiles/__main__/expression/builtin_time_test.go:849 Error: "3.099621454s" is not less than or equal to "3s" Test: TestNowAndUTCTimestamp --- FAIL: TestNowAndUTCTimestamp (3.05s) ``` https://prow.pingcap.net/view/gs/pingcapprow/logs/bazel_test_tidb/1564103182567608320#1:build-log.txt%3A2832 <!-- a step by step guide for reproducing the bug. --> ### 2. What did you expect to see? (Required) ### 3. What did you see instead (Required) ### 4. What is your TiDB version? (Required) <!-- Paste the output of SELECT tidb_version() -->
test
unstable test in the testnowandutctimestamp bug report please answer these questions before submitting your issue thanks minimal reproduce step required run testnowandutctimestamp builtin time test go error trace home jenkins tidb tmp sandbox processwrapper sandbox execroot main bazel out fastbuild bin expression expression test expression test runfiles main expression builtin time test go error is not less than or equal to test testnowandutctimestamp fail testnowandutctimestamp what did you expect to see required what did you see instead required what is your tidb version required
1
776,903
27,264,758,444
IssuesEvent
2023-02-22 17:11:07
ascheid/itsg33-pbmm-issue-gen
https://api.github.com/repos/ascheid/itsg33-pbmm-issue-gen
opened
SC-7(11): Boundary Protection | Restrict Incoming Communications Traffic
Priority: P2 Class: Technical ITSG-33 Suggested Assignment: IT Projects Control: SC-7
# Control Definition BOUNDARY PROTECTION | RESTRICT INCOMING COMMUNICATIONS TRAFFIC The information system only allows incoming communications from [Assignment: organization-defined authorized sources] routed to [Assignment: organization-defined authorized destinations]. # Class Technical # Supplemental Guidance This control enhancement provides determinations that source and destination address pairs represent authorized/allowed communications. Such determinations can be based on several factors including, for example, the presence of source/destination address pairs in lists of authorized/allowed communications, the absence of address pairs in lists of unauthorized/disallowed pairs, or meeting more general rules for authorized/allowed source/destination pairs. Related control: AC-3. # General Guide This security control/enhancement should be addressed where applicable and if practical to do so. # Suggested Assignment IT Projects
1.0
SC-7(11): Boundary Protection | Restrict Incoming Communications Traffic - # Control Definition BOUNDARY PROTECTION | RESTRICT INCOMING COMMUNICATIONS TRAFFIC The information system only allows incoming communications from [Assignment: organization-defined authorized sources] routed to [Assignment: organization-defined authorized destinations]. # Class Technical # Supplemental Guidance This control enhancement provides determinations that source and destination address pairs represent authorized/allowed communications. Such determinations can be based on several factors including, for example, the presence of source/destination address pairs in lists of authorized/allowed communications, the absence of address pairs in lists of unauthorized/disallowed pairs, or meeting more general rules for authorized/allowed source/destination pairs. Related control: AC-3. # General Guide This security control/enhancement should be addressed where applicable and if practical to do so. # Suggested Assignment IT Projects
non_test
sc boundary protection restrict incoming communications traffic control definition boundary protection restrict incoming communications traffic the information system only allows incoming communications from routed to class technical supplemental guidance this control enhancement provides determinations that source and destination address pairs represent authorized allowed communications such determinations can be based on several factors including for example the presence of source destination address pairs in lists of authorized allowed communications the absence of address pairs in lists of unauthorized disallowed pairs or meeting more general rules for authorized allowed source destination pairs related control ac general guide this security control enhancement should be addressed where applicable and if practical to do so suggested assignment it projects
0
284,620
24,611,193,486
IssuesEvent
2022-10-14 21:42:52
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
opened
DISABLED test_binary_op_scalar_fastpath__foreach_mul_cuda_bfloat16 (__main__.TestForeachCUDA)
triaged module: flaky-tests skipped module: mta
Platforms: linux This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_binary_op_scalar_fastpath__foreach_mul_cuda_bfloat16&suite=TestForeachCUDA) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/8897924861). Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 failures and 1 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT BE ALARMED IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_binary_op_scalar_fastpath__foreach_mul_cuda_bfloat16` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
1.0
DISABLED test_binary_op_scalar_fastpath__foreach_mul_cuda_bfloat16 (__main__.TestForeachCUDA) - Platforms: linux This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_binary_op_scalar_fastpath__foreach_mul_cuda_bfloat16&suite=TestForeachCUDA) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/8897924861). Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 failures and 1 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT BE ALARMED IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_binary_op_scalar_fastpath__foreach_mul_cuda_bfloat16` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
test
disabled test binary op scalar fastpath foreach mul cuda main testforeachcuda platforms linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not be alarmed if the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test binary op scalar fastpath foreach mul cuda there should be several instances run as flaky tests are rerun in ci from which you can study the logs
1
579,586
17,195,248,357
IssuesEvent
2021-07-16 16:21:22
magento/magento2
https://api.github.com/repos/magento/magento2
opened
Update laminas/laminas-code composer dependency to version 4.4.2
Priority: P2 Project: Platform Health
Update laminas/laminas-code composer dependency to version 4.4.2
1.0
Update laminas/laminas-code composer dependency to version 4.4.2 - Update laminas/laminas-code composer dependency to version 4.4.2
non_test
update laminas laminas code composer dependency to version update laminas laminas code composer dependency to version
0
529,021
15,378,998,290
IssuesEvent
2021-03-02 19:03:52
internetarchive/openlibrary
https://api.github.com/repos/internetarchive/openlibrary
closed
Can collection carousel display be sorted by # of editions or availability?
Lead: @mekarpeles Needs: Community Discussion Priority: 3 Type: Question
### Question Can collection carousel display be sorted by # of editions or availability? ### Additional context For example, if the results of a query include public domain, lending library, and print-disabled, the ability to sort or filter on the type best suited for the collection would be helpful. See https://openlibrary.org/collections/happy-new-year-2021 where many 'not in library' results appear when there are also borrowable works available. ### Issue resolution criteria This can be closed if a solution is provided or if it is not going to be implemented. ### Stakeholders @seabelis @mekarpeles
1.0
Can collection carousel display be sorted by # of editions or availability? - ### Question Can collection carousel display be sorted by # of editions or availability? ### Additional context For example, if the results of a query include public domain, lending library, and print-disabled, the ability to sort or filter on the type best suited for the collection would be helpful. See https://openlibrary.org/collections/happy-new-year-2021 where many 'not in library' results appear when there are also borrowable works available. ### Issue resolution criteria This can be closed if a solution is provided or if it is not going to be implemented. ### Stakeholders @seabelis @mekarpeles
non_test
can collection carousel display be sorted by of editions or availability question can collection carousel display be sorted by of editions or availability additional context for example if the results of a query include public domain lending library and print disabled the ability to sort or filter on the type best suited for the collection would be helpful see where many not in library results appear when there are also borrowable works available issue resolution criteria this can be closed if a solution is provided or if it is not going to be implemented stakeholders seabelis mekarpeles
0
318,229
27,296,533,454
IssuesEvent
2023-02-23 20:52:55
MPMG-DCC-UFMG/F01
https://api.github.com/repos/MPMG-DCC-UFMG/F01
closed
Teste de generalizacao para a tag Servidores - Dados dos Servidores - Luminárias
generalization test development
DoD: Realizar o teste de Generalização do validador da tag Servidores - Dados dos Servidores para o Município de Luminárias.
1.0
Teste de generalizacao para a tag Servidores - Dados dos Servidores - Luminárias - DoD: Realizar o teste de Generalização do validador da tag Servidores - Dados dos Servidores para o Município de Luminárias.
test
teste de generalizacao para a tag servidores dados dos servidores luminárias dod realizar o teste de generalização do validador da tag servidores dados dos servidores para o município de luminárias
1
21,422
3,899,075,999
IssuesEvent
2016-04-17 14:22:44
BobbyDarkbean/consumer-producer-solutions
https://api.github.com/repos/BobbyDarkbean/consumer-producer-solutions
closed
CP-20: add test extension
enhancement implementation prototype test
Server module mock extension. Implement interfaces: IRequestDecoder, IReplyEncoder, ISocketController, IConnectionTaskChart, IConnectionTaskFactory and provide concrete tasks.
1.0
CP-20: add test extension - Server module mock extension. Implement interfaces: IRequestDecoder, IReplyEncoder, ISocketController, IConnectionTaskChart, IConnectionTaskFactory and provide concrete tasks.
test
cp add test extension server module mock extension implement interfaces irequestdecoder ireplyencoder isocketcontroller iconnectiontaskchart iconnectiontaskfactory and provide concrete tasks
1
66,796
7,018,622,796
IssuesEvent
2017-12-21 14:25:23
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
teamcity: failed tests on master: Examples-ORMs/TestSQLAlchemy, Examples-ORMs/TestSQLAlchemy/FirstRun, Examples-ORMs/TestSQLAlchemy/SecondRun, Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart, Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Order, Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Products, Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Customers, Examples-ORMs/TestSQLAlchemy, Examples-ORMs/TestSQLAlchemy/FirstRun, Examples-ORMs/TestSQLAlchemy/SecondRun, Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart, Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Order, Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Products, Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Customers
Robot test-failure
The following tests appear to have failed: [#451827](https://teamcity.cockroachdb.com/viewLog.html?buildId=451827): ``` --- FAIL: Examples-ORMs/TestSQLAlchemy (183.020s) None --- FAIL: Examples-ORMs/TestSQLAlchemy/FirstRun (180.770s) main_test.go:160: Get http://localhost:6543/ping: dial tcp 127.0.0.1:6543: getsockopt: connection refused --- FAIL: Examples-ORMs/TestSQLAlchemy/SecondRun (1.520s) ------- Stderr: ------- 2017/12/20 21:23:31 process 14415 started: ../cockroach start --logtostderr --insecure --host=localhost --port=0 --http-port=0 --store=/tmp/cockroach-testserver816887418 --listening-url-file=/tmp/cockroach-testserver816887418/listen-url --- FAIL: Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart (0.000s) None --- FAIL: Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Order (0.010s) test_driver.go:213: expecting products from api after creation to be [{0 0x8d4a70 123.4}], found [] --- FAIL: Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Products (0.010s) test_driver.go:213: expecting products from api after creation to be [{0 0x8d4a70 123.4}], found [] --- FAIL: Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Customers (0.020s) test_driver.go:213: expecting products from api after creation to be [{0 0x8d4a70 123.4}], found [] --- FAIL: Examples-ORMs/TestSQLAlchemy (183.020s) None --- FAIL: Examples-ORMs/TestSQLAlchemy/FirstRun (180.770s) main_test.go:160: Get http://localhost:6543/ping: dial tcp 127.0.0.1:6543: getsockopt: connection refused --- FAIL: Examples-ORMs/TestSQLAlchemy/SecondRun (1.520s) ------- Stderr: ------- 2017/12/20 21:23:31 process 14415 started: ../cockroach start --logtostderr --insecure --host=localhost --port=0 --http-port=0 --store=/tmp/cockroach-testserver816887418 --listening-url-file=/tmp/cockroach-testserver816887418/listen-url --- FAIL: Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart (0.000s) None --- FAIL: Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Order (0.010s) test_driver.go:213: expecting products from api after creation to be [{0 0x8d4a70 123.4}], found [] --- FAIL: Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Products (0.010s) test_driver.go:213: expecting products from api after creation to be [{0 0x8d4a70 123.4}], found [] --- FAIL: Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Customers (0.020s) test_driver.go:213: expecting products from api after creation to be [{0 0x8d4a70 123.4}], found [] ``` Please assign, take a look and update the issue accordingly.
1.0
teamcity: failed tests on master: Examples-ORMs/TestSQLAlchemy, Examples-ORMs/TestSQLAlchemy/FirstRun, Examples-ORMs/TestSQLAlchemy/SecondRun, Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart, Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Order, Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Products, Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Customers, Examples-ORMs/TestSQLAlchemy, Examples-ORMs/TestSQLAlchemy/FirstRun, Examples-ORMs/TestSQLAlchemy/SecondRun, Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart, Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Order, Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Products, Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Customers - The following tests appear to have failed: [#451827](https://teamcity.cockroachdb.com/viewLog.html?buildId=451827): ``` --- FAIL: Examples-ORMs/TestSQLAlchemy (183.020s) None --- FAIL: Examples-ORMs/TestSQLAlchemy/FirstRun (180.770s) main_test.go:160: Get http://localhost:6543/ping: dial tcp 127.0.0.1:6543: getsockopt: connection refused --- FAIL: Examples-ORMs/TestSQLAlchemy/SecondRun (1.520s) ------- Stderr: ------- 2017/12/20 21:23:31 process 14415 started: ../cockroach start --logtostderr --insecure --host=localhost --port=0 --http-port=0 --store=/tmp/cockroach-testserver816887418 --listening-url-file=/tmp/cockroach-testserver816887418/listen-url --- FAIL: Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart (0.000s) None --- FAIL: Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Order (0.010s) test_driver.go:213: expecting products from api after creation to be [{0 0x8d4a70 123.4}], found [] --- FAIL: Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Products (0.010s) test_driver.go:213: expecting products from api after creation to be [{0 0x8d4a70 123.4}], found [] --- FAIL: Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Customers (0.020s) test_driver.go:213: expecting products from api after creation to be [{0 0x8d4a70 123.4}], found [] --- FAIL: Examples-ORMs/TestSQLAlchemy (183.020s) None --- FAIL: Examples-ORMs/TestSQLAlchemy/FirstRun (180.770s) main_test.go:160: Get http://localhost:6543/ping: dial tcp 127.0.0.1:6543: getsockopt: connection refused --- FAIL: Examples-ORMs/TestSQLAlchemy/SecondRun (1.520s) ------- Stderr: ------- 2017/12/20 21:23:31 process 14415 started: ../cockroach start --logtostderr --insecure --host=localhost --port=0 --http-port=0 --store=/tmp/cockroach-testserver816887418 --listening-url-file=/tmp/cockroach-testserver816887418/listen-url --- FAIL: Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart (0.000s) None --- FAIL: Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Order (0.010s) test_driver.go:213: expecting products from api after creation to be [{0 0x8d4a70 123.4}], found [] --- FAIL: Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Products (0.010s) test_driver.go:213: expecting products from api after creation to be [{0 0x8d4a70 123.4}], found [] --- FAIL: Examples-ORMs/TestSQLAlchemy/SecondRun/RetrieveFromAPIAfterRestart/Customers (0.020s) test_driver.go:213: expecting products from api after creation to be [{0 0x8d4a70 123.4}], found [] ``` Please assign, take a look and update the issue accordingly.
test
teamcity failed tests on master examples orms testsqlalchemy examples orms testsqlalchemy firstrun examples orms testsqlalchemy secondrun examples orms testsqlalchemy secondrun retrievefromapiafterrestart examples orms testsqlalchemy secondrun retrievefromapiafterrestart order examples orms testsqlalchemy secondrun retrievefromapiafterrestart products examples orms testsqlalchemy secondrun retrievefromapiafterrestart customers examples orms testsqlalchemy examples orms testsqlalchemy firstrun examples orms testsqlalchemy secondrun examples orms testsqlalchemy secondrun retrievefromapiafterrestart examples orms testsqlalchemy secondrun retrievefromapiafterrestart order examples orms testsqlalchemy secondrun retrievefromapiafterrestart products examples orms testsqlalchemy secondrun retrievefromapiafterrestart customers the following tests appear to have failed fail examples orms testsqlalchemy none fail examples orms testsqlalchemy firstrun main test go get dial tcp getsockopt connection refused fail examples orms testsqlalchemy secondrun stderr process started cockroach start logtostderr insecure host localhost port http port store tmp cockroach listening url file tmp cockroach listen url fail examples orms testsqlalchemy secondrun retrievefromapiafterrestart none fail examples orms testsqlalchemy secondrun retrievefromapiafterrestart order test driver go expecting products from api after creation to be found fail examples orms testsqlalchemy secondrun retrievefromapiafterrestart products test driver go expecting products from api after creation to be found fail examples orms testsqlalchemy secondrun retrievefromapiafterrestart customers test driver go expecting products from api after creation to be found fail examples orms testsqlalchemy none fail examples orms testsqlalchemy firstrun main test go get dial tcp getsockopt connection refused fail examples orms testsqlalchemy secondrun stderr process started cockroach start logtostderr insecure host localhost port http port store tmp cockroach listening url file tmp cockroach listen url fail examples orms testsqlalchemy secondrun retrievefromapiafterrestart none fail examples orms testsqlalchemy secondrun retrievefromapiafterrestart order test driver go expecting products from api after creation to be found fail examples orms testsqlalchemy secondrun retrievefromapiafterrestart products test driver go expecting products from api after creation to be found fail examples orms testsqlalchemy secondrun retrievefromapiafterrestart customers test driver go expecting products from api after creation to be found please assign take a look and update the issue accordingly
1
322,961
27,656,502,593
IssuesEvent
2023-03-12 01:57:36
frigid14/stationware
https://api.github.com/repos/frigid14/stationware
closed
Dying between challenges permakills you
bug priority: before playtest difficult
Uhhh uhmm uhmmm. I don't really know a good solution tbqh. The main issue is that the respawning uses the challenge's players, which doesn't include dead people. The main issue is not making this cbt when trying to administrate.
1.0
Dying between challenges permakills you - Uhhh uhmm uhmmm. I don't really know a good solution tbqh. The main issue is that the respawning uses the challenge's players, which doesn't include dead people. The main issue is not making this cbt when trying to administrate.
test
dying between challenges permakills you uhhh uhmm uhmmm i don t really know a good solution tbqh the main issue is that the respawning uses the challenge s players which doesn t include dead people the main issue is not making this cbt when trying to administrate
1
221,604
17,360,218,221
IssuesEvent
2021-07-29 19:28:13
urapadmin/kiosk
https://api.github.com/repos/urapadmin/kiosk
closed
new config key: visibility/collected_material_show_weight
filemaker test-stage
just because we also have show_quantity
1.0
new config key: visibility/collected_material_show_weight - just because we also have show_quantity
test
new config key visibility collected material show weight just because we also have show quantity
1
34,562
7,844,000,100
IssuesEvent
2018-06-19 08:19:15
An-Sar/PrimalCore
https://api.github.com/repos/An-Sar/PrimalCore
closed
Major lag on every new world
Code Review World Gen duplicate
Everytime I create a new world, the world generates normally, but after a few minutes it starts to lag a lot, eventually leading to Minecraft stopping responding. https://paste.ee/p/OPsLu
1.0
Major lag on every new world - Everytime I create a new world, the world generates normally, but after a few minutes it starts to lag a lot, eventually leading to Minecraft stopping responding. https://paste.ee/p/OPsLu
non_test
major lag on every new world everytime i create a new world the world generates normally but after a few minutes it starts to lag a lot eventually leading to minecraft stopping responding
0
305,539
26,391,904,774
IssuesEvent
2023-01-12 16:15:59
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Fix miscellaneous_ops.test_torch_tril_indices
PyTorch Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3897647796/jobs/6655531507" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/3897647796/jobs/6655531507" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3897647796/jobs/6655531507" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/3897647796/jobs/6655531507" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>Not found</summary> Not found </details> <details> <summary>Not found</summary> Not found </details> <details> <summary>Not found</summary> Not found </details> <details> <summary>Not found</summary> Not found </details>
1.0
Fix miscellaneous_ops.test_torch_tril_indices - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3897647796/jobs/6655531507" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/3897647796/jobs/6655531507" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3897647796/jobs/6655531507" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/3897647796/jobs/6655531507" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>Not found</summary> Not found </details> <details> <summary>Not found</summary> Not found </details> <details> <summary>Not found</summary> Not found </details> <details> <summary>Not found</summary> Not found </details>
test
fix miscellaneous ops test torch tril indices tensorflow img src torch img src numpy img src jax img src not found not found not found not found not found not found not found not found
1
116,387
9,850,857,245
IssuesEvent
2019-06-19 09:08:27
zonemaster/zonemaster
https://api.github.com/repos/zonemaster/zonemaster
closed
Add test case that on queries with OPT
test spec
If a name server gets a query with an OPT section, it should respond with an OPT section, unless it does not support OPT, and then is should respond with FORMERR. Create a Test Case that checks this.
1.0
Add test case that on queries with OPT - If a name server gets a query with an OPT section, it should respond with an OPT section, unless it does not support OPT, and then is should respond with FORMERR. Create a Test Case that checks this.
test
add test case that on queries with opt if a name server gets a query with an opt section it should respond with an opt section unless it does not support opt and then is should respond with formerr create a test case that checks this
1
204,001
15,397,418,349
IssuesEvent
2021-03-03 22:07:18
Vivid-Project/frontend
https://api.github.com/repos/Vivid-Project/frontend
closed
FE End to End testing
Frontend Testing
- [ ] dashboard to add new dream form - [ ] add a new dream from to dashboard - [ ] dashboard to all dream view - [ ] all dreams view to dashboard - [ ] dashboard to charts view - [ ] charts view to dashboard - [ ] dashboard search to results - [ ] results to dashboard
1.0
FE End to End testing - - [ ] dashboard to add new dream form - [ ] add a new dream from to dashboard - [ ] dashboard to all dream view - [ ] all dreams view to dashboard - [ ] dashboard to charts view - [ ] charts view to dashboard - [ ] dashboard search to results - [ ] results to dashboard
test
fe end to end testing dashboard to add new dream form add a new dream from to dashboard dashboard to all dream view all dreams view to dashboard dashboard to charts view charts view to dashboard dashboard search to results results to dashboard
1
671,729
22,773,569,134
IssuesEvent
2022-07-08 12:27:15
geosolutions-it/MapStore2
https://api.github.com/repos/geosolutions-it/MapStore2
reopened
Request of BILTerrainProvider layer are not using proxy url
bug Priority: High Accepted C027-COMUNE_FI-2021-SUPPORT 3D
## Description <!-- Add here a few sentences describing the bug. --> The implementation of WMSLayer and TerrainLayer in cesium are providing a configuration for proxy not supported by the BILTerrainProvider class and for this reason the proxy url is never applied. Currently BILTerrainProvider accept proxy as root options and not part of the Cesium.Resource instance - [TerrainLayer](https://github.com/geosolutions-it/MapStore2/blob/master/web/client/components/map/cesium/plugins/TerrainLayer.js#L53-L57) - [WMSLayer](https://github.com/geosolutions-it/MapStore2/blob/master/web/client/components/map/cesium/plugins/WMSLayer.js#L143-L147) ## How to reproduce <!-- A list of steps to reproduce the bug --> - add a terrain layer as per [doc](https://mapstore.readthedocs.io/en/latest/developer-guide/maps-configuration/#terrain) with addtionalLayers property of the Map plugin. This configuration can be applied also in a context to a Map plugin - ensure the terrain layer has a different hostname from the terrain endpoint - open the map viewer and switch to 3D *Expected Result* <!-- Describe here the expected result --> Request to the terrain layer should use the proxy url if not included in the useCORS array of proxyUrl configuration *Current Result* <!-- Describe here the current behavior --> The request of the WMS terrain provider are always direct and they are not using proxy - [x] Not browser related <details><summary> <b>Browser info</b> </summary> <!-- If browser related, please compile the following table --> <!-- If your browser is not in the list please add a new row to the table with the version --> (use this site: <a href="https://www.whatsmybrowser.org/">https://www.whatsmybrowser.org/</a> for non expert users) | Browser Affected | Version | |---|---| |Internet Explorer| | |Edge| | |Chrome| | |Firefox| | |Safari| | </details> ## Other useful information <!-- error stack trace, screenshot, videos, or link to repository code are welcome -->
1.0
Request of BILTerrainProvider layer are not using proxy url - ## Description <!-- Add here a few sentences describing the bug. --> The implementation of WMSLayer and TerrainLayer in cesium are providing a configuration for proxy not supported by the BILTerrainProvider class and for this reason the proxy url is never applied. Currently BILTerrainProvider accept proxy as root options and not part of the Cesium.Resource instance - [TerrainLayer](https://github.com/geosolutions-it/MapStore2/blob/master/web/client/components/map/cesium/plugins/TerrainLayer.js#L53-L57) - [WMSLayer](https://github.com/geosolutions-it/MapStore2/blob/master/web/client/components/map/cesium/plugins/WMSLayer.js#L143-L147) ## How to reproduce <!-- A list of steps to reproduce the bug --> - add a terrain layer as per [doc](https://mapstore.readthedocs.io/en/latest/developer-guide/maps-configuration/#terrain) with addtionalLayers property of the Map plugin. This configuration can be applied also in a context to a Map plugin - ensure the terrain layer has a different hostname from the terrain endpoint - open the map viewer and switch to 3D *Expected Result* <!-- Describe here the expected result --> Request to the terrain layer should use the proxy url if not included in the useCORS array of proxyUrl configuration *Current Result* <!-- Describe here the current behavior --> The request of the WMS terrain provider are always direct and they are not using proxy - [x] Not browser related <details><summary> <b>Browser info</b> </summary> <!-- If browser related, please compile the following table --> <!-- If your browser is not in the list please add a new row to the table with the version --> (use this site: <a href="https://www.whatsmybrowser.org/">https://www.whatsmybrowser.org/</a> for non expert users) | Browser Affected | Version | |---|---| |Internet Explorer| | |Edge| | |Chrome| | |Firefox| | |Safari| | </details> ## Other useful information <!-- error stack trace, screenshot, videos, or link to repository code are welcome -->
non_test
request of bilterrainprovider layer are not using proxy url description the implementation of wmslayer and terrainlayer in cesium are providing a configuration for proxy not supported by the bilterrainprovider class and for this reason the proxy url is never applied currently bilterrainprovider accept proxy as root options and not part of the cesium resource instance how to reproduce add a terrain layer as per with addtionallayers property of the map plugin this configuration can be applied also in a context to a map plugin ensure the terrain layer has a different hostname from the terrain endpoint open the map viewer and switch to expected result request to the terrain layer should use the proxy url if not included in the usecors array of proxyurl configuration current result the request of the wms terrain provider are always direct and they are not using proxy not browser related browser info use this site a href for non expert users browser affected version internet explorer edge chrome firefox safari other useful information
0
220,805
17,261,950,456
IssuesEvent
2021-07-22 08:52:41
FlowCrypt/flowcrypt-android
https://api.github.com/repos/FlowCrypt/flowcrypt-android
opened
Test decryption errors after moving to use PGPainless for the message decryption
android_testing
It would be nice to have such tests. I think it's very important to prevent security bugs (for example after upgrading PGPainless to a newer version). We should be sure that a received content is parsed to expected `blocks` and Android shows the expected UI.
1.0
Test decryption errors after moving to use PGPainless for the message decryption - It would be nice to have such tests. I think it's very important to prevent security bugs (for example after upgrading PGPainless to a newer version). We should be sure that a received content is parsed to expected `blocks` and Android shows the expected UI.
test
test decryption errors after moving to use pgpainless for the message decryption it would be nice to have such tests i think it s very important to prevent security bugs for example after upgrading pgpainless to a newer version we should be sure that a received content is parsed to expected blocks and android shows the expected ui
1
140,280
5,399,734,518
IssuesEvent
2017-02-27 20:12:12
GRIS-UdeM/SpatGRIS
https://api.github.com/repos/GRIS-UdeM/SpatGRIS
opened
DP Buffer size Incorrect
bug High priority
![screen shot 2017-02-27 at 15 05 42](https://cloud.githubusercontent.com/assets/6116482/23377595/9c32192a-fcfe-11e6-80ab-03cdd0e82f16.png) Attention dans DP, le plugin croit qu'il travaille à 2048 mais en réalité DP est à 512. Cela fait planter DP quand on est en mode 12*12 ou 16*16.
1.0
DP Buffer size Incorrect - ![screen shot 2017-02-27 at 15 05 42](https://cloud.githubusercontent.com/assets/6116482/23377595/9c32192a-fcfe-11e6-80ab-03cdd0e82f16.png) Attention dans DP, le plugin croit qu'il travaille à 2048 mais en réalité DP est à 512. Cela fait planter DP quand on est en mode 12*12 ou 16*16.
non_test
dp buffer size incorrect attention dans dp le plugin croit qu il travaille à mais en réalité dp est à cela fait planter dp quand on est en mode ou
0
250,442
21,299,718,629
IssuesEvent
2022-04-15 00:20:29
NMGRL/pychron
https://api.github.com/repos/NMGRL/pychron
closed
Add ability to move label on isochron
Enhancement Testing Required Data Presentation
When plotting highly radiogenic data, labels for step commonly overlap spatially. Add ability to move labels for graphic clarity.
1.0
Add ability to move label on isochron - When plotting highly radiogenic data, labels for step commonly overlap spatially. Add ability to move labels for graphic clarity.
test
add ability to move label on isochron when plotting highly radiogenic data labels for step commonly overlap spatially add ability to move labels for graphic clarity
1
211,898
16,464,257,410
IssuesEvent
2021-05-22 04:31:01
neuropsychology/NeuroKit
https://api.github.com/repos/neuropsychology/NeuroKit
closed
Docs: Example gallery
documentation :scroll: inactive 👻
Would be nice to have an example gallery such as [here](https://sphinx-nbexamples.readthedocs.io/en/latest/examples/index.html). Other example: http://visbrain.org/index.html
1.0
Docs: Example gallery - Would be nice to have an example gallery such as [here](https://sphinx-nbexamples.readthedocs.io/en/latest/examples/index.html). Other example: http://visbrain.org/index.html
non_test
docs example gallery would be nice to have an example gallery such as other example
0
178,441
6,608,821,410
IssuesEvent
2017-09-19 12:33:59
RobotLocomotion/drake
https://api.github.com/repos/RobotLocomotion/drake
opened
CI coverage of pre-built drake visualizer
priority: medium team: kitware type: continuous integration
Drake's CI should offer coverage of pre-built drake visualizer -- for example, that is launches without faults, that it runs for a few seconds without crashing, that it is able to locate data resources. It is not unusual for users to experience problems running this software, so we should have CI coverage to point to that demonstrates that it works (and helps us keep regressions off of master). Relates to #5621 as the feature to be tested in CI. Relates to #6834 as an example of a scenario that should be tested in CI.
1.0
CI coverage of pre-built drake visualizer - Drake's CI should offer coverage of pre-built drake visualizer -- for example, that is launches without faults, that it runs for a few seconds without crashing, that it is able to locate data resources. It is not unusual for users to experience problems running this software, so we should have CI coverage to point to that demonstrates that it works (and helps us keep regressions off of master). Relates to #5621 as the feature to be tested in CI. Relates to #6834 as an example of a scenario that should be tested in CI.
non_test
ci coverage of pre built drake visualizer drake s ci should offer coverage of pre built drake visualizer for example that is launches without faults that it runs for a few seconds without crashing that it is able to locate data resources it is not unusual for users to experience problems running this software so we should have ci coverage to point to that demonstrates that it works and helps us keep regressions off of master relates to as the feature to be tested in ci relates to as an example of a scenario that should be tested in ci
0
22,182
3,940,716,154
IssuesEvent
2016-04-27 02:44:25
extnet/Ext.NET
https://api.github.com/repos/extnet/Ext.NET
closed
GridView's DisableSelection="false" doesn't disable selection
3.x fixed-in-latest-extjs sencha
http://forums.ext.net/showthread.php?53651 http://www.sencha.com/forum/showthread.php?298197 **Update:** This is allegedly fixed since ExtJS 5.1.2.
1.0
GridView's DisableSelection="false" doesn't disable selection - http://forums.ext.net/showthread.php?53651 http://www.sencha.com/forum/showthread.php?298197 **Update:** This is allegedly fixed since ExtJS 5.1.2.
test
gridview s disableselection false doesn t disable selection update this is allegedly fixed since extjs
1
52,859
6,283,865,739
IssuesEvent
2017-07-19 05:42:18
intel-analytics/BigDL
https://api.github.com/repos/intel-analytics/BigDL
opened
Pip install python bigdl need sudo
0.2 release test high priority
If I run the install commands in doc ``` pip install --upgrade pip pip install BigDL==0.2.0.dev3 # for Python 2.7 ``` It will throw exception ``` Downloading/unpacking pip from https://pypi.python.org/packages/b6/ac/7015eb97dc749283ffdec1c3a88ddb8ae03b8fad0f0e611408f196358da3/pip-9.0.1-py2.py3-none-any.whl#md5=297dbd16ef53bcef0447d245815f5144 Downloading pip-9.0.1-py2.py3-none-any.whl (1.3MB): 1.3MB downloaded Installing collected packages: pip Found existing installation: pip 1.5.4 Not uninstalling pip at /usr/lib/python2.7/dist-packages, owned by OS Can't roll back pip; was not uninstalled Cleaning up... Downloading/unpacking BigDL==0.2.0.dev3 Could not find any downloads that satisfy the requirement BigDL==0.2.0.dev3 Cleaning up... No distributions at all found for BigDL==0.2.0.dev3 Storing debug log for failure in /tmp/tmpMCxkeC ``` use sudo help fix this problem for me ``` sudo pip install --upgrade pip sudo pip install BigDL==0.2.0.dev3 # for Python 2.7 ```
1.0
Pip install python bigdl need sudo - If I run the install commands in doc ``` pip install --upgrade pip pip install BigDL==0.2.0.dev3 # for Python 2.7 ``` It will throw exception ``` Downloading/unpacking pip from https://pypi.python.org/packages/b6/ac/7015eb97dc749283ffdec1c3a88ddb8ae03b8fad0f0e611408f196358da3/pip-9.0.1-py2.py3-none-any.whl#md5=297dbd16ef53bcef0447d245815f5144 Downloading pip-9.0.1-py2.py3-none-any.whl (1.3MB): 1.3MB downloaded Installing collected packages: pip Found existing installation: pip 1.5.4 Not uninstalling pip at /usr/lib/python2.7/dist-packages, owned by OS Can't roll back pip; was not uninstalled Cleaning up... Downloading/unpacking BigDL==0.2.0.dev3 Could not find any downloads that satisfy the requirement BigDL==0.2.0.dev3 Cleaning up... No distributions at all found for BigDL==0.2.0.dev3 Storing debug log for failure in /tmp/tmpMCxkeC ``` use sudo help fix this problem for me ``` sudo pip install --upgrade pip sudo pip install BigDL==0.2.0.dev3 # for Python 2.7 ```
test
pip install python bigdl need sudo if i run the install commands in doc pip install upgrade pip pip install bigdl for python it will throw exception downloading unpacking pip from downloading pip none any whl downloaded installing collected packages pip found existing installation pip not uninstalling pip at usr lib dist packages owned by os can t roll back pip was not uninstalled cleaning up downloading unpacking bigdl could not find any downloads that satisfy the requirement bigdl cleaning up no distributions at all found for bigdl storing debug log for failure in tmp tmpmcxkec use sudo help fix this problem for me sudo pip install upgrade pip sudo pip install bigdl for python
1
80,923
7,761,168,756
IssuesEvent
2018-06-01 08:59:33
Nineclown/The-Convenient-ATM
https://api.github.com/repos/Nineclown/The-Convenient-ATM
closed
Definition: 분실된 카드의 재발급을 신청하는 기능
Failed Test pass
재발급의 의미를 더 정확히 정의해야한다. 재발급을 하면 카드 목록에 다시 그 카드가 추가되어야 하는건지 아니면 단순히 재발급이란 액션을 취한건지 명확해야한다. T220: 분실된 카드의 재발급을 신청하는 기능 https://jaehyun379.testrail.io/index.php?/tests/view/220
1.0
Definition: 분실된 카드의 재발급을 신청하는 기능 - 재발급의 의미를 더 정확히 정의해야한다. 재발급을 하면 카드 목록에 다시 그 카드가 추가되어야 하는건지 아니면 단순히 재발급이란 액션을 취한건지 명확해야한다. T220: 분실된 카드의 재발급을 신청하는 기능 https://jaehyun379.testrail.io/index.php?/tests/view/220
test
definition 분실된 카드의 재발급을 신청하는 기능 재발급의 의미를 더 정확히 정의해야한다 재발급을 하면 카드 목록에 다시 그 카드가 추가되어야 하는건지 아니면 단순히 재발급이란 액션을 취한건지 명확해야한다 분실된 카드의 재발급을 신청하는 기능
1
431,992
12,487,403,901
IssuesEvent
2020-05-31 08:53:23
on3iro/aeons-end-randomizer
https://api.github.com/repos/on3iro/aeons-end-randomizer
closed
RFE: Language translation
Priority: High feature
### Is your feature request related to a problem? Please describe. I'm a French player and using French core box. It takes to time to find out card as name doesn't match. ### Describe the solution you'd like As additionnal detail, I would like to say that some French players are mixing language content (even if there are not stricly compatible, card back is different). So: **Quick Win** Just add `Expansion` for each language. It's the simplest solution I think, maybe having a way to quick filter on `Expansions` screen could help. ### Describe alternatives you've considered** **`Expansion` language** For each `Expansion` add a language selection. ### Additional context Whatever the solution, I think using flag icon would be nice !
1.0
RFE: Language translation - ### Is your feature request related to a problem? Please describe. I'm a French player and using French core box. It takes to time to find out card as name doesn't match. ### Describe the solution you'd like As additionnal detail, I would like to say that some French players are mixing language content (even if there are not stricly compatible, card back is different). So: **Quick Win** Just add `Expansion` for each language. It's the simplest solution I think, maybe having a way to quick filter on `Expansions` screen could help. ### Describe alternatives you've considered** **`Expansion` language** For each `Expansion` add a language selection. ### Additional context Whatever the solution, I think using flag icon would be nice !
non_test
rfe language translation is your feature request related to a problem please describe i m a french player and using french core box it takes to time to find out card as name doesn t match describe the solution you d like as additionnal detail i would like to say that some french players are mixing language content even if there are not stricly compatible card back is different so quick win just add expansion for each language it s the simplest solution i think maybe having a way to quick filter on expansions screen could help describe alternatives you ve considered expansion language for each expansion add a language selection additional context whatever the solution i think using flag icon would be nice
0
296,559
25,559,179,603
IssuesEvent
2022-11-30 09:28:34
wazuh/wazuh
https://api.github.com/repos/wazuh/wazuh
opened
Release 4.4.0 - Alpha 1 - E2E UX tests - Monitoring Docker
module/docker type/test/manual team/framework release test/4.4.0
The following issue aims to run the specified test for the current release candidate, report the results, and open new issues for any encountered errors. ## Test information | | | |-------------------------|--------------------------------------------| | **Test name** | Monitoring Docker | | **Category** | Log data analysis | | **Deployment option** |Single indexer and dashboard, multiple servers: Installation assistant| | **Main release issue** |https://github.com/wazuh/wazuh/issues/15505| | **Release candidate #** |https://github.com/wazuh/wazuh/issues/15519| ## Test description Configure Docker listener in a Wazuh Agent Test that this proof of concept works as expected for the current release under test: https://documentation.wazuh.com/current/proof-of-concept-guide/monitoring-docker.html#poc-monitoring-docker Test that this blog post works as expected for the current release under test: https://wazuh.com/blog/monitoring-docker-container-logs-with-wazuh/ ## Test report procedure All test results must have one of the following statuses: | | | |---------------------------------|--------------------------------------------| | :green_circle: | All checks passed. | | :red_circle: | There is at least one failed result. | | :yellow_circle: | There is at least one expected failure or skipped test and no failures. | Any failing test must be properly addressed with a new issue, detailing the error and the possible cause. An extended report of the test results must be attached as a ZIP or TXT file. Please attach any documents, screenshots, or tables to the issue update with the results. This report can be used by the auditors to dig deeper into any possible failures and details. ## Conclusions <!-- All tests have been executed and the results can be found [here](). | | | | | |----------------|-------------|---------------------|----------------| | **Status** | **Test** | **Failure type** | **Notes** | | | | | | All tests have passed and the fails have been reported or justified. Therefore, I conclude that this issue is finished and OK for this release candidate. --> ## Auditors validation The definition of done for this one is the validation of the conclusions and the test results from all auditors. All checks from below must be accepted in order to close this issue. - [ ] @davidjiglesias
2.0
Release 4.4.0 - Alpha 1 - E2E UX tests - Monitoring Docker - The following issue aims to run the specified test for the current release candidate, report the results, and open new issues for any encountered errors. ## Test information | | | |-------------------------|--------------------------------------------| | **Test name** | Monitoring Docker | | **Category** | Log data analysis | | **Deployment option** |Single indexer and dashboard, multiple servers: Installation assistant| | **Main release issue** |https://github.com/wazuh/wazuh/issues/15505| | **Release candidate #** |https://github.com/wazuh/wazuh/issues/15519| ## Test description Configure Docker listener in a Wazuh Agent Test that this proof of concept works as expected for the current release under test: https://documentation.wazuh.com/current/proof-of-concept-guide/monitoring-docker.html#poc-monitoring-docker Test that this blog post works as expected for the current release under test: https://wazuh.com/blog/monitoring-docker-container-logs-with-wazuh/ ## Test report procedure All test results must have one of the following statuses: | | | |---------------------------------|--------------------------------------------| | :green_circle: | All checks passed. | | :red_circle: | There is at least one failed result. | | :yellow_circle: | There is at least one expected failure or skipped test and no failures. | Any failing test must be properly addressed with a new issue, detailing the error and the possible cause. An extended report of the test results must be attached as a ZIP or TXT file. Please attach any documents, screenshots, or tables to the issue update with the results. This report can be used by the auditors to dig deeper into any possible failures and details. ## Conclusions <!-- All tests have been executed and the results can be found [here](). | | | | | |----------------|-------------|---------------------|----------------| | **Status** | **Test** | **Failure type** | **Notes** | | | | | | All tests have passed and the fails have been reported or justified. Therefore, I conclude that this issue is finished and OK for this release candidate. --> ## Auditors validation The definition of done for this one is the validation of the conclusions and the test results from all auditors. All checks from below must be accepted in order to close this issue. - [ ] @davidjiglesias
test
release alpha ux tests monitoring docker the following issue aims to run the specified test for the current release candidate report the results and open new issues for any encountered errors test information test name monitoring docker category log data analysis deployment option single indexer and dashboard multiple servers installation assistant main release issue release candidate test description configure docker listener in a wazuh agent test that this proof of concept works as expected for the current release under test test that this blog post works as expected for the current release under test test report procedure all test results must have one of the following statuses green circle all checks passed red circle there is at least one failed result yellow circle there is at least one expected failure or skipped test and no failures any failing test must be properly addressed with a new issue detailing the error and the possible cause an extended report of the test results must be attached as a zip or txt file please attach any documents screenshots or tables to the issue update with the results this report can be used by the auditors to dig deeper into any possible failures and details conclusions all tests have been executed and the results can be found status test failure type notes all tests have passed and the fails have been reported or justified therefore i conclude that this issue is finished and ok for this release candidate auditors validation the definition of done for this one is the validation of the conclusions and the test results from all auditors all checks from below must be accepted in order to close this issue davidjiglesias
1
325,211
27,856,522,096
IssuesEvent
2023-03-20 23:41:26
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
opened
Test: find in output by default
testplan-item
Refs: https://github.com/microsoft/vscode/issues/174969 - [ ] anyOS - [ ] anyOS Complexity: 3 Authors: @rebornix, @amunger --- This month we made "Find in Output" option turned on by default, this means when you open a notebook, we will try to render cell outputs on idle and when you cmd+f to open the Find Widget, we will try to search the keyword in both cell inputs and outputs * Please test it works as expected, open a notebook which contains output, and you can reveal to matches in outputs. * When you have `notebook.output.scrolling` turned on, it's still functioning as expected _Set up_ To test above changes, you would need to have a working Python/Jupyter environment. You can use the Dev Container for this test, for example: * Install Dev Container remote extension * From the remote picker, choose "Create Dev Container...". From the docker container images, choose Jupyter Data Science Notebook, or Anaconda. * Once connected to the Container, install the Prerelease version of Jupyter and Python extensions * Create an ipynb file to start testing. Sample code to generate long outputs ``` import time for i in range(200): time.sleep(.1) print(i) print("hello") ```
1.0
Test: find in output by default - Refs: https://github.com/microsoft/vscode/issues/174969 - [ ] anyOS - [ ] anyOS Complexity: 3 Authors: @rebornix, @amunger --- This month we made "Find in Output" option turned on by default, this means when you open a notebook, we will try to render cell outputs on idle and when you cmd+f to open the Find Widget, we will try to search the keyword in both cell inputs and outputs * Please test it works as expected, open a notebook which contains output, and you can reveal to matches in outputs. * When you have `notebook.output.scrolling` turned on, it's still functioning as expected _Set up_ To test above changes, you would need to have a working Python/Jupyter environment. You can use the Dev Container for this test, for example: * Install Dev Container remote extension * From the remote picker, choose "Create Dev Container...". From the docker container images, choose Jupyter Data Science Notebook, or Anaconda. * Once connected to the Container, install the Prerelease version of Jupyter and Python extensions * Create an ipynb file to start testing. Sample code to generate long outputs ``` import time for i in range(200): time.sleep(.1) print(i) print("hello") ```
test
test find in output by default refs anyos anyos complexity authors rebornix amunger this month we made find in output option turned on by default this means when you open a notebook we will try to render cell outputs on idle and when you cmd f to open the find widget we will try to search the keyword in both cell inputs and outputs please test it works as expected open a notebook which contains output and you can reveal to matches in outputs when you have notebook output scrolling turned on it s still functioning as expected set up to test above changes you would need to have a working python jupyter environment you can use the dev container for this test for example install dev container remote extension from the remote picker choose create dev container from the docker container images choose jupyter data science notebook or anaconda once connected to the container install the prerelease version of jupyter and python extensions create an ipynb file to start testing sample code to generate long outputs import time for i in range time sleep print i print hello
1
91,374
8,303,852,620
IssuesEvent
2018-09-21 19:00:26
tschottdorf/cockroach
https://api.github.com/repos/tschottdorf/cockroach
closed
teamcity: failed test: test/TestImportPgDump, t, -
C-test-failure O-robot
The following tests appear to have failed on release-banana. You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+test/TestImportPgDump, t, -). [#864629](https://teamcity.cockroachdb.com/viewLog.html?buildId=864629): ``` e---- FAIL: test/TestImportPgDump (0.000s) Test ended in panic. ------- Stdout: ------- W180827 20:41:52.746991 50862 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180827 20:41:52.757923 50862 server/server.go:830 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled I180827 20:41:52.758132 50862 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180827 20:41:52.758156 50862 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180827 20:41:52.761168 50862 server/config.go:496 [n?] 1 storage engine initialized I180827 20:41:52.761191 50862 server/config.go:499 [n?] RocksDB cache size: 128 MiB I180827 20:41:52.761204 50862 server/config.go:499 [n?] store 0: in-memory, size 0 B I180827 20:41:52.767725 50862 server/node.go:373 [n?] **** cluster d5e53e69-a109-4eb6-91bf-29e74ae744ba has been created I180827 20:41:52.767752 50862 server/server.go:1401 [n?] **** add additional nodes by specifying --join=127.0.0.1:41477 I180827 20:41:52.767936 50862 gossip/gossip.go:382 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:41477" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:12 > build_tag:"v2.1.0-alpha.20180702-2025-gf1e7bb1" started_at:1535402512767856449 I180827 20:41:52.770338 50862 storage/store.go:1541 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available I180827 20:41:52.770546 50862 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=1, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00} I180827 20:41:52.770626 50862 storage/stores.go:242 [n1] read 0 node addresses from persistent storage I180827 20:41:52.770721 50862 server/node.go:697 [n1] connecting to gossip network to verify cluster ID... I180827 20:41:52.770760 50862 server/node.go:722 [n1] node connected via gossip and verified as part of cluster "d5e53e69-a109-4eb6-91bf-29e74ae744ba" I180827 20:41:52.770788 50862 server/node.go:546 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes [] I180827 20:41:52.771023 50862 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180827 20:41:52.771066 50862 server/server.go:1807 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined I180827 20:41:52.771159 50862 server/server.go:1538 [n1] starting https server at 127.0.0.1:42563 (use: 127.0.0.1:42563) I180827 20:41:52.771188 50862 server/server.go:1540 [n1] starting grpc/postgres server at 127.0.0.1:41477 I180827 20:41:52.771209 50862 server/server.go:1541 [n1] advertising CockroachDB node at 127.0.0.1:41477 I180827 20:41:52.775258 51089 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180827 20:41:52.776337 50925 storage/replica_command.go:298 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2] I180827 20:41:52.788832 51094 storage/replica_command.go:298 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3] W180827 20:41:52.790188 51128 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "unnamed" id=ec083bbe key=/Table/SystemConfigSpan/Start rw=true pri=0.01126188 iso=SERIALIZABLE stat=PENDING epo=0 ts=1535402512.772758792,0 orig=1535402512.772758792,0 max=1535402512.772758792,0 wto=false rop=false seq=6 I180827 20:41:52.790695 51118 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root} I180827 20:41:52.795125 51100 storage/replica_command.go:298 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4] I180827 20:41:52.800783 51143 storage/replica_command.go:298 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5] I180827 20:41:52.807906 51165 storage/replica_command.go:298 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6] I180827 20:41:52.811784 51141 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:2.0-12 User:root} I180827 20:41:52.818164 50799 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root} I180827 20:41:52.821094 51188 storage/replica_command.go:298 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7] I180827 20:41:52.830709 51176 storage/replica_command.go:298 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8] I180827 20:41:52.839374 51187 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:045a1c98-219f-445b-bd6b-d481f04d6b0d User:root} I180827 20:41:52.849534 51154 storage/replica_command.go:298 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9] I180827 20:41:52.855898 51218 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root} I180827 20:41:52.861462 51240 storage/replica_command.go:298 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10] I180827 20:41:52.868342 51268 storage/replica_command.go:298 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11] I180827 20:41:52.872706 51256 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root} I180827 20:41:52.874819 51264 storage/replica_command.go:298 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12] I180827 20:41:52.876403 50862 server/server.go:1594 [n1] done ensuring all necessary migrations have run I180827 20:41:52.876433 50862 server/server.go:1597 [n1] serving sql connections I180827 20:41:52.879108 51233 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version I180827 20:41:52.879639 51235 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:41477} Attrs: Locality: ServerVersion:2.0-12 BuildTag:v2.1.0-alpha.20180702-2025-gf1e7bb1 StartedAt:1535402512767856449 LocalityAddress:[]} ClusterID:d5e53e69-a109-4eb6-91bf-29e74ae744ba StartedAt:1535402512767856449 LastUp:1535402512767856449} I180827 20:41:52.880318 51302 storage/replica_command.go:298 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13] I180827 20:41:52.927701 50819 storage/replica_command.go:298 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14] I180827 20:41:52.940165 51323 storage/replica_command.go:298 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15] I180827 20:41:52.948539 51355 storage/replica_command.go:298 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16] I180827 20:41:52.953658 51380 storage/replica_command.go:298 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17] I180827 20:41:52.961237 51137 storage/replica_command.go:298 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18] I180827 20:41:52.966548 50832 storage/replica_command.go:298 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19] I180827 20:41:52.977113 51362 storage/replica_command.go:298 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20] I180827 20:41:53.041315 51440 storage/replica_command.go:298 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21] I180827 20:41:53.047478 51414 storage/replica_command.go:298 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22] W180827 20:41:53.081214 50862 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180827 20:41:53.089127 50862 server/server.go:830 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled I180827 20:41:53.089322 50862 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180827 20:41:53.089338 50862 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180827 20:41:53.102793 50862 server/config.go:496 [n?] 1 storage engine initialized I180827 20:41:53.102863 50862 server/config.go:499 [n?] RocksDB cache size: 128 MiB I180827 20:41:53.102878 50862 server/config.go:499 [n?] store 0: in-memory, size 0 B W180827 20:41:53.102953 50862 gossip/gossip.go:1371 [n?] no incoming or outgoing connections I180827 20:41:53.103001 50862 server/server.go:1403 [n?] no stores bootstrapped and --join flag specified, awaiting init command. I180827 20:41:53.115344 51458 gossip/client.go:129 [n?] started gossip client to 127.0.0.1:41477 I180827 20:41:53.125579 51530 gossip/server.go:217 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:36113} I180827 20:41:53.127987 50862 server/node.go:697 [n?] connecting to gossip network to verify cluster ID... I180827 20:41:53.128034 50862 server/node.go:722 [n?] node connected via gossip and verified as part of cluster "d5e53e69-a109-4eb6-91bf-29e74ae744ba" I180827 20:41:53.128397 51575 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180827 20:41:53.134920 51574 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180827 20:41:53.135628 50862 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180827 20:41:53.136461 50862 server/node.go:428 [n?] new node allocated ID 2 I180827 20:41:53.136541 50862 gossip/gossip.go:382 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:36113" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:12 > build_tag:"v2.1.0-alpha.20180702-2025-gf1e7bb1" started_at:1535402513136479434 I180827 20:41:53.136591 50862 storage/stores.go:242 [n2] read 0 node addresses from persistent storage I180827 20:41:53.136624 50862 storage/stores.go:261 [n2] wrote 1 node addresses to persistent storage I180827 20:41:53.137485 51552 storage/stores.go:261 [n1] wrote 1 node addresses to persistent storage I180827 20:41:53.139442 50862 server/node.go:672 [n2] bootstrapped store [n2,s2] I180827 20:41:53.139577 50862 server/node.go:546 [n2] node=2: started with [] engine(s) and attributes [] I180827 20:41:53.140140 50862 server/status/recorder.go:652 [n2] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180827 20:41:53.140166 50862 server/server.go:1807 [n2] Could not start heap profiler worker due to: directory to store profiles could not be determined I180827 20:41:53.140233 50862 server/server.go:1538 [n2] starting https server at 127.0.0.1:39947 (use: 127.0.0.1:39947) I180827 20:41:53.140246 50862 server/server.go:1540 [n2] starting grpc/postgres server at 127.0.0.1:36113 I180827 20:41:53.140256 50862 server/server.go:1541 [n2] advertising CockroachDB node at 127.0.0.1:36113 I180827 20:41:53.140624 51685 server/status/recorder.go:652 [n2,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180827 20:41:53.153945 50862 server/server.go:1594 [n2] done ensuring all necessary migrations have run I180827 20:41:53.153974 50862 server/server.go:1597 [n2] serving sql connections W180827 20:41:53.165268 50862 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180827 20:41:53.185802 51467 server/server_update.go:67 [n2] no need to upgrade, cluster already at the newest version I180827 20:41:53.186848 51469 sql/event_log.go:126 [n2] Event: "node_join", target: 2, info: {Descriptor:{NodeID:2 Address:{NetworkField:tcp AddressField:127.0.0.1:36113} Attrs: Locality: ServerVersion:2.0-12 BuildTag:v2.1.0-alpha.20180702-2025-gf1e7bb1 StartedAt:1535402513136479434 LocalityAddress:[]} ClusterID:d5e53e69-a109-4eb6-91bf-29e74ae744ba StartedAt:1535402513136479434 LastUp:1535402513136479434} I180827 20:41:53.189622 50862 server/server.go:830 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled I180827 20:41:53.189776 50862 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180827 20:41:53.189808 50862 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180827 20:41:53.207782 50862 server/config.go:496 [n?] 1 storage engine initialized I180827 20:41:53.207807 50862 server/config.go:499 [n?] RocksDB cache size: 128 MiB I180827 20:41:53.207815 50862 server/config.go:499 [n?] store 0: in-memory, size 0 B W180827 20:41:53.207911 50862 gossip/gossip.go:1371 [n?] no incoming or outgoing connections I180827 20:41:53.207947 50862 server/server.go:1403 [n?] no stores bootstrapped and --join flag specified, awaiting init command. I180827 20:41:53.211471 51475 rpc/nodedialer/nodedialer.go:92 [ct-client] connection to n2 established I180827 20:41:53.223653 51740 gossip/client.go:129 [n?] started gossip client to 127.0.0.1:41477 I180827 20:41:53.223954 51816 gossip/server.go:217 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:46463} I180827 20:41:53.224401 50862 server/node.go:697 [n?] connecting to gossip network to verify cluster ID... I180827 20:41:53.224432 50862 server/node.go:722 [n?] node connected via gossip and verified as part of cluster "d5e53e69-a109-4eb6-91bf-29e74ae744ba" I180827 20:41:53.224690 51837 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180827 20:41:53.225445 51836 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180827 20:41:53.226030 50862 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180827 20:41:53.226699 50862 server/node.go:428 [n?] new node allocated ID 3 I180827 20:41:53.226763 50862 gossip/gossip.go:382 [n3] NodeDescriptor set to node_id:3 address:<network_field:"tcp" address_field:"127.0.0.1:46463" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:12 > build_tag:"v2.1.0-alpha.20180702-2025-gf1e7bb1" started_at:1535402513226706701 I180827 20:41:53.226805 50862 storage/stores.go:242 [n3] read 0 node addresses from persistent storage I180827 20:41:53.226851 50862 storage/stores.go:261 [n3] wrote 2 node addresses to persistent storage I180827 20:41:53.227563 51809 storage/stores.go:261 [n1] wrote 2 node addresses to persistent storage I180827 20:41:53.227869 51810 storage/stores.go:261 [n2] wrote 2 node addresses to persistent storage I180827 20:41:53.228504 50862 server/node.go:672 [n3] bootstrapped store [n3,s3] I180827 20:41:53.229044 50862 server/node.go:546 [n3] node=3: started with [] engine(s) and attributes [] I180827 20:41:53.229696 50862 server/status/recorder.go:652 [n3] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180827 20:41:53.229749 50862 server/server.go:1807 [n3] Could not start heap profiler worker due to: directory to store profiles could not be determined I180827 20:41:53.235251 50862 server/server.go:1538 [n3] starting https server at 127.0.0.1:43307 (use: 127.0.0.1:43307) I180827 20:41:53.235271 50862 server/server.go:1540 [n3] starting grpc/postgres server at 127.0.0.1:46463 I180827 20:41:53.235283 50862 server/server.go:1541 [n3] advertising CockroachDB node at 127.0.0.1:46463 I180827 20:41:53.240284 50862 server/server.go:1594 [n3] done ensuring all necessary migrations have run I180827 20:41:53.240307 50862 server/server.go:1597 [n3] serving sql connections I180827 20:41:53.243124 51945 server/status/recorder.go:652 [n3,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180827 20:41:53.248117 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r20/1:/Table/{23-50}] sending preemptive snapshot 59e1afc9 at applied index 16 I180827 20:41:53.249136 50862 testutils/testcluster/testcluster.go:536 [n1,s1] has 22 underreplicated ranges I180827 20:41:53.251012 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r20/1:/Table/{23-50}] streamed snapshot to (n2,s2):?: kv pairs: 12, log entries: 6, rate-limit: 8.0 MiB/sec, 3ms I180827 20:41:53.251369 51983 storage/replica_raftstorage.go:784 [n2,s2,r20/?:{-}] applying preemptive snapshot at index 16 (id=59e1afc9, encoded size=2241, 1 rocksdb batches, 6 log entries) I180827 20:41:53.254056 51839 server/server_update.go:67 [n3] no need to upgrade, cluster already at the newest version I180827 20:41:53.255122 51841 sql/event_log.go:126 [n3] Event: "node_join", target: 3, info: {Descriptor:{NodeID:3 Address:{NetworkField:tcp AddressField:127.0.0.1:46463} Attrs: Locality: ServerVersion:2.0-12 BuildTag:v2.1.0-alpha.20180702-2025-gf1e7bb1 StartedAt:1535402513226706701 LocalityAddress:[]} ClusterID:d5e53e69-a109-4eb6-91bf-29e74ae744ba StartedAt:1535402513226706701 LastUp:1535402513226706701} I180827 20:41:53.256061 51983 storage/replica_raftstorage.go:790 [n2,s2,r20/?:/Table/{23-50}] applied preemptive snapshot in 1ms [clear=0ms batch=0ms entries=0ms commit=1ms] I180827 20:41:53.256605 50930 storage/replica_command.go:812 [replicate,n1,s1,r20/1:/Table/{23-50}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r20:/Table/{23-50} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.259565 50930 storage/replica.go:3743 [n1,s1,r20/1:/Table/{23-50}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3 I180827 20:41:53.261627 51625 rpc/nodedialer/nodedialer.go:92 [n2] connection to n1 established I180827 20:41:53.264544 50862 testutils/testcluster/testcluster.go:536 [n1,s1] has 22 underreplicated ranges I180827 20:41:53.286630 50930 rpc/nodedialer/nodedialer.go:92 [replicate,n1,s1,r21/1:/Table/5{0-1}] connection to n3 established I180827 20:41:53.287245 50862 testutils/testcluster/testcluster.go:536 [n1,s1] has 22 underreplicated ranges I180827 20:41:53.287799 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r21/1:/Table/5{0-1}] sending preemptive snapshot de08568a at applied index 18 I180827 20:41:53.288157 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r21/1:/Table/5{0-1}] streamed snapshot to (n3,s3):?: kv pairs: 8, log entries: 8, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.288623 51959 storage/replica_raftstorage.go:784 [n3,s3,r21/?:{-}] applying preemptive snapshot at index 18 (id=de08568a, encoded size=2646, 1 rocksdb batches, 8 log entries) I180827 20:41:53.289814 51959 storage/replica_raftstorage.go:790 [n3,s3,r21/?:/Table/5{0-1}] applied preemptive snapshot in 1ms [clear=0ms batch=0ms entries=0ms commit=1ms] I180827 20:41:53.290329 50930 storage/replica_command.go:812 [replicate,n1,s1,r21/1:/Table/5{0-1}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r21:/Table/5{0-1} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.293678 50930 storage/replica.go:3743 [n1,s1,r21/1:/Table/5{0-1}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.294953 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r22/1:/{Table/51-Max}] sending preemptive snapshot a84e7278 at applied index 12 I180827 20:41:53.295229 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r22/1:/{Table/51-Max}] streamed snapshot to (n3,s3):?: kv pairs: 7, log entries: 2, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.295441 51883 rpc/nodedialer/nodedialer.go:92 [n3] connection to n1 established I180827 20:41:53.295585 51953 storage/replica_raftstorage.go:784 [n3,s3,r22/?:{-}] applying preemptive snapshot at index 12 (id=a84e7278, encoded size=386, 1 rocksdb batches, 2 log entries) I180827 20:41:53.295717 51953 storage/replica_raftstorage.go:790 [n3,s3,r22/?:/{Table/51-Max}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.295955 50930 storage/replica_command.go:812 [replicate,n1,s1,r22/1:/{Table/51-Max}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r22:/{Table/51-Max} [(n1,s1):1, next=2, gen=0] I180827 20:41:53.298097 50930 storage/replica.go:3743 [n1,s1,r22/1:/{Table/51-Max}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.301122 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r8/1:/Table/1{1-2}] sending preemptive snapshot 201bdccc at applied index 18 I180827 20:41:53.301565 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r8/1:/Table/1{1-2}] streamed snapshot to (n3,s3):?: kv pairs: 9, log entries: 8, rate-limit: 8.0 MiB/sec, 3ms I180827 20:41:53.306578 52088 storage/replica_raftstorage.go:784 [n3,s3,r8/?:{-}] applying preemptive snapshot at index 18 (id=201bdccc, encoded size=4352, 1 rocksdb batches, 8 log entries) I180827 20:41:53.306868 52088 storage/replica_raftstorage.go:790 [n3,s3,r8/?:/Table/1{1-2}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.307601 50930 storage/replica_command.go:812 [replicate,n1,s1,r8/1:/Table/1{1-2}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r8:/Table/1{1-2} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.311873 50930 storage/replica.go:3743 [n1,s1,r8/1:/Table/1{1-2}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.314134 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r17/1:/Table/2{0-1}] sending preemptive snapshot 53116eb2 at applied index 16 I180827 20:41:53.314317 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r17/1:/Table/2{0-1}] streamed snapshot to (n3,s3):?: kv pairs: 8, log entries: 6, rate-limit: 8.0 MiB/sec, 2ms I180827 20:41:53.314683 52103 storage/replica_raftstorage.go:784 [n3,s3,r17/?:{-}] applying preemptive snapshot at index 16 (id=53116eb2, encoded size=2105, 1 rocksdb batches, 6 log entries) I180827 20:41:53.314887 52103 storage/replica_raftstorage.go:790 [n3,s3,r17/?:/Table/2{0-1}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.315401 50930 storage/replica_command.go:812 [replicate,n1,s1,r17/1:/Table/2{0-1}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r17:/Table/2{0-1} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.318398 50930 storage/replica.go:3743 [n1,s1,r17/1:/Table/2{0-1}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.319436 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r16/1:/Table/{19-20}] sending preemptive snapshot e0be8540 at applied index 16 I180827 20:41:53.319691 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r16/1:/Table/{19-20}] streamed snapshot to (n2,s2):?: kv pairs: 8, log entries: 6, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.320127 52072 storage/replica_raftstorage.go:784 [n2,s2,r16/?:{-}] applying preemptive snapshot at index 16 (id=e0be8540, encoded size=2109, 1 rocksdb batches, 6 log entries) I180827 20:41:53.320339 52072 storage/replica_raftstorage.go:790 [n2,s2,r16/?:/Table/{19-20}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.320816 50930 storage/replica_command.go:812 [replicate,n1,s1,r16/1:/Table/{19-20}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r16:/Table/{19-20} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.323849 50930 storage/replica.go:3743 [n1,s1,r16/1:/Table/{19-20}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3 I180827 20:41:53.326208 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r15/1:/Table/1{8-9}] sending preemptive snapshot d259ae5c at applied index 16 I180827 20:41:53.326404 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r15/1:/Table/1{8-9}] streamed snapshot to (n2,s2):?: kv pairs: 8, log entries: 6, rate-limit: 8.0 MiB/sec, 2ms I180827 20:41:53.326731 52116 storage/replica_raftstorage.go:784 [n2,s2,r15/?:{-}] applying preemptive snapshot at index 16 (id=d259ae5c, encoded size=2276, 1 rocksdb batches, 6 log entries) I180827 20:41:53.326923 50862 testutils/testcluster/testcluster.go:536 [n1,s1] has 22 underreplicated ranges I180827 20:41:53.326953 52116 storage/replica_raftstorage.go:790 [n2,s2,r15/?:/Table/1{8-9}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.334514 50930 storage/replica_command.go:812 [replicate,n1,s1,r15/1:/Table/1{8-9}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r15:/Table/1{8-9} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.337656 50930 storage/replica.go:3743 [n1,s1,r15/1:/Table/1{8-9}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3 I180827 20:41:53.338767 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r14/1:/Table/1{7-8}] sending preemptive snapshot 9d0058d5 at applied index 16 I180827 20:41:53.339034 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r14/1:/Table/1{7-8}] streamed snapshot to (n2,s2):?: kv pairs: 8, log entries: 6, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.339612 52090 storage/replica_raftstorage.go:784 [n2,s2,r14/?:{-}] applying preemptive snapshot at index 16 (id=9d0058d5, encoded size=2276, 1 rocksdb batches, 6 log entries) I180827 20:41:53.339831 52090 storage/replica_raftstorage.go:790 [n2,s2,r14/?:/Table/1{7-8}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.340173 50930 storage/replica_command.go:812 [replicate,n1,s1,r14/1:/Table/1{7-8}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r14:/Table/1{7-8} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.343121 50930 storage/replica.go:3743 [n1,s1,r14/1:/Table/1{7-8}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3 I180827 20:41:53.345432 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r9/1:/Table/1{2-3}] sending preemptive snapshot 0eea2d20 at applied index 26 I180827 20:41:53.345859 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r9/1:/Table/1{2-3}] streamed snapshot to (n2,s2):?: kv pairs: 53, log entries: 16, rate-limit: 8.0 MiB/sec, 2ms I180827 20:41:53.347137 52066 storage/replica_raftstorage.go:784 [n2,s2,r9/?:{-}] applying preemptive snapshot at index 26 (id=0eea2d20, encoded size=15139, 1 rocksdb batches, 16 log entries) I180827 20:41:53.347467 52066 storage/replica_raftstorage.go:790 [n2,s2,r9/?:/Table/1{2-3}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.348208 50930 storage/replica_command.go:812 [replicate,n1,s1,r9/1:/Table/1{2-3}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r9:/Table/1{2-3} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.352166 50930 storage/replica.go:3743 [n1,s1,r9/1:/Table/1{2-3}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3 I180827 20:41:53.353188 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r4/1:/System/{NodeLive…-tsd}] sending preemptive snapshot 0cdee511 at applied index 39 I180827 20:41:53.353765 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r4/1:/System/{NodeLive…-tsd}] streamed snapshot to (n2,s2):?: kv pairs: 36, log entries: 29, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.354286 51723 storage/replica_raftstorage.go:784 [n2,s2,r4/?:{-}] applying preemptive snapshot at index 39 (id=0cdee511, encoded size=98384, 1 rocksdb batches, 29 log entries) I180827 20:41:53.354994 51723 storage/replica_raftstorage.go:790 [n2,s2,r4/?:/System/{NodeLive…-tsd}] applied preemptive snapshot in 1ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.355529 50930 storage/replica_command.go:812 [replicate,n1,s1,r4/1:/System/{NodeLive…-tsd}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r4:/System/{NodeLivenessMax-tsd} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.358523 50930 storage/replica.go:3743 [n1,s1,r4/1:/System/{NodeLive…-tsd}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3 I180827 20:41:53.360250 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r3/1:/System/NodeLiveness{-Max}] sending preemptive snapshot 965d58b1 at applied index 19 I180827 20:41:53.360436 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r3/1:/System/NodeLiveness{-Max}] streamed snapshot to (n3,s3):?: kv pairs: 10, log entries: 9, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.360789 52150 storage/replica_raftstorage.go:784 [n3,s3,r3/?:{-}] applying preemptive snapshot at index 19 (id=965d58b1, encoded size=4003, 1 rocksdb batches, 9 log entries) I180827 20:41:53.361043 52150 storage/replica_raftstorage.go:790 [n3,s3,r3/?:/System/NodeLiveness{-Max}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.361522 50930 storage/replica_command.go:812 [replicate,n1,s1,r3/1:/System/NodeLiveness{-Max}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r3:/System/NodeLiveness{-Max} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.364392 50930 storage/replica.go:3743 [n1,s1,r3/1:/System/NodeLiveness{-Max}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.366422 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r12/1:/Table/1{5-6}] sending preemptive snapshot 811af376 at applied index 16 I180827 20:41:53.366638 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r12/1:/Table/1{5-6}] streamed snapshot to (n3,s3):?: kv pairs: 8, log entries: 6, rate-limit: 8.0 MiB/sec, 2ms I180827 20:41:53.367089 52137 storage/replica_raftstorage.go:784 [n3,s3,r12/?:{-}] applying preemptive snapshot at index 16 (id=811af376, encoded size=2276, 1 rocksdb batches, 6 log entries) I180827 20:41:53.367359 52137 storage/replica_raftstorage.go:790 [n3,s3,r12/?:/Table/1{5-6}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.368127 50930 storage/replica_command.go:812 [replicate,n1,s1,r12/1:/Table/1{5-6}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r12:/Table/1{5-6} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.371691 50930 storage/replica.go:3743 [n1,s1,r12/1:/Table/1{5-6}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.374563 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r19/1:/Table/2{2-3}] sending preemptive snapshot 9cd02555 at applied index 16 I180827 20:41:53.374760 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r19/1:/Table/2{2-3}] streamed snapshot to (n3,s3):?: kv pairs: 8, log entries: 6, rate-limit: 8.0 MiB/sec, 2ms I180827 20:41:53.375252 52080 storage/replica_raftstorage.go:784 [n3,s3,r19/?:{-}] applying preemptive snapshot at index 16 (id=9cd02555, encoded size=2276, 1 rocksdb batches, 6 log entries) I180827 20:41:53.375582 52080 storage/replica_raftstorage.go:790 [n3,s3,r19/?:/Table/2{2-3}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.375950 50930 storage/replica_command.go:812 [replicate,n1,s1,r19/1:/Table/2{2-3}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r19:/Table/2{2-3} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.381819 50930 storage/replica.go:3743 [n1,s1,r19/1:/Table/2{2-3}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.386461 52091 rpc/nodedialer/nodedialer.go:92 [ct-client] connection to n3 established I180827 20:41:53.386637 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r10/1:/Table/1{3-4}] sending preemptive snapshot a16f4b15 at applied index 64 I180827 20:41:53.388005 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r10/1:/Table/1{3-4}] streamed snapshot to (n3,s3):?: kv pairs: 204, log entries: 54, rate-limit: 8.0 MiB/sec, 4ms I180827 20:41:53.388536 52181 storage/replica_raftstorage.go:784 [n3,s3,r10/?:{-}] applying preemptive snapshot at index 64 (id=a16f4b15, encoded size=62836, 1 rocksdb batches, 54 log entries) I180827 20:41:53.389154 52181 storage/replica_raftstorage.go:790 [n3,s3,r10/?:/Table/1{3-4}] applied preemptive snapshot in 1ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.389513 50930 storage/replica_command.go:812 [replicate,n1,s1,r10/1:/Table/1{3-4}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r10:/Table/1{3-4} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.392649 50930 storage/replica.go:3743 [n1,s1,r10/1:/Table/1{3-4}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.394122 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r2/1:/System/{-NodeLive…}] sending preemptive snapshot 69adabc1 at applied index 23 I180827 20:41:53.394365 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r2/1:/System/{-NodeLive…}] streamed snapshot to (n2,s2):?: kv pairs: 7, log entries: 13, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.394729 52213 storage/replica_raftstorage.go:784 [n2,s2,r2/?:{-}] applying preemptive snapshot at index 23 (id=69adabc1, encoded size=6277, 1 rocksdb batches, 13 log entries) I180827 20:41:53.394981 52213 storage/replica_raftstorage.go:790 [n2,s2,r2/?:/System/{-NodeLive…}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.395465 50930 storage/replica_command.go:812 [replicate,n1,s1,r2/1:/System/{-NodeLive…}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r2:/System/{-NodeLiveness} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.398757 50930 storage/replica.go:3743 [n1,s1,r2/1:/System/{-NodeLive…}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3 I180827 20:41:53.399709 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r18/1:/Table/2{1-2}] sending preemptive snapshot e9df2a4a at applied index 16 I180827 20:41:53.400036 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r18/1:/Table/2{1-2}] streamed snapshot to (n3,s3):?: kv pairs: 8, log entries: 6, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.400391 52185 storage/replica_raftstorage.go:784 [n3,s3,r18/?:{-}] applying preemptive snapshot at index 16 (id=e9df2a4a, encoded size=2272, 1 rocksdb batches, 6 log entries) I180827 20:41:53.400594 52185 storage/replica_raftstorage.go:790 [n3,s3,r18/?:/Table/2{1-2}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.400882 50930 storage/replica_command.go:812 [replicate,n1,s1,r18/1:/Table/2{1-2}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r18:/Table/2{1-2} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.407636 50930 storage/replica.go:3743 [n1,s1,r18/1:/Table/2{1-2}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.408861 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r13/1:/Table/1{6-7}] sending preemptive snapshot 6f914d55 at applied index 16 I180827 20:41:53.409071 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r13/1:/Table/1{6-7}] streamed snapshot to (n2,s2):?: kv pairs: 8, log entries: 6, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.409426 52218 storage/replica_raftstorage.go:784 [n2,s2,r13/?:{-}] applying preemptive snapshot at index 16 (id=6f914d55, encoded size=2276, 1 rocksdb batches, 6 log entries) I180827 20:41:53.409616 52218 storage/replica_raftstorage.go:790 [n2,s2,r13/?:/Table/1{6-7}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.409970 50930 storage/replica_command.go:812 [replicate,n1,s1,r13/1:/Table/1{6-7}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r13:/Table/1{6-7} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.411262 50862 testutils/testcluster/testcluster.go:536 [n1,s1] has 22 underreplicated ranges I180827 20:41:53.412831 50930 storage/replica.go:3743 [n1,s1,r13/1:/Table/1{6-7}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3 I180827 20:41:53.414081 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r11/1:/Table/1{4-5}] sending preemptive snapshot cca961c1 at applied index 16 I180827 20:41:53.414277 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r11/1:/Table/1{4-5}] streamed snapshot to (n3,s3):?: kv pairs: 8, log entries: 6, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.414576 52199 storage/replica_raftstorage.go:784 [n3,s3,r11/?:{-}] applying preemptive snapshot at index 16 (id=cca961c1, encoded size=2272, 1 rocksdb batches, 6 log entries) I180827 20:41:53.414816 52199 storage/replica_raftstorage.go:790 [n3,s3,r11/?:/Table/1{4-5}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.415293 50930 storage/replica_command.go:812 [replicate,n1,s1,r11/1:/Table/1{4-5}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r11:/Table/1{4-5} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.418111 50930 storage/replica.go:3743 [n1,s1,r11/1:/Table/1{4-5}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.419054 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r5/1:/System/ts{d-e}] sending preemptive snapshot 3c3a015f at applied index 27 I180827 20:41:53.423022 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r5/1:/System/ts{d-e}] streamed snapshot to (n3,s3):?: kv pairs: 1391, log entries: 2, rate-limit: 8.0 MiB/sec, 4ms I180827 20:41:53.423893 52201 storage/replica_raftstorage.go:784 [n3,s3,r5/?:{-}] applying preemptive snapshot at index 27 (id=3c3a015f, encoded size=194658, 1 rocksdb batches, 2 log entries) I180827 20:41:53.429501 52201 storage/replica_raftstorage.go:790 [n3,s3,r5/?:/System/ts{d-e}] applied preemptive snapshot in 6ms [clear=0ms batch=0ms entries=2ms commit=4ms] I180827 20:41:53.433500 50930 storage/replica_command.go:812 [replicate,n1,s1,r5/1:/System/ts{d-e}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r5:/System/ts{d-e} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.437580 50930 storage/replica.go:3743 [n1,s1,r5/1:/System/ts{d-e}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.440575 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r6/1:/{System/tse-Table/System…}] sending preemptive snapshot cbd412df at applied index 21 I180827 20:41:53.440794 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r6/1:/{System/tse-Table/System…}] streamed snapshot to (n3,s3):?: kv pairs: 8, log entries: 11, rate-limit: 8.0 MiB/sec, 2ms I180827 20:41:53.441181 52260 storage/replica_raftstorage.go:784 [n3,s3,r6/?:{-}] applying preemptive snapshot at index 21 (id=cbd412df, encoded size=4339, 1 rocksdb batches, 11 log entries) I180827 20:41:53.441400 52260 storage/replica_raftstorage.go:790 [n3,s3,r6/?:/{System/tse-Table/System…}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.441676 50930 storage/replica_command.go:812 [replicate,n1,s1,r6/1:/{System/tse-Table/System…}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r6:/{System/tse-Table/SystemConfigSpan/Start} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.448564 52224 rpc/nodedialer/nodedialer.go:92 [ct-client] connection to n2 established I180827 20:41:53.461587 50930 storage/replica.go:3743 [n1,s1,r6/1:/{System/tse-Table/System…}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.463345 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r7/1:/Table/{SystemCon…-11}] sending preemptive snapshot 114f4385 at applied index 29 I180827 20:41:53.464896 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r7/1:/Table/{SystemCon…-11}] streamed snapshot to (n2,s2):?: kv pairs: 59, log entries: 19, rate-limit: 8.0 MiB/sec, 3ms I180827 20:41:53.465343 52280 storage/replica_raftstorage.go:784 [n2,s2,r7/?:{-}] applying preemptive snapshot at index 29 (id=114f4385, encoded size=16646, 1 rocksdb batches, 19 log entries) I180827 20:41:53.465821 52280 storage/replica_raftstorage.go:790 [n2,s2,r7/?:/Table/{SystemCon…-11}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.466988 50930 storage/replica_command.go:812 [replicate,n1,s1,r7/1:/Table/{SystemCon…-11}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r7:/Table/{SystemConfigSpan/Start-11} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.472743 50930 storage/replica.go:3743 [n1,s1,r7/1:/Table/{SystemCon…-11}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3 I180827 20:41:53.474632 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r1/1:/{Min-System/}] sending preemptive snapshot 0a244018 at applied index 114 I180827 20:41:53.475250 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r1/1:/{Min-System/}] streamed snapshot to (n2,s2):?: kv pairs: 73, log entries: 90, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.475827 52267 storage/replica_raftstorage.go:784 [n2,s2,r1/?:{-}] applying preemptive snapshot at index 114 (id=0a244018, encoded size=40271, 1 rocksdb batches, 90 log entries) I180827 20:41:53.476525 52267 storage/replica_raftstorage.go:790 [n2,s2,r1/?:/{Min-System/}] applied preemptive snapshot in 1ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.476869 50930 storage/replica_command.go:812 [replicate,n1,s1,r1/1:/{Min-System/}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r1:/{Min-System/} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.482912 50930 storage/replica.go:3743 [n1,s1,r1/1:/{Min-System/}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3 I180827 20:41:53.483281 50930 storage/queue.go:873 [n1,replicate] purgatory is now empty I180827 20:41:53.485684 52286 storage/store_snapshot.go:615 [replicate,n1,s1,r20/1:/Table/{23-50}] sending preemptive snapshot f1426c69 at applied index 19 I180827 20:41:53.487316 52286 storage/store_snapshot.go:657 [replicate,n1,s1,r20/1:/Table/{23-50}] streamed snapshot to (n3,s3):?: kv pairs: 13, log entries: 9, rate-limit: 8.0 MiB/sec, 4ms I180827 20:41:53.487681 52252 storage/replica_raftstorage.go:784 [n3,s3,r20/?:{-}] applying preemptive snapshot at index 19 (id=f1426c69, encoded size=3273, 1 rocksdb batches, 9 log entries) I180827 20:41:53.487932 52252 storage/replica_raftstorage.go:790 [n3,s3,r20/?:/Table/{23-50}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.488311 52286 storage/replica_command.go:812 [replicate,n1,s1,r20/1:/Table/{23-50}] change replicas (ADD_REPLICA (n3,s3):3): read existing descriptor r20:/Table/{23-50} [(n1,s1):1, (n2,s2):2, next=3, gen=1] I180827 20:41:53.503580 52286 storage/replica.go:3743 [n1,s1,r20/1:/Table/{23-50}] proposing ADD_REPLICA((n3,s3):3): updated=[(n1,s1):1 (n2,s2):2 (n3,s3):3] next=4 I180827 20:41:53.505707 52235 storage/store_snapshot.go:615 [replicate,n1,s1,r1/1:/{Min-System/}] sending preemptive snapshot 99036b07 at applied index 119 I180827 20:41:53.506514 52235 storage/store_snapshot.go:657 [replicate,n1,s1,r1/1:/{Min-System/}] streamed snapshot to (n3,s3):?: kv pairs: 78, log entries: 95, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.507282 52188 storage/replica_raftstorage.go:784 [n3,s3,r1/?:{-}] applying preemptive snapshot at index 119 (id=99036b07, encoded size=42101, 1 rocksdb batches, 95 log entries) I180827 20:41:53.508109 52188 storage/replica_raftstorage.go:790 [n3,s3,r1/?:/{Min-System/}] applied preemptive snapshot in 1ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.508641 52235 storage/replica_command.go:812 [replicate,n1,s1,r1/1:/{Min-System/}] change replicas (ADD_REPLICA (n3,s3):3): read existing descriptor r1:/{Min-System/} [(n1,s1):1, (n2,s2):2, next=3, gen=1] I180827 20:41:53.512524 52235 storage/replica.go:3743 [n1,s1,r1/1:/{Min-System/}] proposing ADD_REPLICA((n3,s3):3): updated=[(n1,s1):1 (n2,s2):2 (n3,s3):3] next=4 I180827 20:41:53.513999 52209 storage/store_snapshot.go:615 [replicate,n1,s1,r7/1:/Table/{SystemCon…-11}] sending preemptive snapshot bb53109c at applied index 32 I180827 20:41:53.514379 52209 storage/store_snapshot.go:657 [replicate,n1,s1,r7/1:/Table/{SystemCon…-11}] streamed snapshot to (n3,s3):?: kv pairs: 60, log entries: 22, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.514821 52292 storage/replica_raftstorage.go:784 [n3,s3,r7/?:{-}] applying preemptive snapshot at index 32 (id=bb53109c, encoded size=17687, 1 rocksdb batches, 22 log entries) I180827 20:41:53.515905 52292 storage/replica_raftstorage.go:790 [n3,s3,r7/?:/Table/{SystemCon…-11}] applied preemptive snapshot in 1ms [clear=1ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.516367 52209 storage/replica_command.go:812 [replicate,n1,s1,r7/1:/Table/{SystemCon…-11}] change replicas (ADD_REPLICA (n3,s3):3): read existing descriptor r7:/Table/{SystemConfigSpan/Start-11} [(n1,s1):1, (n2,s2):2, next=3, gen=1] I180827 20:41:53.520158 52209 storage/replica.go:3743 [n1,s1,r7/1:/Table/{SystemCon…-11}] proposing ADD_REPLICA((n3,s3):3): updated=[(n1,s1):1 (n2,s2):2 (n3,s3):3] next=4 I180827 20:41:53.521958 52312 storage/store_snapshot.go:615 [replicate,n1,s1,r6/1:/{System/tse-Table/System…}] sending preemptive snapshot 2ca43612 at applied index 24 I180827 20:41:53.522776 52312 storage/store_snapshot.go:657 [replicate,n1,s1,r6/1:/{System/tse-Table/System…}] streamed snapshot to (n2,s2):?: kv pairs: 9, log entries: 14, rate-limit: 8.0 MiB/sec, 2ms I180827 20:41:53.523128 52239 storage/replica_raftstorage.go:784 [n2,s2,r6/?:{-}] applying preemptive snapshot at index 24 (id=2ca43612, encoded size=5410, 1 rocksdb batches, 14 log entries) I180827 20:41:53.523377 52239 storage/replica_raftstorage.go:790 [n2,s2,r6/?:/{System/tse-Table/System…}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.523701 52312 storage/replica_command.go:812 [replicate,n1,s1,r6/1:/{System/tse-Table/System…}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r6:/{System/tse-Table/SystemConfigSpan/Start} [(n1,s1):1, (n3,s3):2, next=3, gen=1] I180827 20:41:53.525176 50862 testutils/testcluster/testcluster.go:536 [n1,s1] has 19 underreplicated ranges I180827 20:41:53.527482 52312 storage/replica.go:3743 [n1,s1,r6/1:/{System/tse-Table/System…}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3] next=4 I180827 20:41:53.528875 52327 storage/store_snapshot.go:615 [replicate,n1,s1,r5/1:/System/ts{d-e}] sending preemptive snapshot 731be2ae at applied index 30 I180827 20:41:53.532860 52327 storage/store_snapshot.go:657 [replicate,n1,s1,r5/1:/System/ts{d-e}] streamed snapshot to (n2,s2):?: kv pairs: 1392, log entries: 5, rate-limit: 8.0 MiB/sec, 4ms I180827 20:41:53.533361 52316 storage/replica_raftstorage.go:784 [n2,s2,r5/?:{-}] applying preemptive snapshot at index 30 (id=731be2ae, encoded size=195741, 1 rocksdb batches, 5 log entries) I180827 20:41:53.535834 52316 storage/replica_raftstorage.go:790 [n2,s2,r5/?:/System/ts{d-e}] applied preemptive snapshot in 2ms [clear=0ms batch=0ms entries=0ms commit=2ms] I180827 20:41:53.536253 52327 storage/replica_command.go:812 [replicate,n1,s1,r5/1:/System/ts{d-e}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r5:/System/ts{d-e} [(n1,s1):1, (n3,s3):2, next=3, gen=1] I180827 20:41:53.540576 52327 storage/replica.go:3743 [n1,s1,r5/1:/System/ts{d-e}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3] next=4 I180827 20:41:53.545804 52341 storage/store_snapshot.go:615 [replicate,n1,s1,r11/1:/Table/1{4-5}] sending preemptive snapshot 7497a95f at applied index 19 I180827 20:41:53.546108 52341 storage/store_snapshot.go:657 [replicate,n1,s1,r11/1:/Table/1{4-5}] streamed snapshot to (n2,s2):?: kv pairs: 9, log entries: 9, rate-limit: 8.0 MiB/sec, 4ms I180827 20:41:53.546590 52275 storage/replica_raftstorage.go:784 [n2,s2,r11/?:{-}] applying preemptive snapshot at index 19 (id=7497a95f, encoded size=3304, 1 rocksdb batches, 9 log entries) I180827 20:41:53.546960 52275 storage/replica_raftstorage.go:790 [n2,s2,r11/?:/Table/1{4-5}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.547386 52341 storage/replica_command.go:812 [replicate,n1,s1,r11/1:/Table/1{4-5}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r11:/Table/1{4-5} [(n1,s1):1, (n3,s3):2, next=3, gen=1] I180827 20:41:53.551568 52341 storage/replica.go:3743 [n1,s1,r11/1:/Table/1{4-5}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3] next=4 I180827 20:41:53.554959 52323 storage/store_snapshot.go:615 [replicate,n1,s1,r13/1:/Table/1{6-7}] sending preemptive snapshot 5932a5bd at applied index 19 I180827 20:41:53.555353 52323 storage/store_snapshot.go:657 [replicate,n1,s1,r13/1:/Table/1{6-7}] streamed snapshot to (n3,s3):?: kv pairs: 9, log entries: 9, rate-limit: 8.0 MiB/sec, 3ms I180827 20:41:53.555743 52329 storage/replica_raftstorage.go:784 [n3,s3,r13/?:{-}] applying preemptive snapshot at index 19 (id=5932a5bd, encoded size=3308, 1 rocksdb batches, 9 log entries) I180827 20:41:53.556103 52329 storage/replica_raftstorage.go:790 [n3,s3,r13/?:/Table/1{6-7}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.556489 52323 storage/replica_command.go:812 [replicate,n1,s1,r13/1:/Table/1{6-7}] change replicas (ADD_REPLICA (n3,s3):3): read existing descriptor r13:/Table/1{6-7} [(n1,s1):1, (n2,s2):2, next=3, gen=1] I180827 20:41:53.563494 52323 storage/replica.go:3743 [n1,s1,r13/1:/Table/1{6-7}] proposing ADD_REPLICA((n3,s3):3): updated=[(n1,s1):1 (n2,s2):2 (n3,s3):3] next=4 I180827 20:41:53.566866 52362 storage/store_snapshot.go:615 [replicate,n1,s1,r18/1:/Table/2{1-2}] sending preemptive snapshot c74baa54 at applied index 19 I180827 20:41:53.568042 52362 storage/store_snapshot.go:657 [replicate,n1,s1,r18/1:/Table/2{1-2}] streamed snapshot to (n2,s2):?: kv pairs: 9, log entries: 9, rate-limit: 8.0 MiB/sec, 3ms I180827 20:41:53.568417 52366 storage/replica_raftstorage.go:784 [n2,s2,r18/?:{-}] applying preemptive snapshot at index 19 (id=c74baa54, encoded size=3304, 1 rocksdb batches, 9 log entries) I180827 20:41:53.568651 52366 storage/replica_raftstorage.go:790 [n2,s2,r18/?:/Table/2{1-2}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.568954 52362 storage/replica_command.go:812 [replicate,n1,s1,r18/1:/Table/2{1-2}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r18:/Table/2{1-2} [(n1,s1):1, (n3,s3):2, next=3, gen=1] I180827 20:41:53.572711 52362 storage/replica.go:3743 [n1,s1,r18/1:/Table/2{1-2}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3] next=4 I180827 20:41:53.574655 52190 storage/store_snapshot.go:615 [replicate,n1,s1,r2/1:/System/{-NodeLive…}] sending preemptive snapshot d4c499ea at applied index 26 I180827 20:41:53.574962 52190 storage/store_snapshot.go:657 [replicate,n1,s1,r2/1:/System/{-NodeLive…}] streamed snapshot to (n3,s3):?: kv pairs: 8, log entries: 16, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.575725 52299 storage/replica_raftstorage.go:784 [n3,s3,r2/?:{-}] applying preemptive snapshot at index 26 (id=d4c499ea, encoded size=7349, 1 rocksdb batches, 16 log entries) I180827 20:41:53.576022 52299 storage/replica_raftstorage.go:790 [n3,s3,r2/?:/System/{-NodeLive…}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.576405 52190 storage/replica_command.go:812 [replicate,n1,s1,r2/1:/System/{-NodeLive…}] change replicas (ADD_REPLICA (n3,s3):3): read existing descriptor r2:/System/{-NodeLiveness} [(n1,s1):1, (n2,s2):2, next=3, gen=1] I180827 20:41:53.579762 52190 storage/replica.go:3743 [n1,s1,r2/1:/System/{-NodeLive…}] proposing ADD_REPLICA((n3,s3):3): updated=[(n1,s1):1 (n2,s2):2 (n3,s3):3] next=4 I180827 20:41:53.583103 52372 storage/store_snapshot.go:615 [replicate,n1,s1,r10/1:/Table/1{3-4}] sending preemptive snapshot dbe83d06 at applied index 103 I180827 20:41:53.583765 52372 storage/store_snapshot.go:657 [replicate,n1,s1,r10/1:/Table/1{3-4}] streamed snapshot to (n2,s2):?: kv pairs: 295, log entries: 10, rate-limit: 8.0 MiB/sec, 2ms I180827 20:41:53.584214 52369 storage/replica_raftstorage.go:784 [n2,s2,r10/?:{-}] applying preemptive snapshot at index 103 (id=dbe83d06, encoded size=38018, 1 rocksdb batches, 10 log entries) I180827 20:41:53.584577 52369 storage/replica_raftstorage.go:790 [n2,s2,r10/?:/Table/1{3-4}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.584963 52372 storage/replica_command.go:812 [replicate,n1,s1,r10/1:/Table/1{3-4}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r10:/Table/1{3-4} [(n1,s1):1, (n3,s3):2, next=3, gen=1] I180827 20:41:53.588661 52372 storage/replica.go:3743 [n1,s1,r10/1:/Table/1{3-4}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3] next=4 I180827 20:41:53.590522 52331 storage/store_snapshot.go:615 [replicate,n1,s1,r19/1:/Table/2{2-3}] sending preemptive snapshot ba0de389 at applied index 19 I180827 20:41:53.596120 52331 storage/store_snapshot.go:657 [replicate,n1,s1,r19/1:/Table/2{2-3}] streamed snapshot to (n2,s2):?: kv pairs: 9, log entries: 9, rate-limit: 8.0 MiB/sec, 6ms I180827 20:41:53.597215 52259 storage/replica_raftstorage.go:784 [n2,s2,r19/?:{-}] applying preemptive snapshot at index 19 (id=ba0de389, encoded size=3308, 1 rocksdb batches, 9 log entries) I180827 20:41:53.597484 52259 storage/replica_raftstorage.go:790 [n2,s2,r19/?:/Table/2{2-3}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.597898 52331 storage/replica_command.go:812 [replicate,n1,s1,r19/1:/Table/2{2-3}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r19:/Table/2{2-3} [(n1,s1):1, (n3,s3):2, next=3, gen=1] I180827 20:41:53.601937 52331 storage/replica.go:3743 [n1,s1,r19/1:/Table/2{2-3}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3] next=4 I180827 20:41:53.604379 52303 storage/store_snapshot.go:615 [replicate,n1,s1,r12/1:/Table/1{5-6}] sending preemptive snapshot 60969a90 at applied index 19 I180827 20:41:53.606558 52303 storage/store_snapshot.go:657 [replicate,n1,s1,r12/1:/Table/1{5-6}] streamed snapshot to (n2,s2):?: kv pairs: 9, log entries: 9, rate-limit: 8.0 MiB/sec, 3ms I180827 20:41:53.606975 52307 storage/replica_raftstorage.go:784 [n2,s2,r12/?:{-}] applying preemptive snapshot at index 19 (id=60969a90, encoded size=3308, 1 rocksdb batches, 9 log entries) I180827 20:41:53.607211 52307 storage/replica_raftstorage.go:790 [n2,s2,r12/?:/Table/1{5-6}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.608276 52303 storage/replica_command.go:812 [replicate,n1,s1,r12/1:/Table/1{5-6}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r12:/Table/1{5-6} [(n1,s1):1, (n3,s3):2, next=3, gen=1] I180827 20:41:53.612795 52303 storage/replica.go:3743 [n1,s1,r12/1:/Table/1{5-6}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3] next=4 I180827 20:41:53.615927 52392 storage/store_snapshot.go:615 [replicate,n1,s1,r3/1:/System/NodeLiveness{-Max}] sending preemptive snapshot 3d427041 at applied index 22 I180827 20:41:53.617153 52392 storage/store_snapshot.go:657 [replicate,n1,s1,r3/1:/System/NodeLiveness{-Max}] streamed snapshot to (n2,s2):?: kv pairs: 11, log entries: 12, rate-limit: 8.0 MiB/sec, 3ms I180827 20:41:53.618565 52405 storage/replica_raftstorage.go:784 [n2,s2,r3/?:{-}] applying preemptive snapshot at index 22 (id=3d427041, encoded size=5215, 1 rocksdb batches, 12 log entries) I180827 20:41:53.619140 52405 storage/replica_raftstorage.go:790 [n2,s2,r3/?:/System/NodeLiveness{-Max}] applied preemptive snapshot in 1ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.621854 50862 testutils/testcluster/testcluster.go:536 [n1,s1] has 10 underreplicated ranges I180827 20:41:53.635001 52392 storage/replica_command.go:812 [replicate,n1,s1,r3/1:/System/NodeLiveness{-Max}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r3:/System/NodeLiveness{-Max} [(n1,s1):1, (n3,s3):2, next=3, gen=1] I180827 20:41:53.638726 52392 storage/replica.go:3743 [n1,s1,r3/1:/System/NodeLiveness{-Max}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3] next=4 I180827 20:41:53.643490 52420 storage/store_snapshot.go:615 [replicate,n1,s1,r4/1:/System/{NodeLive…-tsd}] sending preemptive snapshot a2505d74 at applied index 42 I180827 20:41:53.644245 52420 storage/store_snapshot.go:657 [replicate,n1,s1,r4/1:/System/{NodeLive…-tsd}] streamed snapshot to (n3,s3):?: kv pairs: 37, log entries: 32, rate-limit: 8.0 MiB/sec, 3ms I180827 20:41:53.644709 52436 storage/replica_raftstorage.go:784 [n3,s3,r4/?:{-}] applying preemptive snapshot at index 42 (id=a2505d74, encoded size=99568, 1 rocksdb batches, 32 log entries) I180827 20:41:53.645176 52436 storage/replica_raftstorage.go:790 [n3,s3,r4/?:/System/{NodeLive…-tsd}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.645567 52420 storage/replica_command.go:812 [replicate,n1,s1,r4/1:/System/{NodeLive…-tsd}] change replicas (ADD_REPLICA (n3,s3):3): read existing descriptor r4:/System/{NodeLivenessMax-tsd} [(n1,s1):1, (n2,s2):2, next=3, gen=1] I180827 20:41:53.649433 52420 storage/replica.go:3743 [n1,s1,r4/1:/System/{NodeLive…-tsd}] proposing ADD_REPLICA((n3,s3):3): updated=[(n1,s1):1 (n2,s2):2 (n3,s3):3] next=4 I180827 20:41:53.652117 52398 storage/store_snapshot.go:615 [replicate,n1,s1,r9/1:/Table/1{2-3}] sending preemptive snapshot 6dc7ffcb at applied index 29 I180827 20:41:53.653729 52398 storage/store_snapshot.go:657 [replicate,n1,s1,r9/1:/Table/1{2-3}] streamed snapshot to (n3,s3):?: kv pairs: 54, log entries: 19, rate-limit: 8.0 MiB/sec, 3ms I180827 20:41:53.654216 52402 storage/replica_raftstorage.go:784 [n3,s3,r9/?:{-}] applying preemptive snapshot at index 29 (id=6dc7ffcb, encoded size=16171, 1 ``` Please assign, take a look and update the issue accordingly.
1.0
teamcity: failed test: test/TestImportPgDump, t, - - The following tests appear to have failed on release-banana. You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+test/TestImportPgDump, t, -). [#864629](https://teamcity.cockroachdb.com/viewLog.html?buildId=864629): ``` e---- FAIL: test/TestImportPgDump (0.000s) Test ended in panic. ------- Stdout: ------- W180827 20:41:52.746991 50862 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180827 20:41:52.757923 50862 server/server.go:830 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled I180827 20:41:52.758132 50862 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180827 20:41:52.758156 50862 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180827 20:41:52.761168 50862 server/config.go:496 [n?] 1 storage engine initialized I180827 20:41:52.761191 50862 server/config.go:499 [n?] RocksDB cache size: 128 MiB I180827 20:41:52.761204 50862 server/config.go:499 [n?] store 0: in-memory, size 0 B I180827 20:41:52.767725 50862 server/node.go:373 [n?] **** cluster d5e53e69-a109-4eb6-91bf-29e74ae744ba has been created I180827 20:41:52.767752 50862 server/server.go:1401 [n?] **** add additional nodes by specifying --join=127.0.0.1:41477 I180827 20:41:52.767936 50862 gossip/gossip.go:382 [n1] NodeDescriptor set to node_id:1 address:<network_field:"tcp" address_field:"127.0.0.1:41477" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:12 > build_tag:"v2.1.0-alpha.20180702-2025-gf1e7bb1" started_at:1535402512767856449 I180827 20:41:52.770338 50862 storage/store.go:1541 [n1,s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available I180827 20:41:52.770546 50862 server/node.go:476 [n1] initialized store [n1,s1]: disk (capacity=512 MiB, available=512 MiB, used=0 B, logicalBytes=6.9 KiB), ranges=1, leases=1, queries=0.00, writes=0.00, bytesPerReplica={p10=7103.00 p25=7103.00 p50=7103.00 p75=7103.00 p90=7103.00 pMax=7103.00}, writesPerReplica={p10=0.00 p25=0.00 p50=0.00 p75=0.00 p90=0.00 pMax=0.00} I180827 20:41:52.770626 50862 storage/stores.go:242 [n1] read 0 node addresses from persistent storage I180827 20:41:52.770721 50862 server/node.go:697 [n1] connecting to gossip network to verify cluster ID... I180827 20:41:52.770760 50862 server/node.go:722 [n1] node connected via gossip and verified as part of cluster "d5e53e69-a109-4eb6-91bf-29e74ae744ba" I180827 20:41:52.770788 50862 server/node.go:546 [n1] node=1: started with [<no-attributes>=<in-mem>] engine(s) and attributes [] I180827 20:41:52.771023 50862 server/status/recorder.go:652 [n1] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180827 20:41:52.771066 50862 server/server.go:1807 [n1] Could not start heap profiler worker due to: directory to store profiles could not be determined I180827 20:41:52.771159 50862 server/server.go:1538 [n1] starting https server at 127.0.0.1:42563 (use: 127.0.0.1:42563) I180827 20:41:52.771188 50862 server/server.go:1540 [n1] starting grpc/postgres server at 127.0.0.1:41477 I180827 20:41:52.771209 50862 server/server.go:1541 [n1] advertising CockroachDB node at 127.0.0.1:41477 I180827 20:41:52.775258 51089 server/status/recorder.go:652 [n1,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180827 20:41:52.776337 50925 storage/replica_command.go:298 [split,n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2] I180827 20:41:52.788832 51094 storage/replica_command.go:298 [split,n1,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3] W180827 20:41:52.790188 51128 storage/intent_resolver.go:668 [n1,s1] failed to push during intent resolution: failed to push "unnamed" id=ec083bbe key=/Table/SystemConfigSpan/Start rw=true pri=0.01126188 iso=SERIALIZABLE stat=PENDING epo=0 ts=1535402512.772758792,0 orig=1535402512.772758792,0 max=1535402512.772758792,0 wto=false rop=false seq=6 I180827 20:41:52.790695 51118 sql/event_log.go:126 [n1,intExec=optInToDiagnosticsStatReporting] Event: "set_cluster_setting", target: 0, info: {SettingName:diagnostics.reporting.enabled Value:true User:root} I180827 20:41:52.795125 51100 storage/replica_command.go:298 [split,n1,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4] I180827 20:41:52.800783 51143 storage/replica_command.go:298 [split,n1,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5] I180827 20:41:52.807906 51165 storage/replica_command.go:298 [split,n1,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6] I180827 20:41:52.811784 51141 sql/event_log.go:126 [n1,intExec=set-setting] Event: "set_cluster_setting", target: 0, info: {SettingName:version Value:2.0-12 User:root} I180827 20:41:52.818164 50799 sql/event_log.go:126 [n1,intExec=disableNetTrace] Event: "set_cluster_setting", target: 0, info: {SettingName:trace.debug.enable Value:false User:root} I180827 20:41:52.821094 51188 storage/replica_command.go:298 [split,n1,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7] I180827 20:41:52.830709 51176 storage/replica_command.go:298 [split,n1,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8] I180827 20:41:52.839374 51187 sql/event_log.go:126 [n1,intExec=initializeClusterSecret] Event: "set_cluster_setting", target: 0, info: {SettingName:cluster.secret Value:045a1c98-219f-445b-bd6b-d481f04d6b0d User:root} I180827 20:41:52.849534 51154 storage/replica_command.go:298 [split,n1,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9] I180827 20:41:52.855898 51218 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 50, info: {DatabaseName:defaultdb Statement:CREATE DATABASE IF NOT EXISTS defaultdb User:root} I180827 20:41:52.861462 51240 storage/replica_command.go:298 [split,n1,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10] I180827 20:41:52.868342 51268 storage/replica_command.go:298 [split,n1,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11] I180827 20:41:52.872706 51256 sql/event_log.go:126 [n1,intExec=create-default-db] Event: "create_database", target: 51, info: {DatabaseName:postgres Statement:CREATE DATABASE IF NOT EXISTS postgres User:root} I180827 20:41:52.874819 51264 storage/replica_command.go:298 [split,n1,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12] I180827 20:41:52.876403 50862 server/server.go:1594 [n1] done ensuring all necessary migrations have run I180827 20:41:52.876433 50862 server/server.go:1597 [n1] serving sql connections I180827 20:41:52.879108 51233 server/server_update.go:67 [n1] no need to upgrade, cluster already at the newest version I180827 20:41:52.879639 51235 sql/event_log.go:126 [n1] Event: "node_join", target: 1, info: {Descriptor:{NodeID:1 Address:{NetworkField:tcp AddressField:127.0.0.1:41477} Attrs: Locality: ServerVersion:2.0-12 BuildTag:v2.1.0-alpha.20180702-2025-gf1e7bb1 StartedAt:1535402512767856449 LocalityAddress:[]} ClusterID:d5e53e69-a109-4eb6-91bf-29e74ae744ba StartedAt:1535402512767856449 LastUp:1535402512767856449} I180827 20:41:52.880318 51302 storage/replica_command.go:298 [split,n1,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13] I180827 20:41:52.927701 50819 storage/replica_command.go:298 [split,n1,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14] I180827 20:41:52.940165 51323 storage/replica_command.go:298 [split,n1,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15] I180827 20:41:52.948539 51355 storage/replica_command.go:298 [split,n1,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16] I180827 20:41:52.953658 51380 storage/replica_command.go:298 [split,n1,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17] I180827 20:41:52.961237 51137 storage/replica_command.go:298 [split,n1,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18] I180827 20:41:52.966548 50832 storage/replica_command.go:298 [split,n1,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19] I180827 20:41:52.977113 51362 storage/replica_command.go:298 [split,n1,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20] I180827 20:41:53.041315 51440 storage/replica_command.go:298 [split,n1,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21] I180827 20:41:53.047478 51414 storage/replica_command.go:298 [split,n1,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22] W180827 20:41:53.081214 50862 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180827 20:41:53.089127 50862 server/server.go:830 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled I180827 20:41:53.089322 50862 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180827 20:41:53.089338 50862 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180827 20:41:53.102793 50862 server/config.go:496 [n?] 1 storage engine initialized I180827 20:41:53.102863 50862 server/config.go:499 [n?] RocksDB cache size: 128 MiB I180827 20:41:53.102878 50862 server/config.go:499 [n?] store 0: in-memory, size 0 B W180827 20:41:53.102953 50862 gossip/gossip.go:1371 [n?] no incoming or outgoing connections I180827 20:41:53.103001 50862 server/server.go:1403 [n?] no stores bootstrapped and --join flag specified, awaiting init command. I180827 20:41:53.115344 51458 gossip/client.go:129 [n?] started gossip client to 127.0.0.1:41477 I180827 20:41:53.125579 51530 gossip/server.go:217 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:36113} I180827 20:41:53.127987 50862 server/node.go:697 [n?] connecting to gossip network to verify cluster ID... I180827 20:41:53.128034 50862 server/node.go:722 [n?] node connected via gossip and verified as part of cluster "d5e53e69-a109-4eb6-91bf-29e74ae744ba" I180827 20:41:53.128397 51575 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180827 20:41:53.134920 51574 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180827 20:41:53.135628 50862 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180827 20:41:53.136461 50862 server/node.go:428 [n?] new node allocated ID 2 I180827 20:41:53.136541 50862 gossip/gossip.go:382 [n2] NodeDescriptor set to node_id:2 address:<network_field:"tcp" address_field:"127.0.0.1:36113" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:12 > build_tag:"v2.1.0-alpha.20180702-2025-gf1e7bb1" started_at:1535402513136479434 I180827 20:41:53.136591 50862 storage/stores.go:242 [n2] read 0 node addresses from persistent storage I180827 20:41:53.136624 50862 storage/stores.go:261 [n2] wrote 1 node addresses to persistent storage I180827 20:41:53.137485 51552 storage/stores.go:261 [n1] wrote 1 node addresses to persistent storage I180827 20:41:53.139442 50862 server/node.go:672 [n2] bootstrapped store [n2,s2] I180827 20:41:53.139577 50862 server/node.go:546 [n2] node=2: started with [] engine(s) and attributes [] I180827 20:41:53.140140 50862 server/status/recorder.go:652 [n2] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180827 20:41:53.140166 50862 server/server.go:1807 [n2] Could not start heap profiler worker due to: directory to store profiles could not be determined I180827 20:41:53.140233 50862 server/server.go:1538 [n2] starting https server at 127.0.0.1:39947 (use: 127.0.0.1:39947) I180827 20:41:53.140246 50862 server/server.go:1540 [n2] starting grpc/postgres server at 127.0.0.1:36113 I180827 20:41:53.140256 50862 server/server.go:1541 [n2] advertising CockroachDB node at 127.0.0.1:36113 I180827 20:41:53.140624 51685 server/status/recorder.go:652 [n2,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180827 20:41:53.153945 50862 server/server.go:1594 [n2] done ensuring all necessary migrations have run I180827 20:41:53.153974 50862 server/server.go:1597 [n2] serving sql connections W180827 20:41:53.165268 50862 server/status/runtime.go:294 [n?] Could not parse build timestamp: parsing time "" as "2006/01/02 15:04:05": cannot parse "" as "2006" I180827 20:41:53.185802 51467 server/server_update.go:67 [n2] no need to upgrade, cluster already at the newest version I180827 20:41:53.186848 51469 sql/event_log.go:126 [n2] Event: "node_join", target: 2, info: {Descriptor:{NodeID:2 Address:{NetworkField:tcp AddressField:127.0.0.1:36113} Attrs: Locality: ServerVersion:2.0-12 BuildTag:v2.1.0-alpha.20180702-2025-gf1e7bb1 StartedAt:1535402513136479434 LocalityAddress:[]} ClusterID:d5e53e69-a109-4eb6-91bf-29e74ae744ba StartedAt:1535402513136479434 LastUp:1535402513136479434} I180827 20:41:53.189622 50862 server/server.go:830 [n?] monitoring forward clock jumps based on server.clock.forward_jump_check_enabled I180827 20:41:53.189776 50862 base/addr_validation.go:260 [n?] server certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180827 20:41:53.189808 50862 base/addr_validation.go:300 [n?] web UI certificate addresses: IP=127.0.0.1,::1; DNS=localhost,*.local; CN=node I180827 20:41:53.207782 50862 server/config.go:496 [n?] 1 storage engine initialized I180827 20:41:53.207807 50862 server/config.go:499 [n?] RocksDB cache size: 128 MiB I180827 20:41:53.207815 50862 server/config.go:499 [n?] store 0: in-memory, size 0 B W180827 20:41:53.207911 50862 gossip/gossip.go:1371 [n?] no incoming or outgoing connections I180827 20:41:53.207947 50862 server/server.go:1403 [n?] no stores bootstrapped and --join flag specified, awaiting init command. I180827 20:41:53.211471 51475 rpc/nodedialer/nodedialer.go:92 [ct-client] connection to n2 established I180827 20:41:53.223653 51740 gossip/client.go:129 [n?] started gossip client to 127.0.0.1:41477 I180827 20:41:53.223954 51816 gossip/server.go:217 [n1] received initial cluster-verification connection from {tcp 127.0.0.1:46463} I180827 20:41:53.224401 50862 server/node.go:697 [n?] connecting to gossip network to verify cluster ID... I180827 20:41:53.224432 50862 server/node.go:722 [n?] node connected via gossip and verified as part of cluster "d5e53e69-a109-4eb6-91bf-29e74ae744ba" I180827 20:41:53.224690 51837 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180827 20:41:53.225445 51836 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180827 20:41:53.226030 50862 kv/dist_sender.go:345 [n?] unable to determine this node's attributes for replica selection; node is most likely bootstrapping I180827 20:41:53.226699 50862 server/node.go:428 [n?] new node allocated ID 3 I180827 20:41:53.226763 50862 gossip/gossip.go:382 [n3] NodeDescriptor set to node_id:3 address:<network_field:"tcp" address_field:"127.0.0.1:46463" > attrs:<> locality:<> ServerVersion:<major_val:2 minor_val:0 patch:0 unstable:12 > build_tag:"v2.1.0-alpha.20180702-2025-gf1e7bb1" started_at:1535402513226706701 I180827 20:41:53.226805 50862 storage/stores.go:242 [n3] read 0 node addresses from persistent storage I180827 20:41:53.226851 50862 storage/stores.go:261 [n3] wrote 2 node addresses to persistent storage I180827 20:41:53.227563 51809 storage/stores.go:261 [n1] wrote 2 node addresses to persistent storage I180827 20:41:53.227869 51810 storage/stores.go:261 [n2] wrote 2 node addresses to persistent storage I180827 20:41:53.228504 50862 server/node.go:672 [n3] bootstrapped store [n3,s3] I180827 20:41:53.229044 50862 server/node.go:546 [n3] node=3: started with [] engine(s) and attributes [] I180827 20:41:53.229696 50862 server/status/recorder.go:652 [n3] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180827 20:41:53.229749 50862 server/server.go:1807 [n3] Could not start heap profiler worker due to: directory to store profiles could not be determined I180827 20:41:53.235251 50862 server/server.go:1538 [n3] starting https server at 127.0.0.1:43307 (use: 127.0.0.1:43307) I180827 20:41:53.235271 50862 server/server.go:1540 [n3] starting grpc/postgres server at 127.0.0.1:46463 I180827 20:41:53.235283 50862 server/server.go:1541 [n3] advertising CockroachDB node at 127.0.0.1:46463 I180827 20:41:53.240284 50862 server/server.go:1594 [n3] done ensuring all necessary migrations have run I180827 20:41:53.240307 50862 server/server.go:1597 [n3] serving sql connections I180827 20:41:53.243124 51945 server/status/recorder.go:652 [n3,summaries] available memory from cgroups (8.0 EiB) exceeds system memory 16 GiB, using system memory I180827 20:41:53.248117 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r20/1:/Table/{23-50}] sending preemptive snapshot 59e1afc9 at applied index 16 I180827 20:41:53.249136 50862 testutils/testcluster/testcluster.go:536 [n1,s1] has 22 underreplicated ranges I180827 20:41:53.251012 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r20/1:/Table/{23-50}] streamed snapshot to (n2,s2):?: kv pairs: 12, log entries: 6, rate-limit: 8.0 MiB/sec, 3ms I180827 20:41:53.251369 51983 storage/replica_raftstorage.go:784 [n2,s2,r20/?:{-}] applying preemptive snapshot at index 16 (id=59e1afc9, encoded size=2241, 1 rocksdb batches, 6 log entries) I180827 20:41:53.254056 51839 server/server_update.go:67 [n3] no need to upgrade, cluster already at the newest version I180827 20:41:53.255122 51841 sql/event_log.go:126 [n3] Event: "node_join", target: 3, info: {Descriptor:{NodeID:3 Address:{NetworkField:tcp AddressField:127.0.0.1:46463} Attrs: Locality: ServerVersion:2.0-12 BuildTag:v2.1.0-alpha.20180702-2025-gf1e7bb1 StartedAt:1535402513226706701 LocalityAddress:[]} ClusterID:d5e53e69-a109-4eb6-91bf-29e74ae744ba StartedAt:1535402513226706701 LastUp:1535402513226706701} I180827 20:41:53.256061 51983 storage/replica_raftstorage.go:790 [n2,s2,r20/?:/Table/{23-50}] applied preemptive snapshot in 1ms [clear=0ms batch=0ms entries=0ms commit=1ms] I180827 20:41:53.256605 50930 storage/replica_command.go:812 [replicate,n1,s1,r20/1:/Table/{23-50}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r20:/Table/{23-50} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.259565 50930 storage/replica.go:3743 [n1,s1,r20/1:/Table/{23-50}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3 I180827 20:41:53.261627 51625 rpc/nodedialer/nodedialer.go:92 [n2] connection to n1 established I180827 20:41:53.264544 50862 testutils/testcluster/testcluster.go:536 [n1,s1] has 22 underreplicated ranges I180827 20:41:53.286630 50930 rpc/nodedialer/nodedialer.go:92 [replicate,n1,s1,r21/1:/Table/5{0-1}] connection to n3 established I180827 20:41:53.287245 50862 testutils/testcluster/testcluster.go:536 [n1,s1] has 22 underreplicated ranges I180827 20:41:53.287799 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r21/1:/Table/5{0-1}] sending preemptive snapshot de08568a at applied index 18 I180827 20:41:53.288157 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r21/1:/Table/5{0-1}] streamed snapshot to (n3,s3):?: kv pairs: 8, log entries: 8, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.288623 51959 storage/replica_raftstorage.go:784 [n3,s3,r21/?:{-}] applying preemptive snapshot at index 18 (id=de08568a, encoded size=2646, 1 rocksdb batches, 8 log entries) I180827 20:41:53.289814 51959 storage/replica_raftstorage.go:790 [n3,s3,r21/?:/Table/5{0-1}] applied preemptive snapshot in 1ms [clear=0ms batch=0ms entries=0ms commit=1ms] I180827 20:41:53.290329 50930 storage/replica_command.go:812 [replicate,n1,s1,r21/1:/Table/5{0-1}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r21:/Table/5{0-1} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.293678 50930 storage/replica.go:3743 [n1,s1,r21/1:/Table/5{0-1}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.294953 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r22/1:/{Table/51-Max}] sending preemptive snapshot a84e7278 at applied index 12 I180827 20:41:53.295229 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r22/1:/{Table/51-Max}] streamed snapshot to (n3,s3):?: kv pairs: 7, log entries: 2, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.295441 51883 rpc/nodedialer/nodedialer.go:92 [n3] connection to n1 established I180827 20:41:53.295585 51953 storage/replica_raftstorage.go:784 [n3,s3,r22/?:{-}] applying preemptive snapshot at index 12 (id=a84e7278, encoded size=386, 1 rocksdb batches, 2 log entries) I180827 20:41:53.295717 51953 storage/replica_raftstorage.go:790 [n3,s3,r22/?:/{Table/51-Max}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.295955 50930 storage/replica_command.go:812 [replicate,n1,s1,r22/1:/{Table/51-Max}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r22:/{Table/51-Max} [(n1,s1):1, next=2, gen=0] I180827 20:41:53.298097 50930 storage/replica.go:3743 [n1,s1,r22/1:/{Table/51-Max}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.301122 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r8/1:/Table/1{1-2}] sending preemptive snapshot 201bdccc at applied index 18 I180827 20:41:53.301565 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r8/1:/Table/1{1-2}] streamed snapshot to (n3,s3):?: kv pairs: 9, log entries: 8, rate-limit: 8.0 MiB/sec, 3ms I180827 20:41:53.306578 52088 storage/replica_raftstorage.go:784 [n3,s3,r8/?:{-}] applying preemptive snapshot at index 18 (id=201bdccc, encoded size=4352, 1 rocksdb batches, 8 log entries) I180827 20:41:53.306868 52088 storage/replica_raftstorage.go:790 [n3,s3,r8/?:/Table/1{1-2}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.307601 50930 storage/replica_command.go:812 [replicate,n1,s1,r8/1:/Table/1{1-2}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r8:/Table/1{1-2} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.311873 50930 storage/replica.go:3743 [n1,s1,r8/1:/Table/1{1-2}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.314134 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r17/1:/Table/2{0-1}] sending preemptive snapshot 53116eb2 at applied index 16 I180827 20:41:53.314317 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r17/1:/Table/2{0-1}] streamed snapshot to (n3,s3):?: kv pairs: 8, log entries: 6, rate-limit: 8.0 MiB/sec, 2ms I180827 20:41:53.314683 52103 storage/replica_raftstorage.go:784 [n3,s3,r17/?:{-}] applying preemptive snapshot at index 16 (id=53116eb2, encoded size=2105, 1 rocksdb batches, 6 log entries) I180827 20:41:53.314887 52103 storage/replica_raftstorage.go:790 [n3,s3,r17/?:/Table/2{0-1}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.315401 50930 storage/replica_command.go:812 [replicate,n1,s1,r17/1:/Table/2{0-1}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r17:/Table/2{0-1} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.318398 50930 storage/replica.go:3743 [n1,s1,r17/1:/Table/2{0-1}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.319436 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r16/1:/Table/{19-20}] sending preemptive snapshot e0be8540 at applied index 16 I180827 20:41:53.319691 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r16/1:/Table/{19-20}] streamed snapshot to (n2,s2):?: kv pairs: 8, log entries: 6, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.320127 52072 storage/replica_raftstorage.go:784 [n2,s2,r16/?:{-}] applying preemptive snapshot at index 16 (id=e0be8540, encoded size=2109, 1 rocksdb batches, 6 log entries) I180827 20:41:53.320339 52072 storage/replica_raftstorage.go:790 [n2,s2,r16/?:/Table/{19-20}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.320816 50930 storage/replica_command.go:812 [replicate,n1,s1,r16/1:/Table/{19-20}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r16:/Table/{19-20} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.323849 50930 storage/replica.go:3743 [n1,s1,r16/1:/Table/{19-20}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3 I180827 20:41:53.326208 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r15/1:/Table/1{8-9}] sending preemptive snapshot d259ae5c at applied index 16 I180827 20:41:53.326404 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r15/1:/Table/1{8-9}] streamed snapshot to (n2,s2):?: kv pairs: 8, log entries: 6, rate-limit: 8.0 MiB/sec, 2ms I180827 20:41:53.326731 52116 storage/replica_raftstorage.go:784 [n2,s2,r15/?:{-}] applying preemptive snapshot at index 16 (id=d259ae5c, encoded size=2276, 1 rocksdb batches, 6 log entries) I180827 20:41:53.326923 50862 testutils/testcluster/testcluster.go:536 [n1,s1] has 22 underreplicated ranges I180827 20:41:53.326953 52116 storage/replica_raftstorage.go:790 [n2,s2,r15/?:/Table/1{8-9}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.334514 50930 storage/replica_command.go:812 [replicate,n1,s1,r15/1:/Table/1{8-9}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r15:/Table/1{8-9} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.337656 50930 storage/replica.go:3743 [n1,s1,r15/1:/Table/1{8-9}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3 I180827 20:41:53.338767 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r14/1:/Table/1{7-8}] sending preemptive snapshot 9d0058d5 at applied index 16 I180827 20:41:53.339034 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r14/1:/Table/1{7-8}] streamed snapshot to (n2,s2):?: kv pairs: 8, log entries: 6, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.339612 52090 storage/replica_raftstorage.go:784 [n2,s2,r14/?:{-}] applying preemptive snapshot at index 16 (id=9d0058d5, encoded size=2276, 1 rocksdb batches, 6 log entries) I180827 20:41:53.339831 52090 storage/replica_raftstorage.go:790 [n2,s2,r14/?:/Table/1{7-8}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.340173 50930 storage/replica_command.go:812 [replicate,n1,s1,r14/1:/Table/1{7-8}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r14:/Table/1{7-8} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.343121 50930 storage/replica.go:3743 [n1,s1,r14/1:/Table/1{7-8}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3 I180827 20:41:53.345432 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r9/1:/Table/1{2-3}] sending preemptive snapshot 0eea2d20 at applied index 26 I180827 20:41:53.345859 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r9/1:/Table/1{2-3}] streamed snapshot to (n2,s2):?: kv pairs: 53, log entries: 16, rate-limit: 8.0 MiB/sec, 2ms I180827 20:41:53.347137 52066 storage/replica_raftstorage.go:784 [n2,s2,r9/?:{-}] applying preemptive snapshot at index 26 (id=0eea2d20, encoded size=15139, 1 rocksdb batches, 16 log entries) I180827 20:41:53.347467 52066 storage/replica_raftstorage.go:790 [n2,s2,r9/?:/Table/1{2-3}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.348208 50930 storage/replica_command.go:812 [replicate,n1,s1,r9/1:/Table/1{2-3}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r9:/Table/1{2-3} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.352166 50930 storage/replica.go:3743 [n1,s1,r9/1:/Table/1{2-3}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3 I180827 20:41:53.353188 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r4/1:/System/{NodeLive…-tsd}] sending preemptive snapshot 0cdee511 at applied index 39 I180827 20:41:53.353765 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r4/1:/System/{NodeLive…-tsd}] streamed snapshot to (n2,s2):?: kv pairs: 36, log entries: 29, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.354286 51723 storage/replica_raftstorage.go:784 [n2,s2,r4/?:{-}] applying preemptive snapshot at index 39 (id=0cdee511, encoded size=98384, 1 rocksdb batches, 29 log entries) I180827 20:41:53.354994 51723 storage/replica_raftstorage.go:790 [n2,s2,r4/?:/System/{NodeLive…-tsd}] applied preemptive snapshot in 1ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.355529 50930 storage/replica_command.go:812 [replicate,n1,s1,r4/1:/System/{NodeLive…-tsd}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r4:/System/{NodeLivenessMax-tsd} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.358523 50930 storage/replica.go:3743 [n1,s1,r4/1:/System/{NodeLive…-tsd}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3 I180827 20:41:53.360250 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r3/1:/System/NodeLiveness{-Max}] sending preemptive snapshot 965d58b1 at applied index 19 I180827 20:41:53.360436 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r3/1:/System/NodeLiveness{-Max}] streamed snapshot to (n3,s3):?: kv pairs: 10, log entries: 9, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.360789 52150 storage/replica_raftstorage.go:784 [n3,s3,r3/?:{-}] applying preemptive snapshot at index 19 (id=965d58b1, encoded size=4003, 1 rocksdb batches, 9 log entries) I180827 20:41:53.361043 52150 storage/replica_raftstorage.go:790 [n3,s3,r3/?:/System/NodeLiveness{-Max}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.361522 50930 storage/replica_command.go:812 [replicate,n1,s1,r3/1:/System/NodeLiveness{-Max}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r3:/System/NodeLiveness{-Max} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.364392 50930 storage/replica.go:3743 [n1,s1,r3/1:/System/NodeLiveness{-Max}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.366422 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r12/1:/Table/1{5-6}] sending preemptive snapshot 811af376 at applied index 16 I180827 20:41:53.366638 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r12/1:/Table/1{5-6}] streamed snapshot to (n3,s3):?: kv pairs: 8, log entries: 6, rate-limit: 8.0 MiB/sec, 2ms I180827 20:41:53.367089 52137 storage/replica_raftstorage.go:784 [n3,s3,r12/?:{-}] applying preemptive snapshot at index 16 (id=811af376, encoded size=2276, 1 rocksdb batches, 6 log entries) I180827 20:41:53.367359 52137 storage/replica_raftstorage.go:790 [n3,s3,r12/?:/Table/1{5-6}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.368127 50930 storage/replica_command.go:812 [replicate,n1,s1,r12/1:/Table/1{5-6}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r12:/Table/1{5-6} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.371691 50930 storage/replica.go:3743 [n1,s1,r12/1:/Table/1{5-6}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.374563 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r19/1:/Table/2{2-3}] sending preemptive snapshot 9cd02555 at applied index 16 I180827 20:41:53.374760 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r19/1:/Table/2{2-3}] streamed snapshot to (n3,s3):?: kv pairs: 8, log entries: 6, rate-limit: 8.0 MiB/sec, 2ms I180827 20:41:53.375252 52080 storage/replica_raftstorage.go:784 [n3,s3,r19/?:{-}] applying preemptive snapshot at index 16 (id=9cd02555, encoded size=2276, 1 rocksdb batches, 6 log entries) I180827 20:41:53.375582 52080 storage/replica_raftstorage.go:790 [n3,s3,r19/?:/Table/2{2-3}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.375950 50930 storage/replica_command.go:812 [replicate,n1,s1,r19/1:/Table/2{2-3}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r19:/Table/2{2-3} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.381819 50930 storage/replica.go:3743 [n1,s1,r19/1:/Table/2{2-3}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.386461 52091 rpc/nodedialer/nodedialer.go:92 [ct-client] connection to n3 established I180827 20:41:53.386637 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r10/1:/Table/1{3-4}] sending preemptive snapshot a16f4b15 at applied index 64 I180827 20:41:53.388005 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r10/1:/Table/1{3-4}] streamed snapshot to (n3,s3):?: kv pairs: 204, log entries: 54, rate-limit: 8.0 MiB/sec, 4ms I180827 20:41:53.388536 52181 storage/replica_raftstorage.go:784 [n3,s3,r10/?:{-}] applying preemptive snapshot at index 64 (id=a16f4b15, encoded size=62836, 1 rocksdb batches, 54 log entries) I180827 20:41:53.389154 52181 storage/replica_raftstorage.go:790 [n3,s3,r10/?:/Table/1{3-4}] applied preemptive snapshot in 1ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.389513 50930 storage/replica_command.go:812 [replicate,n1,s1,r10/1:/Table/1{3-4}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r10:/Table/1{3-4} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.392649 50930 storage/replica.go:3743 [n1,s1,r10/1:/Table/1{3-4}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.394122 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r2/1:/System/{-NodeLive…}] sending preemptive snapshot 69adabc1 at applied index 23 I180827 20:41:53.394365 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r2/1:/System/{-NodeLive…}] streamed snapshot to (n2,s2):?: kv pairs: 7, log entries: 13, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.394729 52213 storage/replica_raftstorage.go:784 [n2,s2,r2/?:{-}] applying preemptive snapshot at index 23 (id=69adabc1, encoded size=6277, 1 rocksdb batches, 13 log entries) I180827 20:41:53.394981 52213 storage/replica_raftstorage.go:790 [n2,s2,r2/?:/System/{-NodeLive…}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.395465 50930 storage/replica_command.go:812 [replicate,n1,s1,r2/1:/System/{-NodeLive…}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r2:/System/{-NodeLiveness} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.398757 50930 storage/replica.go:3743 [n1,s1,r2/1:/System/{-NodeLive…}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3 I180827 20:41:53.399709 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r18/1:/Table/2{1-2}] sending preemptive snapshot e9df2a4a at applied index 16 I180827 20:41:53.400036 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r18/1:/Table/2{1-2}] streamed snapshot to (n3,s3):?: kv pairs: 8, log entries: 6, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.400391 52185 storage/replica_raftstorage.go:784 [n3,s3,r18/?:{-}] applying preemptive snapshot at index 16 (id=e9df2a4a, encoded size=2272, 1 rocksdb batches, 6 log entries) I180827 20:41:53.400594 52185 storage/replica_raftstorage.go:790 [n3,s3,r18/?:/Table/2{1-2}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.400882 50930 storage/replica_command.go:812 [replicate,n1,s1,r18/1:/Table/2{1-2}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r18:/Table/2{1-2} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.407636 50930 storage/replica.go:3743 [n1,s1,r18/1:/Table/2{1-2}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.408861 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r13/1:/Table/1{6-7}] sending preemptive snapshot 6f914d55 at applied index 16 I180827 20:41:53.409071 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r13/1:/Table/1{6-7}] streamed snapshot to (n2,s2):?: kv pairs: 8, log entries: 6, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.409426 52218 storage/replica_raftstorage.go:784 [n2,s2,r13/?:{-}] applying preemptive snapshot at index 16 (id=6f914d55, encoded size=2276, 1 rocksdb batches, 6 log entries) I180827 20:41:53.409616 52218 storage/replica_raftstorage.go:790 [n2,s2,r13/?:/Table/1{6-7}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.409970 50930 storage/replica_command.go:812 [replicate,n1,s1,r13/1:/Table/1{6-7}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r13:/Table/1{6-7} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.411262 50862 testutils/testcluster/testcluster.go:536 [n1,s1] has 22 underreplicated ranges I180827 20:41:53.412831 50930 storage/replica.go:3743 [n1,s1,r13/1:/Table/1{6-7}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3 I180827 20:41:53.414081 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r11/1:/Table/1{4-5}] sending preemptive snapshot cca961c1 at applied index 16 I180827 20:41:53.414277 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r11/1:/Table/1{4-5}] streamed snapshot to (n3,s3):?: kv pairs: 8, log entries: 6, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.414576 52199 storage/replica_raftstorage.go:784 [n3,s3,r11/?:{-}] applying preemptive snapshot at index 16 (id=cca961c1, encoded size=2272, 1 rocksdb batches, 6 log entries) I180827 20:41:53.414816 52199 storage/replica_raftstorage.go:790 [n3,s3,r11/?:/Table/1{4-5}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.415293 50930 storage/replica_command.go:812 [replicate,n1,s1,r11/1:/Table/1{4-5}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r11:/Table/1{4-5} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.418111 50930 storage/replica.go:3743 [n1,s1,r11/1:/Table/1{4-5}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.419054 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r5/1:/System/ts{d-e}] sending preemptive snapshot 3c3a015f at applied index 27 I180827 20:41:53.423022 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r5/1:/System/ts{d-e}] streamed snapshot to (n3,s3):?: kv pairs: 1391, log entries: 2, rate-limit: 8.0 MiB/sec, 4ms I180827 20:41:53.423893 52201 storage/replica_raftstorage.go:784 [n3,s3,r5/?:{-}] applying preemptive snapshot at index 27 (id=3c3a015f, encoded size=194658, 1 rocksdb batches, 2 log entries) I180827 20:41:53.429501 52201 storage/replica_raftstorage.go:790 [n3,s3,r5/?:/System/ts{d-e}] applied preemptive snapshot in 6ms [clear=0ms batch=0ms entries=2ms commit=4ms] I180827 20:41:53.433500 50930 storage/replica_command.go:812 [replicate,n1,s1,r5/1:/System/ts{d-e}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r5:/System/ts{d-e} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.437580 50930 storage/replica.go:3743 [n1,s1,r5/1:/System/ts{d-e}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.440575 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r6/1:/{System/tse-Table/System…}] sending preemptive snapshot cbd412df at applied index 21 I180827 20:41:53.440794 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r6/1:/{System/tse-Table/System…}] streamed snapshot to (n3,s3):?: kv pairs: 8, log entries: 11, rate-limit: 8.0 MiB/sec, 2ms I180827 20:41:53.441181 52260 storage/replica_raftstorage.go:784 [n3,s3,r6/?:{-}] applying preemptive snapshot at index 21 (id=cbd412df, encoded size=4339, 1 rocksdb batches, 11 log entries) I180827 20:41:53.441400 52260 storage/replica_raftstorage.go:790 [n3,s3,r6/?:/{System/tse-Table/System…}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.441676 50930 storage/replica_command.go:812 [replicate,n1,s1,r6/1:/{System/tse-Table/System…}] change replicas (ADD_REPLICA (n3,s3):2): read existing descriptor r6:/{System/tse-Table/SystemConfigSpan/Start} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.448564 52224 rpc/nodedialer/nodedialer.go:92 [ct-client] connection to n2 established I180827 20:41:53.461587 50930 storage/replica.go:3743 [n1,s1,r6/1:/{System/tse-Table/System…}] proposing ADD_REPLICA((n3,s3):2): updated=[(n1,s1):1 (n3,s3):2] next=3 I180827 20:41:53.463345 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r7/1:/Table/{SystemCon…-11}] sending preemptive snapshot 114f4385 at applied index 29 I180827 20:41:53.464896 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r7/1:/Table/{SystemCon…-11}] streamed snapshot to (n2,s2):?: kv pairs: 59, log entries: 19, rate-limit: 8.0 MiB/sec, 3ms I180827 20:41:53.465343 52280 storage/replica_raftstorage.go:784 [n2,s2,r7/?:{-}] applying preemptive snapshot at index 29 (id=114f4385, encoded size=16646, 1 rocksdb batches, 19 log entries) I180827 20:41:53.465821 52280 storage/replica_raftstorage.go:790 [n2,s2,r7/?:/Table/{SystemCon…-11}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.466988 50930 storage/replica_command.go:812 [replicate,n1,s1,r7/1:/Table/{SystemCon…-11}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r7:/Table/{SystemConfigSpan/Start-11} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.472743 50930 storage/replica.go:3743 [n1,s1,r7/1:/Table/{SystemCon…-11}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3 I180827 20:41:53.474632 50930 storage/store_snapshot.go:615 [replicate,n1,s1,r1/1:/{Min-System/}] sending preemptive snapshot 0a244018 at applied index 114 I180827 20:41:53.475250 50930 storage/store_snapshot.go:657 [replicate,n1,s1,r1/1:/{Min-System/}] streamed snapshot to (n2,s2):?: kv pairs: 73, log entries: 90, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.475827 52267 storage/replica_raftstorage.go:784 [n2,s2,r1/?:{-}] applying preemptive snapshot at index 114 (id=0a244018, encoded size=40271, 1 rocksdb batches, 90 log entries) I180827 20:41:53.476525 52267 storage/replica_raftstorage.go:790 [n2,s2,r1/?:/{Min-System/}] applied preemptive snapshot in 1ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.476869 50930 storage/replica_command.go:812 [replicate,n1,s1,r1/1:/{Min-System/}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r1:/{Min-System/} [(n1,s1):1, next=2, gen=1] I180827 20:41:53.482912 50930 storage/replica.go:3743 [n1,s1,r1/1:/{Min-System/}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3 I180827 20:41:53.483281 50930 storage/queue.go:873 [n1,replicate] purgatory is now empty I180827 20:41:53.485684 52286 storage/store_snapshot.go:615 [replicate,n1,s1,r20/1:/Table/{23-50}] sending preemptive snapshot f1426c69 at applied index 19 I180827 20:41:53.487316 52286 storage/store_snapshot.go:657 [replicate,n1,s1,r20/1:/Table/{23-50}] streamed snapshot to (n3,s3):?: kv pairs: 13, log entries: 9, rate-limit: 8.0 MiB/sec, 4ms I180827 20:41:53.487681 52252 storage/replica_raftstorage.go:784 [n3,s3,r20/?:{-}] applying preemptive snapshot at index 19 (id=f1426c69, encoded size=3273, 1 rocksdb batches, 9 log entries) I180827 20:41:53.487932 52252 storage/replica_raftstorage.go:790 [n3,s3,r20/?:/Table/{23-50}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.488311 52286 storage/replica_command.go:812 [replicate,n1,s1,r20/1:/Table/{23-50}] change replicas (ADD_REPLICA (n3,s3):3): read existing descriptor r20:/Table/{23-50} [(n1,s1):1, (n2,s2):2, next=3, gen=1] I180827 20:41:53.503580 52286 storage/replica.go:3743 [n1,s1,r20/1:/Table/{23-50}] proposing ADD_REPLICA((n3,s3):3): updated=[(n1,s1):1 (n2,s2):2 (n3,s3):3] next=4 I180827 20:41:53.505707 52235 storage/store_snapshot.go:615 [replicate,n1,s1,r1/1:/{Min-System/}] sending preemptive snapshot 99036b07 at applied index 119 I180827 20:41:53.506514 52235 storage/store_snapshot.go:657 [replicate,n1,s1,r1/1:/{Min-System/}] streamed snapshot to (n3,s3):?: kv pairs: 78, log entries: 95, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.507282 52188 storage/replica_raftstorage.go:784 [n3,s3,r1/?:{-}] applying preemptive snapshot at index 119 (id=99036b07, encoded size=42101, 1 rocksdb batches, 95 log entries) I180827 20:41:53.508109 52188 storage/replica_raftstorage.go:790 [n3,s3,r1/?:/{Min-System/}] applied preemptive snapshot in 1ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.508641 52235 storage/replica_command.go:812 [replicate,n1,s1,r1/1:/{Min-System/}] change replicas (ADD_REPLICA (n3,s3):3): read existing descriptor r1:/{Min-System/} [(n1,s1):1, (n2,s2):2, next=3, gen=1] I180827 20:41:53.512524 52235 storage/replica.go:3743 [n1,s1,r1/1:/{Min-System/}] proposing ADD_REPLICA((n3,s3):3): updated=[(n1,s1):1 (n2,s2):2 (n3,s3):3] next=4 I180827 20:41:53.513999 52209 storage/store_snapshot.go:615 [replicate,n1,s1,r7/1:/Table/{SystemCon…-11}] sending preemptive snapshot bb53109c at applied index 32 I180827 20:41:53.514379 52209 storage/store_snapshot.go:657 [replicate,n1,s1,r7/1:/Table/{SystemCon…-11}] streamed snapshot to (n3,s3):?: kv pairs: 60, log entries: 22, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.514821 52292 storage/replica_raftstorage.go:784 [n3,s3,r7/?:{-}] applying preemptive snapshot at index 32 (id=bb53109c, encoded size=17687, 1 rocksdb batches, 22 log entries) I180827 20:41:53.515905 52292 storage/replica_raftstorage.go:790 [n3,s3,r7/?:/Table/{SystemCon…-11}] applied preemptive snapshot in 1ms [clear=1ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.516367 52209 storage/replica_command.go:812 [replicate,n1,s1,r7/1:/Table/{SystemCon…-11}] change replicas (ADD_REPLICA (n3,s3):3): read existing descriptor r7:/Table/{SystemConfigSpan/Start-11} [(n1,s1):1, (n2,s2):2, next=3, gen=1] I180827 20:41:53.520158 52209 storage/replica.go:3743 [n1,s1,r7/1:/Table/{SystemCon…-11}] proposing ADD_REPLICA((n3,s3):3): updated=[(n1,s1):1 (n2,s2):2 (n3,s3):3] next=4 I180827 20:41:53.521958 52312 storage/store_snapshot.go:615 [replicate,n1,s1,r6/1:/{System/tse-Table/System…}] sending preemptive snapshot 2ca43612 at applied index 24 I180827 20:41:53.522776 52312 storage/store_snapshot.go:657 [replicate,n1,s1,r6/1:/{System/tse-Table/System…}] streamed snapshot to (n2,s2):?: kv pairs: 9, log entries: 14, rate-limit: 8.0 MiB/sec, 2ms I180827 20:41:53.523128 52239 storage/replica_raftstorage.go:784 [n2,s2,r6/?:{-}] applying preemptive snapshot at index 24 (id=2ca43612, encoded size=5410, 1 rocksdb batches, 14 log entries) I180827 20:41:53.523377 52239 storage/replica_raftstorage.go:790 [n2,s2,r6/?:/{System/tse-Table/System…}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.523701 52312 storage/replica_command.go:812 [replicate,n1,s1,r6/1:/{System/tse-Table/System…}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r6:/{System/tse-Table/SystemConfigSpan/Start} [(n1,s1):1, (n3,s3):2, next=3, gen=1] I180827 20:41:53.525176 50862 testutils/testcluster/testcluster.go:536 [n1,s1] has 19 underreplicated ranges I180827 20:41:53.527482 52312 storage/replica.go:3743 [n1,s1,r6/1:/{System/tse-Table/System…}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3] next=4 I180827 20:41:53.528875 52327 storage/store_snapshot.go:615 [replicate,n1,s1,r5/1:/System/ts{d-e}] sending preemptive snapshot 731be2ae at applied index 30 I180827 20:41:53.532860 52327 storage/store_snapshot.go:657 [replicate,n1,s1,r5/1:/System/ts{d-e}] streamed snapshot to (n2,s2):?: kv pairs: 1392, log entries: 5, rate-limit: 8.0 MiB/sec, 4ms I180827 20:41:53.533361 52316 storage/replica_raftstorage.go:784 [n2,s2,r5/?:{-}] applying preemptive snapshot at index 30 (id=731be2ae, encoded size=195741, 1 rocksdb batches, 5 log entries) I180827 20:41:53.535834 52316 storage/replica_raftstorage.go:790 [n2,s2,r5/?:/System/ts{d-e}] applied preemptive snapshot in 2ms [clear=0ms batch=0ms entries=0ms commit=2ms] I180827 20:41:53.536253 52327 storage/replica_command.go:812 [replicate,n1,s1,r5/1:/System/ts{d-e}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r5:/System/ts{d-e} [(n1,s1):1, (n3,s3):2, next=3, gen=1] I180827 20:41:53.540576 52327 storage/replica.go:3743 [n1,s1,r5/1:/System/ts{d-e}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3] next=4 I180827 20:41:53.545804 52341 storage/store_snapshot.go:615 [replicate,n1,s1,r11/1:/Table/1{4-5}] sending preemptive snapshot 7497a95f at applied index 19 I180827 20:41:53.546108 52341 storage/store_snapshot.go:657 [replicate,n1,s1,r11/1:/Table/1{4-5}] streamed snapshot to (n2,s2):?: kv pairs: 9, log entries: 9, rate-limit: 8.0 MiB/sec, 4ms I180827 20:41:53.546590 52275 storage/replica_raftstorage.go:784 [n2,s2,r11/?:{-}] applying preemptive snapshot at index 19 (id=7497a95f, encoded size=3304, 1 rocksdb batches, 9 log entries) I180827 20:41:53.546960 52275 storage/replica_raftstorage.go:790 [n2,s2,r11/?:/Table/1{4-5}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.547386 52341 storage/replica_command.go:812 [replicate,n1,s1,r11/1:/Table/1{4-5}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r11:/Table/1{4-5} [(n1,s1):1, (n3,s3):2, next=3, gen=1] I180827 20:41:53.551568 52341 storage/replica.go:3743 [n1,s1,r11/1:/Table/1{4-5}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3] next=4 I180827 20:41:53.554959 52323 storage/store_snapshot.go:615 [replicate,n1,s1,r13/1:/Table/1{6-7}] sending preemptive snapshot 5932a5bd at applied index 19 I180827 20:41:53.555353 52323 storage/store_snapshot.go:657 [replicate,n1,s1,r13/1:/Table/1{6-7}] streamed snapshot to (n3,s3):?: kv pairs: 9, log entries: 9, rate-limit: 8.0 MiB/sec, 3ms I180827 20:41:53.555743 52329 storage/replica_raftstorage.go:784 [n3,s3,r13/?:{-}] applying preemptive snapshot at index 19 (id=5932a5bd, encoded size=3308, 1 rocksdb batches, 9 log entries) I180827 20:41:53.556103 52329 storage/replica_raftstorage.go:790 [n3,s3,r13/?:/Table/1{6-7}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.556489 52323 storage/replica_command.go:812 [replicate,n1,s1,r13/1:/Table/1{6-7}] change replicas (ADD_REPLICA (n3,s3):3): read existing descriptor r13:/Table/1{6-7} [(n1,s1):1, (n2,s2):2, next=3, gen=1] I180827 20:41:53.563494 52323 storage/replica.go:3743 [n1,s1,r13/1:/Table/1{6-7}] proposing ADD_REPLICA((n3,s3):3): updated=[(n1,s1):1 (n2,s2):2 (n3,s3):3] next=4 I180827 20:41:53.566866 52362 storage/store_snapshot.go:615 [replicate,n1,s1,r18/1:/Table/2{1-2}] sending preemptive snapshot c74baa54 at applied index 19 I180827 20:41:53.568042 52362 storage/store_snapshot.go:657 [replicate,n1,s1,r18/1:/Table/2{1-2}] streamed snapshot to (n2,s2):?: kv pairs: 9, log entries: 9, rate-limit: 8.0 MiB/sec, 3ms I180827 20:41:53.568417 52366 storage/replica_raftstorage.go:784 [n2,s2,r18/?:{-}] applying preemptive snapshot at index 19 (id=c74baa54, encoded size=3304, 1 rocksdb batches, 9 log entries) I180827 20:41:53.568651 52366 storage/replica_raftstorage.go:790 [n2,s2,r18/?:/Table/2{1-2}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.568954 52362 storage/replica_command.go:812 [replicate,n1,s1,r18/1:/Table/2{1-2}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r18:/Table/2{1-2} [(n1,s1):1, (n3,s3):2, next=3, gen=1] I180827 20:41:53.572711 52362 storage/replica.go:3743 [n1,s1,r18/1:/Table/2{1-2}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3] next=4 I180827 20:41:53.574655 52190 storage/store_snapshot.go:615 [replicate,n1,s1,r2/1:/System/{-NodeLive…}] sending preemptive snapshot d4c499ea at applied index 26 I180827 20:41:53.574962 52190 storage/store_snapshot.go:657 [replicate,n1,s1,r2/1:/System/{-NodeLive…}] streamed snapshot to (n3,s3):?: kv pairs: 8, log entries: 16, rate-limit: 8.0 MiB/sec, 1ms I180827 20:41:53.575725 52299 storage/replica_raftstorage.go:784 [n3,s3,r2/?:{-}] applying preemptive snapshot at index 26 (id=d4c499ea, encoded size=7349, 1 rocksdb batches, 16 log entries) I180827 20:41:53.576022 52299 storage/replica_raftstorage.go:790 [n3,s3,r2/?:/System/{-NodeLive…}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.576405 52190 storage/replica_command.go:812 [replicate,n1,s1,r2/1:/System/{-NodeLive…}] change replicas (ADD_REPLICA (n3,s3):3): read existing descriptor r2:/System/{-NodeLiveness} [(n1,s1):1, (n2,s2):2, next=3, gen=1] I180827 20:41:53.579762 52190 storage/replica.go:3743 [n1,s1,r2/1:/System/{-NodeLive…}] proposing ADD_REPLICA((n3,s3):3): updated=[(n1,s1):1 (n2,s2):2 (n3,s3):3] next=4 I180827 20:41:53.583103 52372 storage/store_snapshot.go:615 [replicate,n1,s1,r10/1:/Table/1{3-4}] sending preemptive snapshot dbe83d06 at applied index 103 I180827 20:41:53.583765 52372 storage/store_snapshot.go:657 [replicate,n1,s1,r10/1:/Table/1{3-4}] streamed snapshot to (n2,s2):?: kv pairs: 295, log entries: 10, rate-limit: 8.0 MiB/sec, 2ms I180827 20:41:53.584214 52369 storage/replica_raftstorage.go:784 [n2,s2,r10/?:{-}] applying preemptive snapshot at index 103 (id=dbe83d06, encoded size=38018, 1 rocksdb batches, 10 log entries) I180827 20:41:53.584577 52369 storage/replica_raftstorage.go:790 [n2,s2,r10/?:/Table/1{3-4}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.584963 52372 storage/replica_command.go:812 [replicate,n1,s1,r10/1:/Table/1{3-4}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r10:/Table/1{3-4} [(n1,s1):1, (n3,s3):2, next=3, gen=1] I180827 20:41:53.588661 52372 storage/replica.go:3743 [n1,s1,r10/1:/Table/1{3-4}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3] next=4 I180827 20:41:53.590522 52331 storage/store_snapshot.go:615 [replicate,n1,s1,r19/1:/Table/2{2-3}] sending preemptive snapshot ba0de389 at applied index 19 I180827 20:41:53.596120 52331 storage/store_snapshot.go:657 [replicate,n1,s1,r19/1:/Table/2{2-3}] streamed snapshot to (n2,s2):?: kv pairs: 9, log entries: 9, rate-limit: 8.0 MiB/sec, 6ms I180827 20:41:53.597215 52259 storage/replica_raftstorage.go:784 [n2,s2,r19/?:{-}] applying preemptive snapshot at index 19 (id=ba0de389, encoded size=3308, 1 rocksdb batches, 9 log entries) I180827 20:41:53.597484 52259 storage/replica_raftstorage.go:790 [n2,s2,r19/?:/Table/2{2-3}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.597898 52331 storage/replica_command.go:812 [replicate,n1,s1,r19/1:/Table/2{2-3}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r19:/Table/2{2-3} [(n1,s1):1, (n3,s3):2, next=3, gen=1] I180827 20:41:53.601937 52331 storage/replica.go:3743 [n1,s1,r19/1:/Table/2{2-3}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3] next=4 I180827 20:41:53.604379 52303 storage/store_snapshot.go:615 [replicate,n1,s1,r12/1:/Table/1{5-6}] sending preemptive snapshot 60969a90 at applied index 19 I180827 20:41:53.606558 52303 storage/store_snapshot.go:657 [replicate,n1,s1,r12/1:/Table/1{5-6}] streamed snapshot to (n2,s2):?: kv pairs: 9, log entries: 9, rate-limit: 8.0 MiB/sec, 3ms I180827 20:41:53.606975 52307 storage/replica_raftstorage.go:784 [n2,s2,r12/?:{-}] applying preemptive snapshot at index 19 (id=60969a90, encoded size=3308, 1 rocksdb batches, 9 log entries) I180827 20:41:53.607211 52307 storage/replica_raftstorage.go:790 [n2,s2,r12/?:/Table/1{5-6}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.608276 52303 storage/replica_command.go:812 [replicate,n1,s1,r12/1:/Table/1{5-6}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r12:/Table/1{5-6} [(n1,s1):1, (n3,s3):2, next=3, gen=1] I180827 20:41:53.612795 52303 storage/replica.go:3743 [n1,s1,r12/1:/Table/1{5-6}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3] next=4 I180827 20:41:53.615927 52392 storage/store_snapshot.go:615 [replicate,n1,s1,r3/1:/System/NodeLiveness{-Max}] sending preemptive snapshot 3d427041 at applied index 22 I180827 20:41:53.617153 52392 storage/store_snapshot.go:657 [replicate,n1,s1,r3/1:/System/NodeLiveness{-Max}] streamed snapshot to (n2,s2):?: kv pairs: 11, log entries: 12, rate-limit: 8.0 MiB/sec, 3ms I180827 20:41:53.618565 52405 storage/replica_raftstorage.go:784 [n2,s2,r3/?:{-}] applying preemptive snapshot at index 22 (id=3d427041, encoded size=5215, 1 rocksdb batches, 12 log entries) I180827 20:41:53.619140 52405 storage/replica_raftstorage.go:790 [n2,s2,r3/?:/System/NodeLiveness{-Max}] applied preemptive snapshot in 1ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.621854 50862 testutils/testcluster/testcluster.go:536 [n1,s1] has 10 underreplicated ranges I180827 20:41:53.635001 52392 storage/replica_command.go:812 [replicate,n1,s1,r3/1:/System/NodeLiveness{-Max}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r3:/System/NodeLiveness{-Max} [(n1,s1):1, (n3,s3):2, next=3, gen=1] I180827 20:41:53.638726 52392 storage/replica.go:3743 [n1,s1,r3/1:/System/NodeLiveness{-Max}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3] next=4 I180827 20:41:53.643490 52420 storage/store_snapshot.go:615 [replicate,n1,s1,r4/1:/System/{NodeLive…-tsd}] sending preemptive snapshot a2505d74 at applied index 42 I180827 20:41:53.644245 52420 storage/store_snapshot.go:657 [replicate,n1,s1,r4/1:/System/{NodeLive…-tsd}] streamed snapshot to (n3,s3):?: kv pairs: 37, log entries: 32, rate-limit: 8.0 MiB/sec, 3ms I180827 20:41:53.644709 52436 storage/replica_raftstorage.go:784 [n3,s3,r4/?:{-}] applying preemptive snapshot at index 42 (id=a2505d74, encoded size=99568, 1 rocksdb batches, 32 log entries) I180827 20:41:53.645176 52436 storage/replica_raftstorage.go:790 [n3,s3,r4/?:/System/{NodeLive…-tsd}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms] I180827 20:41:53.645567 52420 storage/replica_command.go:812 [replicate,n1,s1,r4/1:/System/{NodeLive…-tsd}] change replicas (ADD_REPLICA (n3,s3):3): read existing descriptor r4:/System/{NodeLivenessMax-tsd} [(n1,s1):1, (n2,s2):2, next=3, gen=1] I180827 20:41:53.649433 52420 storage/replica.go:3743 [n1,s1,r4/1:/System/{NodeLive…-tsd}] proposing ADD_REPLICA((n3,s3):3): updated=[(n1,s1):1 (n2,s2):2 (n3,s3):3] next=4 I180827 20:41:53.652117 52398 storage/store_snapshot.go:615 [replicate,n1,s1,r9/1:/Table/1{2-3}] sending preemptive snapshot 6dc7ffcb at applied index 29 I180827 20:41:53.653729 52398 storage/store_snapshot.go:657 [replicate,n1,s1,r9/1:/Table/1{2-3}] streamed snapshot to (n3,s3):?: kv pairs: 54, log entries: 19, rate-limit: 8.0 MiB/sec, 3ms I180827 20:41:53.654216 52402 storage/replica_raftstorage.go:784 [n3,s3,r9/?:{-}] applying preemptive snapshot at index 29 (id=6dc7ffcb, encoded size=16171, 1 ``` Please assign, take a look and update the issue accordingly.
test
teamcity failed test test testimportpgdump t the following tests appear to have failed on release banana you may want to check t e fail test testimportpgdump test ended in panic stdout server status runtime go could not parse build timestamp parsing time as cannot parse as server server go monitoring forward clock jumps based on server clock forward jump check enabled base addr validation go server certificate addresses ip dns localhost local cn node base addr validation go web ui certificate addresses ip dns localhost local cn node server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b server node go cluster has been created server server go add additional nodes by specifying join gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag alpha started at storage store go failed initial metrics computation system config not yet available server node go initialized store disk capacity mib available mib used b logicalbytes kib ranges leases queries writes bytesperreplica pmax writesperreplica pmax storage stores go read node addresses from persistent storage server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster server node go node started with engine s and attributes server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go could not start heap profiler worker due to directory to store profiles could not be determined server server go starting https server at use server server go starting grpc postgres server at server server go advertising cockroachdb node at server status recorder go available memory from cgroups eib exceeds system memory gib using system memory storage replica command go initiating a split of this range at key system storage replica command go initiating a split of this range at key system nodeliveness storage intent resolver go failed to push during intent resolution failed to push unnamed id key table systemconfigspan start rw true pri iso serializable stat pending epo ts orig max wto false rop false seq sql event log go event set cluster setting target info settingname diagnostics reporting enabled value true user root storage replica command go initiating a split of this range at key system nodelivenessmax storage replica command go initiating a split of this range at key system tsd storage replica command go initiating a split of this range at key system tse sql event log go event set cluster setting target info settingname version value user root sql event log go event set cluster setting target info settingname trace debug enable value false user root storage replica command go initiating a split of this range at key table systemconfigspan start storage replica command go initiating a split of this range at key table sql event log go event set cluster setting target info settingname cluster secret value user root storage replica command go initiating a split of this range at key table sql event log go event create database target info databasename defaultdb statement create database if not exists defaultdb user root storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event create database target info databasename postgres statement create database if not exists postgres user root storage replica command go initiating a split of this range at key table server server go done ensuring all necessary migrations have run server server go serving sql connections server server update go no need to upgrade cluster already at the newest version sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion buildtag alpha startedat localityaddress clusterid startedat lastup storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table server status runtime go could not parse build timestamp parsing time as cannot parse as server server go monitoring forward clock jumps based on server clock forward jump check enabled base addr validation go server certificate addresses ip dns localhost local cn node base addr validation go web ui certificate addresses ip dns localhost local cn node server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b gossip gossip go no incoming or outgoing connections server server go no stores bootstrapped and join flag specified awaiting init command gossip client go started gossip client to gossip server go received initial cluster verification connection from tcp server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping server node go new node allocated id gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag alpha started at storage stores go read node addresses from persistent storage storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage server node go bootstrapped store server node go node started with engine s and attributes server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go could not start heap profiler worker due to directory to store profiles could not be determined server server go starting https server at use server server go starting grpc postgres server at server server go advertising cockroachdb node at server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go done ensuring all necessary migrations have run server server go serving sql connections server status runtime go could not parse build timestamp parsing time as cannot parse as server server update go no need to upgrade cluster already at the newest version sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion buildtag alpha startedat localityaddress clusterid startedat lastup server server go monitoring forward clock jumps based on server clock forward jump check enabled base addr validation go server certificate addresses ip dns localhost local cn node base addr validation go web ui certificate addresses ip dns localhost local cn node server config go storage engine initialized server config go rocksdb cache size mib server config go store in memory size b gossip gossip go no incoming or outgoing connections server server go no stores bootstrapped and join flag specified awaiting init command rpc nodedialer nodedialer go connection to established gossip client go started gossip client to gossip server go received initial cluster verification connection from tcp server node go connecting to gossip network to verify cluster id server node go node connected via gossip and verified as part of cluster kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping kv dist sender go unable to determine this node s attributes for replica selection node is most likely bootstrapping server node go new node allocated id gossip gossip go nodedescriptor set to node id address attrs locality serverversion build tag alpha started at storage stores go read node addresses from persistent storage storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage storage stores go wrote node addresses to persistent storage server node go bootstrapped store server node go node started with engine s and attributes server status recorder go available memory from cgroups eib exceeds system memory gib using system memory server server go could not start heap profiler worker due to directory to store profiles could not be determined server server go starting https server at use server server go starting grpc postgres server at server server go advertising cockroachdb node at server server go done ensuring all necessary migrations have run server server go serving sql connections server status recorder go available memory from cgroups eib exceeds system memory gib using system memory storage store snapshot go sending preemptive snapshot at applied index testutils testcluster testcluster go has underreplicated ranges storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries server server update go no need to upgrade cluster already at the newest version sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion buildtag alpha startedat localityaddress clusterid startedat lastup storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next rpc nodedialer nodedialer go connection to established testutils testcluster testcluster go has underreplicated ranges rpc nodedialer nodedialer go connection to established testutils testcluster testcluster go has underreplicated ranges storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec rpc nodedialer nodedialer go connection to established storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table max storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries testutils testcluster testcluster go has underreplicated ranges storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor system nodelivenessmax tsd storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor system nodeliveness max storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next rpc nodedialer nodedialer go connection to established storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor system nodeliveness storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table testutils testcluster testcluster go has underreplicated ranges storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor system ts d e storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor system tse table systemconfigspan start rpc nodedialer nodedialer go connection to established storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table systemconfigspan start storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor min system storage replica go proposing add replica updated next storage queue go purgatory is now empty storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor min system storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table systemconfigspan start storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor system tse table systemconfigspan start testutils testcluster testcluster go has underreplicated ranges storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor system ts d e storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor system nodeliveness storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in testutils testcluster testcluster go has underreplicated ranges storage replica command go change replicas add replica read existing descriptor system nodeliveness max storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor system nodelivenessmax tsd storage replica go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size please assign take a look and update the issue accordingly
1
253,980
27,338,744,700
IssuesEvent
2023-02-26 14:41:41
asaf-mend-test/lvp-is-amazing
https://api.github.com/repos/asaf-mend-test/lvp-is-amazing
closed
bootstrap-4.0.0-beta.tgz: 2 vulnerabilities (highest severity is: 6.1) - autoclosed
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-4.0.0-beta.tgz</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://registry.npmjs.org/bootstrap/-/bootstrap-4.0.0-beta.tgz">https://registry.npmjs.org/bootstrap/-/bootstrap-4.0.0-beta.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/bootstrap/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/asaf-mend-test/lvp-is-amazing/commit/e263f4c15af0c3b33faf1d0c1660af18d97b9e79">e263f4c15af0c3b33faf1d0c1660af18d97b9e79</a></p></details> #### Oops, something went wrong. We couldn’t find a fix. Support token-263e18ac2ab6b46f7bf997d902db838711677416400043_1 ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (bootstrap version) | Fix PR available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2019-8331](https://www.mend.io/vulnerability-database/CVE-2019-8331) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-4.0.0-beta.tgz | Direct | N/A | &#10060; | | [CVE-2016-10735](https://www.mend.io/vulnerability-database/CVE-2016-10735) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-4.0.0-beta.tgz | Direct | N/A | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-8331</summary> ### Vulnerable Library - <b>bootstrap-4.0.0-beta.tgz</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://registry.npmjs.org/bootstrap/-/bootstrap-4.0.0-beta.tgz">https://registry.npmjs.org/bootstrap/-/bootstrap-4.0.0-beta.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/bootstrap/package.json</p> <p> Dependency Hierarchy: - :x: **bootstrap-4.0.0-beta.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/asaf-mend-test/lvp-is-amazing/commit/e263f4c15af0c3b33faf1d0c1660af18d97b9e79">e263f4c15af0c3b33faf1d0c1660af18d97b9e79</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute. <p>Publish Date: 2019-02-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-8331>CVE-2019-8331</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-02-20</p> <p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p> </p> <p></p> <p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2016-10735</summary> ### Vulnerable Library - <b>bootstrap-4.0.0-beta.tgz</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://registry.npmjs.org/bootstrap/-/bootstrap-4.0.0-beta.tgz">https://registry.npmjs.org/bootstrap/-/bootstrap-4.0.0-beta.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/bootstrap/package.json</p> <p> Dependency Hierarchy: - :x: **bootstrap-4.0.0-beta.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/asaf-mend-test/lvp-is-amazing/commit/e263f4c15af0c3b33faf1d0c1660af18d97b9e79">e263f4c15af0c3b33faf1d0c1660af18d97b9e79</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-10735>CVE-2016-10735</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10735">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10735</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: bootstrap - 3.4.0, 4.0.0-beta.2</p> </p> <p></p> <p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p> </details>
True
bootstrap-4.0.0-beta.tgz: 2 vulnerabilities (highest severity is: 6.1) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-4.0.0-beta.tgz</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://registry.npmjs.org/bootstrap/-/bootstrap-4.0.0-beta.tgz">https://registry.npmjs.org/bootstrap/-/bootstrap-4.0.0-beta.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/bootstrap/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/asaf-mend-test/lvp-is-amazing/commit/e263f4c15af0c3b33faf1d0c1660af18d97b9e79">e263f4c15af0c3b33faf1d0c1660af18d97b9e79</a></p></details> #### Oops, something went wrong. We couldn’t find a fix. Support token-263e18ac2ab6b46f7bf997d902db838711677416400043_1 ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (bootstrap version) | Fix PR available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2019-8331](https://www.mend.io/vulnerability-database/CVE-2019-8331) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-4.0.0-beta.tgz | Direct | N/A | &#10060; | | [CVE-2016-10735](https://www.mend.io/vulnerability-database/CVE-2016-10735) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-4.0.0-beta.tgz | Direct | N/A | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-8331</summary> ### Vulnerable Library - <b>bootstrap-4.0.0-beta.tgz</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://registry.npmjs.org/bootstrap/-/bootstrap-4.0.0-beta.tgz">https://registry.npmjs.org/bootstrap/-/bootstrap-4.0.0-beta.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/bootstrap/package.json</p> <p> Dependency Hierarchy: - :x: **bootstrap-4.0.0-beta.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/asaf-mend-test/lvp-is-amazing/commit/e263f4c15af0c3b33faf1d0c1660af18d97b9e79">e263f4c15af0c3b33faf1d0c1660af18d97b9e79</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute. <p>Publish Date: 2019-02-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-8331>CVE-2019-8331</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-02-20</p> <p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p> </p> <p></p> <p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2016-10735</summary> ### Vulnerable Library - <b>bootstrap-4.0.0-beta.tgz</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://registry.npmjs.org/bootstrap/-/bootstrap-4.0.0-beta.tgz">https://registry.npmjs.org/bootstrap/-/bootstrap-4.0.0-beta.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/bootstrap/package.json</p> <p> Dependency Hierarchy: - :x: **bootstrap-4.0.0-beta.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/asaf-mend-test/lvp-is-amazing/commit/e263f4c15af0c3b33faf1d0c1660af18d97b9e79">e263f4c15af0c3b33faf1d0c1660af18d97b9e79</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-10735>CVE-2016-10735</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10735">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10735</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: bootstrap - 3.4.0, 4.0.0-beta.2</p> </p> <p></p> <p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p> </details>
non_test
bootstrap beta tgz vulnerabilities highest severity is autoclosed vulnerable library bootstrap beta tgz the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file package json path to vulnerable library node modules bootstrap package json found in head commit a href oops something went wrong we couldn’t find a fix support token vulnerabilities cve severity cvss dependency type fixed in bootstrap version fix pr available medium bootstrap beta tgz direct n a medium bootstrap beta tgz direct n a details cve vulnerable library bootstrap beta tgz the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file package json path to vulnerable library node modules bootstrap package json dependency hierarchy x bootstrap beta tgz vulnerable library found in head commit a href found in base branch main vulnerability details in bootstrap before and x before xss is possible in the tooltip or popover data template attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution bootstrap bootstrap sass in order to enable automatic remediation please create cve vulnerable library bootstrap beta tgz the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file package json path to vulnerable library node modules bootstrap package json dependency hierarchy x bootstrap beta tgz vulnerable library found in head commit a href found in base branch main vulnerability details in bootstrap x before and x beta before beta xss is possible in the data target attribute a different vulnerability than cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap beta in order to enable automatic remediation please create
0
93,350
15,885,816,705
IssuesEvent
2021-04-09 21:16:02
turkdevops/grafana
https://api.github.com/repos/turkdevops/grafana
opened
WS-2019-0427 (Medium) detected in elliptic-6.5.1.tgz
security vulnerability
## WS-2019-0427 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.5.1.tgz</b></p></summary> <p>EC cryptography</p> <p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.5.1.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.5.1.tgz</a></p> <p>Path to dependency file: grafana/node_modules/elliptic/package.json</p> <p>Path to vulnerable library: grafana/node_modules/elliptic/package.json</p> <p> Dependency Hierarchy: - webpack-4.29.6.tgz (Root Library) - node-libs-browser-2.2.1.tgz - crypto-browserify-3.12.0.tgz - create-ecdh-4.0.3.tgz - :x: **elliptic-6.5.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/turkdevops/grafana/commit/e866e218913576b37ba96fa7aae4d6cbcb61df30">e866e218913576b37ba96fa7aae4d6cbcb61df30</a></p> <p>Found in base branch: <b>datasource-meta</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The function getNAF() in elliptic library has information leakage. This issue is mitigated in version 6.5.2 <p>Publish Date: 2019-11-22 <p>URL: <a href=https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a>WS-2019-0427</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a">https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a</a></p> <p>Release Date: 2020-05-24</p> <p>Fix Resolution: v6.5.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2019-0427 (Medium) detected in elliptic-6.5.1.tgz - ## WS-2019-0427 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.5.1.tgz</b></p></summary> <p>EC cryptography</p> <p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.5.1.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.5.1.tgz</a></p> <p>Path to dependency file: grafana/node_modules/elliptic/package.json</p> <p>Path to vulnerable library: grafana/node_modules/elliptic/package.json</p> <p> Dependency Hierarchy: - webpack-4.29.6.tgz (Root Library) - node-libs-browser-2.2.1.tgz - crypto-browserify-3.12.0.tgz - create-ecdh-4.0.3.tgz - :x: **elliptic-6.5.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/turkdevops/grafana/commit/e866e218913576b37ba96fa7aae4d6cbcb61df30">e866e218913576b37ba96fa7aae4d6cbcb61df30</a></p> <p>Found in base branch: <b>datasource-meta</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The function getNAF() in elliptic library has information leakage. This issue is mitigated in version 6.5.2 <p>Publish Date: 2019-11-22 <p>URL: <a href=https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a>WS-2019-0427</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a">https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a</a></p> <p>Release Date: 2020-05-24</p> <p>Fix Resolution: v6.5.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
ws medium detected in elliptic tgz ws medium severity vulnerability vulnerable library elliptic tgz ec cryptography library home page a href path to dependency file grafana node modules elliptic package json path to vulnerable library grafana node modules elliptic package json dependency hierarchy webpack tgz root library node libs browser tgz crypto browserify tgz create ecdh tgz x elliptic tgz vulnerable library found in head commit a href found in base branch datasource meta vulnerability details the function getnaf in elliptic library has information leakage this issue is mitigated in version publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
319,442
27,373,692,977
IssuesEvent
2023-02-28 02:57:01
void-linux/void-packages
https://api.github.com/repos/void-linux/void-packages
opened
vscode doesn't apply non-English Display Language
bug needs-testing
### Is this a new report? Yes ### System Info Void newest ### Package(s) Affected vscode-1.75.1_1 ### Does a report exist for this bug with the project's home (upstream) and/or another distro? https://github.com/microsoft/vscode/issues/154417 ### Expected behaviour could apply display language other than english ### Actual behaviour could NOT apply display language other than english ### Steps to reproduce open vscode install language extention such as Chinese Display Langue. Press `ctrl + shift + p` and choose to configure the display language to zh-cn. reboot vscode. Display language was NOT changed.
1.0
vscode doesn't apply non-English Display Language - ### Is this a new report? Yes ### System Info Void newest ### Package(s) Affected vscode-1.75.1_1 ### Does a report exist for this bug with the project's home (upstream) and/or another distro? https://github.com/microsoft/vscode/issues/154417 ### Expected behaviour could apply display language other than english ### Actual behaviour could NOT apply display language other than english ### Steps to reproduce open vscode install language extention such as Chinese Display Langue. Press `ctrl + shift + p` and choose to configure the display language to zh-cn. reboot vscode. Display language was NOT changed.
test
vscode doesn t apply non english display language is this a new report yes system info void newest package s affected vscode does a report exist for this bug with the project s home upstream and or another distro expected behaviour could apply display language other than english actual behaviour could not apply display language other than english steps to reproduce open vscode install language extention such as chinese display langue press ctrl shift p and choose to configure the display language to zh cn reboot vscode display language was not changed
1
1,262
5,353,855,504
IssuesEvent
2017-02-20 07:54:14
espeak-ng/espeak-ng
https://api.github.com/repos/espeak-ng/espeak-ng
closed
Merge the android branch into master.
maintainability portability resolved/fixed
Now that espeak-ng has diverged from espeak, it makes sense to have the android branch merged into the main development line. This will make it easier to maintain the Android support in the future and keep it up-to-date. - [x] Merge the android code into the master branch. - [x] Fix building the JNI and libespeak-ng code for Android. - [x] Incorporate the Android README text into the README file. - [x] Add ucd-tools as a git subtree. - [x] Use ucd-tools to create a compatibility shim for the wide-character APIs.
True
Merge the android branch into master. - Now that espeak-ng has diverged from espeak, it makes sense to have the android branch merged into the main development line. This will make it easier to maintain the Android support in the future and keep it up-to-date. - [x] Merge the android code into the master branch. - [x] Fix building the JNI and libespeak-ng code for Android. - [x] Incorporate the Android README text into the README file. - [x] Add ucd-tools as a git subtree. - [x] Use ucd-tools to create a compatibility shim for the wide-character APIs.
non_test
merge the android branch into master now that espeak ng has diverged from espeak it makes sense to have the android branch merged into the main development line this will make it easier to maintain the android support in the future and keep it up to date merge the android code into the master branch fix building the jni and libespeak ng code for android incorporate the android readme text into the readme file add ucd tools as a git subtree use ucd tools to create a compatibility shim for the wide character apis
0
28,958
11,706,038,472
IssuesEvent
2020-03-07 19:32:11
vlaship/spark-streaming
https://api.github.com/repos/vlaship/spark-streaming
opened
CVE-2019-14892 (Medium) detected in jackson-databind-2.6.5.jar
security vulnerability
## CVE-2019-14892 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/spark-streaming/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.5/d50be1723a09befd903887099ff2014ea9020333/jackson-databind-2.6.5.jar,/root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.5/d50be1723a09befd903887099ff2014ea9020333/jackson-databind-2.6.5.jar</p> <p> Dependency Hierarchy: - spark-streaming_2.11-2.2.0.jar (Root Library) - spark-core_2.11-2.2.0.jar - :x: **jackson-databind-2.6.5.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vlaship/spark-streaming/commit/8193fe8b5a8089e67347f5851a02c41ea5056fb0">8193fe8b5a8089e67347f5851a02c41ea5056fb0</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was discovered in jackson-databind in versions before 2.9.10, 2.8.11.5 and 2.6.7.3, where it would permit polymorphic deserialization of a malicious object using commons-configuration 1 and 2 JNDI classes. An attacker could use this flaw to execute arbitrary code. <p>Publish Date: 2020-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14892>CVE-2019-14892</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14892">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14892</a></p> <p>Release Date: 2020-03-02</p> <p>Fix Resolution: com.fasterxml.jackson:jackson-base:jackson-databind-2.8.11.5,com.fasterxml.jackson:jackson-base:jackson-databind-2.9.10,com.fasterxml.jackson:jackson-base:jackson-databind-2.10.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-14892 (Medium) detected in jackson-databind-2.6.5.jar - ## CVE-2019-14892 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/spark-streaming/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.5/d50be1723a09befd903887099ff2014ea9020333/jackson-databind-2.6.5.jar,/root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.5/d50be1723a09befd903887099ff2014ea9020333/jackson-databind-2.6.5.jar</p> <p> Dependency Hierarchy: - spark-streaming_2.11-2.2.0.jar (Root Library) - spark-core_2.11-2.2.0.jar - :x: **jackson-databind-2.6.5.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vlaship/spark-streaming/commit/8193fe8b5a8089e67347f5851a02c41ea5056fb0">8193fe8b5a8089e67347f5851a02c41ea5056fb0</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was discovered in jackson-databind in versions before 2.9.10, 2.8.11.5 and 2.6.7.3, where it would permit polymorphic deserialization of a malicious object using commons-configuration 1 and 2 JNDI classes. An attacker could use this flaw to execute arbitrary code. <p>Publish Date: 2020-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14892>CVE-2019-14892</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14892">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14892</a></p> <p>Release Date: 2020-03-02</p> <p>Fix Resolution: com.fasterxml.jackson:jackson-base:jackson-databind-2.8.11.5,com.fasterxml.jackson:jackson-base:jackson-databind-2.9.10,com.fasterxml.jackson:jackson-base:jackson-databind-2.10.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in jackson databind jar cve medium severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws scm spark streaming build gradle path to vulnerable library root gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar root gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spark streaming jar root library spark core jar x jackson databind jar vulnerable library found in head commit a href vulnerability details a flaw was discovered in jackson databind in versions before and where it would permit polymorphic deserialization of a malicious object using commons configuration and jndi classes an attacker could use this flaw to execute arbitrary code publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson jackson base jackson databind com fasterxml jackson jackson base jackson databind com fasterxml jackson jackson base jackson databind step up your open source security game with whitesource
0
136,307
30,519,700,595
IssuesEvent
2023-07-19 07:11:05
h4sh5/pypi-auto-scanner
https://api.github.com/repos/h4sh5/pypi-auto-scanner
opened
apache-tvm 0.14.dev54 has 9 GuardDog issues
guarddog code-execution exec-base64
https://pypi.org/project/apache-tvm https://inspector.pypi.io/project/apache-tvm ```{ "dependency": "apache-tvm", "version": "0.14.dev54", "result": { "issues": 9, "errors": {}, "results": { "exec-base64": [ { "location": "tvm/3rdparty/dmlc-core/tracker/dmlc_tracker/launcher.py/tvm/3rdparty/dmlc-core/tracker/dmlc_tracker/launcher.py:80", "code": " ret = subprocess.call(args=sys.argv[1:], env=env)", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" }, { "location": "tvm/3rdparty/dmlc-core/tracker/dmlc_tracker/local.py/tvm/3rdparty/dmlc-core/tracker/dmlc_tracker/local.py:33", "code": " ret = subprocess.call(cmdline, shell=True, env=env)", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" }, { "location": "tvm/3rdparty/dmlc-core/tracker/dmlc_tracker/local.py/tvm/3rdparty/dmlc-core/tracker/dmlc_tracker/local.py:35", "code": " ret = subprocess.call(cmdline, shell=True, executable='bash', env=env)", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" }, { "location": "tvm/contrib/nvcc.py/tvm/contrib/nvcc.py:105", "code": " proc = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" }, { "location": "tvm/contrib/xcode.py/tvm/contrib/xcode.py:154", "code": " proc = subprocess.Popen(\n \" \".join(cmd1) + \";\" + \" \".join(cmd2),\n shell=True,\n stdout=subprocess.PIPE,\n stderr=subprocess.STDOUT,\n )", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" }, { "location": "tvm/rust/tvm/examples/resnet/src/build_resnet.py/tvm/rust/tvm/examples/resnet/src/build_resnet.py:111", "code": " synset = eval(data)", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" }, { "location": "tvm/tir/schedule/testing.py/tvm/tir/schedule/testing.py:73", "code": " exec(py_trace, vars_dict, {\"sch\": new_sch}) # pylint: disable=exec-used", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" } ], "code-execution": [ { "location": "tvm/3rdparty/cutlass/tools/library/scripts/pycutlass/setup.py/tvm/3rdparty/cutlass/tools/library/scripts/pycutlass/setup.py:52", "code": " os.system(\"./build.sh librmm rmm\")", "message": "This package is executing OS commands in the setup.py file" }, { "location": "tvm/3rdparty/cutlass_fpA_intB_gemm/cutlass/tools/library/scripts/pycutlass/setup.py/tvm/3rdparty/cutlass_fpA_intB_gemm/cutlass/tools/library/scripts/pycutlass/setup.py:20", "code": " os.system(\"./build.sh librmm rmm\")", "message": "This package is executing OS commands in the setup.py file" } ] }, "path": "/tmp/tmpqsbnzic2/apache-tvm" } }```
1.0
apache-tvm 0.14.dev54 has 9 GuardDog issues - https://pypi.org/project/apache-tvm https://inspector.pypi.io/project/apache-tvm ```{ "dependency": "apache-tvm", "version": "0.14.dev54", "result": { "issues": 9, "errors": {}, "results": { "exec-base64": [ { "location": "tvm/3rdparty/dmlc-core/tracker/dmlc_tracker/launcher.py/tvm/3rdparty/dmlc-core/tracker/dmlc_tracker/launcher.py:80", "code": " ret = subprocess.call(args=sys.argv[1:], env=env)", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" }, { "location": "tvm/3rdparty/dmlc-core/tracker/dmlc_tracker/local.py/tvm/3rdparty/dmlc-core/tracker/dmlc_tracker/local.py:33", "code": " ret = subprocess.call(cmdline, shell=True, env=env)", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" }, { "location": "tvm/3rdparty/dmlc-core/tracker/dmlc_tracker/local.py/tvm/3rdparty/dmlc-core/tracker/dmlc_tracker/local.py:35", "code": " ret = subprocess.call(cmdline, shell=True, executable='bash', env=env)", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" }, { "location": "tvm/contrib/nvcc.py/tvm/contrib/nvcc.py:105", "code": " proc = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" }, { "location": "tvm/contrib/xcode.py/tvm/contrib/xcode.py:154", "code": " proc = subprocess.Popen(\n \" \".join(cmd1) + \";\" + \" \".join(cmd2),\n shell=True,\n stdout=subprocess.PIPE,\n stderr=subprocess.STDOUT,\n )", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" }, { "location": "tvm/rust/tvm/examples/resnet/src/build_resnet.py/tvm/rust/tvm/examples/resnet/src/build_resnet.py:111", "code": " synset = eval(data)", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" }, { "location": "tvm/tir/schedule/testing.py/tvm/tir/schedule/testing.py:73", "code": " exec(py_trace, vars_dict, {\"sch\": new_sch}) # pylint: disable=exec-used", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" } ], "code-execution": [ { "location": "tvm/3rdparty/cutlass/tools/library/scripts/pycutlass/setup.py/tvm/3rdparty/cutlass/tools/library/scripts/pycutlass/setup.py:52", "code": " os.system(\"./build.sh librmm rmm\")", "message": "This package is executing OS commands in the setup.py file" }, { "location": "tvm/3rdparty/cutlass_fpA_intB_gemm/cutlass/tools/library/scripts/pycutlass/setup.py/tvm/3rdparty/cutlass_fpA_intB_gemm/cutlass/tools/library/scripts/pycutlass/setup.py:20", "code": " os.system(\"./build.sh librmm rmm\")", "message": "This package is executing OS commands in the setup.py file" } ] }, "path": "/tmp/tmpqsbnzic2/apache-tvm" } }```
non_test
apache tvm has guarddog issues dependency apache tvm version result issues errors results exec location tvm dmlc core tracker dmlc tracker launcher py tvm dmlc core tracker dmlc tracker launcher py code ret subprocess call args sys argv env env message this package contains a call to the eval function with a encoded string as argument nthis is a common method used to hide a malicious payload in a module as static analysis will not decode the nstring n location tvm dmlc core tracker dmlc tracker local py tvm dmlc core tracker dmlc tracker local py code ret subprocess call cmdline shell true env env message this package contains a call to the eval function with a encoded string as argument nthis is a common method used to hide a malicious payload in a module as static analysis will not decode the nstring n location tvm dmlc core tracker dmlc tracker local py tvm dmlc core tracker dmlc tracker local py code ret subprocess call cmdline shell true executable bash env env message this package contains a call to the eval function with a encoded string as argument nthis is a common method used to hide a malicious payload in a module as static analysis will not decode the nstring n location tvm contrib nvcc py tvm contrib nvcc py code proc subprocess popen cmd stdout subprocess pipe stderr subprocess stdout message this package contains a call to the eval function with a encoded string as argument nthis is a common method used to hide a malicious payload in a module as static analysis will not decode the nstring n location tvm contrib xcode py tvm contrib xcode py code proc subprocess popen n join join n shell true n stdout subprocess pipe n stderr subprocess stdout n message this package contains a call to the eval function with a encoded string as argument nthis is a common method used to hide a malicious payload in a module as static analysis will not decode the nstring n location tvm rust tvm examples resnet src build resnet py tvm rust tvm examples resnet src build resnet py code synset eval data message this package contains a call to the eval function with a encoded string as argument nthis is a common method used to hide a malicious payload in a module as static analysis will not decode the nstring n location tvm tir schedule testing py tvm tir schedule testing py code exec py trace vars dict sch new sch pylint disable exec used message this package contains a call to the eval function with a encoded string as argument nthis is a common method used to hide a malicious payload in a module as static analysis will not decode the nstring n code execution location tvm cutlass tools library scripts pycutlass setup py tvm cutlass tools library scripts pycutlass setup py code os system build sh librmm rmm message this package is executing os commands in the setup py file location tvm cutlass fpa intb gemm cutlass tools library scripts pycutlass setup py tvm cutlass fpa intb gemm cutlass tools library scripts pycutlass setup py code os system build sh librmm rmm message this package is executing os commands in the setup py file path tmp apache tvm
0
105,654
23,088,899,758
IssuesEvent
2022-07-26 13:45:18
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
closed
worker: potentially use AWS's default credential provider instead of static
estimate/1d team/batchers user-code-execution
Use AWS default credential chain provider rather than just the [current static provider](https://github.com/sourcegraph/sourcegraph/blob/ce2005995a8126960451c125605605dd54696f49/enterprise/cmd/worker/internal/executorqueue/aws_reporter.go#L96). In https://github.com/sourcegraph/accounts/issues/565 's case, it could eliminate the need to inject AWS keys via env vars.
1.0
worker: potentially use AWS's default credential provider instead of static - Use AWS default credential chain provider rather than just the [current static provider](https://github.com/sourcegraph/sourcegraph/blob/ce2005995a8126960451c125605605dd54696f49/enterprise/cmd/worker/internal/executorqueue/aws_reporter.go#L96). In https://github.com/sourcegraph/accounts/issues/565 's case, it could eliminate the need to inject AWS keys via env vars.
non_test
worker potentially use aws s default credential provider instead of static use aws default credential chain provider rather than just the in s case it could eliminate the need to inject aws keys via env vars
0
198,389
14,977,593,353
IssuesEvent
2021-01-28 09:42:45
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
At least Java 11 is required to build elasticsearch gradle tools
>test-failure
<!-- Please fill out the following information, and ensure you have attempted to reproduce locally --> **Build scan**: **Repro line**: Build file 'elasticsearch/buildSrc/build.gradle' line: 61 **Reproduces locally?**: **Applicable branches**: master **Failure history**: <!-- Link to build stats and possible indication of when this started failing and how often it fails <https://build-stats.elastic.co/app/kibana> --> **Failure excerpt**: What went wrong: A problem occurred evaluating project ':buildSrc'. At least Java 11 is required to build elasticsearch gradle tools idea:2020.1.2 i want to build elasticsearch to idea,and set JDK to 15,however,this error occurred。 thank you very much
1.0
At least Java 11 is required to build elasticsearch gradle tools - <!-- Please fill out the following information, and ensure you have attempted to reproduce locally --> **Build scan**: **Repro line**: Build file 'elasticsearch/buildSrc/build.gradle' line: 61 **Reproduces locally?**: **Applicable branches**: master **Failure history**: <!-- Link to build stats and possible indication of when this started failing and how often it fails <https://build-stats.elastic.co/app/kibana> --> **Failure excerpt**: What went wrong: A problem occurred evaluating project ':buildSrc'. At least Java 11 is required to build elasticsearch gradle tools idea:2020.1.2 i want to build elasticsearch to idea,and set JDK to 15,however,this error occurred。 thank you very much
test
at least java is required to build elasticsearch gradle tools please fill out the following information and ensure you have attempted to reproduce locally build scan repro line build file elasticsearch buildsrc build gradle line reproduces locally applicable branches master failure history link to build stats and possible indication of when this started failing and how often it fails failure excerpt what went wrong a problem occurred evaluating project buildsrc at least java is required to build elasticsearch gradle tools idea: i want to build elasticsearch to idea and set jdk to however this error occurred。 thank you very much
1
66,218
8,895,637,612
IssuesEvent
2019-01-16 09:16:13
nbs-system/naxsi
https://api.github.com/repos/nbs-system/naxsi
opened
Make the testsuite run on various linux flavors
documentation good first issue help wanted testsuite
As suggested by @jvoisin here, we should make sure our test suite runs on various distributions : https://github.com/nbs-system/snuffleupagus/commit/df84a0fc60eb7d937b27a4ac540cb637a55972bf
1.0
Make the testsuite run on various linux flavors - As suggested by @jvoisin here, we should make sure our test suite runs on various distributions : https://github.com/nbs-system/snuffleupagus/commit/df84a0fc60eb7d937b27a4ac540cb637a55972bf
non_test
make the testsuite run on various linux flavors as suggested by jvoisin here we should make sure our test suite runs on various distributions
0
338,910
30,328,218,205
IssuesEvent
2023-07-11 03:01:13
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Fix math.test_tensorflow_scalar_mul
TensorFlow Frontend Sub Task Failing Test
| | | |---|---| |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5515219518/jobs/10055314850"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5515219518/jobs/10055314850"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5498239750/jobs/10019558175"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5498239750/jobs/10019558175"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5498239750/jobs/10019558175"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix math.test_tensorflow_scalar_mul - | | | |---|---| |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5515219518/jobs/10055314850"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5515219518/jobs/10055314850"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5498239750/jobs/10019558175"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5498239750/jobs/10019558175"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5498239750/jobs/10019558175"><img src=https://img.shields.io/badge/-success-success></a>
test
fix math test tensorflow scalar mul jax a href src numpy a href src tensorflow a href src torch a href src paddle a href src
1
251,202
21,445,266,654
IssuesEvent
2022-04-25 05:14:43
mautic/mautic
https://api.github.com/repos/mautic/mautic
closed
Multiple areas of Mautic throw a 500 error
T1 bug ready-to-test regression reports essential dynamic-content email
## Bug Description I came across this issue when running the Cypress DWC tests and found it to also be reported in Emails (see #9646 #9755 and /https://forum.mautic.org/t/notices-trying-to-access-array-offset-on-value-of-type-null/18008). I found it by trying to open a DWC item that was created but others find it with using certain email templates (with the legacy builder) as listed in the forum thread. | Q | A | --- | --- | Mautic version | 3.3.1 | PHP version | 7.4 | Browser | Chrome ### Steps to reproduce 1. Create dynamic content 2. Click to edit dynamic content 3. Notice a 503 error and log entry below ### Log errors `mautic.NOTICE: PHP Notice - Trying to access array offset on value of type null - in file /var/www/html/app/bundles/EmailBundle/Views/FormTheme/Email/_emailform_dynamicContent_entry_filters_entry_filters_entry_widget.html.php - at line 14 {"mauticTemplateVars":{"app":"[object] (Symfony\\Bundle\\FrameworkBundle\\Templating\\GlobalVariables: {})","value":null,"attr":{"class":"form-control"},"form":"[object] (Symfony\\Component\\Form\\FormView: null)","id":"emailform_dynamicContent_0_filters_0_filters___name__","name":"__name__","full_name":"emailform[dynamicContent][0][filters][0][filters][__name__]","disabled":false,"label":false,"label_format":null,"multipart":false,"block_prefixes":["form","dynamic_content_filter_entry_filters","_emailform_dynamicContent_entry_filters_entry_filters_entry"],"unique_block_prefix":"_emailform_dynamicContent_entry_filters_entry_filters_entry","translation_domain":null,"cache_key":"_emailform_dynamicContent_entry_filters_entry_filters_entry_dynamic_content_filter_entry_filters","errors":"[object] (Symfony\\Component\\Form\\FormErrorIterator: )","valid":true,"data":null,"required":true,"size":null,"label_attr":[],"compound":true,"method":"POST","action":"","submitted":false,"mauticContent":"email","mauticTemplate":"MauticEmailBundle:FormTheme\\Email:_emailform_dynamicContent_entry_filters_entry_filters_entry_widget.html.php","view":"[object] (Mautic\\CoreBundle\\Templating\\Engine\\PhpEngine: {})"},"app":"[object] (Symfony\\Bundle\\FrameworkBundle\\Templating\\GlobalVariables: {})","value":null,"attr":{"class":"form-control"},"form":"[object] (Symfony\\Component\\Form\\FormView: null)","id":"emailform_dynamicContent_0_filters_0_filters___name__","name":"__name__","full_name":"emailform[dynamicContent][0][filters][0][filters][__name__]","disabled":false,"label":false,"label_format":null,"multipart":false,"block_prefixes":["form","dynamic_content_filter_entry_filters","_emailform_dynamicContent_entry_filters_entry_filters_entry"],"unique_block_prefix":"_emailform_dynamicContent_entry_filters_entry_filters_entry","translation_domain":null,"cache_key":"_emailform_dynamicContent_entry_filters_entry_filters_entry_dynamic_content_filter_entry_filters","errors":"[object] (Symfony\\Component\\Form\\FormErrorIterator: )","valid":true,"data":null,"required":true,"size":null,"label_attr":[],"compound":true,"method":"POST","action":"","submitted":false,"mauticContent":"email","mauticTemplate":"MauticEmailBundle:FormTheme\\Email:_emailform_dynamicContent_entry_filters_entry_filters_entry_widget.html.php","view":"[object] (Mautic\\CoreBundle\\Templating\\Engine\\PhpEngine: {})","isPrototype":true,"filterType":"","filterGroup":""} [] ` <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/96997535-multiple-areas-of-mautic-throw-a-500-error?utm_campaign=plugin&utm_content=tracker%2F5355074&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F5355074&utm_medium=issues&utm_source=github). </bountysource-plugin>
1.0
Multiple areas of Mautic throw a 500 error - ## Bug Description I came across this issue when running the Cypress DWC tests and found it to also be reported in Emails (see #9646 #9755 and /https://forum.mautic.org/t/notices-trying-to-access-array-offset-on-value-of-type-null/18008). I found it by trying to open a DWC item that was created but others find it with using certain email templates (with the legacy builder) as listed in the forum thread. | Q | A | --- | --- | Mautic version | 3.3.1 | PHP version | 7.4 | Browser | Chrome ### Steps to reproduce 1. Create dynamic content 2. Click to edit dynamic content 3. Notice a 503 error and log entry below ### Log errors `mautic.NOTICE: PHP Notice - Trying to access array offset on value of type null - in file /var/www/html/app/bundles/EmailBundle/Views/FormTheme/Email/_emailform_dynamicContent_entry_filters_entry_filters_entry_widget.html.php - at line 14 {"mauticTemplateVars":{"app":"[object] (Symfony\\Bundle\\FrameworkBundle\\Templating\\GlobalVariables: {})","value":null,"attr":{"class":"form-control"},"form":"[object] (Symfony\\Component\\Form\\FormView: null)","id":"emailform_dynamicContent_0_filters_0_filters___name__","name":"__name__","full_name":"emailform[dynamicContent][0][filters][0][filters][__name__]","disabled":false,"label":false,"label_format":null,"multipart":false,"block_prefixes":["form","dynamic_content_filter_entry_filters","_emailform_dynamicContent_entry_filters_entry_filters_entry"],"unique_block_prefix":"_emailform_dynamicContent_entry_filters_entry_filters_entry","translation_domain":null,"cache_key":"_emailform_dynamicContent_entry_filters_entry_filters_entry_dynamic_content_filter_entry_filters","errors":"[object] (Symfony\\Component\\Form\\FormErrorIterator: )","valid":true,"data":null,"required":true,"size":null,"label_attr":[],"compound":true,"method":"POST","action":"","submitted":false,"mauticContent":"email","mauticTemplate":"MauticEmailBundle:FormTheme\\Email:_emailform_dynamicContent_entry_filters_entry_filters_entry_widget.html.php","view":"[object] (Mautic\\CoreBundle\\Templating\\Engine\\PhpEngine: {})"},"app":"[object] (Symfony\\Bundle\\FrameworkBundle\\Templating\\GlobalVariables: {})","value":null,"attr":{"class":"form-control"},"form":"[object] (Symfony\\Component\\Form\\FormView: null)","id":"emailform_dynamicContent_0_filters_0_filters___name__","name":"__name__","full_name":"emailform[dynamicContent][0][filters][0][filters][__name__]","disabled":false,"label":false,"label_format":null,"multipart":false,"block_prefixes":["form","dynamic_content_filter_entry_filters","_emailform_dynamicContent_entry_filters_entry_filters_entry"],"unique_block_prefix":"_emailform_dynamicContent_entry_filters_entry_filters_entry","translation_domain":null,"cache_key":"_emailform_dynamicContent_entry_filters_entry_filters_entry_dynamic_content_filter_entry_filters","errors":"[object] (Symfony\\Component\\Form\\FormErrorIterator: )","valid":true,"data":null,"required":true,"size":null,"label_attr":[],"compound":true,"method":"POST","action":"","submitted":false,"mauticContent":"email","mauticTemplate":"MauticEmailBundle:FormTheme\\Email:_emailform_dynamicContent_entry_filters_entry_filters_entry_widget.html.php","view":"[object] (Mautic\\CoreBundle\\Templating\\Engine\\PhpEngine: {})","isPrototype":true,"filterType":"","filterGroup":""} [] ` <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/96997535-multiple-areas-of-mautic-throw-a-500-error?utm_campaign=plugin&utm_content=tracker%2F5355074&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F5355074&utm_medium=issues&utm_source=github). </bountysource-plugin>
test
multiple areas of mautic throw a error bug description i came across this issue when running the cypress dwc tests and found it to also be reported in emails see and i found it by trying to open a dwc item that was created but others find it with using certain email templates with the legacy builder as listed in the forum thread q a mautic version php version browser chrome steps to reproduce create dynamic content click to edit dynamic content notice a error and log entry below log errors mautic notice php notice trying to access array offset on value of type null in file var www html app bundles emailbundle views formtheme email emailform dynamiccontent entry filters entry filters entry widget html php at line mautictemplatevars app symfony bundle frameworkbundle templating globalvariables value null attr class form control form symfony component form formview null id emailform dynamiccontent filters filters name name name full name emailform disabled false label false label format null multipart false block prefixes unique block prefix emailform dynamiccontent entry filters entry filters entry translation domain null cache key emailform dynamiccontent entry filters entry filters entry dynamic content filter entry filters errors symfony component form formerroriterator valid true data null required true size null label attr compound true method post action submitted false mauticcontent email mautictemplate mauticemailbundle formtheme email emailform dynamiccontent entry filters entry filters entry widget html php view mautic corebundle templating engine phpengine app symfony bundle frameworkbundle templating globalvariables value null attr class form control form symfony component form formview null id emailform dynamiccontent filters filters name name name full name emailform disabled false label false label format null multipart false block prefixes unique block prefix emailform dynamiccontent entry filters entry filters entry translation domain null cache key emailform dynamiccontent entry filters entry filters entry dynamic content filter entry filters errors symfony component form formerroriterator valid true data null required true size null label attr compound true method post action submitted false mauticcontent email mautictemplate mauticemailbundle formtheme email emailform dynamiccontent entry filters entry filters entry widget html php view mautic corebundle templating engine phpengine isprototype true filtertype filtergroup want to back this issue we accept bounties via
1
337,513
30,248,909,276
IssuesEvent
2023-07-06 18:47:29
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
opened
Fix jax_numpy_manipulation.test_jax_moveaxis
JAX Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5477575940"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5473974053"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5473078102"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix jax_numpy_manipulation.test_jax_moveaxis - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5477575940"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5473974053"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5473078102"><img src=https://img.shields.io/badge/-success-success></a>
test
fix jax numpy manipulation test jax moveaxis tensorflow a href src jax a href src numpy a href src
1
16,106
3,507,538,990
IssuesEvent
2016-01-08 13:54:41
exc-asia-and-europe/ziziphus
https://api.github.com/repos/exc-asia-and-europe/ziziphus
closed
Parsing names from VIAF
AgentSet Edit Mode in progress LocationSet To be tested ... View Mode
Some institutions seem to have two separate tags for their names. This seems to be the case when they're national institutions, in the current incident it is the national archives. When searching for "national archives" the results list in Ziziphus only offers nation names and you have to hover over the i-button to get the full information. ![image](https://cloud.githubusercontent.com/assets/3437957/11422247/bf3b62ba-943a-11e5-8f73-9890a4ea5c20.png) For the "United States. National Archives and Records Administration" [VIAF-ID: #265208613] the xml looks like this: ![image](https://cloud.githubusercontent.com/assets/3437957/11422278/f38faaf8-943a-11e5-80f0-6e5d82e3de14.png) So there are two separate tags - one for the nation and one for the institution's name. Is it in any way possible to get both these tags shown in the results list in Ziziphus?
1.0
Parsing names from VIAF - Some institutions seem to have two separate tags for their names. This seems to be the case when they're national institutions, in the current incident it is the national archives. When searching for "national archives" the results list in Ziziphus only offers nation names and you have to hover over the i-button to get the full information. ![image](https://cloud.githubusercontent.com/assets/3437957/11422247/bf3b62ba-943a-11e5-8f73-9890a4ea5c20.png) For the "United States. National Archives and Records Administration" [VIAF-ID: #265208613] the xml looks like this: ![image](https://cloud.githubusercontent.com/assets/3437957/11422278/f38faaf8-943a-11e5-80f0-6e5d82e3de14.png) So there are two separate tags - one for the nation and one for the institution's name. Is it in any way possible to get both these tags shown in the results list in Ziziphus?
test
parsing names from viaf some institutions seem to have two separate tags for their names this seems to be the case when they re national institutions in the current incident it is the national archives when searching for national archives the results list in ziziphus only offers nation names and you have to hover over the i button to get the full information for the united states national archives and records administration the xml looks like this so there are two separate tags one for the nation and one for the institution s name is it in any way possible to get both these tags shown in the results list in ziziphus
1
82,873
3,619,695,692
IssuesEvent
2016-02-08 16:57:54
miracle091/transmission-remote-dotnet
https://api.github.com/repos/miracle091/transmission-remote-dotnet
closed
Add more trackers to existing torrent
Priority-Medium Type-Enhancement
``` Would be very good to have the possibility to add more trackers to an existing torrent. ``` Original issue reported on code.google.com by `djpi...@gmail.com` on 26 Feb 2012 at 11:09
1.0
Add more trackers to existing torrent - ``` Would be very good to have the possibility to add more trackers to an existing torrent. ``` Original issue reported on code.google.com by `djpi...@gmail.com` on 26 Feb 2012 at 11:09
non_test
add more trackers to existing torrent would be very good to have the possibility to add more trackers to an existing torrent original issue reported on code google com by djpi gmail com on feb at
0
712,721
24,505,416,958
IssuesEvent
2022-10-10 15:56:14
oceanprotocol/df-py
https://api.github.com/repos/oceanprotocol/df-py
opened
Reduce claim txs down to 1
Priority: Low
### Problem - Users need to complete multiple txs/claims to complete their rewards - Low stake users are frustrated because they are paying larger fees relative to how much they have invested - If we add any other reward streams in the future, the number of claims grows - It's impossible for us to reconcile the claim from FeeDistributor.vy w/ DFRewards.sol, at a minimum we'll have 2 claim tx's ### Candidate Solution - We still have core functionality that still need to be delivered. This is an optimization that improves the UX for everyone, however, especially the low-stake users. - Build a proxy contract that reduces all claims into a single TX, across all contracts ### DoD: - [ ] Front end ticket has been created when this issue starts - [ ] Proxy contract has been deployed - [ ] Front end has been updated
1.0
Reduce claim txs down to 1 - ### Problem - Users need to complete multiple txs/claims to complete their rewards - Low stake users are frustrated because they are paying larger fees relative to how much they have invested - If we add any other reward streams in the future, the number of claims grows - It's impossible for us to reconcile the claim from FeeDistributor.vy w/ DFRewards.sol, at a minimum we'll have 2 claim tx's ### Candidate Solution - We still have core functionality that still need to be delivered. This is an optimization that improves the UX for everyone, however, especially the low-stake users. - Build a proxy contract that reduces all claims into a single TX, across all contracts ### DoD: - [ ] Front end ticket has been created when this issue starts - [ ] Proxy contract has been deployed - [ ] Front end has been updated
non_test
reduce claim txs down to problem users need to complete multiple txs claims to complete their rewards low stake users are frustrated because they are paying larger fees relative to how much they have invested if we add any other reward streams in the future the number of claims grows it s impossible for us to reconcile the claim from feedistributor vy w dfrewards sol at a minimum we ll have claim tx s candidate solution we still have core functionality that still need to be delivered this is an optimization that improves the ux for everyone however especially the low stake users build a proxy contract that reduces all claims into a single tx across all contracts dod front end ticket has been created when this issue starts proxy contract has been deployed front end has been updated
0
39,628
10,368,075,926
IssuesEvent
2019-09-07 14:04:46
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
[CLI] Usage of double quotes in unknown command message
Area/BuildTools Priority/Blocker Type/Bug
**Description:** <!-- Give a brief description of the issue --> $title It would be better if we use single quotes here <img width="705" alt="Screen Shot 2019-09-03 at 1 58 18 PM" src="https://user-images.githubusercontent.com/24229626/64157305-1cf67780-ce54-11e9-8ba5-b5258d4d95b9.png"> **Steps to reproduce:** `➜ ballerina abc` **Affected Versions:** Ballerina 1.0.0-beta Language specification 2019R3 Ballerina tool 1.0.0 **OS, DB, other environment details and versions:** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
1.0
[CLI] Usage of double quotes in unknown command message - **Description:** <!-- Give a brief description of the issue --> $title It would be better if we use single quotes here <img width="705" alt="Screen Shot 2019-09-03 at 1 58 18 PM" src="https://user-images.githubusercontent.com/24229626/64157305-1cf67780-ce54-11e9-8ba5-b5258d4d95b9.png"> **Steps to reproduce:** `➜ ballerina abc` **Affected Versions:** Ballerina 1.0.0-beta Language specification 2019R3 Ballerina tool 1.0.0 **OS, DB, other environment details and versions:** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
non_test
usage of double quotes in unknown command message description title it would be better if we use single quotes here img width alt screen shot at pm src steps to reproduce ➜ ballerina abc affected versions ballerina beta language specification ballerina tool os db other environment details and versions related issues optional suggested labels optional suggested assignees optional
0
71,000
18,373,481,611
IssuesEvent
2021-10-11 04:53:54
rust-lang/rust
https://api.github.com/repos/rust-lang/rust
closed
Build failed due to Python27\lib\subprocess.py
O-windows A-rustbuild T-infra
I am trying to build Rust from source on Windows, but meet an error: ![image](https://user-images.githubusercontent.com/43683886/51110498-db1cb080-1833-11e9-935d-b28687f56dd0.png) Who could help me to solve this? Thanks.
1.0
Build failed due to Python27\lib\subprocess.py - I am trying to build Rust from source on Windows, but meet an error: ![image](https://user-images.githubusercontent.com/43683886/51110498-db1cb080-1833-11e9-935d-b28687f56dd0.png) Who could help me to solve this? Thanks.
non_test
build failed due to lib subprocess py i am trying to build rust from source on windows but meet an error who could help me to solve this thanks
0
187,177
14,427,197,677
IssuesEvent
2020-12-06 02:30:40
kalexmills/github-vet-tests-dec2020
https://api.github.com/repos/kalexmills/github-vet-tests-dec2020
closed
soggiest/ferrarin: vendor/k8s.io/kubernetes/plugin/pkg/scheduler/algorithm/predicates/predicates_test.go; 45 LoC
fresh small test vendored
Found a possible issue in [soggiest/ferrarin](https://www.github.com/soggiest/ferrarin) at [vendor/k8s.io/kubernetes/plugin/pkg/scheduler/algorithm/predicates/predicates_test.go](https://github.com/soggiest/ferrarin/blob/ece6d947b1d107f2d7690debb647466465bb51d3/vendor/k8s.io/kubernetes/plugin/pkg/scheduler/algorithm/predicates/predicates_test.go#L2702-L2746) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to node at line 2716 may start a goroutine [Click here to see the code in its original context.](https://github.com/soggiest/ferrarin/blob/ece6d947b1d107f2d7690debb647466465bb51d3/vendor/k8s.io/kubernetes/plugin/pkg/scheduler/algorithm/predicates/predicates_test.go#L2702-L2746) <details> <summary>Click here to show the 45 line(s) of Go which triggered the analyzer.</summary> ```go for _, node := range test.nodes { var podsOnNode []*api.Pod for _, pod := range test.pods { if pod.Spec.NodeName == node.Name { podsOnNode = append(podsOnNode, pod) } } testFit := PodAffinityChecker{ info: nodeListInfo, podLister: algorithm.FakePodLister(test.pods), failureDomains: priorityutil.Topologies{DefaultKeys: strings.Split(api.DefaultFailureDomains, ",")}, } nodeInfo := schedulercache.NewNodeInfo(podsOnNode...) nodeInfo.SetNode(&node) nodeInfoMap := map[string]*schedulercache.NodeInfo{node.Name: nodeInfo} fits, reasons, err := testFit.InterPodAffinityMatches(test.pod, PredicateMetadata(test.pod, nodeInfoMap), nodeInfo) if err != nil { t.Errorf("%s: unexpected error %v", test.test, err) } if !fits && !reflect.DeepEqual(reasons, affinityExpectedFailureReasons) { t.Errorf("%s: unexpected failure reasons: %v", test.test, reasons) } affinity, err := api.GetAffinityFromPodAnnotations(test.pod.ObjectMeta.Annotations) if err != nil { t.Errorf("%s: unexpected error: %v", test.test, err) } if affinity != nil && affinity.NodeAffinity != nil { nodeInfo := schedulercache.NewNodeInfo() nodeInfo.SetNode(&node) nodeInfoMap := map[string]*schedulercache.NodeInfo{node.Name: nodeInfo} fits2, reasons, err := PodSelectorMatches(test.pod, PredicateMetadata(test.pod, nodeInfoMap), nodeInfo) if err != nil { t.Errorf("%s: unexpected error: %v", test.test, err) } if !fits2 && !reflect.DeepEqual(reasons, selectorExpectedFailureReasons) { t.Errorf("%s: unexpected failure reasons: %v, want: %v", test.test, reasons, selectorExpectedFailureReasons) } fits = fits && fits2 } if fits != test.fits[node.Name] { t.Errorf("%s: expected %v for %s got %v", test.test, test.fits[node.Name], node.Name, fits) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: ece6d947b1d107f2d7690debb647466465bb51d3
1.0
soggiest/ferrarin: vendor/k8s.io/kubernetes/plugin/pkg/scheduler/algorithm/predicates/predicates_test.go; 45 LoC - Found a possible issue in [soggiest/ferrarin](https://www.github.com/soggiest/ferrarin) at [vendor/k8s.io/kubernetes/plugin/pkg/scheduler/algorithm/predicates/predicates_test.go](https://github.com/soggiest/ferrarin/blob/ece6d947b1d107f2d7690debb647466465bb51d3/vendor/k8s.io/kubernetes/plugin/pkg/scheduler/algorithm/predicates/predicates_test.go#L2702-L2746) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to node at line 2716 may start a goroutine [Click here to see the code in its original context.](https://github.com/soggiest/ferrarin/blob/ece6d947b1d107f2d7690debb647466465bb51d3/vendor/k8s.io/kubernetes/plugin/pkg/scheduler/algorithm/predicates/predicates_test.go#L2702-L2746) <details> <summary>Click here to show the 45 line(s) of Go which triggered the analyzer.</summary> ```go for _, node := range test.nodes { var podsOnNode []*api.Pod for _, pod := range test.pods { if pod.Spec.NodeName == node.Name { podsOnNode = append(podsOnNode, pod) } } testFit := PodAffinityChecker{ info: nodeListInfo, podLister: algorithm.FakePodLister(test.pods), failureDomains: priorityutil.Topologies{DefaultKeys: strings.Split(api.DefaultFailureDomains, ",")}, } nodeInfo := schedulercache.NewNodeInfo(podsOnNode...) nodeInfo.SetNode(&node) nodeInfoMap := map[string]*schedulercache.NodeInfo{node.Name: nodeInfo} fits, reasons, err := testFit.InterPodAffinityMatches(test.pod, PredicateMetadata(test.pod, nodeInfoMap), nodeInfo) if err != nil { t.Errorf("%s: unexpected error %v", test.test, err) } if !fits && !reflect.DeepEqual(reasons, affinityExpectedFailureReasons) { t.Errorf("%s: unexpected failure reasons: %v", test.test, reasons) } affinity, err := api.GetAffinityFromPodAnnotations(test.pod.ObjectMeta.Annotations) if err != nil { t.Errorf("%s: unexpected error: %v", test.test, err) } if affinity != nil && affinity.NodeAffinity != nil { nodeInfo := schedulercache.NewNodeInfo() nodeInfo.SetNode(&node) nodeInfoMap := map[string]*schedulercache.NodeInfo{node.Name: nodeInfo} fits2, reasons, err := PodSelectorMatches(test.pod, PredicateMetadata(test.pod, nodeInfoMap), nodeInfo) if err != nil { t.Errorf("%s: unexpected error: %v", test.test, err) } if !fits2 && !reflect.DeepEqual(reasons, selectorExpectedFailureReasons) { t.Errorf("%s: unexpected failure reasons: %v, want: %v", test.test, reasons, selectorExpectedFailureReasons) } fits = fits && fits2 } if fits != test.fits[node.Name] { t.Errorf("%s: expected %v for %s got %v", test.test, test.fits[node.Name], node.Name, fits) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: ece6d947b1d107f2d7690debb647466465bb51d3
test
soggiest ferrarin vendor io kubernetes plugin pkg scheduler algorithm predicates predicates test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to node at line may start a goroutine click here to show the line s of go which triggered the analyzer go for node range test nodes var podsonnode api pod for pod range test pods if pod spec nodename node name podsonnode append podsonnode pod testfit podaffinitychecker info nodelistinfo podlister algorithm fakepodlister test pods failuredomains priorityutil topologies defaultkeys strings split api defaultfailuredomains nodeinfo schedulercache newnodeinfo podsonnode nodeinfo setnode node nodeinfomap map schedulercache nodeinfo node name nodeinfo fits reasons err testfit interpodaffinitymatches test pod predicatemetadata test pod nodeinfomap nodeinfo if err nil t errorf s unexpected error v test test err if fits reflect deepequal reasons affinityexpectedfailurereasons t errorf s unexpected failure reasons v test test reasons affinity err api getaffinityfrompodannotations test pod objectmeta annotations if err nil t errorf s unexpected error v test test err if affinity nil affinity nodeaffinity nil nodeinfo schedulercache newnodeinfo nodeinfo setnode node nodeinfomap map schedulercache nodeinfo node name nodeinfo reasons err podselectormatches test pod predicatemetadata test pod nodeinfomap nodeinfo if err nil t errorf s unexpected error v test test err if reflect deepequal reasons selectorexpectedfailurereasons t errorf s unexpected failure reasons v want v test test reasons selectorexpectedfailurereasons fits fits if fits test fits t errorf s expected v for s got v test test test fits node name fits leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
141,378
21,508,682,221
IssuesEvent
2022-04-28 00:19:03
microsoft/TypeScript
https://api.github.com/repos/microsoft/TypeScript
closed
error TS2612 Property will overwrite base property not raised if "strictNullCheck" is true
Design Limitation
# Bug Report If you enable [useDefineForClassFields](https://www.typescriptlang.org/tsconfig#useDefineForClassFields) then a new error is possible TS26126 If you have code that trigger TS2612 Property 'xxx' will overwrite the base propert in 'yyy'. It this is intentional, add an initializer. Othertwise, add a "declare" modifier or remove the redundant declaration. The error do not appear if you enable strictNullCheck. Without strictNullCheck you'll see this error. If you put strict=true or strictNullCheck=true the error is not reported anymore ### 🔎 Search Terms TS2612 strictNullCheck ### 🕗 Version & Regression Information Noticed this on typescript 4.6.3 Tested on typescript@nightly 4.7.0-dev.20220408 and it behave the same (no error on strict mode) Tested on 4.3.5 and it behave the same (no error on strict mode) TS2612 was introduced in 4.3 so previous version do not make sense ### ⏯ Playground Link This link contain a minimal code to show the problem but you need to enable/disable strictNullCheck and [useDefineForClassFields](https://www.typescriptlang.org/tsconfig#useDefineForClassFields) to see it in action [Playground link with relevant code](https://www.typescriptlang.org/play?ssl=33&ssc=2&pln=10&pc=1#code/PTAEHUFMBsGMHsC2lQBd5oBYoCoE8AHSAZVgCcBLA1UABWgEM8BzM+AVwDsATAGiwoBnUENANQAd0gAjQRVSQAUCEmYKsTKGYUAbpGF4OY0BoadYKdJMoL+gzAzIoz3UNEiPOofEVKVqAHSKymAAmkYI7NCuqGqcANag8ABmIjQUXrFOKBJMggBcISGgoAC0oACCbvCwDKgU8JkY7p7ehCTkVDQS2E6gnPCxGcwmZqDSTgzxxWWVoASMFmgYkAAeRJTInN3ymj4d-jSCeNsMq-wuoPaOltigAKoASgAywhK7SbGQZIIz5VWCFzSeCrZagNYbChbHaxUDcCjJZLfSDbExIAgUdxkUBIursJzCFJtXydajBVDtUAAIQYghQAF5QABvRQlErSWmQfJXVCUTjMADcigAvkLFBSiKAACLfXSQVyMml00AAMmZrLZ3Flem43MEvOGQuFwVgjEEwiqLLZ8zYBG5SsgYutCE4+rI7Fg6DIAAoCLaAPz2zkASnV1utQ0EAT98AIoEZMbj-v9zPGnO5ACIOXSM6BRRqSsbjYpTbThFTwasFDwLWG2YnuTLKDqnWyXW6PV7fbbG9r5aGreGSoJ2BtvcGhUOSpHo7b4zbY5O2UXgiocM0PGQvIh4H0GMD2DQvm4zMx2AxmJB+Kb1Il9-A9CIvBmAKKrBiIBb6XO70AZ8AOKgADkwgAHKQBIGZBCoADyXxkO8dL8JeRyoI4CiuNIeCgE4O46MMWD6CgCCINCwiXMeEi7tEIgGBw2ILEwrAcDwQQlEAA) ### 💻 Code <!-- Please post the relevant code sample here as well--> ```ts type Base = { base: string; }; type Derived = Base & { derived: string; } class A { prop: Base; constructor(prop?: Base) { this.prop = prop ?? { base: "base" }; } } class B extends A { // This should trigger TS2612 but does not do it with strictNullCheck enabled prop: Derived; constructor(prop: Derived) { super(); this.prop = prop; } } ``` ### 🙁 Actual behavior When compiled with: * target - esnext * useDefineConfigForClass - true * strictNullCheck - true the code does not raise any error ### 🙂 Expected behavior It should raise the TS2612 error
1.0
error TS2612 Property will overwrite base property not raised if "strictNullCheck" is true - # Bug Report If you enable [useDefineForClassFields](https://www.typescriptlang.org/tsconfig#useDefineForClassFields) then a new error is possible TS26126 If you have code that trigger TS2612 Property 'xxx' will overwrite the base propert in 'yyy'. It this is intentional, add an initializer. Othertwise, add a "declare" modifier or remove the redundant declaration. The error do not appear if you enable strictNullCheck. Without strictNullCheck you'll see this error. If you put strict=true or strictNullCheck=true the error is not reported anymore ### 🔎 Search Terms TS2612 strictNullCheck ### 🕗 Version & Regression Information Noticed this on typescript 4.6.3 Tested on typescript@nightly 4.7.0-dev.20220408 and it behave the same (no error on strict mode) Tested on 4.3.5 and it behave the same (no error on strict mode) TS2612 was introduced in 4.3 so previous version do not make sense ### ⏯ Playground Link This link contain a minimal code to show the problem but you need to enable/disable strictNullCheck and [useDefineForClassFields](https://www.typescriptlang.org/tsconfig#useDefineForClassFields) to see it in action [Playground link with relevant code](https://www.typescriptlang.org/play?ssl=33&ssc=2&pln=10&pc=1#code/PTAEHUFMBsGMHsC2lQBd5oBYoCoE8AHSAZVgCcBLA1UABWgEM8BzM+AVwDsATAGiwoBnUENANQAd0gAjQRVSQAUCEmYKsTKGYUAbpGF4OY0BoadYKdJMoL+gzAzIoz3UNEiPOofEVKVqAHSKymAAmkYI7NCuqGqcANag8ABmIjQUXrFOKBJMggBcISGgoAC0oACCbvCwDKgU8JkY7p7ehCTkVDQS2E6gnPCxGcwmZqDSTgzxxWWVoASMFmgYkAAeRJTInN3ymj4d-jSCeNsMq-wuoPaOltigAKoASgAywhK7SbGQZIIz5VWCFzSeCrZagNYbChbHaxUDcCjJZLfSDbExIAgUdxkUBIursJzCFJtXydajBVDtUAAIQYghQAF5QABvRQlErSWmQfJXVCUTjMADcigAvkLFBSiKAACLfXSQVyMml00AAMmZrLZ3Flem43MEvOGQuFwVgjEEwiqLLZ8zYBG5SsgYutCE4+rI7Fg6DIAAoCLaAPz2zkASnV1utQ0EAT98AIoEZMbj-v9zPGnO5ACIOXSM6BRRqSsbjYpTbThFTwasFDwLWG2YnuTLKDqnWyXW6PV7fbbG9r5aGreGSoJ2BtvcGhUOSpHo7b4zbY5O2UXgiocM0PGQvIh4H0GMD2DQvm4zMx2AxmJB+Kb1Il9-A9CIvBmAKKrBiIBb6XO70AZ8AOKgADkwgAHKQBIGZBCoADyXxkO8dL8JeRyoI4CiuNIeCgE4O46MMWD6CgCCINCwiXMeEi7tEIgGBw2ILEwrAcDwQQlEAA) ### 💻 Code <!-- Please post the relevant code sample here as well--> ```ts type Base = { base: string; }; type Derived = Base & { derived: string; } class A { prop: Base; constructor(prop?: Base) { this.prop = prop ?? { base: "base" }; } } class B extends A { // This should trigger TS2612 but does not do it with strictNullCheck enabled prop: Derived; constructor(prop: Derived) { super(); this.prop = prop; } } ``` ### 🙁 Actual behavior When compiled with: * target - esnext * useDefineConfigForClass - true * strictNullCheck - true the code does not raise any error ### 🙂 Expected behavior It should raise the TS2612 error
non_test
error property will overwrite base property not raised if strictnullcheck is true bug report if you enable then a new error is possible if you have code that trigger property xxx will overwrite the base propert in yyy it this is intentional add an initializer othertwise add a declare modifier or remove the redundant declaration the error do not appear if you enable strictnullcheck without strictnullcheck you ll see this error if you put strict true or strictnullcheck true the error is not reported anymore 🔎 search terms strictnullcheck 🕗 version regression information noticed this on typescript tested on typescript nightly dev and it behave the same no error on strict mode tested on and it behave the same no error on strict mode was introduced in so previous version do not make sense ⏯ playground link this link contain a minimal code to show the problem but you need to enable disable strictnullcheck and to see it in action 💻 code ts type base base string type derived base derived string class a prop base constructor prop base this prop prop base base class b extends a this should trigger but does not do it with strictnullcheck enabled prop derived constructor prop derived super this prop prop 🙁 actual behavior when compiled with target esnext usedefineconfigforclass true strictnullcheck true the code does not raise any error 🙂 expected behavior it should raise the error
0
104,433
8,972,411,053
IssuesEvent
2019-01-29 18:12:54
phetsims/energy-forms-and-changes
https://api.github.com/repos/phetsims/energy-forms-and-changes
closed
Liquid containers can pick up thermometers when dropped
priority:5-deferred status:fixed-pending-testing type:bug
**Test device:** Dell Laptop **Operating System:** Win 10 **Browser:** chrome **Problem description:** For https://github.com/phetsims/QA/issues/258. Probably connected to https://github.com/phetsims/energy-forms-and-changes/issues/67. When dropping a beaker, if it passes through a thermometer on the way down, it may pick it up. It may also occur if you drop the container while overlapped with the thermometer, but not attached to it. This is not the behavior of the blocks. **Steps to reproduce:** 1. Put a thermometer above the water or oil while it sits on the bench 2. Pick up the liquid and wave it through the thermometer 3. Put liquid above the thermometer and drop **Screenshots:** ![liquiddrop](https://user-images.githubusercontent.com/41024075/51213955-af540600-18d9-11e9-8bbe-398a8d974908.gif) Troubleshooting information (do not edit): <details> Name: ‪Energy Forms And Changes‬ URL: https://phet-dev.colorado.edu/html/energy-forms-and-changes/1.0.0-dev.15/phet/energy-forms-and-changes_en_phet.html Version: 1.0.0-dev.15 2019-01-15 20:46:49 UTC Features missing: touch Flags: pixelRatioScaling User Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/71.0.3578.98 Safari/537.36 Language: en-US Window: 1536x732 Pixel Ratio: 2.5/1 WebGL: WebGL 1.0 (OpenGL ES 2.0 Chromium) GLSL: WebGL GLSL ES 1.0 (OpenGL ES GLSL ES 1.0 Chromium) Vendor: WebKit (WebKit WebGL) Vertex: attribs: 16 varying: 30 uniform: 4096 Texture: size: 16384 imageUnits: 16 (vertex: 16, combined: 32) Max viewport: 16384x16384 OES_texture_float: true Dependencies JSON: {} </details>
1.0
Liquid containers can pick up thermometers when dropped - **Test device:** Dell Laptop **Operating System:** Win 10 **Browser:** chrome **Problem description:** For https://github.com/phetsims/QA/issues/258. Probably connected to https://github.com/phetsims/energy-forms-and-changes/issues/67. When dropping a beaker, if it passes through a thermometer on the way down, it may pick it up. It may also occur if you drop the container while overlapped with the thermometer, but not attached to it. This is not the behavior of the blocks. **Steps to reproduce:** 1. Put a thermometer above the water or oil while it sits on the bench 2. Pick up the liquid and wave it through the thermometer 3. Put liquid above the thermometer and drop **Screenshots:** ![liquiddrop](https://user-images.githubusercontent.com/41024075/51213955-af540600-18d9-11e9-8bbe-398a8d974908.gif) Troubleshooting information (do not edit): <details> Name: ‪Energy Forms And Changes‬ URL: https://phet-dev.colorado.edu/html/energy-forms-and-changes/1.0.0-dev.15/phet/energy-forms-and-changes_en_phet.html Version: 1.0.0-dev.15 2019-01-15 20:46:49 UTC Features missing: touch Flags: pixelRatioScaling User Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/71.0.3578.98 Safari/537.36 Language: en-US Window: 1536x732 Pixel Ratio: 2.5/1 WebGL: WebGL 1.0 (OpenGL ES 2.0 Chromium) GLSL: WebGL GLSL ES 1.0 (OpenGL ES GLSL ES 1.0 Chromium) Vendor: WebKit (WebKit WebGL) Vertex: attribs: 16 varying: 30 uniform: 4096 Texture: size: 16384 imageUnits: 16 (vertex: 16, combined: 32) Max viewport: 16384x16384 OES_texture_float: true Dependencies JSON: {} </details>
test
liquid containers can pick up thermometers when dropped test device dell laptop operating system win browser chrome problem description for probably connected to when dropping a beaker if it passes through a thermometer on the way down it may pick it up it may also occur if you drop the container while overlapped with the thermometer but not attached to it this is not the behavior of the blocks steps to reproduce put a thermometer above the water or oil while it sits on the bench pick up the liquid and wave it through the thermometer put liquid above the thermometer and drop screenshots troubleshooting information do not edit name ‪energy forms and changes‬ url version dev utc features missing touch flags pixelratioscaling user agent mozilla windows nt applewebkit khtml like gecko chrome safari language en us window pixel ratio webgl webgl opengl es chromium glsl webgl glsl es opengl es glsl es chromium vendor webkit webkit webgl vertex attribs varying uniform texture size imageunits vertex combined max viewport oes texture float true dependencies json
1
217
2,581,733,982
IssuesEvent
2015-02-14 11:20:54
StyleCI/StyleCI
https://api.github.com/repos/StyleCI/StyleCI
closed
HTTP Missuse
security
GET requests must not cause the sever to perform any actions by definition. We're currently misusing them. I don't think we can avoid this for the oauth callback, but for everything else, we need to change this at some point before we open this to the public. We're also got a CSRF issue there too...
True
HTTP Missuse - GET requests must not cause the sever to perform any actions by definition. We're currently misusing them. I don't think we can avoid this for the oauth callback, but for everything else, we need to change this at some point before we open this to the public. We're also got a CSRF issue there too...
non_test
http missuse get requests must not cause the sever to perform any actions by definition we re currently misusing them i don t think we can avoid this for the oauth callback but for everything else we need to change this at some point before we open this to the public we re also got a csrf issue there too
0
190,470
14,546,973,142
IssuesEvent
2020-12-15 22:07:26
pachyderm/pachyderm
https://api.github.com/repos/pachyderm/pachyderm
closed
TestExtractRestore and TestExtractRestoreNoObjects
test flake
These tests seem to regularly time out in CI. Intuitively, they're large tests, so it makes some sense that they'd be brittle, but in at least one recent failure the timeout seemed to be due to a job stuck in merging (with a lot of restarts): ``` ====================================================================== pachctl list pipeline ---------------------------------------------------------------------- NAME VERSION INPUT CREATED STATE / LAST JOB DESCRIPTION TestExtractRestoreObjects-P2-d13e443debd3 1 TestExtractRestoreObjects-P1-7d9d768f9c53:/* 16 minutes ago running / success TestExtractRestoreObjects-P1-7d9d768f9c53 1 TestExtractRestoreObjects-P0-c3e186fa5ec8:/* 16 minutes ago running / success TestExtractRestoreObjects-P0-c3e186fa5ec8 1 TestExtractRestoreObjects-in-7ee4a3106153:/* 16 minutes ago running / merging ====================================================================== pachctl list job ---------------------------------------------------------------------- ID PIPELINE STARTED DURATION RESTART PROGRESS DL UL STATE 372f3107272b4a249b8c5002fcd3fe6a TestExtractRestoreObjects-P0-c3e186fa5ec8 16 minutes ago - 73 1 + 1 / 2 80MiB 80MiB merging 0dee5b9b497d4a1caba5ee63c2f21c43 TestExtractRestoreObjects-P2-d13e443debd3 16 minutes ago 5 seconds 0 2 + 0 / 2 80MiB 80MiB success b13164c50b7a41c99ab95a7c68e5c29b TestExtractRestoreObjects-P1-7d9d768f9c53 16 minutes ago 6 seconds 0 2 + 0 / 2 80MiB 80MiB success 993d7ebd9490426797c4fafbddc9d231 TestExtractRestoreObjects-P0-c3e186fa5ec8 16 minutes ago 7 seconds 0 2 + 0 / 2 80MiB 80MiB success ``` It's possible, therefore, that this test is also caused by https://github.com/pachyderm/pachyderm/issues/5448. While the pipeline that's stuck doesn't have multiple workers, the fact that there's multiple pipelines means workers from each pipeline are sharing the same hashtree cache as well, so the same problem could arise. We should revisit this issue once that's fixed, since this might go away.
1.0
TestExtractRestore and TestExtractRestoreNoObjects - These tests seem to regularly time out in CI. Intuitively, they're large tests, so it makes some sense that they'd be brittle, but in at least one recent failure the timeout seemed to be due to a job stuck in merging (with a lot of restarts): ``` ====================================================================== pachctl list pipeline ---------------------------------------------------------------------- NAME VERSION INPUT CREATED STATE / LAST JOB DESCRIPTION TestExtractRestoreObjects-P2-d13e443debd3 1 TestExtractRestoreObjects-P1-7d9d768f9c53:/* 16 minutes ago running / success TestExtractRestoreObjects-P1-7d9d768f9c53 1 TestExtractRestoreObjects-P0-c3e186fa5ec8:/* 16 minutes ago running / success TestExtractRestoreObjects-P0-c3e186fa5ec8 1 TestExtractRestoreObjects-in-7ee4a3106153:/* 16 minutes ago running / merging ====================================================================== pachctl list job ---------------------------------------------------------------------- ID PIPELINE STARTED DURATION RESTART PROGRESS DL UL STATE 372f3107272b4a249b8c5002fcd3fe6a TestExtractRestoreObjects-P0-c3e186fa5ec8 16 minutes ago - 73 1 + 1 / 2 80MiB 80MiB merging 0dee5b9b497d4a1caba5ee63c2f21c43 TestExtractRestoreObjects-P2-d13e443debd3 16 minutes ago 5 seconds 0 2 + 0 / 2 80MiB 80MiB success b13164c50b7a41c99ab95a7c68e5c29b TestExtractRestoreObjects-P1-7d9d768f9c53 16 minutes ago 6 seconds 0 2 + 0 / 2 80MiB 80MiB success 993d7ebd9490426797c4fafbddc9d231 TestExtractRestoreObjects-P0-c3e186fa5ec8 16 minutes ago 7 seconds 0 2 + 0 / 2 80MiB 80MiB success ``` It's possible, therefore, that this test is also caused by https://github.com/pachyderm/pachyderm/issues/5448. While the pipeline that's stuck doesn't have multiple workers, the fact that there's multiple pipelines means workers from each pipeline are sharing the same hashtree cache as well, so the same problem could arise. We should revisit this issue once that's fixed, since this might go away.
test
testextractrestore and testextractrestorenoobjects these tests seem to regularly time out in ci intuitively they re large tests so it makes some sense that they d be brittle but in at least one recent failure the timeout seemed to be due to a job stuck in merging with a lot of restarts pachctl list pipeline name version input created state last job description testextractrestoreobjects testextractrestoreobjects minutes ago running success testextractrestoreobjects testextractrestoreobjects minutes ago running success testextractrestoreobjects testextractrestoreobjects in minutes ago running merging pachctl list job id pipeline started duration restart progress dl ul state testextractrestoreobjects minutes ago merging testextractrestoreobjects minutes ago seconds success testextractrestoreobjects minutes ago seconds success testextractrestoreobjects minutes ago seconds success it s possible therefore that this test is also caused by while the pipeline that s stuck doesn t have multiple workers the fact that there s multiple pipelines means workers from each pipeline are sharing the same hashtree cache as well so the same problem could arise we should revisit this issue once that s fixed since this might go away
1
184,655
14,289,809,606
IssuesEvent
2020-11-23 19:51:46
github-vet/rangeclosure-findings
https://api.github.com/repos/github-vet/rangeclosure-findings
closed
jerrinsg/go-pmem: src/encoding/base32/base32_test.go; 18 LoC
fresh small test
Found a possible issue in [jerrinsg/go-pmem](https://www.github.com/jerrinsg/go-pmem) at [src/encoding/base32/base32_test.go](https://github.com/jerrinsg/go-pmem/blob/35da56fed65ecec7fce3310d537b250238d3c1a8/src/encoding/base32/base32_test.go#L592-L609) The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements which capture loop variables. [Click here to see the code in its original context.](https://github.com/jerrinsg/go-pmem/blob/35da56fed65ecec7fce3310d537b250238d3c1a8/src/encoding/base32/base32_test.go#L592-L609) <details> <summary>Click here to show the 18 line(s) of Go which triggered the analyzer.</summary> ```go for _, chunks := range testcase.chunkCombinations { pr, pw := io.Pipe() // Write the encoded chunks into the pipe go func() { for _, chunk := range chunks { pw.Write([]byte(chunk)) } pw.Close() }() decoder := NewDecoder(StdEncoding, pr) _, err := ioutil.ReadAll(decoder) if err != testcase.expected { t.Errorf("Expected %v, got %v; case %s %+v", testcase.expected, err, testcase.prefix, chunks) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 35da56fed65ecec7fce3310d537b250238d3c1a8
1.0
jerrinsg/go-pmem: src/encoding/base32/base32_test.go; 18 LoC - Found a possible issue in [jerrinsg/go-pmem](https://www.github.com/jerrinsg/go-pmem) at [src/encoding/base32/base32_test.go](https://github.com/jerrinsg/go-pmem/blob/35da56fed65ecec7fce3310d537b250238d3c1a8/src/encoding/base32/base32_test.go#L592-L609) The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements which capture loop variables. [Click here to see the code in its original context.](https://github.com/jerrinsg/go-pmem/blob/35da56fed65ecec7fce3310d537b250238d3c1a8/src/encoding/base32/base32_test.go#L592-L609) <details> <summary>Click here to show the 18 line(s) of Go which triggered the analyzer.</summary> ```go for _, chunks := range testcase.chunkCombinations { pr, pw := io.Pipe() // Write the encoded chunks into the pipe go func() { for _, chunk := range chunks { pw.Write([]byte(chunk)) } pw.Close() }() decoder := NewDecoder(StdEncoding, pr) _, err := ioutil.ReadAll(decoder) if err != testcase.expected { t.Errorf("Expected %v, got %v; case %s %+v", testcase.expected, err, testcase.prefix, chunks) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 35da56fed65ecec7fce3310d537b250238d3c1a8
test
jerrinsg go pmem src encoding test go loc found a possible issue in at the below snippet of go code triggered static analysis which searches for goroutines and or defer statements which capture loop variables click here to show the line s of go which triggered the analyzer go for chunks range testcase chunkcombinations pr pw io pipe write the encoded chunks into the pipe go func for chunk range chunks pw write byte chunk pw close decoder newdecoder stdencoding pr err ioutil readall decoder if err testcase expected t errorf expected v got v case s v testcase expected err testcase prefix chunks leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
74,224
7,389,764,652
IssuesEvent
2018-03-16 09:54:03
pkrog/biodb
https://api.github.com/repos/pkrog/biodb
opened
Improve MassCsvFile testing db
masscsvfile testing
Replace current fake MassCsvFile testing database file by an extract of Massbank.
1.0
Improve MassCsvFile testing db - Replace current fake MassCsvFile testing database file by an extract of Massbank.
test
improve masscsvfile testing db replace current fake masscsvfile testing database file by an extract of massbank
1
41,503
5,366,645,633
IssuesEvent
2017-02-22 00:47:54
Microsoft/vscode
https://api.github.com/repos/Microsoft/vscode
opened
Test: Type Definition Provider
testplan-item
**OS** - [ ] any Complexity: 1 1. Open the VSCode Codebase and open `extensions/markdown/src/extension.ts` 2. Right click on a variable like `telemetryReporter` in `var telemetryReporter: TelemetryReporter | null;` 3. You should see an option to go to the type definition in the context menu 4. Try going to the definition. 5. This should open `telemetryReporter.d.ts` and go to the `TelemetryReporter` 6. Try using go to type definition in other places in the VSCode codebase and ensure it jumps to the defining type
1.0
Test: Type Definition Provider - **OS** - [ ] any Complexity: 1 1. Open the VSCode Codebase and open `extensions/markdown/src/extension.ts` 2. Right click on a variable like `telemetryReporter` in `var telemetryReporter: TelemetryReporter | null;` 3. You should see an option to go to the type definition in the context menu 4. Try going to the definition. 5. This should open `telemetryReporter.d.ts` and go to the `TelemetryReporter` 6. Try using go to type definition in other places in the VSCode codebase and ensure it jumps to the defining type
test
test type definition provider os any complexity open the vscode codebase and open extensions markdown src extension ts right click on a variable like telemetryreporter in var telemetryreporter telemetryreporter null you should see an option to go to the type definition in the context menu try going to the definition this should open telemetryreporter d ts and go to the telemetryreporter try using go to type definition in other places in the vscode codebase and ensure it jumps to the defining type
1
569,645
17,015,640,520
IssuesEvent
2021-07-02 11:37:25
adirh3/Fluent-Search
https://api.github.com/repos/adirh3/Fluent-Search
closed
Pinned results navigation error
Low Priority bug
**Describe the bug** I have to press the right arrow key two times to go from the **first pinned result** to the **second one** **To Reproduce** Steps to reproduce the behavior: 1. Pin two results on the home screen 2. Trigger FS 3. Press the _down arrow key_ 4. navigate to the **second pinned result** using the _right arrow_ 5. see the error **Expected behavior** I should be able to navigate to the **second pinned** result with 1 press **Desktop (please complete the following information):** - Windows 10 Version: 20H1 - Fluent Search Version 0.9.88.2
1.0
Pinned results navigation error - **Describe the bug** I have to press the right arrow key two times to go from the **first pinned result** to the **second one** **To Reproduce** Steps to reproduce the behavior: 1. Pin two results on the home screen 2. Trigger FS 3. Press the _down arrow key_ 4. navigate to the **second pinned result** using the _right arrow_ 5. see the error **Expected behavior** I should be able to navigate to the **second pinned** result with 1 press **Desktop (please complete the following information):** - Windows 10 Version: 20H1 - Fluent Search Version 0.9.88.2
non_test
pinned results navigation error describe the bug i have to press the right arrow key two times to go from the first pinned result to the second one to reproduce steps to reproduce the behavior pin two results on the home screen trigger fs press the down arrow key navigate to the second pinned result using the right arrow see the error expected behavior i should be able to navigate to the second pinned result with press desktop please complete the following information windows version fluent search version
0
58,426
6,598,075,345
IssuesEvent
2017-09-16 00:01:26
QubesOS/updates-status
https://api.github.com/repos/QubesOS/updates-status
closed
app-linux-usb-proxy v1.0.11 (r4.0)
r4.0-dom0-testing r4.0-fc24-testing r4.0-fc25-testing r4.0-jessie-testing r4.0-stretch-testing
Update of app-linux-usb-proxy to v1.0.11 for Qubes r4.0, see comments below for details. Built from: https://github.com/QubesOS/qubes-app-linux-usb-proxy/commit/d137852633dc9e495d7fcfceafca2eba513e5cdd [Changes since previous version](https://github.com/QubesOS/qubes-app-linux-usb-proxy/compare/v1.0.10...v1.0.11): QubesOS/qubes-app-linux-usb-proxy@d137852 version 1.0.11 QubesOS/qubes-app-linux-usb-proxy@9e1e6b9 core3: ignore non-ascii characters in device description Referenced issues: QubesOS/qubes-issues#2972 If you're release manager, you can issue GPG-inline signed command: * `Upload app-linux-usb-proxy d137852633dc9e495d7fcfceafca2eba513e5cdd r4.0 current repo` (available 7 days from now) * `Upload app-linux-usb-proxy d137852633dc9e495d7fcfceafca2eba513e5cdd r4.0 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now) * `Upload app-linux-usb-proxy d137852633dc9e495d7fcfceafca2eba513e5cdd r4.0 security-testing repo` Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
5.0
app-linux-usb-proxy v1.0.11 (r4.0) - Update of app-linux-usb-proxy to v1.0.11 for Qubes r4.0, see comments below for details. Built from: https://github.com/QubesOS/qubes-app-linux-usb-proxy/commit/d137852633dc9e495d7fcfceafca2eba513e5cdd [Changes since previous version](https://github.com/QubesOS/qubes-app-linux-usb-proxy/compare/v1.0.10...v1.0.11): QubesOS/qubes-app-linux-usb-proxy@d137852 version 1.0.11 QubesOS/qubes-app-linux-usb-proxy@9e1e6b9 core3: ignore non-ascii characters in device description Referenced issues: QubesOS/qubes-issues#2972 If you're release manager, you can issue GPG-inline signed command: * `Upload app-linux-usb-proxy d137852633dc9e495d7fcfceafca2eba513e5cdd r4.0 current repo` (available 7 days from now) * `Upload app-linux-usb-proxy d137852633dc9e495d7fcfceafca2eba513e5cdd r4.0 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now) * `Upload app-linux-usb-proxy d137852633dc9e495d7fcfceafca2eba513e5cdd r4.0 security-testing repo` Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
test
app linux usb proxy update of app linux usb proxy to for qubes see comments below for details built from qubesos qubes app linux usb proxy version qubesos qubes app linux usb proxy ignore non ascii characters in device description referenced issues qubesos qubes issues if you re release manager you can issue gpg inline signed command upload app linux usb proxy current repo available days from now upload app linux usb proxy current dists repo you can choose subset of distributions like vm vm available days from now upload app linux usb proxy security testing repo above commands will work only if packages in current testing repository were built from given commit i e no new version superseded it
1
66,133
12,727,970,490
IssuesEvent
2020-06-25 00:58:55
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
menu + anchor results in error
No Code Attached Yet
### Steps to reproduce the issue copy menu link url on your joomla website set menu type to: URL paste url to link attach anchor #ananchor to url the url looks like this: http://www.aaa.de/aaa/aaa/index.php?option=com_content&view=article&id=40&Itemid=104#section4 set error reporting to development click menu link ### Expected result on menu item click, jump to anchor "point" in page ### Actual result the method works (menu item click results in jumping to anchor but one gets: Notice: Undefined index: option in /var/www/vhosts/aaa.de/aaa.de/aaa/aaa/components/com_content/views/article/view.html.php on line 245 Notice: Undefined index: option in /var/www/vhosts/aaa.de/aaa.de/aaa/aaa/components/com_content/views/article/view.html.php on line 253 Notice: Undefined index: option in /var/www/vhosts/aaa.de/aaa.de/aaa/aaa/components/com_content/views/article/view.html.php on line 253 ### System information (as much as possible) Joomla! 3.9.8 windows 10 intel 2.6ghz, 4 gb ram google chrome Version 83.0.4103.116 (Offizieller Build) (64-Bit) ### Additional comments
1.0
menu + anchor results in error - ### Steps to reproduce the issue copy menu link url on your joomla website set menu type to: URL paste url to link attach anchor #ananchor to url the url looks like this: http://www.aaa.de/aaa/aaa/index.php?option=com_content&view=article&id=40&Itemid=104#section4 set error reporting to development click menu link ### Expected result on menu item click, jump to anchor "point" in page ### Actual result the method works (menu item click results in jumping to anchor but one gets: Notice: Undefined index: option in /var/www/vhosts/aaa.de/aaa.de/aaa/aaa/components/com_content/views/article/view.html.php on line 245 Notice: Undefined index: option in /var/www/vhosts/aaa.de/aaa.de/aaa/aaa/components/com_content/views/article/view.html.php on line 253 Notice: Undefined index: option in /var/www/vhosts/aaa.de/aaa.de/aaa/aaa/components/com_content/views/article/view.html.php on line 253 ### System information (as much as possible) Joomla! 3.9.8 windows 10 intel 2.6ghz, 4 gb ram google chrome Version 83.0.4103.116 (Offizieller Build) (64-Bit) ### Additional comments
non_test
menu anchor results in error steps to reproduce the issue copy menu link url on your joomla website set menu type to url paste url to link attach anchor ananchor to url the url looks like this set error reporting to development click menu link expected result on menu item click jump to anchor point in page actual result the method works menu item click results in jumping to anchor but one gets notice undefined index option in var www vhosts aaa de aaa de aaa aaa components com content views article view html php on line notice undefined index option in var www vhosts aaa de aaa de aaa aaa components com content views article view html php on line notice undefined index option in var www vhosts aaa de aaa de aaa aaa components com content views article view html php on line system information as much as possible joomla windows intel gb ram google chrome version offizieller build bit additional comments
0
308,794
26,631,951,725
IssuesEvent
2023-01-24 18:34:34
Azure/azure-sdk-for-js
https://api.github.com/repos/Azure/azure-sdk-for-js
closed
[Key Vault - Administration] Deploy test resources failing in nightly runs
Client KeyVault test-reliability
Key Vault - Administration nightly test runs are failing with: > Error message: > VERBOSE: Running registered exit actions > VERBOSE: Logging out of service principal '***' > VERBOSE: Performing the operation "New-TestResources.ps1" on target "***". > Write-Error: /mnt/vss/_work/1/s/eng/common/TestResources/New-TestResources.ps1:755 > Line | > 755 | &$postDeploymentScript -ResourceGroupName $ResourceGroupN … > | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > | Failed to export security domain data. > > ##[error]PowerShell exited with code '1'. For more details check here: https://dev.azure.com/azure-sdk/internal/_build/results?buildId=1813663&view=results
1.0
[Key Vault - Administration] Deploy test resources failing in nightly runs - Key Vault - Administration nightly test runs are failing with: > Error message: > VERBOSE: Running registered exit actions > VERBOSE: Logging out of service principal '***' > VERBOSE: Performing the operation "New-TestResources.ps1" on target "***". > Write-Error: /mnt/vss/_work/1/s/eng/common/TestResources/New-TestResources.ps1:755 > Line | > 755 | &$postDeploymentScript -ResourceGroupName $ResourceGroupN … > | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > | Failed to export security domain data. > > ##[error]PowerShell exited with code '1'. For more details check here: https://dev.azure.com/azure-sdk/internal/_build/results?buildId=1813663&view=results
test
deploy test resources failing in nightly runs key vault administration nightly test runs are failing with error message verbose running registered exit actions verbose logging out of service principal verbose performing the operation new testresources on target write error mnt vss work s eng common testresources new testresources line postdeploymentscript resourcegroupname resourcegroupn … failed to export security domain data powershell exited with code for more details check here
1
387,636
11,463,853,344
IssuesEvent
2020-02-07 16:48:02
fritzing/fritzing-app
https://api.github.com/repos/fritzing/fritzing-app
closed
menu color
Priority-High bug imported
_From [jazzcomm...@gmail.com](https://code.google.com/u/102214720157054660916/) on January 05, 2012 22:39:28_ What steps will reproduce the problem? 1.linux version 2. 3. the menu bar is dark gray, and the menu's(file edit and so on) are black so there is too little contrast to see the menus. please change menubar background to lighter shade of gray. that's all thanks for the nice software greetings from Saigon, Vietnam Attach your sketch file and/or custom part files to the bug report. If the sketch uses custom parts, save the sketch as shareable (under the file menu). What is the expected output? What do you see instead? What version of Fritzing are you using? On what operating system? Please use labels and text to provide additional information. _Original issue: http://code.google.com/p/fritzing/issues/detail?id=1752_
1.0
menu color - _From [jazzcomm...@gmail.com](https://code.google.com/u/102214720157054660916/) on January 05, 2012 22:39:28_ What steps will reproduce the problem? 1.linux version 2. 3. the menu bar is dark gray, and the menu's(file edit and so on) are black so there is too little contrast to see the menus. please change menubar background to lighter shade of gray. that's all thanks for the nice software greetings from Saigon, Vietnam Attach your sketch file and/or custom part files to the bug report. If the sketch uses custom parts, save the sketch as shareable (under the file menu). What is the expected output? What do you see instead? What version of Fritzing are you using? On what operating system? Please use labels and text to provide additional information. _Original issue: http://code.google.com/p/fritzing/issues/detail?id=1752_
non_test
menu color from on january what steps will reproduce the problem linux version the menu bar is dark gray and the menu s file edit and so on are black so there is too little contrast to see the menus please change menubar background to lighter shade of gray that s all thanks for the nice software greetings from saigon vietnam attach your sketch file and or custom part files to the bug report if the sketch uses custom parts save the sketch as shareable under the file menu what is the expected output what do you see instead what version of fritzing are you using on what operating system please use labels and text to provide additional information original issue
0
65,448
6,963,980,002
IssuesEvent
2017-12-08 19:38:59
filippotessaro/unitn_finder
https://api.github.com/repos/filippotessaro/unitn_finder
closed
Ricerche specifiche
enhancement test
Inserendo richieste specifiche (come ad esempio mail fabio) il bot ritorna tutti i professori di nome Fabio con la loro mail, la stessa cosa avviene se si ricerca il numero di telefono, ecc. Se invece si ricerca un determinato dato specificando nome e cognome di un professore il bot ritorna il dato ricercato solo del professore specificato accompagnato da nome e cognome del professore. HTD: Se scrivo Fabio Casati mail il bot mi restituisce solo i dati richiesti quindi Fabio Casati e la sua mail.
1.0
Ricerche specifiche - Inserendo richieste specifiche (come ad esempio mail fabio) il bot ritorna tutti i professori di nome Fabio con la loro mail, la stessa cosa avviene se si ricerca il numero di telefono, ecc. Se invece si ricerca un determinato dato specificando nome e cognome di un professore il bot ritorna il dato ricercato solo del professore specificato accompagnato da nome e cognome del professore. HTD: Se scrivo Fabio Casati mail il bot mi restituisce solo i dati richiesti quindi Fabio Casati e la sua mail.
test
ricerche specifiche inserendo richieste specifiche come ad esempio mail fabio il bot ritorna tutti i professori di nome fabio con la loro mail la stessa cosa avviene se si ricerca il numero di telefono ecc se invece si ricerca un determinato dato specificando nome e cognome di un professore il bot ritorna il dato ricercato solo del professore specificato accompagnato da nome e cognome del professore htd se scrivo fabio casati mail il bot mi restituisce solo i dati richiesti quindi fabio casati e la sua mail
1
97,213
12,219,433,573
IssuesEvent
2020-05-01 21:43:49
phetsims/energy-skate-park
https://api.github.com/repos/phetsims/energy-skate-park
closed
Add skater with adjustable mass to Intro screen?
design:general status:ready-for-review
Kathy suggested adding a Custom skater with adjustable mass to the Intro screen. This would add quite a bit of flexibility, and may be desirable functionality for PhET-iO clients. When selecting "Custom", a NumberControl with a range of 1-100 kg would appear below the Skater ComboBox. We may need an additional piece of skater artwork for the Custom option, or we could consider using the default Skater 1. Let's discuss at design meeting.
1.0
Add skater with adjustable mass to Intro screen? - Kathy suggested adding a Custom skater with adjustable mass to the Intro screen. This would add quite a bit of flexibility, and may be desirable functionality for PhET-iO clients. When selecting "Custom", a NumberControl with a range of 1-100 kg would appear below the Skater ComboBox. We may need an additional piece of skater artwork for the Custom option, or we could consider using the default Skater 1. Let's discuss at design meeting.
non_test
add skater with adjustable mass to intro screen kathy suggested adding a custom skater with adjustable mass to the intro screen this would add quite a bit of flexibility and may be desirable functionality for phet io clients when selecting custom a numbercontrol with a range of kg would appear below the skater combobox we may need an additional piece of skater artwork for the custom option or we could consider using the default skater let s discuss at design meeting
0
163,749
12,743,018,583
IssuesEvent
2020-06-26 09:34:35
victorbucutea/parking-space
https://api.github.com/repos/victorbucutea/parking-space
closed
Valabilitate expirata, dar valabila
bug fixed - to retest
1.Acceseaza https://go-park-staging.herokuapp.com/app/index.html#!/map/myposts Observat Un loc de parcare are afisat "Valabilitate expirata", dar mai jos arata ca mai este valabil inca 6 zile. Asteptat Daca are Valabilitatea expirata, nu ar mai trebui sa fie afisat ca mai este valabil inca 6 zile (sau cateva ore dupa caz) Testat pe Desktop si mobile Utilizator: ovidiu.ovidiu0@gmail.com Parola: ovidiu22 ![105587318_2381832255450529_7137476204282992381_n](https://user-images.githubusercontent.com/33632578/85789361-23cdde00-b737-11ea-954b-3dd5b52a830e.jpg)
1.0
Valabilitate expirata, dar valabila - 1.Acceseaza https://go-park-staging.herokuapp.com/app/index.html#!/map/myposts Observat Un loc de parcare are afisat "Valabilitate expirata", dar mai jos arata ca mai este valabil inca 6 zile. Asteptat Daca are Valabilitatea expirata, nu ar mai trebui sa fie afisat ca mai este valabil inca 6 zile (sau cateva ore dupa caz) Testat pe Desktop si mobile Utilizator: ovidiu.ovidiu0@gmail.com Parola: ovidiu22 ![105587318_2381832255450529_7137476204282992381_n](https://user-images.githubusercontent.com/33632578/85789361-23cdde00-b737-11ea-954b-3dd5b52a830e.jpg)
test
valabilitate expirata dar valabila acceseaza observat un loc de parcare are afisat valabilitate expirata dar mai jos arata ca mai este valabil inca zile asteptat daca are valabilitatea expirata nu ar mai trebui sa fie afisat ca mai este valabil inca zile sau cateva ore dupa caz testat pe desktop si mobile utilizator ovidiu gmail com parola
1
440,387
30,744,144,712
IssuesEvent
2023-07-28 13:48:32
kwurst/GitKit-FarmData2
https://api.github.com/repos/kwurst/GitKit-FarmData2
opened
Make Update plural
documentation Round1 Typos
Change "Update" to "Updates" in the phrase "Update to any of the FarmData2 documentation are welcome." in the first sentence of the "Documentation" section of the `CONTRIBUTING.md` file.
1.0
Make Update plural - Change "Update" to "Updates" in the phrase "Update to any of the FarmData2 documentation are welcome." in the first sentence of the "Documentation" section of the `CONTRIBUTING.md` file.
non_test
make update plural change update to updates in the phrase update to any of the documentation are welcome in the first sentence of the documentation section of the contributing md file
0
45,262
5,707,111,100
IssuesEvent
2017-04-18 13:09:35
nkp-osf/VDK
https://api.github.com/repos/nkp-osf/VDK
closed
Filtr: stav ve zdroji
K testovani
Bylo by potřeba vytvořit nový filtr "stav ve zdroji" (např. stav vyřazeno ve zdroji NKF)
1.0
Filtr: stav ve zdroji - Bylo by potřeba vytvořit nový filtr "stav ve zdroji" (např. stav vyřazeno ve zdroji NKF)
test
filtr stav ve zdroji bylo by potřeba vytvořit nový filtr stav ve zdroji např stav vyřazeno ve zdroji nkf
1