Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
43,838
5,566,972,280
IssuesEvent
2017-03-27 00:58:10
wphicks/Writing3D
https://api.github.com/repos/wphicks/Writing3D
closed
Vertical text alignment handled differently from legacy system
legacy testing-needed
In CW, text (and other) objects have horiz-align and vert-align properties, defaulting to "(center)". W3D either has different defaults or interprets these differently. horiz-alignment seems OK. With respect to vert-alignment, W3D currently treats the Y position parameter as referring to the ? ascender height of the font (should refer to font height / 2).
1.0
Vertical text alignment handled differently from legacy system - In CW, text (and other) objects have horiz-align and vert-align properties, defaulting to "(center)". W3D either has different defaults or interprets these differently. horiz-alignment seems OK. With respect to vert-alignment, W3D currently treats the Y position parameter as referring to the ? ascender height of the font (should refer to font height / 2).
test
vertical text alignment handled differently from legacy system in cw text and other objects have horiz align and vert align properties defaulting to center either has different defaults or interprets these differently horiz alignment seems ok with respect to vert alignment currently treats the y position parameter as referring to the ascender height of the font should refer to font height
1
787,953
27,736,931,757
IssuesEvent
2023-03-15 11:53:44
SzFMV2023-Tavasz/AutomatedCar-A
https://api.github.com/repos/SzFMV2023-Tavasz/AutomatedCar-A
opened
## Mozgatás: hajtáslánc és kormányzás
effort: high priority: critical type: user story
A *mozgatás* modul felelőssége a vezérelt autó (egocar) mozgatása, mozgásának számítása. Ez magába foglalja a hajtáslánc és a kormányzás megvalósítását. A modul a HMI-től kap bemenetet, mindenek előtt váltóállás, gáz- és fékpedálállás valamint kormányelfordulás. A váltó automata, ami azzal jár, hogy a HMI a négy állapot (P, R, N, D) egyikét közli. - **P**ark: Ez az egyik olyan állapot amiben a motor beindítható (a másik az N), ez a váltó alapállása, mechanikusan megszünteti az erőátvitelt. A valóságban nem helyettesíti kéziféket, de mivel azt nem kell implementálni a feladat során, lehet úgy tekinteni, hogy P-ben a kézifék is be van húzva. Az autó nem mozdul ebben az állásban. - **R**everse: hátramenet - **N**eutral: üres, ebben az állásban sem jut a motorerő tengelyekre, a motor nem gyorsíthatja az autót. Ha üresben gázt adunk, a motor felpörög, méghozzá jelentősen, mivel nincs ami ellene dolgozna. Ha az autónak volt lendülete, az még hajtja tovább. - **D**rive: előremenet. A D-m belül definiálni kell 4-5 belső fokozatot (mint ahogy manuális váltónál is lenne), le kell programozni, hogy valamilyen fordulatszám értékeknél a váltó váltson. Ezen értékek meghatározásához lehet találni motorkarakterisztikákat (lehetőség szerint utcai autó kerüljön kiválasztásra, ne valami sportautó). A belső fokozatok a felhasználó interfész szempontjából transzparensek. A pedál állások \[0-100\] skálán érkeznek. 0: nincs lenyomva, 1: 1%-ig van lenyomva, 100: tövig (100%-ig) le van nyomva. Minél jobban le van nyomva a gázpedál, annál több teljesítményt kell a motornak kiadnia. A motor belső működését nem kell részletekbe menően implementálni, pláne nem egy belső égésű motorét, egy elektromos hajtáslánc egyszerűbb. A szoftverben üzemanyag-fogyasztást (vagy akkumulátor töltöttséget) és hatótávokat nem kell kezelni. A kormány jellemzően valamilyen áttétel segítségévél befolyásolja a kormányzott kerekek helyzetét. Ez ebben a szoftverben sokkal egyszerűbb is lehet. Ha 60°-ot tengelyelfordulást feltételezünk és a kormány „nulla” állásból +/- 60-at mozdulhat el, akkor lényegében 1:1-es „áttételünk” van. Ha az input +/- 100-as skálán adja meg a kormányelfordulás mértékét, akkor azt kell a tengelyelfordulásra képzeni. A bemeneti skálával kapcsolatban a HMI csapattal kell egyeztetni. A úgymond kimenete egy mozgásvektor, vagyis az, hogy a következő ciklusban az autó (referenciapontjának) X, Y koordinátáit mennyivel kell módosítani. A pedálállás és a motor korábbi állapotának függvényében meghatározásra kerül a sebesség, a kormányállás valamint az egocar korábbi orientációjának függvényében meghatározásra kerül, hogy módosul-e az autó iránya, a kettő eredőjeként a teljes vektor. A modullal kapcsolatban kihívás a hajtáslánc működésének és a mozgás és kanyarodás fizikájának megértése, implementálása. A fizika tekintetében sem kell elaprózni a dolgokat. Két erő elégséges: a motorerő mint gyorsítja a járművet (ez a gázpedállal szabályozható) és egy fékező erő, amelyet a fékpedállal lehet szabályozni, valamint ezen felül egy konstans fékező erőnek is lennie kell, tehát a fékező erő akkor sem nulla, ha a fékpedál állása nulla. Ez utóbbit mindegy minek nevezzük (légellenállás, csúszási-súrlódási erő, a kettő eredője, stb.), nem kell cicomázni, de legyen. Külön feladat észben tartatni és célszerűen előre felkészülni arra, hogy a 3. sprintes vezetéstámogató modulok (LKA, AAC, AEB, PP) közvetlenül az egocar hajtásláncára és a kormányra hatnak. Fel kell készíteni a modult ilyen, „nem a HMI-ről” érkező inputok kezelésére is, amelyek ráadásul magasabb prioritásúak. Pl. ha a vészfékező rendszer „lenyomja” a fékpedált, akkor magasabb prioritással kell kezelni mint a HMI-ről érkező pedálállást. ### Definition of Done - [ ] Az autó gázpedál állásától függően gyorsul - [ ] a fék- és gázpedál állapota a billentyű nyomva tartásának idejével szabályozható - [ ] fék- és gázpedál valamint a kormány sem binárisan működik, a billentyű nyomva tartás idejétől függ az input intenzitása - a fék- és gázpedál \[0, 100\] skálán (ℕ), a kormányelforgatás \[-100, 100\] skálán (ℤ) kerül meghatározásra - [ ] fék- és gázpedál valamint a kormány is fokozatosan (1 másodperc) áll vissza alaphelyzetbe a billentyű felengedésével - [ ] az automata váltó 4 állapota szabályozható - szekvenciális váltóról lévén szó, sorban állíthatók a fokozatok: P(ark), R(everse), N(eutral), D(rive) - alaphelyzet: P, „felváltás” után R, majd N, majd D. „Leváltás” ugyanez visszafele. - [ ] A gyorsulás a „belső fokozatok” szerint kerül meghatározásra - [ ] Az autó a gázpedál felengedésével fokozatosan lassul, majd megáll - [ ] Az autó R válóállásban tolat - [ ] Felkészíteni a modult, hogy a vészfékező, az adaptív tempomat vagy a sávtartó automatika is küldhet inputot, melyek magasabb prioritásúak - vészfékező értelemszerűen fékezés inputot - az adaptív tempomat és a parkoló asszisztens gáz és fék inputot is - a sávtartó automatika a kormányállást módosítja - [ ] Autó kanyarodásának biztosítása valóságos fordulókör szerint - ehhez szükséges extra tulajdonságok meghatározása - [ ] A meghatározott mozgásvektor alapján az autó pozíciójának frissítése - ez az AutomatedCar osztály x,y koordinátáinak frissítését jelenti - [ ] Tolatás során is valósághű kanyarodás történik
1.0
## Mozgatás: hajtáslánc és kormányzás - A *mozgatás* modul felelőssége a vezérelt autó (egocar) mozgatása, mozgásának számítása. Ez magába foglalja a hajtáslánc és a kormányzás megvalósítását. A modul a HMI-től kap bemenetet, mindenek előtt váltóállás, gáz- és fékpedálállás valamint kormányelfordulás. A váltó automata, ami azzal jár, hogy a HMI a négy állapot (P, R, N, D) egyikét közli. - **P**ark: Ez az egyik olyan állapot amiben a motor beindítható (a másik az N), ez a váltó alapállása, mechanikusan megszünteti az erőátvitelt. A valóságban nem helyettesíti kéziféket, de mivel azt nem kell implementálni a feladat során, lehet úgy tekinteni, hogy P-ben a kézifék is be van húzva. Az autó nem mozdul ebben az állásban. - **R**everse: hátramenet - **N**eutral: üres, ebben az állásban sem jut a motorerő tengelyekre, a motor nem gyorsíthatja az autót. Ha üresben gázt adunk, a motor felpörög, méghozzá jelentősen, mivel nincs ami ellene dolgozna. Ha az autónak volt lendülete, az még hajtja tovább. - **D**rive: előremenet. A D-m belül definiálni kell 4-5 belső fokozatot (mint ahogy manuális váltónál is lenne), le kell programozni, hogy valamilyen fordulatszám értékeknél a váltó váltson. Ezen értékek meghatározásához lehet találni motorkarakterisztikákat (lehetőség szerint utcai autó kerüljön kiválasztásra, ne valami sportautó). A belső fokozatok a felhasználó interfész szempontjából transzparensek. A pedál állások \[0-100\] skálán érkeznek. 0: nincs lenyomva, 1: 1%-ig van lenyomva, 100: tövig (100%-ig) le van nyomva. Minél jobban le van nyomva a gázpedál, annál több teljesítményt kell a motornak kiadnia. A motor belső működését nem kell részletekbe menően implementálni, pláne nem egy belső égésű motorét, egy elektromos hajtáslánc egyszerűbb. A szoftverben üzemanyag-fogyasztást (vagy akkumulátor töltöttséget) és hatótávokat nem kell kezelni. A kormány jellemzően valamilyen áttétel segítségévél befolyásolja a kormányzott kerekek helyzetét. Ez ebben a szoftverben sokkal egyszerűbb is lehet. Ha 60°-ot tengelyelfordulást feltételezünk és a kormány „nulla” állásból +/- 60-at mozdulhat el, akkor lényegében 1:1-es „áttételünk” van. Ha az input +/- 100-as skálán adja meg a kormányelfordulás mértékét, akkor azt kell a tengelyelfordulásra képzeni. A bemeneti skálával kapcsolatban a HMI csapattal kell egyeztetni. A úgymond kimenete egy mozgásvektor, vagyis az, hogy a következő ciklusban az autó (referenciapontjának) X, Y koordinátáit mennyivel kell módosítani. A pedálállás és a motor korábbi állapotának függvényében meghatározásra kerül a sebesség, a kormányállás valamint az egocar korábbi orientációjának függvényében meghatározásra kerül, hogy módosul-e az autó iránya, a kettő eredőjeként a teljes vektor. A modullal kapcsolatban kihívás a hajtáslánc működésének és a mozgás és kanyarodás fizikájának megértése, implementálása. A fizika tekintetében sem kell elaprózni a dolgokat. Két erő elégséges: a motorerő mint gyorsítja a járművet (ez a gázpedállal szabályozható) és egy fékező erő, amelyet a fékpedállal lehet szabályozni, valamint ezen felül egy konstans fékező erőnek is lennie kell, tehát a fékező erő akkor sem nulla, ha a fékpedál állása nulla. Ez utóbbit mindegy minek nevezzük (légellenállás, csúszási-súrlódási erő, a kettő eredője, stb.), nem kell cicomázni, de legyen. Külön feladat észben tartatni és célszerűen előre felkészülni arra, hogy a 3. sprintes vezetéstámogató modulok (LKA, AAC, AEB, PP) közvetlenül az egocar hajtásláncára és a kormányra hatnak. Fel kell készíteni a modult ilyen, „nem a HMI-ről” érkező inputok kezelésére is, amelyek ráadásul magasabb prioritásúak. Pl. ha a vészfékező rendszer „lenyomja” a fékpedált, akkor magasabb prioritással kell kezelni mint a HMI-ről érkező pedálállást. ### Definition of Done - [ ] Az autó gázpedál állásától függően gyorsul - [ ] a fék- és gázpedál állapota a billentyű nyomva tartásának idejével szabályozható - [ ] fék- és gázpedál valamint a kormány sem binárisan működik, a billentyű nyomva tartás idejétől függ az input intenzitása - a fék- és gázpedál \[0, 100\] skálán (ℕ), a kormányelforgatás \[-100, 100\] skálán (ℤ) kerül meghatározásra - [ ] fék- és gázpedál valamint a kormány is fokozatosan (1 másodperc) áll vissza alaphelyzetbe a billentyű felengedésével - [ ] az automata váltó 4 állapota szabályozható - szekvenciális váltóról lévén szó, sorban állíthatók a fokozatok: P(ark), R(everse), N(eutral), D(rive) - alaphelyzet: P, „felváltás” után R, majd N, majd D. „Leváltás” ugyanez visszafele. - [ ] A gyorsulás a „belső fokozatok” szerint kerül meghatározásra - [ ] Az autó a gázpedál felengedésével fokozatosan lassul, majd megáll - [ ] Az autó R válóállásban tolat - [ ] Felkészíteni a modult, hogy a vészfékező, az adaptív tempomat vagy a sávtartó automatika is küldhet inputot, melyek magasabb prioritásúak - vészfékező értelemszerűen fékezés inputot - az adaptív tempomat és a parkoló asszisztens gáz és fék inputot is - a sávtartó automatika a kormányállást módosítja - [ ] Autó kanyarodásának biztosítása valóságos fordulókör szerint - ehhez szükséges extra tulajdonságok meghatározása - [ ] A meghatározott mozgásvektor alapján az autó pozíciójának frissítése - ez az AutomatedCar osztály x,y koordinátáinak frissítését jelenti - [ ] Tolatás során is valósághű kanyarodás történik
non_test
mozgatás hajtáslánc és kormányzás a mozgatás modul felelőssége a vezérelt autó egocar mozgatása mozgásának számítása ez magába foglalja a hajtáslánc és a kormányzás megvalósítását a modul a hmi től kap bemenetet mindenek előtt váltóállás gáz és fékpedálállás valamint kormányelfordulás a váltó automata ami azzal jár hogy a hmi a négy állapot p r n d egyikét közli p ark ez az egyik olyan állapot amiben a motor beindítható a másik az n ez a váltó alapállása mechanikusan megszünteti az erőátvitelt a valóságban nem helyettesíti kéziféket de mivel azt nem kell implementálni a feladat során lehet úgy tekinteni hogy p ben a kézifék is be van húzva az autó nem mozdul ebben az állásban r everse hátramenet n eutral üres ebben az állásban sem jut a motorerő tengelyekre a motor nem gyorsíthatja az autót ha üresben gázt adunk a motor felpörög méghozzá jelentősen mivel nincs ami ellene dolgozna ha az autónak volt lendülete az még hajtja tovább d rive előremenet a d m belül definiálni kell belső fokozatot mint ahogy manuális váltónál is lenne le kell programozni hogy valamilyen fordulatszám értékeknél a váltó váltson ezen értékek meghatározásához lehet találni motorkarakterisztikákat lehetőség szerint utcai autó kerüljön kiválasztásra ne valami sportautó a belső fokozatok a felhasználó interfész szempontjából transzparensek a pedál állások skálán érkeznek nincs lenyomva ig van lenyomva tövig ig le van nyomva minél jobban le van nyomva a gázpedál annál több teljesítményt kell a motornak kiadnia a motor belső működését nem kell részletekbe menően implementálni pláne nem egy belső égésű motorét egy elektromos hajtáslánc egyszerűbb a szoftverben üzemanyag fogyasztást vagy akkumulátor töltöttséget és hatótávokat nem kell kezelni a kormány jellemzően valamilyen áttétel segítségévél befolyásolja a kormányzott kerekek helyzetét ez ebben a szoftverben sokkal egyszerűbb is lehet ha ° ot tengelyelfordulást feltételezünk és a kormány „nulla” állásból at mozdulhat el akkor lényegében es „áttételünk” van ha az input as skálán adja meg a kormányelfordulás mértékét akkor azt kell a tengelyelfordulásra képzeni a bemeneti skálával kapcsolatban a hmi csapattal kell egyeztetni a úgymond kimenete egy mozgásvektor vagyis az hogy a következő ciklusban az autó referenciapontjának x y koordinátáit mennyivel kell módosítani a pedálállás és a motor korábbi állapotának függvényében meghatározásra kerül a sebesség a kormányállás valamint az egocar korábbi orientációjának függvényében meghatározásra kerül hogy módosul e az autó iránya a kettő eredőjeként a teljes vektor a modullal kapcsolatban kihívás a hajtáslánc működésének és a mozgás és kanyarodás fizikájának megértése implementálása a fizika tekintetében sem kell elaprózni a dolgokat két erő elégséges a motorerő mint gyorsítja a járművet ez a gázpedállal szabályozható és egy fékező erő amelyet a fékpedállal lehet szabályozni valamint ezen felül egy konstans fékező erőnek is lennie kell tehát a fékező erő akkor sem nulla ha a fékpedál állása nulla ez utóbbit mindegy minek nevezzük légellenállás csúszási súrlódási erő a kettő eredője stb nem kell cicomázni de legyen külön feladat észben tartatni és célszerűen előre felkészülni arra hogy a sprintes vezetéstámogató modulok lka aac aeb pp közvetlenül az egocar hajtásláncára és a kormányra hatnak fel kell készíteni a modult ilyen „nem a hmi ről” érkező inputok kezelésére is amelyek ráadásul magasabb prioritásúak pl ha a vészfékező rendszer „lenyomja” a fékpedált akkor magasabb prioritással kell kezelni mint a hmi ről érkező pedálállást definition of done az autó gázpedál állásától függően gyorsul a fék és gázpedál állapota a billentyű nyomva tartásának idejével szabályozható fék és gázpedál valamint a kormány sem binárisan működik a billentyű nyomva tartás idejétől függ az input intenzitása a fék és gázpedál skálán a kormányelforgatás skálán kerül meghatározásra fék és gázpedál valamint a kormány is fokozatosan másodperc áll vissza alaphelyzetbe a billentyű felengedésével az automata váltó állapota szabályozható szekvenciális váltóról lévén szó sorban állíthatók a fokozatok p ark r everse n eutral d rive alaphelyzet p „felváltás” után r majd n majd d „leváltás” ugyanez visszafele a gyorsulás a „belső fokozatok” szerint kerül meghatározásra az autó a gázpedál felengedésével fokozatosan lassul majd megáll az autó r válóállásban tolat felkészíteni a modult hogy a vészfékező az adaptív tempomat vagy a sávtartó automatika is küldhet inputot melyek magasabb prioritásúak vészfékező értelemszerűen fékezés inputot az adaptív tempomat és a parkoló asszisztens gáz és fék inputot is a sávtartó automatika a kormányállást módosítja autó kanyarodásának biztosítása valóságos fordulókör szerint ehhez szükséges extra tulajdonságok meghatározása a meghatározott mozgásvektor alapján az autó pozíciójának frissítése ez az automatedcar osztály x y koordinátáinak frissítését jelenti tolatás során is valósághű kanyarodás történik
0
303,240
26,195,504,277
IssuesEvent
2023-01-03 13:06:26
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
[CI] GeoShapeGeoHexGridAggregatorTests testGeoShapeBounds failing
:Analytics/Geo >test-failure Team:Analytics
**Build scan:** https://gradle-enterprise.elastic.co/s/c2u7pwivm6x2s/tests/:x-pack:plugin:spatial:test/org.elasticsearch.xpack.spatial.search.aggregations.bucket.geogrid.GeoShapeGeoHexGridAggregatorTests/testGeoShapeBounds **Reproduction line:** ``` ./gradlew ':x-pack:plugin:spatial:test' --tests "org.elasticsearch.xpack.spatial.search.aggregations.bucket.geogrid.GeoShapeGeoHexGridAggregatorTests.testGeoShapeBounds" -Dtests.seed=D8A104D02E2813FA -Dtests.locale=tr -Dtests.timezone=America/Creston -Druntime.java=17 ``` **Applicable branches:** main **Reproduces locally?:** Yes **Failure history:** https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.spatial.search.aggregations.bucket.geogrid.GeoShapeGeoHexGridAggregatorTests&tests.test=testGeoShapeBounds **Failure excerpt:** ``` java.lang.AssertionError: Expected: <44L> but: was <46L> at __randomizedtesting.SeedInfo.seed([D8A104D02E2813FA:5177EADB4AD71EB9]:0) at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18) at org.junit.Assert.assertThat(Assert.java:956) at org.junit.Assert.assertThat(Assert.java:923) at org.elasticsearch.xpack.spatial.search.aggregations.bucket.geogrid.GeoShapeGeoGridTestCase.lambda$testGeoShapeBounds$11(GeoShapeGeoGridTestCase.java:184) at org.elasticsearch.xpack.spatial.search.aggregations.bucket.geogrid.GeoShapeGeoGridTestCase.lambda$testCase$14(GeoShapeGeoGridTestCase.java:274) at org.elasticsearch.search.aggregations.AggregatorTestCase.testCase(AggregatorTestCase.java:653) at org.elasticsearch.xpack.spatial.search.aggregations.bucket.geogrid.GeoShapeGeoGridTestCase.testCase(GeoShapeGeoGridTestCase.java:272) at org.elasticsearch.xpack.spatial.search.aggregations.bucket.geogrid.GeoShapeGeoGridTestCase.testCase(GeoShapeGeoGridTestCase.java:245) at org.elasticsearch.xpack.spatial.search.aggregations.bucket.geogrid.GeoShapeGeoGridTestCase.testGeoShapeBounds(GeoShapeGeoGridTestCase.java:174) at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2) at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:568) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45) at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60) at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44) at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60) at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850) at java.lang.Thread.run(Thread.java:833) ```
1.0
[CI] GeoShapeGeoHexGridAggregatorTests testGeoShapeBounds failing - **Build scan:** https://gradle-enterprise.elastic.co/s/c2u7pwivm6x2s/tests/:x-pack:plugin:spatial:test/org.elasticsearch.xpack.spatial.search.aggregations.bucket.geogrid.GeoShapeGeoHexGridAggregatorTests/testGeoShapeBounds **Reproduction line:** ``` ./gradlew ':x-pack:plugin:spatial:test' --tests "org.elasticsearch.xpack.spatial.search.aggregations.bucket.geogrid.GeoShapeGeoHexGridAggregatorTests.testGeoShapeBounds" -Dtests.seed=D8A104D02E2813FA -Dtests.locale=tr -Dtests.timezone=America/Creston -Druntime.java=17 ``` **Applicable branches:** main **Reproduces locally?:** Yes **Failure history:** https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.spatial.search.aggregations.bucket.geogrid.GeoShapeGeoHexGridAggregatorTests&tests.test=testGeoShapeBounds **Failure excerpt:** ``` java.lang.AssertionError: Expected: <44L> but: was <46L> at __randomizedtesting.SeedInfo.seed([D8A104D02E2813FA:5177EADB4AD71EB9]:0) at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18) at org.junit.Assert.assertThat(Assert.java:956) at org.junit.Assert.assertThat(Assert.java:923) at org.elasticsearch.xpack.spatial.search.aggregations.bucket.geogrid.GeoShapeGeoGridTestCase.lambda$testGeoShapeBounds$11(GeoShapeGeoGridTestCase.java:184) at org.elasticsearch.xpack.spatial.search.aggregations.bucket.geogrid.GeoShapeGeoGridTestCase.lambda$testCase$14(GeoShapeGeoGridTestCase.java:274) at org.elasticsearch.search.aggregations.AggregatorTestCase.testCase(AggregatorTestCase.java:653) at org.elasticsearch.xpack.spatial.search.aggregations.bucket.geogrid.GeoShapeGeoGridTestCase.testCase(GeoShapeGeoGridTestCase.java:272) at org.elasticsearch.xpack.spatial.search.aggregations.bucket.geogrid.GeoShapeGeoGridTestCase.testCase(GeoShapeGeoGridTestCase.java:245) at org.elasticsearch.xpack.spatial.search.aggregations.bucket.geogrid.GeoShapeGeoGridTestCase.testGeoShapeBounds(GeoShapeGeoGridTestCase.java:174) at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2) at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:568) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45) at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60) at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44) at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60) at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850) at java.lang.Thread.run(Thread.java:833) ```
test
geoshapegeohexgridaggregatortests testgeoshapebounds failing build scan reproduction line gradlew x pack plugin spatial test tests org elasticsearch xpack spatial search aggregations bucket geogrid geoshapegeohexgridaggregatortests testgeoshapebounds dtests seed dtests locale tr dtests timezone america creston druntime java applicable branches main reproduces locally yes failure history failure excerpt java lang assertionerror expected but was at randomizedtesting seedinfo seed at org hamcrest matcherassert assertthat matcherassert java at org junit assert assertthat assert java at org junit assert assertthat assert java at org elasticsearch xpack spatial search aggregations bucket geogrid geoshapegeogridtestcase lambda testgeoshapebounds geoshapegeogridtestcase java at org elasticsearch xpack spatial search aggregations bucket geogrid geoshapegeogridtestcase lambda testcase geoshapegeogridtestcase java at org elasticsearch search aggregations aggregatortestcase testcase aggregatortestcase java at org elasticsearch xpack spatial search aggregations bucket geogrid geoshapegeogridtestcase testcase geoshapegeogridtestcase java at org elasticsearch xpack spatial search aggregations bucket geogrid geoshapegeogridtestcase testcase geoshapegeogridtestcase java at org elasticsearch xpack spatial search aggregations bucket geogrid geoshapegeogridtestcase testgeoshapebounds geoshapegeogridtestcase java at jdk internal reflect nativemethodaccessorimpl nativemethodaccessorimpl java at jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulesetupteardownchained evaluate testrulesetupteardownchained java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulethreadandtestname evaluate testrulethreadandtestname java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol forktimeoutingtask threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol evaluate threadleakcontrol java at com carrotsearch randomizedtesting randomizedrunner runsingletest randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulestoreclassname evaluate testrulestoreclassname java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testruleassertionsrequired evaluate testruleassertionsrequired java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testruleignoretestsuites evaluate testruleignoretestsuites java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol lambda forktimeoutingtask threadleakcontrol java at java lang thread run thread java
1
17,961
24,799,073,895
IssuesEvent
2022-10-24 19:57:59
programmer2514/BetterDiscord-CollapsibleUI
https://api.github.com/repos/programmer2514/BetterDiscord-CollapsibleUI
closed
When some other BetterDiscord plugin gets update - CollapsibleUI stops working until you toggle it off/on
bug compatibility
Steps for reproduction. CollapsibleUI version 5.7.2. BD Core version 1.8.2 1. While everything works fine - wait for some plugin to get updated so BetterDiscord notification pops up ![image](https://user-images.githubusercontent.com/33165440/196707210-61d9c76d-14ae-4920-b849-8863e96c4857.png) 2. Click "All" - plugins above get updates just fine. 3. CollapsibleUI's buttons are gone ![image](https://user-images.githubusercontent.com/33165440/196707594-494d7a3e-c771-4d4c-913e-20b87aff26ca.png) 4. But the plugin is still on in settings ![image](https://user-images.githubusercontent.com/33165440/196707481-91f858a9-63d7-499a-a793-c38406464da4.png) 5. Toggling off/on brings it back to working state PS. this issue is a summary of the [support thread](https://discord.com/channels/86004744966914048/1030944619711713310/1030944619711713310) in forum channel of BetterDiscord server
True
When some other BetterDiscord plugin gets update - CollapsibleUI stops working until you toggle it off/on - Steps for reproduction. CollapsibleUI version 5.7.2. BD Core version 1.8.2 1. While everything works fine - wait for some plugin to get updated so BetterDiscord notification pops up ![image](https://user-images.githubusercontent.com/33165440/196707210-61d9c76d-14ae-4920-b849-8863e96c4857.png) 2. Click "All" - plugins above get updates just fine. 3. CollapsibleUI's buttons are gone ![image](https://user-images.githubusercontent.com/33165440/196707594-494d7a3e-c771-4d4c-913e-20b87aff26ca.png) 4. But the plugin is still on in settings ![image](https://user-images.githubusercontent.com/33165440/196707481-91f858a9-63d7-499a-a793-c38406464da4.png) 5. Toggling off/on brings it back to working state PS. this issue is a summary of the [support thread](https://discord.com/channels/86004744966914048/1030944619711713310/1030944619711713310) in forum channel of BetterDiscord server
non_test
when some other betterdiscord plugin gets update collapsibleui stops working until you toggle it off on steps for reproduction collapsibleui version bd core version while everything works fine wait for some plugin to get updated so betterdiscord notification pops up click all plugins above get updates just fine collapsibleui s buttons are gone but the plugin is still on in settings toggling off on brings it back to working state ps this issue is a summary of the in forum channel of betterdiscord server
0
186,617
14,401,912,059
IssuesEvent
2020-12-03 14:18:30
NVIDIA/spark-rapids
https://api.github.com/repos/NVIDIA/spark-rapids
closed
[FEA] commonize the version checking in the integration tests
P2 feature request test
**Is your feature request related to a problem? Please describe.** currently we have a couple ways to checks versions. 1. --runtime-env option 2. is_spark_300 and is_before_spark_310 in spark_session.py I think we could get rid of both of those and use the shim layer versioning checks like we do in the unit tests. We should investigate and try to make things common.
1.0
[FEA] commonize the version checking in the integration tests - **Is your feature request related to a problem? Please describe.** currently we have a couple ways to checks versions. 1. --runtime-env option 2. is_spark_300 and is_before_spark_310 in spark_session.py I think we could get rid of both of those and use the shim layer versioning checks like we do in the unit tests. We should investigate and try to make things common.
test
commonize the version checking in the integration tests is your feature request related to a problem please describe currently we have a couple ways to checks versions runtime env option is spark and is before spark in spark session py i think we could get rid of both of those and use the shim layer versioning checks like we do in the unit tests we should investigate and try to make things common
1
137,662
11,149,591,541
IssuesEvent
2019-12-23 19:15:58
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
runSecurityScans is missing RBAC
[zube]: To Test alpha area/scan-tool kind/bug-qa team/ca
**What kind of request is this (question/bug/enhancement/feature request):** bug **Steps to reproduce (least amount of steps as possible):** As a cluster member, do a view in API on the cluster, and click on `runSecurityScan` from the actions listed on the page. **Expected Result:** A cluster member should NOT be able to perform `runSecurityScan` action on the cluster **Actual Result:** A cluster member is able to perform `runSecurityScan` action on the cluster **Note:** But the user is NOT able to view the scan reports. **Other details that may be helpful:** **Environment information** - Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): master-head - latest - Installation option (single install/HA): single <!-- If the reported issue is regarding a created cluster, please provide requested info below --> **Cluster information** - Cluster type (Hosted/Infrastructure Provider/Custom/Imported): custom - Kubernetes version (use `kubectl version`): ``` 1.16 ```
1.0
runSecurityScans is missing RBAC - **What kind of request is this (question/bug/enhancement/feature request):** bug **Steps to reproduce (least amount of steps as possible):** As a cluster member, do a view in API on the cluster, and click on `runSecurityScan` from the actions listed on the page. **Expected Result:** A cluster member should NOT be able to perform `runSecurityScan` action on the cluster **Actual Result:** A cluster member is able to perform `runSecurityScan` action on the cluster **Note:** But the user is NOT able to view the scan reports. **Other details that may be helpful:** **Environment information** - Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): master-head - latest - Installation option (single install/HA): single <!-- If the reported issue is regarding a created cluster, please provide requested info below --> **Cluster information** - Cluster type (Hosted/Infrastructure Provider/Custom/Imported): custom - Kubernetes version (use `kubectl version`): ``` 1.16 ```
test
runsecurityscans is missing rbac what kind of request is this question bug enhancement feature request bug steps to reproduce least amount of steps as possible as a cluster member do a view in api on the cluster and click on runsecurityscan from the actions listed on the page expected result a cluster member should not be able to perform runsecurityscan action on the cluster actual result a cluster member is able to perform runsecurityscan action on the cluster note but the user is not able to view the scan reports other details that may be helpful environment information rancher version rancher rancher rancher server image tag or shown bottom left in the ui master head latest installation option single install ha single if the reported issue is regarding a created cluster please provide requested info below cluster information cluster type hosted infrastructure provider custom imported custom kubernetes version use kubectl version
1
403,533
27,422,374,942
IssuesEvent
2023-03-01 17:36:00
hashicorp/terraform-provider-aws
https://api.github.com/repos/hashicorp/terraform-provider-aws
closed
[Docs]:
documentation service/acm needs-triage
### Documentation Link https://registry.terraform.io/providers/hashicorp/aws/latest/docs/resources/acm_certificate ### Description In the the attribute reference section, the ID is described as ARN value which is not correct. [id](https://registry.terraform.io/providers/hashicorp/aws/latest/docs/resources/acm_certificate#id) - ARN of the certificate This should be changed to [id](https://registry.terraform.io/providers/hashicorp/aws/latest/docs/resources/acm_certificate#id) - ID of the certificate ### References _No response_ ### Would you like to implement a fix? Yes
1.0
[Docs]: - ### Documentation Link https://registry.terraform.io/providers/hashicorp/aws/latest/docs/resources/acm_certificate ### Description In the the attribute reference section, the ID is described as ARN value which is not correct. [id](https://registry.terraform.io/providers/hashicorp/aws/latest/docs/resources/acm_certificate#id) - ARN of the certificate This should be changed to [id](https://registry.terraform.io/providers/hashicorp/aws/latest/docs/resources/acm_certificate#id) - ID of the certificate ### References _No response_ ### Would you like to implement a fix? Yes
non_test
documentation link description in the the attribute reference section the id is described as arn value which is not correct arn of the certificate this should be changed to id of the certificate references no response would you like to implement a fix yes
0
346,950
10,422,136,298
IssuesEvent
2019-09-16 08:18:06
getkirby/kirby
https://api.github.com/repos/getkirby/kirby
closed
collection's append() and prepend() behave differently when only one argument is provided
difficulty: medium 🤞 priority: medium 🔜 type: bug 🐛
**Describe the bug** _Note: this issue discusses append() and prepend() methods of collection class, BUT, when following code links from the $page->append() and $page->prepend() reference pages, one leads to 'kirby/src/Cms/Collection.php' while the other leads to 'kirby/src/Toolkit/Collection.php'. I do not know the implications of this, but I assume it is correct, and those are the methods we are using when we do $pages->append($page) or prepend($page)_ The methods append() and prepend() from the collection class behave differently when only one argument (for example a page via $pages->append($page)) is provided. append() checks if the provided element is an object, and if it is, it looks for the object's ID and uses it as key when adding the object to the data array; while prepend() uses array_unshift() effectively adding the element (object or not) without a key, so the key becomes zero (0). [kirby/src/Cms/Collection.php](https://github.com/getkirby/kirby/blob/3.2.4/src/Cms/Collection.php#L102) ``` /** * Appends an element to the data array * * @param mixed $key * @param mixed $item * @return Kirby\Cms\Collection */ public function append(...$args) { if (count($args) === 1) { if (is_object($args[0]) === true) { $this->data[$args[0]->id()] = $args[0]; } else { $this->data[] = $args[0]; } } elseif (count($args) === 2) { $this->set($args[0], $args[1]); } return $this; } ``` [kirby/src/Toolkit/Collection.php](https://github.com/getkirby/kirby/blob/3.2.4/src/Toolkit/Collection.php#L709) ``` /** * Prepends an element to the data array * * @param mixed $key * @param mixed $item * @return self */ public function prepend(...$args) { if (count($args) === 1) { array_unshift($this->data, $args[0]); } elseif (count($args) === 2) { $data = $this->data; $this->data = []; $this->set($args[0], $args[1]); $this->data += $data; } return $this; } ``` **To Reproduce** Steps to reproduce the behavior: 1. Append a page to a $pages object, without providing a key, and dump 2. Prepend a page to a $pages object, without providing a key, and dump 3. Compare the dumps 4. Notice in append's dump the key is the page's ID, while in prepend's dump. the key is zero (0) **Expected behavior** Either both work in the same way, checking for an object and looking for its key if it not provided, or neither do, and then possibly throw an Exception, as I believe K2 did. A possible rewrite of prepend() (bear with my low php skills, please) : ``` public function prepend(...$args) { if (count($args) === 1) { if (is_object($args[0]) === true) { $this->prepend($args[0]->id(), $args[0]); } else { array_unshift($this->data, $args[0]); } } elseif (count($args) === 2) { $data = $this->data; $this->data = []; $this->set($args[0], $args[1]); $this->data += $data; } return $this; } ``` Alternatively this difference could be explained in the reference page of the methods. Right now this difference is not visible unless one checks the code, and it already confused other people than myself, see [here](https://github.com/getkirby/kirby/issues/1653). **Kirby Version** "3.1.2-rc.1" but the quoted code is the same in the last version in this repo **Desktop (please complete the following information):** - OS: Linux Mint cinnaon 19.2 **Additional context** [Related forum question.](https://forum.getkirby.com/t/pages-add-vs-append-and-prepend/15608/8 )
1.0
collection's append() and prepend() behave differently when only one argument is provided - **Describe the bug** _Note: this issue discusses append() and prepend() methods of collection class, BUT, when following code links from the $page->append() and $page->prepend() reference pages, one leads to 'kirby/src/Cms/Collection.php' while the other leads to 'kirby/src/Toolkit/Collection.php'. I do not know the implications of this, but I assume it is correct, and those are the methods we are using when we do $pages->append($page) or prepend($page)_ The methods append() and prepend() from the collection class behave differently when only one argument (for example a page via $pages->append($page)) is provided. append() checks if the provided element is an object, and if it is, it looks for the object's ID and uses it as key when adding the object to the data array; while prepend() uses array_unshift() effectively adding the element (object or not) without a key, so the key becomes zero (0). [kirby/src/Cms/Collection.php](https://github.com/getkirby/kirby/blob/3.2.4/src/Cms/Collection.php#L102) ``` /** * Appends an element to the data array * * @param mixed $key * @param mixed $item * @return Kirby\Cms\Collection */ public function append(...$args) { if (count($args) === 1) { if (is_object($args[0]) === true) { $this->data[$args[0]->id()] = $args[0]; } else { $this->data[] = $args[0]; } } elseif (count($args) === 2) { $this->set($args[0], $args[1]); } return $this; } ``` [kirby/src/Toolkit/Collection.php](https://github.com/getkirby/kirby/blob/3.2.4/src/Toolkit/Collection.php#L709) ``` /** * Prepends an element to the data array * * @param mixed $key * @param mixed $item * @return self */ public function prepend(...$args) { if (count($args) === 1) { array_unshift($this->data, $args[0]); } elseif (count($args) === 2) { $data = $this->data; $this->data = []; $this->set($args[0], $args[1]); $this->data += $data; } return $this; } ``` **To Reproduce** Steps to reproduce the behavior: 1. Append a page to a $pages object, without providing a key, and dump 2. Prepend a page to a $pages object, without providing a key, and dump 3. Compare the dumps 4. Notice in append's dump the key is the page's ID, while in prepend's dump. the key is zero (0) **Expected behavior** Either both work in the same way, checking for an object and looking for its key if it not provided, or neither do, and then possibly throw an Exception, as I believe K2 did. A possible rewrite of prepend() (bear with my low php skills, please) : ``` public function prepend(...$args) { if (count($args) === 1) { if (is_object($args[0]) === true) { $this->prepend($args[0]->id(), $args[0]); } else { array_unshift($this->data, $args[0]); } } elseif (count($args) === 2) { $data = $this->data; $this->data = []; $this->set($args[0], $args[1]); $this->data += $data; } return $this; } ``` Alternatively this difference could be explained in the reference page of the methods. Right now this difference is not visible unless one checks the code, and it already confused other people than myself, see [here](https://github.com/getkirby/kirby/issues/1653). **Kirby Version** "3.1.2-rc.1" but the quoted code is the same in the last version in this repo **Desktop (please complete the following information):** - OS: Linux Mint cinnaon 19.2 **Additional context** [Related forum question.](https://forum.getkirby.com/t/pages-add-vs-append-and-prepend/15608/8 )
non_test
collection s append and prepend behave differently when only one argument is provided describe the bug note this issue discusses append and prepend methods of collection class but when following code links from the page append and page prepend reference pages one leads to kirby src cms collection php while the other leads to kirby src toolkit collection php i do not know the implications of this but i assume it is correct and those are the methods we are using when we do pages append page or prepend page the methods append and prepend from the collection class behave differently when only one argument for example a page via pages append page is provided append checks if the provided element is an object and if it is it looks for the object s id and uses it as key when adding the object to the data array while prepend uses array unshift effectively adding the element object or not without a key so the key becomes zero appends an element to the data array param mixed key param mixed item return kirby cms collection public function append args if count args if is object args true this data id args else this data args elseif count args this set args args return this prepends an element to the data array param mixed key param mixed item return self public function prepend args if count args array unshift this data args elseif count args data this data this data this set args args this data data return this to reproduce steps to reproduce the behavior append a page to a pages object without providing a key and dump prepend a page to a pages object without providing a key and dump compare the dumps notice in append s dump the key is the page s id while in prepend s dump the key is zero expected behavior either both work in the same way checking for an object and looking for its key if it not provided or neither do and then possibly throw an exception as i believe did a possible rewrite of prepend bear with my low php skills please public function prepend args if count args if is object args true this prepend args id args else array unshift this data args elseif count args data this data this data this set args args this data data return this alternatively this difference could be explained in the reference page of the methods right now this difference is not visible unless one checks the code and it already confused other people than myself see kirby version rc but the quoted code is the same in the last version in this repo desktop please complete the following information os linux mint cinnaon additional context
0
194,749
6,898,424,625
IssuesEvent
2017-11-24 09:28:22
ballerinalang/composer
https://api.github.com/repos/ballerinalang/composer
closed
Add placeholders for all the properties in the annotations
Priority/Normal Severity/Minor Type/Bug
Add placeholders for all the properties in the annotations so that the user knows what to enter and the format to enter. eg: for path property - add the path with a slash in front ![path2](https://user-images.githubusercontent.com/1845370/31331058-8a6f58d2-acfe-11e7-956c-b1f906f7443a.png) ![cons](https://user-images.githubusercontent.com/1845370/31331049-83c19cb6-acfe-11e7-80c6-3d8d564da128.png)
1.0
Add placeholders for all the properties in the annotations - Add placeholders for all the properties in the annotations so that the user knows what to enter and the format to enter. eg: for path property - add the path with a slash in front ![path2](https://user-images.githubusercontent.com/1845370/31331058-8a6f58d2-acfe-11e7-956c-b1f906f7443a.png) ![cons](https://user-images.githubusercontent.com/1845370/31331049-83c19cb6-acfe-11e7-80c6-3d8d564da128.png)
non_test
add placeholders for all the properties in the annotations add placeholders for all the properties in the annotations so that the user knows what to enter and the format to enter eg for path property add the path with a slash in front
0
201,626
15,215,960,204
IssuesEvent
2021-02-17 14:59:29
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
[Failing Test] [sig-storage] CSI Volumes pd-csi driver tests failing
kind/failing-test needs-triage priority/important-soon sig/storage
<!-- Please only use this template for submitting reports about continuously failing tests or jobs in Kubernetes CI --> **Which jobs are failing**: `ci-kubernetes-e2e-gci-gce-serial` **Which test(s) are failing**: - `[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works` - `[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works` - `[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it` - `[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it` **Since when has it been failing**: Feb 6 16:57PST 5fa704c6a **Testgrid link**: https://testgrid.k8s.io/sig-release-master-informing#gce-cos-master-serial&include-filter-by-regex=CSI%20Volumes **Reason for failure**: https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-kubernetes-e2e-gci-gce-serial/1358218259307958272#1:build-log.txt%3A1759 ``` I0207 01:13:51.104] STEP: creating a StorageClass volume-expand-40955dxt2 I0207 01:13:51.149] STEP: creating a claim I0207 01:13:51.150] Feb 7 01:13:51.149: INFO: Warning: Making PVC: VolumeMode specified as invalid empty string, treating as nil I0207 01:13:51.254] STEP: Creating a pod with dynamically provisioned volume I0207 01:13:51.964] WARNING: pod log: csi-gce-pd-node-m2c4h/csi-driver-registrar: the server does not allow this method on the requested resource ( pods/log csi-gce-pd-node-m2c4h) I0207 01:13:52.005] WARNING: pod log: csi-gce-pd-node-m2c4h/gce-pd-driver: the server does not allow this method on the requested resource ( pods/log csi-gce-pd-node-m2c4h) ``` This warning repeats until we get to ... here https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-kubernetes-e2e-gci-gce-serial/1358218259307958272#1:build-log.txt%3A1773 ``` I0207 01:14:29.413] STEP: Deleting the previously created pod I0207 01:14:29.413] Feb 7 01:14:29.413: INFO: Deleting pod "pod-0016e560-e9c7-4d49-bb59-03de4438d5a8" in namespace "volume-expand-4095" I0207 01:14:29.454] Feb 7 01:14:29.454: INFO: Wait up to 5m0s for pod "pod-0016e560-e9c7-4d49-bb59-03de4438d5a8" to be fully deleted I0207 01:14:39.532] STEP: Expanding current pvc I0207 01:14:39.533] Feb 7 01:14:39.532: INFO: currentPvcSize {{5368709120 0} {<nil>} 5Gi BinarySI}, newSize {{6442450944 0} {<nil>} BinarySI} I0207 01:14:39.619] STEP: Waiting for cloudprovider resize to finish I0207 01:19:20.523] WARNING: pod log: csi-gce-pd-node-m2c4h/csi-driver-registrar: the server does not allow this method on the requested resource ( pods/log csi-gce-pd-node-m2c4h) I0207 01:19:20.563] WARNING: pod log: csi-gce-pd-node-m2c4h/gce-pd-driver: the server does not allow this method on the requested resource ( pods/log csi-gce-pd-node-m2c4h) ``` And it just repeats those two warnings until it times out. **Anything else we need to know**: Most likely related to #98805 ? /sig storage /priority important-soon
1.0
[Failing Test] [sig-storage] CSI Volumes pd-csi driver tests failing - <!-- Please only use this template for submitting reports about continuously failing tests or jobs in Kubernetes CI --> **Which jobs are failing**: `ci-kubernetes-e2e-gci-gce-serial` **Which test(s) are failing**: - `[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works` - `[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works` - `[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it` - `[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it` **Since when has it been failing**: Feb 6 16:57PST 5fa704c6a **Testgrid link**: https://testgrid.k8s.io/sig-release-master-informing#gce-cos-master-serial&include-filter-by-regex=CSI%20Volumes **Reason for failure**: https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-kubernetes-e2e-gci-gce-serial/1358218259307958272#1:build-log.txt%3A1759 ``` I0207 01:13:51.104] STEP: creating a StorageClass volume-expand-40955dxt2 I0207 01:13:51.149] STEP: creating a claim I0207 01:13:51.150] Feb 7 01:13:51.149: INFO: Warning: Making PVC: VolumeMode specified as invalid empty string, treating as nil I0207 01:13:51.254] STEP: Creating a pod with dynamically provisioned volume I0207 01:13:51.964] WARNING: pod log: csi-gce-pd-node-m2c4h/csi-driver-registrar: the server does not allow this method on the requested resource ( pods/log csi-gce-pd-node-m2c4h) I0207 01:13:52.005] WARNING: pod log: csi-gce-pd-node-m2c4h/gce-pd-driver: the server does not allow this method on the requested resource ( pods/log csi-gce-pd-node-m2c4h) ``` This warning repeats until we get to ... here https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-kubernetes-e2e-gci-gce-serial/1358218259307958272#1:build-log.txt%3A1773 ``` I0207 01:14:29.413] STEP: Deleting the previously created pod I0207 01:14:29.413] Feb 7 01:14:29.413: INFO: Deleting pod "pod-0016e560-e9c7-4d49-bb59-03de4438d5a8" in namespace "volume-expand-4095" I0207 01:14:29.454] Feb 7 01:14:29.454: INFO: Wait up to 5m0s for pod "pod-0016e560-e9c7-4d49-bb59-03de4438d5a8" to be fully deleted I0207 01:14:39.532] STEP: Expanding current pvc I0207 01:14:39.533] Feb 7 01:14:39.532: INFO: currentPvcSize {{5368709120 0} {<nil>} 5Gi BinarySI}, newSize {{6442450944 0} {<nil>} BinarySI} I0207 01:14:39.619] STEP: Waiting for cloudprovider resize to finish I0207 01:19:20.523] WARNING: pod log: csi-gce-pd-node-m2c4h/csi-driver-registrar: the server does not allow this method on the requested resource ( pods/log csi-gce-pd-node-m2c4h) I0207 01:19:20.563] WARNING: pod log: csi-gce-pd-node-m2c4h/gce-pd-driver: the server does not allow this method on the requested resource ( pods/log csi-gce-pd-node-m2c4h) ``` And it just repeats those two warnings until it times out. **Anything else we need to know**: Most likely related to #98805 ? /sig storage /priority important-soon
test
csi volumes pd csi driver tests failing which jobs are failing ci kubernetes gci gce serial which test s are failing csi volumes volume expand verify if offline pvc expansion works csi volumes volume expand verify if offline pvc expansion works csi volumes volume expand should resize volume when pvc is edited while pod is using it csi volumes volume expand should resize volume when pvc is edited while pod is using it since when has it been failing feb testgrid link reason for failure step creating a storageclass volume expand step creating a claim feb info warning making pvc volumemode specified as invalid empty string treating as nil step creating a pod with dynamically provisioned volume warning pod log csi gce pd node csi driver registrar the server does not allow this method on the requested resource pods log csi gce pd node warning pod log csi gce pd node gce pd driver the server does not allow this method on the requested resource pods log csi gce pd node this warning repeats until we get to here step deleting the previously created pod feb info deleting pod pod in namespace volume expand feb info wait up to for pod pod to be fully deleted step expanding current pvc feb info currentpvcsize binarysi newsize binarysi step waiting for cloudprovider resize to finish warning pod log csi gce pd node csi driver registrar the server does not allow this method on the requested resource pods log csi gce pd node warning pod log csi gce pd node gce pd driver the server does not allow this method on the requested resource pods log csi gce pd node and it just repeats those two warnings until it times out anything else we need to know most likely related to sig storage priority important soon
1
300,454
25,969,850,335
IssuesEvent
2022-12-19 10:18:51
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: import/tpch/nodes=8 failed
C-test-failure O-robot O-roachtest release-blocker branch-release-22.1
roachtest.import/tpch/nodes=8 [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=8014310&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=8014310&tab=artifacts#/import/tpch/nodes=8) on release-22.1 @ [d84033d9ee0dc4f35901bcc2e2311b6288e84569](https://github.com/cockroachdb/cockroach/commits/d84033d9ee0dc4f35901bcc2e2311b6288e84569): ``` The test failed on branch=release-22.1, cloud=gce: test artifacts and logs in: /artifacts/import/tpch/nodes=8/run_1 monitor.go:127,import.go:312,test_runner.go:883: monitor failure: monitor command failure: unexpected node event: 6: dead (exit status 7) (1) attached stack trace -- stack trace: | main.(*monitorImpl).WaitE | main/pkg/cmd/roachtest/monitor.go:115 | main.(*monitorImpl).Wait | main/pkg/cmd/roachtest/monitor.go:123 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerImportTPCH.func1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/import.go:312 | [...repeated from below...] Wraps: (2) monitor failure Wraps: (3) attached stack trace -- stack trace: | main.(*monitorImpl).wait.func3 | main/pkg/cmd/roachtest/monitor.go:202 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1581 Wraps: (4) monitor command failure Wraps: (5) unexpected node event: 6: dead (exit status 7) Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *errors.errorString ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #90021 roachtest: import/tpch/nodes=8 failed [C-test-failure O-roachtest O-robot T-disaster-recovery branch-release-22.2] </p> </details> /cc @cockroachdb/disaster-recovery <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*import/tpch/nodes=8.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: import/tpch/nodes=8 failed - roachtest.import/tpch/nodes=8 [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=8014310&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=8014310&tab=artifacts#/import/tpch/nodes=8) on release-22.1 @ [d84033d9ee0dc4f35901bcc2e2311b6288e84569](https://github.com/cockroachdb/cockroach/commits/d84033d9ee0dc4f35901bcc2e2311b6288e84569): ``` The test failed on branch=release-22.1, cloud=gce: test artifacts and logs in: /artifacts/import/tpch/nodes=8/run_1 monitor.go:127,import.go:312,test_runner.go:883: monitor failure: monitor command failure: unexpected node event: 6: dead (exit status 7) (1) attached stack trace -- stack trace: | main.(*monitorImpl).WaitE | main/pkg/cmd/roachtest/monitor.go:115 | main.(*monitorImpl).Wait | main/pkg/cmd/roachtest/monitor.go:123 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerImportTPCH.func1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/import.go:312 | [...repeated from below...] Wraps: (2) monitor failure Wraps: (3) attached stack trace -- stack trace: | main.(*monitorImpl).wait.func3 | main/pkg/cmd/roachtest/monitor.go:202 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1581 Wraps: (4) monitor command failure Wraps: (5) unexpected node event: 6: dead (exit status 7) Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *errors.errorString ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #90021 roachtest: import/tpch/nodes=8 failed [C-test-failure O-roachtest O-robot T-disaster-recovery branch-release-22.2] </p> </details> /cc @cockroachdb/disaster-recovery <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*import/tpch/nodes=8.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
roachtest import tpch nodes failed roachtest import tpch nodes with on release the test failed on branch release cloud gce test artifacts and logs in artifacts import tpch nodes run monitor go import go test runner go monitor failure monitor command failure unexpected node event dead exit status attached stack trace stack trace main monitorimpl waite main pkg cmd roachtest monitor go main monitorimpl wait main pkg cmd roachtest monitor go github com cockroachdb cockroach pkg cmd roachtest tests registerimporttpch github com cockroachdb cockroach pkg cmd roachtest tests import go wraps monitor failure wraps attached stack trace stack trace main monitorimpl wait main pkg cmd roachtest monitor go runtime goexit goroot src runtime asm s wraps monitor command failure wraps unexpected node event dead exit status error types withstack withstack errutil withprefix withstack withstack errutil withprefix errors errorstring help see see same failure on other branches roachtest import tpch nodes failed cc cockroachdb disaster recovery
1
327,429
28,062,461,212
IssuesEvent
2023-03-29 13:27:47
eclipse-openj9/openj9
https://api.github.com/repos/eclipse-openj9/openj9
closed
[FFI/Test] A nested case intended for the exception handling code in upcall method
comp:test project:panama
As mentioned at https://github.com/eclipse-openj9/openj9/pull/16932#issuecomment-1476347858, we will add one test case to verify the nested exception case as follows: ``` callout callin callout callin (throws an exception) ``` to ensure the stacking of the jump buffers is working well as expected. FYI: @gacholio, @tajila, @pshipton
1.0
[FFI/Test] A nested case intended for the exception handling code in upcall method - As mentioned at https://github.com/eclipse-openj9/openj9/pull/16932#issuecomment-1476347858, we will add one test case to verify the nested exception case as follows: ``` callout callin callout callin (throws an exception) ``` to ensure the stacking of the jump buffers is working well as expected. FYI: @gacholio, @tajila, @pshipton
test
a nested case intended for the exception handling code in upcall method as mentioned at we will add one test case to verify the nested exception case as follows callout callin callout callin throws an exception to ensure the stacking of the jump buffers is working well as expected fyi gacholio tajila pshipton
1
52,285
6,226,185,217
IssuesEvent
2017-07-10 17:53:15
aquaaerobic/AquaNereda
https://api.github.com/repos/aquaaerobic/AquaNereda
closed
APD Labor and Burden
Active Resolved \ Please Test
Labor and burden estimates are not being calculated in APD (markup codes 18 and 19). Please review this as a part of updating the price worksheet.
1.0
APD Labor and Burden - Labor and burden estimates are not being calculated in APD (markup codes 18 and 19). Please review this as a part of updating the price worksheet.
test
apd labor and burden labor and burden estimates are not being calculated in apd markup codes and please review this as a part of updating the price worksheet
1
49,970
13,495,488,291
IssuesEvent
2020-09-12 00:02:34
Azure/AKS
https://api.github.com/repos/Azure/AKS
closed
az aks get-credentials should return unique credentials for each user instead of returning shared credentials
client/AzCLI resolution/answer-provided security
For General Availability of AKS, will `az aks get-credentials` enforce kubectl to connect with credentials unique to each AAD user logged in via `az login` instead of returning shared credentials that are the same for all users as it does at the moment?<br><br> I'm concerned that production clusters need to be secure at General Availability such that kubectl credentials to access the cluster can't walk out of the organisation. Ideally, if an AAD account is disabled, the key in the *.kube\config* file should no longer allow cluster access. This might be RBAC-AAD integration, or it might be something simpler for GA. One compromise might be to continue to use shared credentials (for now) but create the ability to easily cycle/refresh the credentials that `az aks get-credentials` returns so that older *.kube\config* file contents can no longer be used to access the cluster.<br><br> Alternatively, is there something I've missed with AKS that can stop *anyone* being able to access the cluster *from anywhere* as long as they have a copy of the *currently non-revocable* credentials in that file? (As mentioned @neilpeterson)
True
az aks get-credentials should return unique credentials for each user instead of returning shared credentials - For General Availability of AKS, will `az aks get-credentials` enforce kubectl to connect with credentials unique to each AAD user logged in via `az login` instead of returning shared credentials that are the same for all users as it does at the moment?<br><br> I'm concerned that production clusters need to be secure at General Availability such that kubectl credentials to access the cluster can't walk out of the organisation. Ideally, if an AAD account is disabled, the key in the *.kube\config* file should no longer allow cluster access. This might be RBAC-AAD integration, or it might be something simpler for GA. One compromise might be to continue to use shared credentials (for now) but create the ability to easily cycle/refresh the credentials that `az aks get-credentials` returns so that older *.kube\config* file contents can no longer be used to access the cluster.<br><br> Alternatively, is there something I've missed with AKS that can stop *anyone* being able to access the cluster *from anywhere* as long as they have a copy of the *currently non-revocable* credentials in that file? (As mentioned @neilpeterson)
non_test
az aks get credentials should return unique credentials for each user instead of returning shared credentials for general availability of aks will az aks get credentials enforce kubectl to connect with credentials unique to each aad user logged in via az login instead of returning shared credentials that are the same for all users as it does at the moment i m concerned that production clusters need to be secure at general availability such that kubectl credentials to access the cluster can t walk out of the organisation ideally if an aad account is disabled the key in the kube config file should no longer allow cluster access this might be rbac aad integration or it might be something simpler for ga one compromise might be to continue to use shared credentials for now but create the ability to easily cycle refresh the credentials that az aks get credentials returns so that older kube config file contents can no longer be used to access the cluster alternatively is there something i ve missed with aks that can stop anyone being able to access the cluster from anywhere as long as they have a copy of the currently non revocable credentials in that file as mentioned neilpeterson
0
400,794
11,780,774,280
IssuesEvent
2020-03-16 20:49:35
kir-dev/pek-next
https://api.github.com/repos/kir-dev/pek-next
opened
Use a state management gem to update statuses
priority: high refactor
A magyar leírás lent található ## English description To update group / SVIE status, currently we modify the DB directly. We should have transitions defined and set the states accordingly. We can use a gem like [aasm](https://github.com/aasm/aasm) for this. --- ## Magyar leírás Jelenleg a kör / SVIE státuszok állítása közvetlen attribute-ok állítgatásából áll. Ehelyett valamilyen state management eszközt pl. [aasm](https://github.com/aasm/aasm) kéne használni.
1.0
Use a state management gem to update statuses - A magyar leírás lent található ## English description To update group / SVIE status, currently we modify the DB directly. We should have transitions defined and set the states accordingly. We can use a gem like [aasm](https://github.com/aasm/aasm) for this. --- ## Magyar leírás Jelenleg a kör / SVIE státuszok állítása közvetlen attribute-ok állítgatásából áll. Ehelyett valamilyen state management eszközt pl. [aasm](https://github.com/aasm/aasm) kéne használni.
non_test
use a state management gem to update statuses a magyar leírás lent található english description to update group svie status currently we modify the db directly we should have transitions defined and set the states accordingly we can use a gem like for this magyar leírás jelenleg a kör svie státuszok állítása közvetlen attribute ok állítgatásából áll ehelyett valamilyen state management eszközt pl kéne használni
0
660,954
22,036,746,761
IssuesEvent
2022-05-28 18:02:21
pystardust/ani-cli
https://api.github.com/repos/pystardust/ani-cli
opened
VLC sound issues
type: bug priority 2: medium
**Metadata** Version: 2.2.0 OS: Gentoo Linux Shell: Bash Anime: Shingeki no Kyojin The Final Season part 2 **Describe the bug** When playing any episode of the final season part 2 of Aot, VLC starts acting weird. Seeking or skipping ahead by any amount breaks sound, and then I am completely unable to change the player position with the video playing. Seems to only happen with this anime. **Steps To Reproduce** 1. Run `ani-cli -d flcl` 2. Choose AOT Final Season part 2 3. Select any episode 4. Skip ahead or seek **Expected behavior** Sound and player position should work fine
1.0
VLC sound issues - **Metadata** Version: 2.2.0 OS: Gentoo Linux Shell: Bash Anime: Shingeki no Kyojin The Final Season part 2 **Describe the bug** When playing any episode of the final season part 2 of Aot, VLC starts acting weird. Seeking or skipping ahead by any amount breaks sound, and then I am completely unable to change the player position with the video playing. Seems to only happen with this anime. **Steps To Reproduce** 1. Run `ani-cli -d flcl` 2. Choose AOT Final Season part 2 3. Select any episode 4. Skip ahead or seek **Expected behavior** Sound and player position should work fine
non_test
vlc sound issues metadata version os gentoo linux shell bash anime shingeki no kyojin the final season part describe the bug when playing any episode of the final season part of aot vlc starts acting weird seeking or skipping ahead by any amount breaks sound and then i am completely unable to change the player position with the video playing seems to only happen with this anime steps to reproduce run ani cli d flcl choose aot final season part select any episode skip ahead or seek expected behavior sound and player position should work fine
0
818,409
30,687,807,679
IssuesEvent
2023-07-26 13:26:35
opendatahub-io/odh-dashboard
https://api.github.com/repos/opendatahub-io/odh-dashboard
closed
[Bug]: Error message for project deletion appears in the modal only in the next deletion attempt
kind/bug priority/normal feature/ds-projects field-priority
### Is there an existing issue for this? - [X] I have searched the existing issues ### Deploy type OpenDataHub core version (eg. `v1.6.0`) ### Version 1.6.0 ### Current Behavior When a user with Edit permissions on a DS project tries to delete it, the modal box gets closed without showing any error messages. If user opens the deletion modal a second time, the error is rendered before confirming the deletion action. GIF representing the scenario: ![deletion-error-msg](https://github.com/opendatahub-io/odh-dashboard/assets/16117276/52d9c549-502e-4997-a47e-140bab421301) ### Expected Behavior To either delete the project or disable the action ### Steps To Reproduce 1. create a DS project with user A 2. share Edit permission to user B 3. try to delete the DS project using user B 4. try again ### Workaround (if any) _No response_ ### What browsers are you seeing the problem on? _No response_ ### Anything else _No response_
2.0
[Bug]: Error message for project deletion appears in the modal only in the next deletion attempt - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Deploy type OpenDataHub core version (eg. `v1.6.0`) ### Version 1.6.0 ### Current Behavior When a user with Edit permissions on a DS project tries to delete it, the modal box gets closed without showing any error messages. If user opens the deletion modal a second time, the error is rendered before confirming the deletion action. GIF representing the scenario: ![deletion-error-msg](https://github.com/opendatahub-io/odh-dashboard/assets/16117276/52d9c549-502e-4997-a47e-140bab421301) ### Expected Behavior To either delete the project or disable the action ### Steps To Reproduce 1. create a DS project with user A 2. share Edit permission to user B 3. try to delete the DS project using user B 4. try again ### Workaround (if any) _No response_ ### What browsers are you seeing the problem on? _No response_ ### Anything else _No response_
non_test
error message for project deletion appears in the modal only in the next deletion attempt is there an existing issue for this i have searched the existing issues deploy type opendatahub core version eg version current behavior when a user with edit permissions on a ds project tries to delete it the modal box gets closed without showing any error messages if user opens the deletion modal a second time the error is rendered before confirming the deletion action gif representing the scenario expected behavior to either delete the project or disable the action steps to reproduce create a ds project with user a share edit permission to user b try to delete the ds project using user b try again workaround if any no response what browsers are you seeing the problem on no response anything else no response
0
142,377
21,721,167,315
IssuesEvent
2022-05-11 00:23:40
Minaira/2d-rpg-unity
https://api.github.com/repos/Minaira/2d-rpg-unity
opened
Initial game design
Design
We will need to come up with initial requirements for this game. The initial plan is to come up with a turn-based RPG, pokemon is a good starting point. We can look into starting with something like this: https://github.com/MichaelAquilina/Some-2D-RPG/blob/master/README.md
1.0
Initial game design - We will need to come up with initial requirements for this game. The initial plan is to come up with a turn-based RPG, pokemon is a good starting point. We can look into starting with something like this: https://github.com/MichaelAquilina/Some-2D-RPG/blob/master/README.md
non_test
initial game design we will need to come up with initial requirements for this game the initial plan is to come up with a turn based rpg pokemon is a good starting point we can look into starting with something like this
0
76,376
15,495,955,169
IssuesEvent
2021-03-11 01:48:01
kadirselcuk/wp-graphql
https://api.github.com/repos/kadirselcuk/wp-graphql
opened
CVE-2020-7608 (Medium) detected in multiple libraries
security vulnerability
## CVE-2020-7608 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-4.2.1.tgz</b>, <b>yargs-parser-5.0.0.tgz</b>, <b>yargs-parser-7.0.0.tgz</b></p></summary> <p> <details><summary><b>yargs-parser-4.2.1.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-4.2.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-4.2.1.tgz</a></p> <p>Path to dependency file: wp-graphql/src/Admin/GraphiQL/app/package.json</p> <p>Path to vulnerable library: wp-graphql/src/Admin/GraphiQL/app/node_modules/webpack-dev-server/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.1.5.tgz (Root Library) - webpack-dev-server-2.11.3.tgz - yargs-6.6.0.tgz - :x: **yargs-parser-4.2.1.tgz** (Vulnerable Library) </details> <details><summary><b>yargs-parser-5.0.0.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-5.0.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-5.0.0.tgz</a></p> <p>Path to dependency file: wp-graphql/src/Admin/GraphiQL/app/package.json</p> <p>Path to vulnerable library: wp-graphql/src/Admin/GraphiQL/app/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.1.5.tgz (Root Library) - jest-20.0.4.tgz - jest-cli-20.0.4.tgz - yargs-7.1.0.tgz - :x: **yargs-parser-5.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>yargs-parser-7.0.0.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz</a></p> <p>Path to dependency file: wp-graphql/src/Admin/GraphiQL/app/package.json</p> <p>Path to vulnerable library: wp-graphql/src/Admin/GraphiQL/app/node_modules/webpack/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.1.5.tgz (Root Library) - webpack-3.8.1.tgz - yargs-8.0.2.tgz - :x: **yargs-parser-7.0.0.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload. <p>Publish Date: 2020-03-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608</a></p> <p>Release Date: 2020-03-16</p> <p>Fix Resolution: v18.1.1;13.1.2;15.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7608 (Medium) detected in multiple libraries - ## CVE-2020-7608 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-4.2.1.tgz</b>, <b>yargs-parser-5.0.0.tgz</b>, <b>yargs-parser-7.0.0.tgz</b></p></summary> <p> <details><summary><b>yargs-parser-4.2.1.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-4.2.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-4.2.1.tgz</a></p> <p>Path to dependency file: wp-graphql/src/Admin/GraphiQL/app/package.json</p> <p>Path to vulnerable library: wp-graphql/src/Admin/GraphiQL/app/node_modules/webpack-dev-server/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.1.5.tgz (Root Library) - webpack-dev-server-2.11.3.tgz - yargs-6.6.0.tgz - :x: **yargs-parser-4.2.1.tgz** (Vulnerable Library) </details> <details><summary><b>yargs-parser-5.0.0.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-5.0.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-5.0.0.tgz</a></p> <p>Path to dependency file: wp-graphql/src/Admin/GraphiQL/app/package.json</p> <p>Path to vulnerable library: wp-graphql/src/Admin/GraphiQL/app/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.1.5.tgz (Root Library) - jest-20.0.4.tgz - jest-cli-20.0.4.tgz - yargs-7.1.0.tgz - :x: **yargs-parser-5.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>yargs-parser-7.0.0.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz</a></p> <p>Path to dependency file: wp-graphql/src/Admin/GraphiQL/app/package.json</p> <p>Path to vulnerable library: wp-graphql/src/Admin/GraphiQL/app/node_modules/webpack/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.1.5.tgz (Root Library) - webpack-3.8.1.tgz - yargs-8.0.2.tgz - :x: **yargs-parser-7.0.0.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload. <p>Publish Date: 2020-03-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608</a></p> <p>Release Date: 2020-03-16</p> <p>Fix Resolution: v18.1.1;13.1.2;15.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries yargs parser tgz yargs parser tgz yargs parser tgz yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file wp graphql src admin graphiql app package json path to vulnerable library wp graphql src admin graphiql app node modules webpack dev server node modules yargs parser package json dependency hierarchy react scripts tgz root library webpack dev server tgz yargs tgz x yargs parser tgz vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file wp graphql src admin graphiql app package json path to vulnerable library wp graphql src admin graphiql app node modules yargs parser package json dependency hierarchy react scripts tgz root library jest tgz jest cli tgz yargs tgz x yargs parser tgz vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file wp graphql src admin graphiql app package json path to vulnerable library wp graphql src admin graphiql app node modules webpack node modules yargs parser package json dependency hierarchy react scripts tgz root library webpack tgz yargs tgz x yargs parser tgz vulnerable library found in base branch master vulnerability details yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
167,909
14,119,134,259
IssuesEvent
2020-11-08 16:22:25
ViRGiL175/android-diner-delivery
https://api.github.com/repos/ViRGiL175/android-diner-delivery
opened
🚁 Скрам-стратегия по разработке
type: documentation
Всё то, что нужно знать об организации работы по проекту. ## Роли ![Роли](https://user-images.githubusercontent.com/64417900/95680394-6df5a500-0bea-11eb-9f2e-db5be1a5d7b7.png) ### Stakeholders - заинтересованные стороны (заказчик/пользователь/автор идеи и тд) - те, для кого проект принесет выгоду - участвуют в Sprint Demo Примечание. Перевод с английского - "акционер". [Подробнее о Stakeholder](https://ru.scrum-time.com/infobase/stakeholders-scrum.php) ### Developer (команда разработки) - пишет код - прислушивается к мнению всей команды, в частности к мнению Product Owner-а - "жалобы" на проблемы отправляет к Scrum Master - активно участвует во всех собраниях [Подробнее о команде разработки в Scrum](https://ru.scrum-time.com/infobase/development-team-scrum.php) ### Product Owner - составляет прозрачный и понятный бэклог - корректирует работу Команды разработки в зависимости от пожеланий Stake Holder-ов - старается оптимизировать работу Команды разработки - "жалобы" на проблемы отправляет к Scrum Master - активно участвует во всех собраниях [Подробнее о Product Owner](https://ru.scrum-time.com/infobase/product-owner.php) ### Scrum Master - помогает составлять бэклог - следит за тем, чтобы команда работала по методологии Scrum; корректирует работу команды, при необходимости - следит за динамикой работы всей команды, старается оптимизировать работу Команды разработки - активно помогает в решении проблем участников команды - проводит все собрания [Подробнее о Scrum Master](https://ru.scrum-time.com/infobase/scrum-master.php) ## Жизненный цикл ### Шаг 1. Создание бэклога Каждый участник должен понимать, с какими объектами, процессами мы работаем и какие принципы соблюдаем. Каждый объект или процесс - ишью с меткой `type:_baclog`. Опционально. Здесь же составляется основная документация по проекту, каждый документ - ишью с меткой `type:_documentation`. Примечание. Бэклог может представлять собой ООП-схему модели проекта. ### Шаг 2. Создание историй Пользовательская история - краткая характеристика пользователя + его пожелание + цель, которой он хочет достигнуть. Например. Когда пеший клиент заходит в Дайнер, он хочет сделать заказ, чтобы быть уверенным, что он получит блюдо, когда оно будет готово. Пользовательская история обязательно содержит критерии выполненности и приоритет. Как узнать, какие истории выполнять в первую очередь? Как понять, что работа над историей закончена. Примечание. Истории создаются на основе бэклога (объектов и процессов). [Подробнее о пользовательских историях](https://unetway.com/tutorial/scrum-users-stories) ### Шаг 3. Sprint Planning 3.1 Product Owner вместе с командой формулирует цель спринта. Какую ценность выполненный спринт принесет проекту? 3.2 Составляется бэклог спринта - список задач (ишью), которые будут реализованы за спринт. Учитывается приоритет и динамика команды. Все задачи отправляются в колонку "To Do" нового проджекта для текущего спринта. Команда оценивает сложность ишью ([скрам-покер](https://ru.scrum-time.com/infobase/planning-poker.php)), добавляет к ним метки `complexity`. Примечание. Длительность Sprint Planning зависит от длительности спринта. 1 неделя спринта = 2 часа Sprint Planning (необязательно четко соблюдать рамки, это рекомендация). [Подробнее о Sprint Planning](https://ru.scrum-time.com/infobase/sprint-planning-meeting.php) ### Шаг 4. Работа над спринтом, Daily Miting По мере выполнения задач ишью перемещаются в колонки "In Progress" и "Done" проджекта текущего спринта. В течении спринта в бэклог добавляются улучшения (ишью с меткой `type:_enhacement`) #### Daily Miting Команда собирается и обсуждает свою работу над спринтом. Каждый из участников команды кратко отвечает на 3 вопроса: - Что я делал после предыдущего Daily Miting? - Что я буду делать до следующего Daily Miting? - Что-то мешает мне работать? Все всех слушают и комментируют, при необходимости. Примечание. Daily Miting проводится для прозрачности, каждому будет полезно представлять картинку в целом. Например, если разработчик не может разобраться в функционале новой библиотеки, то второй сразу же узнает об этом на Daily Miting и сможет помочь. Также все участники команды будут знать, какая работа ведется во всех "сферах", а не только в своей. Для этого же создается проджект с бэклогом спринта. В идеале Daily Miting проводится каждый день, причем стоя, чтобы не затягивать собрание, и длиться не больше 15 минут. [Подробнее о Daily Miting](https://ru.scrum-time.com/infobase/daily-scrum-meeting.php) ### Шаг 5. Собрание. Sprint Demo Встреча с Stakeholder, презентация продукта и завершенного за спринт функционала. Stakeholder оценивает готовность продукта, работоспособность функционала, дает обратную связь. [Подробнее о Sprint Demo](https://ru.scrum-time.com/infobase/sprint-review-meeting.php) ### Шаг 6. Собрание. Retrospectiv Команда обсуждает свои итоги за завершенный спринт, делится впечатлениями: что получилось, что нет и что мешало работе. В течении следующего спринта команда реализует глобальное улучшение - кайдзен. На ретроспективе команда думает, каким будет это улучшение и как оно ускорит работу. [Подробнее о Retrospectiv](https://ru.scrum-time.com/infobase/sprint-retrospective-meeting.php) ### Шаг 7. Возвращение к шагу 3. После завершения работы над текущим спринтом и обсуждения его итогов, команда планирует следующий спринт и цикл повторяется.
1.0
🚁 Скрам-стратегия по разработке - Всё то, что нужно знать об организации работы по проекту. ## Роли ![Роли](https://user-images.githubusercontent.com/64417900/95680394-6df5a500-0bea-11eb-9f2e-db5be1a5d7b7.png) ### Stakeholders - заинтересованные стороны (заказчик/пользователь/автор идеи и тд) - те, для кого проект принесет выгоду - участвуют в Sprint Demo Примечание. Перевод с английского - "акционер". [Подробнее о Stakeholder](https://ru.scrum-time.com/infobase/stakeholders-scrum.php) ### Developer (команда разработки) - пишет код - прислушивается к мнению всей команды, в частности к мнению Product Owner-а - "жалобы" на проблемы отправляет к Scrum Master - активно участвует во всех собраниях [Подробнее о команде разработки в Scrum](https://ru.scrum-time.com/infobase/development-team-scrum.php) ### Product Owner - составляет прозрачный и понятный бэклог - корректирует работу Команды разработки в зависимости от пожеланий Stake Holder-ов - старается оптимизировать работу Команды разработки - "жалобы" на проблемы отправляет к Scrum Master - активно участвует во всех собраниях [Подробнее о Product Owner](https://ru.scrum-time.com/infobase/product-owner.php) ### Scrum Master - помогает составлять бэклог - следит за тем, чтобы команда работала по методологии Scrum; корректирует работу команды, при необходимости - следит за динамикой работы всей команды, старается оптимизировать работу Команды разработки - активно помогает в решении проблем участников команды - проводит все собрания [Подробнее о Scrum Master](https://ru.scrum-time.com/infobase/scrum-master.php) ## Жизненный цикл ### Шаг 1. Создание бэклога Каждый участник должен понимать, с какими объектами, процессами мы работаем и какие принципы соблюдаем. Каждый объект или процесс - ишью с меткой `type:_baclog`. Опционально. Здесь же составляется основная документация по проекту, каждый документ - ишью с меткой `type:_documentation`. Примечание. Бэклог может представлять собой ООП-схему модели проекта. ### Шаг 2. Создание историй Пользовательская история - краткая характеристика пользователя + его пожелание + цель, которой он хочет достигнуть. Например. Когда пеший клиент заходит в Дайнер, он хочет сделать заказ, чтобы быть уверенным, что он получит блюдо, когда оно будет готово. Пользовательская история обязательно содержит критерии выполненности и приоритет. Как узнать, какие истории выполнять в первую очередь? Как понять, что работа над историей закончена. Примечание. Истории создаются на основе бэклога (объектов и процессов). [Подробнее о пользовательских историях](https://unetway.com/tutorial/scrum-users-stories) ### Шаг 3. Sprint Planning 3.1 Product Owner вместе с командой формулирует цель спринта. Какую ценность выполненный спринт принесет проекту? 3.2 Составляется бэклог спринта - список задач (ишью), которые будут реализованы за спринт. Учитывается приоритет и динамика команды. Все задачи отправляются в колонку "To Do" нового проджекта для текущего спринта. Команда оценивает сложность ишью ([скрам-покер](https://ru.scrum-time.com/infobase/planning-poker.php)), добавляет к ним метки `complexity`. Примечание. Длительность Sprint Planning зависит от длительности спринта. 1 неделя спринта = 2 часа Sprint Planning (необязательно четко соблюдать рамки, это рекомендация). [Подробнее о Sprint Planning](https://ru.scrum-time.com/infobase/sprint-planning-meeting.php) ### Шаг 4. Работа над спринтом, Daily Miting По мере выполнения задач ишью перемещаются в колонки "In Progress" и "Done" проджекта текущего спринта. В течении спринта в бэклог добавляются улучшения (ишью с меткой `type:_enhacement`) #### Daily Miting Команда собирается и обсуждает свою работу над спринтом. Каждый из участников команды кратко отвечает на 3 вопроса: - Что я делал после предыдущего Daily Miting? - Что я буду делать до следующего Daily Miting? - Что-то мешает мне работать? Все всех слушают и комментируют, при необходимости. Примечание. Daily Miting проводится для прозрачности, каждому будет полезно представлять картинку в целом. Например, если разработчик не может разобраться в функционале новой библиотеки, то второй сразу же узнает об этом на Daily Miting и сможет помочь. Также все участники команды будут знать, какая работа ведется во всех "сферах", а не только в своей. Для этого же создается проджект с бэклогом спринта. В идеале Daily Miting проводится каждый день, причем стоя, чтобы не затягивать собрание, и длиться не больше 15 минут. [Подробнее о Daily Miting](https://ru.scrum-time.com/infobase/daily-scrum-meeting.php) ### Шаг 5. Собрание. Sprint Demo Встреча с Stakeholder, презентация продукта и завершенного за спринт функционала. Stakeholder оценивает готовность продукта, работоспособность функционала, дает обратную связь. [Подробнее о Sprint Demo](https://ru.scrum-time.com/infobase/sprint-review-meeting.php) ### Шаг 6. Собрание. Retrospectiv Команда обсуждает свои итоги за завершенный спринт, делится впечатлениями: что получилось, что нет и что мешало работе. В течении следующего спринта команда реализует глобальное улучшение - кайдзен. На ретроспективе команда думает, каким будет это улучшение и как оно ускорит работу. [Подробнее о Retrospectiv](https://ru.scrum-time.com/infobase/sprint-retrospective-meeting.php) ### Шаг 7. Возвращение к шагу 3. После завершения работы над текущим спринтом и обсуждения его итогов, команда планирует следующий спринт и цикл повторяется.
non_test
🚁 скрам стратегия по разработке всё то что нужно знать об организации работы по проекту роли stakeholders заинтересованные стороны заказчик пользователь автор идеи и тд те для кого проект принесет выгоду участвуют в sprint demo примечание перевод с английского акционер developer команда разработки пишет код прислушивается к мнению всей команды в частности к мнению product owner а жалобы на проблемы отправляет к scrum master активно участвует во всех собраниях product owner составляет прозрачный и понятный бэклог корректирует работу команды разработки в зависимости от пожеланий stake holder ов старается оптимизировать работу команды разработки жалобы на проблемы отправляет к scrum master активно участвует во всех собраниях scrum master помогает составлять бэклог следит за тем чтобы команда работала по методологии scrum корректирует работу команды при необходимости следит за динамикой работы всей команды старается оптимизировать работу команды разработки активно помогает в решении проблем участников команды проводит все собрания жизненный цикл шаг создание бэклога каждый участник должен понимать с какими объектами процессами мы работаем и какие принципы соблюдаем каждый объект или процесс ишью с меткой type baclog опционально здесь же составляется основная документация по проекту каждый документ ишью с меткой type documentation примечание бэклог может представлять собой ооп схему модели проекта шаг создание историй пользовательская история краткая характеристика пользователя его пожелание цель которой он хочет достигнуть например когда пеший клиент заходит в дайнер он хочет сделать заказ чтобы быть уверенным что он получит блюдо когда оно будет готово пользовательская история обязательно содержит критерии выполненности и приоритет как узнать какие истории выполнять в первую очередь как понять что работа над историей закончена примечание истории создаются на основе бэклога объектов и процессов шаг sprint planning product owner вместе с командой формулирует цель спринта какую ценность выполненный спринт принесет проекту составляется бэклог спринта список задач ишью которые будут реализованы за спринт учитывается приоритет и динамика команды все задачи отправляются в колонку to do нового проджекта для текущего спринта команда оценивает сложность ишью добавляет к ним метки complexity примечание длительность sprint planning зависит от длительности спринта неделя спринта часа sprint planning необязательно четко соблюдать рамки это рекомендация шаг работа над спринтом daily miting по мере выполнения задач ишью перемещаются в колонки in progress и done проджекта текущего спринта в течении спринта в бэклог добавляются улучшения ишью с меткой type enhacement daily miting команда собирается и обсуждает свою работу над спринтом каждый из участников команды кратко отвечает на вопроса что я делал после предыдущего daily miting что я буду делать до следующего daily miting что то мешает мне работать все всех слушают и комментируют при необходимости примечание daily miting проводится для прозрачности каждому будет полезно представлять картинку в целом например если разработчик не может разобраться в функционале новой библиотеки то второй сразу же узнает об этом на daily miting и сможет помочь также все участники команды будут знать какая работа ведется во всех сферах а не только в своей для этого же создается проджект с бэклогом спринта в идеале daily miting проводится каждый день причем стоя чтобы не затягивать собрание и длиться не больше минут шаг собрание sprint demo встреча с stakeholder презентация продукта и завершенного за спринт функционала stakeholder оценивает готовность продукта работоспособность функционала дает обратную связь шаг собрание retrospectiv команда обсуждает свои итоги за завершенный спринт делится впечатлениями что получилось что нет и что мешало работе в течении следующего спринта команда реализует глобальное улучшение кайдзен на ретроспективе команда думает каким будет это улучшение и как оно ускорит работу шаг возвращение к шагу после завершения работы над текущим спринтом и обсуждения его итогов команда планирует следующий спринт и цикл повторяется
0
122,660
17,762,068,188
IssuesEvent
2021-08-29 21:59:36
ghc-dev/Vanessa-Moore
https://api.github.com/repos/ghc-dev/Vanessa-Moore
opened
CVE-2021-3533 (Low) detected in ansible-2.9.9.tar.gz
security vulnerability
## CVE-2021-3533 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ansible-2.9.9.tar.gz</b></p></summary> <p>Radically simple IT automation</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/00/5d/e10b83e0e6056dbd5b4809b451a191395175a57e3175ce04e35d9c5fc2a0/ansible-2.9.9.tar.gz">https://files.pythonhosted.org/packages/00/5d/e10b83e0e6056dbd5b4809b451a191395175a57e3175ce04e35d9c5fc2a0/ansible-2.9.9.tar.gz</a></p> <p>Path to dependency file: Vanessa-Moore/requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> Dependency Hierarchy: - :x: **ansible-2.9.9.tar.gz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Vanessa-Moore/commit/1a39144956acd5ccd1e6bf76ab7ca339691f4ec0">1a39144956acd5ccd1e6bf76ab7ca339691f4ec0</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Vulnerability in ansible when ANSIBLE_ASYNC_DIR defaults to ~/.ansible_async/ but is settable by the user. It can be set by the ansible user to a subdirectory of a world writable directory, for instance ANSIBLE_ASYNC_DIR=/tmp/username-ansible-async/. When this occurs, there is a race condition on the managed machine. A malicious, low privileged account on the remote machine can pre-create /tmp/username-ansible-async and then use various attacks to access the async result data. <p>Publish Date: 2021-05-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3533>CVE-2021-3533</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"ansible","packageVersion":"2.9.9","packageFilePaths":["/requirements.txt"],"isTransitiveDependency":false,"dependencyTree":"ansible:2.9.9","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-3533","vulnerabilityDetails":"Vulnerability in ansible when ANSIBLE_ASYNC_DIR defaults to ~/.ansible_async/ but is settable by the user. It can be set by the ansible user to a subdirectory of a world writable directory, for instance ANSIBLE_ASYNC_DIR\u003d/tmp/username-ansible-async/. When this occurs, there is a race condition on the managed machine. A malicious, low privileged account on the remote machine can pre-create /tmp/username-ansible-async and then use various attacks to access the async result data.\n","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3533","cvss3Severity":"low","cvss3Score":"2.5","cvss3Metrics":{"A":"None","AC":"High","PR":"Low","S":"Unchanged","C":"Low","UI":"None","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-3533 (Low) detected in ansible-2.9.9.tar.gz - ## CVE-2021-3533 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ansible-2.9.9.tar.gz</b></p></summary> <p>Radically simple IT automation</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/00/5d/e10b83e0e6056dbd5b4809b451a191395175a57e3175ce04e35d9c5fc2a0/ansible-2.9.9.tar.gz">https://files.pythonhosted.org/packages/00/5d/e10b83e0e6056dbd5b4809b451a191395175a57e3175ce04e35d9c5fc2a0/ansible-2.9.9.tar.gz</a></p> <p>Path to dependency file: Vanessa-Moore/requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> Dependency Hierarchy: - :x: **ansible-2.9.9.tar.gz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Vanessa-Moore/commit/1a39144956acd5ccd1e6bf76ab7ca339691f4ec0">1a39144956acd5ccd1e6bf76ab7ca339691f4ec0</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Vulnerability in ansible when ANSIBLE_ASYNC_DIR defaults to ~/.ansible_async/ but is settable by the user. It can be set by the ansible user to a subdirectory of a world writable directory, for instance ANSIBLE_ASYNC_DIR=/tmp/username-ansible-async/. When this occurs, there is a race condition on the managed machine. A malicious, low privileged account on the remote machine can pre-create /tmp/username-ansible-async and then use various attacks to access the async result data. <p>Publish Date: 2021-05-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3533>CVE-2021-3533</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"ansible","packageVersion":"2.9.9","packageFilePaths":["/requirements.txt"],"isTransitiveDependency":false,"dependencyTree":"ansible:2.9.9","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-3533","vulnerabilityDetails":"Vulnerability in ansible when ANSIBLE_ASYNC_DIR defaults to ~/.ansible_async/ but is settable by the user. It can be set by the ansible user to a subdirectory of a world writable directory, for instance ANSIBLE_ASYNC_DIR\u003d/tmp/username-ansible-async/. When this occurs, there is a race condition on the managed machine. A malicious, low privileged account on the remote machine can pre-create /tmp/username-ansible-async and then use various attacks to access the async result data.\n","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3533","cvss3Severity":"low","cvss3Score":"2.5","cvss3Metrics":{"A":"None","AC":"High","PR":"Low","S":"Unchanged","C":"Low","UI":"None","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> -->
non_test
cve low detected in ansible tar gz cve low severity vulnerability vulnerable library ansible tar gz radically simple it automation library home page a href path to dependency file vanessa moore requirements txt path to vulnerable library requirements txt dependency hierarchy x ansible tar gz vulnerable library found in head commit a href found in base branch master vulnerability details vulnerability in ansible when ansible async dir defaults to ansible async but is settable by the user it can be set by the ansible user to a subdirectory of a world writable directory for instance ansible async dir tmp username ansible async when this occurs there is a race condition on the managed machine a malicious low privileged account on the remote machine can pre create tmp username ansible async and then use various attacks to access the async result data publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree ansible isminimumfixversionavailable false basebranches vulnerabilityidentifier cve vulnerabilitydetails vulnerability in ansible when ansible async dir defaults to ansible async but is settable by the user it can be set by the ansible user to a subdirectory of a world writable directory for instance ansible async dir tmp username ansible async when this occurs there is a race condition on the managed machine a malicious low privileged account on the remote machine can pre create tmp username ansible async and then use various attacks to access the async result data n vulnerabilityurl
0
590,661
17,784,004,763
IssuesEvent
2021-08-31 08:52:31
knowit/julekalender_frontend
https://api.github.com/repos/knowit/julekalender_frontend
opened
Markdown preview for kommentarer
enhancement priority frontend backend
Markdown-parseren på backend er ganske streng. Mange i fjor bet seg på korrekt format for blokk-elementer. Det hadde vært å greit kunne se kommentaren sin før man submitter. Krever nytt endepunkt på backend (rask jobb).
1.0
Markdown preview for kommentarer - Markdown-parseren på backend er ganske streng. Mange i fjor bet seg på korrekt format for blokk-elementer. Det hadde vært å greit kunne se kommentaren sin før man submitter. Krever nytt endepunkt på backend (rask jobb).
non_test
markdown preview for kommentarer markdown parseren på backend er ganske streng mange i fjor bet seg på korrekt format for blokk elementer det hadde vært å greit kunne se kommentaren sin før man submitter krever nytt endepunkt på backend rask jobb
0
13,303
3,322,185,278
IssuesEvent
2015-11-09 13:17:28
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
Backwards compatibility tests should be run with the security manager
test v2.1.0 v3.0.0
Right now the backwards compatibility tests disable the security manager entirely or the external nodes won't run. We should fix that.
1.0
Backwards compatibility tests should be run with the security manager - Right now the backwards compatibility tests disable the security manager entirely or the external nodes won't run. We should fix that.
test
backwards compatibility tests should be run with the security manager right now the backwards compatibility tests disable the security manager entirely or the external nodes won t run we should fix that
1
253,191
21,660,604,600
IssuesEvent
2022-05-06 18:38:16
damccorm/test-migration-target
https://api.github.com/repos/damccorm/test-migration-target
opened
Figure out TestPipeline.create(PipelineOptions) / TestPipeline.fromOptions(PipelineOptions) story
P3 improvement sdk-java-core clarified test
TestPipeline integrates with the integration testing environment and relies heavily on being able to be configured by the environment and executed on many runners. Tests which rely on mutating PipelineOptions before creating the TestPipeline easily can get the integration wrong by creating PipelineOptions from PipelineOptionsFactory and then calling either TestPipeline.create(options) or TestPipeline.fromOptions(options), thus ignoring any integration environment pipeline options specified. We should fix the exposed methods on TestPipeline to prevent users from making this simple mistake. One suggestion is to create a TestPipeline builder which will give access to a mutable PipelineOptions which the user can edit before calling build() creating a TestPipeline. Imported from Jira [BEAM-672](https://issues.apache.org/jira/browse/BEAM-672). Original Jira may contain additional context. Reported by: lcwik.
1.0
Figure out TestPipeline.create(PipelineOptions) / TestPipeline.fromOptions(PipelineOptions) story - TestPipeline integrates with the integration testing environment and relies heavily on being able to be configured by the environment and executed on many runners. Tests which rely on mutating PipelineOptions before creating the TestPipeline easily can get the integration wrong by creating PipelineOptions from PipelineOptionsFactory and then calling either TestPipeline.create(options) or TestPipeline.fromOptions(options), thus ignoring any integration environment pipeline options specified. We should fix the exposed methods on TestPipeline to prevent users from making this simple mistake. One suggestion is to create a TestPipeline builder which will give access to a mutable PipelineOptions which the user can edit before calling build() creating a TestPipeline. Imported from Jira [BEAM-672](https://issues.apache.org/jira/browse/BEAM-672). Original Jira may contain additional context. Reported by: lcwik.
test
figure out testpipeline create pipelineoptions testpipeline fromoptions pipelineoptions story testpipeline integrates with the integration testing environment and relies heavily on being able to be configured by the environment and executed on many runners tests which rely on mutating pipelineoptions before creating the testpipeline easily can get the integration wrong by creating pipelineoptions from pipelineoptionsfactory and then calling either testpipeline create options or testpipeline fromoptions options thus ignoring any integration environment pipeline options specified we should fix the exposed methods on testpipeline to prevent users from making this simple mistake one suggestion is to create a testpipeline builder which will give access to a mutable pipelineoptions which the user can edit before calling build creating a testpipeline imported from jira original jira may contain additional context reported by lcwik
1
178,412
14,668,809,284
IssuesEvent
2020-12-29 22:21:45
FlaxEngine/FlaxEngine
https://api.github.com/repos/FlaxEngine/FlaxEngine
closed
UI Coordinate Spaces
documentation
**Issue description:** Currently those are the most important methods for tranforming from one coordinate system to another: - `PointToWindow` and `PointFromWindow` - `ScreenToClient` and `ClientToScreen` (note: there are [two](https://github.com/FlaxEngine/FlaxEngine/blob/be1b9857f9d219811febf18e144fa4d1e63ca9fa/Source/Engine/UI/GUI/Control.cs#L1116) sorta [different](https://github.com/FlaxEngine/FlaxEngine/blob/be1b9857f9d219811febf18e144fa4d1e63ca9fa/Source/Engine/Platform/Windows/WindowsWindow.cpp#L395) methods with the same name) - `PointFromParent` and `PointToParent` So, we have the following coordinate spaces (also see https://github.com/FlaxEngine/FlaxDocs/issues/26 ) 1. **Screen** space 2. **Window** space, also called **client** space: Should this one be DPI-adjusted? 3. **Local** **control** space, also called **client** space. Or just "**point**". This feels a bit inconsistent. To keep the number of breaking changes down, I'd suggest renaming the C# `ScreenToClient` and `ClientToScreen` to `PointToScreen` and `PointFromScreen`. Furthermore, I suppose window space (C# side) should be DPI-adjusted. **Flax version:** Flax 1.0
1.0
UI Coordinate Spaces - **Issue description:** Currently those are the most important methods for tranforming from one coordinate system to another: - `PointToWindow` and `PointFromWindow` - `ScreenToClient` and `ClientToScreen` (note: there are [two](https://github.com/FlaxEngine/FlaxEngine/blob/be1b9857f9d219811febf18e144fa4d1e63ca9fa/Source/Engine/UI/GUI/Control.cs#L1116) sorta [different](https://github.com/FlaxEngine/FlaxEngine/blob/be1b9857f9d219811febf18e144fa4d1e63ca9fa/Source/Engine/Platform/Windows/WindowsWindow.cpp#L395) methods with the same name) - `PointFromParent` and `PointToParent` So, we have the following coordinate spaces (also see https://github.com/FlaxEngine/FlaxDocs/issues/26 ) 1. **Screen** space 2. **Window** space, also called **client** space: Should this one be DPI-adjusted? 3. **Local** **control** space, also called **client** space. Or just "**point**". This feels a bit inconsistent. To keep the number of breaking changes down, I'd suggest renaming the C# `ScreenToClient` and `ClientToScreen` to `PointToScreen` and `PointFromScreen`. Furthermore, I suppose window space (C# side) should be DPI-adjusted. **Flax version:** Flax 1.0
non_test
ui coordinate spaces issue description currently those are the most important methods for tranforming from one coordinate system to another pointtowindow and pointfromwindow screentoclient and clienttoscreen note there are sorta methods with the same name pointfromparent and pointtoparent so we have the following coordinate spaces also see screen space window space also called client space should this one be dpi adjusted local control space also called client space or just point this feels a bit inconsistent to keep the number of breaking changes down i d suggest renaming the c screentoclient and clienttoscreen to pointtoscreen and pointfromscreen furthermore i suppose window space c side should be dpi adjusted flax version flax
0
83,689
7,879,593,039
IssuesEvent
2018-06-26 13:49:15
sudar/bulk-delete
https://api.github.com/repos/sudar/bulk-delete
closed
Delete Post Revision - Add tests
in progress tests
## Tests - [x] Add tests to delete revisions for a single post - [x] Add tests to delete revisions for multiple posts
1.0
Delete Post Revision - Add tests - ## Tests - [x] Add tests to delete revisions for a single post - [x] Add tests to delete revisions for multiple posts
test
delete post revision add tests tests add tests to delete revisions for a single post add tests to delete revisions for multiple posts
1
20,940
27,798,557,654
IssuesEvent
2023-03-17 14:17:35
aiidateam/aiida-core
https://api.github.com/repos/aiidateam/aiida-core
closed
Process functions: Infer `help` of input ports from function argument docstring
topic/workflows type/accepted feature priority/nice-to-have topic/processes
The `input` method of the `ProcessSpec` allows defining a `help` message for the input port, which is very useful when defining `WorkChain` and `CalcJob` plugins. However, this API is not directly available for process functions since the process spec is inferred dynamically from the function signature. It would be possible to infer it from the docstring. This would be very useful especially when the process function gets exposed in a workchain. The docstring will then be immediately available from the workchain's process specification and the user does not have to search the source code of the process function.
1.0
Process functions: Infer `help` of input ports from function argument docstring - The `input` method of the `ProcessSpec` allows defining a `help` message for the input port, which is very useful when defining `WorkChain` and `CalcJob` plugins. However, this API is not directly available for process functions since the process spec is inferred dynamically from the function signature. It would be possible to infer it from the docstring. This would be very useful especially when the process function gets exposed in a workchain. The docstring will then be immediately available from the workchain's process specification and the user does not have to search the source code of the process function.
non_test
process functions infer help of input ports from function argument docstring the input method of the processspec allows defining a help message for the input port which is very useful when defining workchain and calcjob plugins however this api is not directly available for process functions since the process spec is inferred dynamically from the function signature it would be possible to infer it from the docstring this would be very useful especially when the process function gets exposed in a workchain the docstring will then be immediately available from the workchain s process specification and the user does not have to search the source code of the process function
0
67,013
7,032,078,428
IssuesEvent
2017-12-26 23:42:40
supercollider/supercollider
https://api.github.com/repos/supercollider/supercollider
opened
Exceptions thrown in UnitTests silently exit
bug comp: UnitTest
adding this to the class library: TestFoo : UnitTest { test_foo { //"this will be silently ignored".throw; } } and running `TestFoo.run` correctly gives the report: UNIT TEST............. There were no failures but if you uncomment the `.throw` line, the unit test silently exits and no report is given. even worse, attempts to `UnitTest.runAll` will suddenly stop at this test and never print a report for the entire suite of unit tests. in the current 3.9 branch, `UnitTest.runAll` silently exits because of a thrown exception in `TestUGen_Duty`.
1.0
Exceptions thrown in UnitTests silently exit - adding this to the class library: TestFoo : UnitTest { test_foo { //"this will be silently ignored".throw; } } and running `TestFoo.run` correctly gives the report: UNIT TEST............. There were no failures but if you uncomment the `.throw` line, the unit test silently exits and no report is given. even worse, attempts to `UnitTest.runAll` will suddenly stop at this test and never print a report for the entire suite of unit tests. in the current 3.9 branch, `UnitTest.runAll` silently exits because of a thrown exception in `TestUGen_Duty`.
test
exceptions thrown in unittests silently exit adding this to the class library testfoo unittest test foo this will be silently ignored throw and running testfoo run correctly gives the report unit test there were no failures but if you uncomment the throw line the unit test silently exits and no report is given even worse attempts to unittest runall will suddenly stop at this test and never print a report for the entire suite of unit tests in the current branch unittest runall silently exits because of a thrown exception in testugen duty
1
7,240
2,889,411,597
IssuesEvent
2015-06-13 11:42:25
acemod/ACE3
https://api.github.com/repos/acemod/ACE3
closed
Vanilla wounding system kicks in often
bug needs testing
ACE3 Version: 3.0.1 **Mods:** * @cba_a3 * @ace3 * @acre2 * @rhs * @leightopforpack * @BWMod **Placed ACE3 Modules:** ``` class ace_common_forceAllSettings { value = 0; typeName = "BOOL"; force = 0; }; class ace_frag_enabled { value = 1; typeName = "BOOL"; force = 0; }; class ace_interaction_EnableTeamManagement { value = 1; typeName = "BOOL"; force = 0; }; class ace_magazinerepack_TimePerAmmo { value = 1.5; typeName = "SCALAR"; force = 0; }; class ace_magazinerepack_TimePerMagazine { value = 2; typeName = "SCALAR"; force = 0; }; class ace_magazinerepack_TimePerBeltLink { value = 8; typeName = "SCALAR"; force = 0; }; class ace_map_BFT_Interval { value = 1; typeName = "SCALAR"; force = 0; }; class ace_map_BFT_Enabled { value = 0; typeName = "BOOL"; force = 0; }; class ace_map_BFT_HideAiGroups { value = 0; typeName = "BOOL"; force = 0; }; class ace_map_mapIllumination { value = 1; typeName = "BOOL"; force = 0; }; class ace_map_mapShake { value = 1; typeName = "BOOL"; force = 0; }; class ace_map_mapLimitZoom { value = 0; typeName = "BOOL"; force = 0; }; class ace_map_mapShowCursorCoordinates { value = 0; typeName = "BOOL"; force = 0; }; class ace_maptools_EveryoneCanDrawOnBriefing { value = 1; typeName = "BOOL"; force = 0; }; class ace_microdagr_MapDataAvailable { value = 1; typeName = "SCALAR"; force = 0; }; class ace_mk6mortar_airResistanceEnabled { value = 1; typeName = "BOOL"; force = 0; }; class ace_mk6mortar_allowComputerRangefinder { value = 0; typeName = "BOOL"; force = 0; }; class ace_mk6mortar_allowCompass { value = 1; typeName = "BOOL"; force = 0; }; class ace_nametags_showCursorTagForVehicles { value = 0; typeName = "BOOL"; force = 0; }; class ace_nametags_PlayerNamesViewDistance { value = 5; typeName = "SCALAR"; force = 0; }; class ace_nametags_PlayerNamesMaxAlpha { value = 0.8; typeName = "SCALAR"; force = 0; }; class ace_respawn_SavePreDeathGear { value = 0; typeName = "BOOL"; force = 0; }; class ace_respawn_RemoveDeadBodiesDisconnected { value = 1; typeName = "BOOL"; force = 0; }; class ace_switchunits_EnableSwitchUnits { value = 0; typeName = "BOOL"; force = 0; }; class ace_switchunits_SwitchToWest { value = 0; typeName = "BOOL"; force = 0; }; class ace_switchunits_SwitchToEast { value = 0; typeName = "BOOL"; force = 0; }; class ace_switchunits_SwitchToIndependent { value = 0; typeName = "BOOL"; force = 0; }; class ace_switchunits_SwitchToCivilian { value = 0; typeName = "BOOL"; force = 0; }; class ace_switchunits_EnableSafeZone { value = 1; typeName = "BOOL"; force = 0; }; class ace_switchunits_SafeZoneRadius { value = 100; typeName = "SCALAR"; force = 0; }; class ace_vehiclelock_DefaultLockpickStrength { value = 10; typeName = "SCALAR"; force = 0; }; class ace_vehiclelock_LockVehicleInventory { value = 0; typeName = "BOOL"; force = 0; }; class ace_vehiclelock_VehicleStartingLockState { value = -1; typeName = "SCALAR"; force = 0; }; class ace_weather_enableServerController { value = 1; typeName = "BOOL"; force = 0; }; class ace_weather_useACEWeather { value = 1; typeName = "BOOL"; force = 0; }; class ace_weather_syncRain { value = 1; typeName = "BOOL"; force = 0; }; class ace_weather_syncWind { value = 1; typeName = "BOOL"; force = 0; }; class ace_weather_syncMisc { value = 1; typeName = "BOOL"; force = 0; }; class ace_weather_serverUpdateInterval { value = 60; typeName = "SCALAR"; force = 0; }; class ace_winddeflection_enabled { value = 1; typeName = "BOOL"; force = 0; }; class ace_winddeflection_vehicleEnabled { value = 1; typeName = "BOOL"; force = 0; }; class ace_winddeflection_simulationInterval { value = 0.05; typeName = "SCALAR"; force = 0; }; class ace_winddeflection_simulationRadius { value = 3000; typeName = "SCALAR"; force = 0; }; class ace_explosives_RequireSpecialist { value = 0; typeName = "BOOL"; force = 0; }; class ace_explosives_PunishNonSpecialists { value = 1; typeName = "BOOL"; force = 0; }; class ace_hearing_EnableCombatDeafness { value = 1; typeName = "BOOL"; force = 0; }; class ace_hearing_EarplugsVolume { value = 0.5; typeName = "SCALAR"; force = 0; }; class ace_hearing_UnconsciousnessVolume { value = 0.2; typeName = "SCALAR"; force = 0; }; class ace_medical_level { value = 2; typeName = "SCALAR"; force = 0; }; class ace_medical_medicSetting { value = 2; typeName = "SCALAR"; force = 0; }; class ace_medical_enableFor { value = 1; typeName = "SCALAR"; force = 0; }; class ace_medical_enableOverdosing { value = 1; typeName = "BOOL"; force = 0; }; class ace_medical_bleedingCoefficient { value = 0.9; typeName = "SCALAR"; force = 0; }; class ace_medical_painCoefficient { value = 0.9; typeName = "SCALAR"; force = 0; }; class ace_medical_enableAirway { value = 0; typeName = "BOOL"; force = 0; }; class ace_medical_enableFractures { value = 0; typeName = "BOOL"; force = 0; }; class ace_medical_enableAdvancedWounds { value = 1; typeName = "BOOL"; force = 0; }; class ace_medical_enableVehicleCrashes { value = 1; typeName = "BOOL"; force = 0; }; class ace_medical_enableScreams { value = 1; typeName = "BOOL"; force = 0; }; class ace_medical_playerDamageThreshold { value = 0.9; typeName = "SCALAR"; force = 0; }; class ace_medical_AIDamageThreshold { value = 1; typeName = "SCALAR"; force = 0; }; class ace_medical_enableUnconsciousnessAI { value = 1; typeName = "SCALAR"; force = 0; }; class ace_medical_preventInstaDeath { value = 0; typeName = "BOOL"; force = 0; }; class ace_medical_enableRevive { value = 0; typeName = "SCALAR"; force = 0; }; class ace_medical_maxReviveTime { value = 120; typeName = "SCALAR"; force = 0; }; class ace_medical_amountOfReviveLives { value = -1; typeName = "SCALAR"; force = 0; }; class ace_medical_allowDeadBodyMovement { value = 0; typeName = "BOOL"; force = 0; }; class ace_medical_allowLitterCreation { value = 1; typeName = "BOOL"; force = 0; }; class ace_medical_litterCleanUpDelay { value = 1800; typeName = "SCALAR"; force = 0; }; class ace_medical_medicSetting_PAK { value = 1; typeName = "SCALAR"; force = 0; }; class ace_medical_medicSetting_SurgicalKit { value = 1; typeName = "SCALAR"; force = 0; }; class ace_medical_consumeItem_PAK { value = 1; typeName = "SCALAR"; force = 0; }; class ace_medical_consumeItem_SurgicalKit { value = 1; typeName = "SCALAR"; force = 0; }; class ace_medical_useLocation_PAK { value = 3; typeName = "SCALAR"; force = 0; }; class ace_medical_useLocation_SurgicalKit { value = 3; typeName = "SCALAR"; force = 0; }; class ace_medical_keepLocalSettingsSynced { value = 1; typeName = "BOOL"; force = 0; }; class ace_medical_healHitPointAfterAdvBandage { value = 1; typeName = "BOOL"; force = 0; }; class ace_medical_allowUnconsciousAnimationOnTreatment { value = 0; typeName = "BOOL"; force = 0; }; class ace_medical_moveUnitsFromGroupOnUnconscious { value = 0; typeName = "BOOL"; force = 0; }; class ace_medical_menuTypeStyle { value = 0; typeName = "SCALAR"; force = 0; }; class ace_advanced_ballistics_enabled { value = 1; typeName = "BOOL"; force = 0; }; class ace_advanced_ballistics_alwaysSimulateForSnipers { value = 1; typeName = "BOOL"; force = 0; }; class ace_advanced_ballistics_disabledInFullAutoMode { value = 0; typeName = "BOOL"; force = 0; }; class ace_advanced_ballistics_onlyActiveForLocalPlayers { value = 1; typeName = "BOOL"; force = 0; }; class ace_advanced_ballistics_ammoTemperatureEnabled { value = 1; typeName = "BOOL"; force = 0; }; class ace_advanced_ballistics_barrelLengthInfluenceEnabled { value = 1; typeName = "BOOL"; force = 0; }; class ace_advanced_ballistics_bulletTraceEnabled { value = 1; typeName = "BOOL"; force = 0; }; class ace_advanced_ballistics_simulationInterval { value = 0; typeName = "SCALAR"; force = 0; }; class ace_advanced_ballistics_simulationRadius { value = 3000; typeName = "SCALAR"; force = 0; }; ``` **Description:** *We had so far many situations, when player after being hit by bullets, or for example smashing on the ground on parachute were often hurt from both ACE3 Adv. medical and vanilla wounding system. Happened so far on both BWMod soldiers and RHS ones. I even tried gear up vanilla NATO to be 1:1 BWMod soldiers or RHS ones, didn't work. It's problematic cause players often complain that they are still hurt even when they are completely healed (ACE3 wounding). Only way to overcome this is to heal up by using vanilla actions near vanilla medevac vehicle. Seems to happen only on MP.* **Steps to reproduce:** * *Use RHS soldier or BWMod one, also advanced medical system* * *Play on MP* * *Get hit by AI or other player* * *Heal yourself up and observe that you are still yelling from pain, screen is blurry (A3 vanilla wounding)* **Where did the issue occur?** *"Multiplayer"* **RPT log file:** *Well nothing related to medical inside mine. It just happens...*
1.0
Vanilla wounding system kicks in often - ACE3 Version: 3.0.1 **Mods:** * @cba_a3 * @ace3 * @acre2 * @rhs * @leightopforpack * @BWMod **Placed ACE3 Modules:** ``` class ace_common_forceAllSettings { value = 0; typeName = "BOOL"; force = 0; }; class ace_frag_enabled { value = 1; typeName = "BOOL"; force = 0; }; class ace_interaction_EnableTeamManagement { value = 1; typeName = "BOOL"; force = 0; }; class ace_magazinerepack_TimePerAmmo { value = 1.5; typeName = "SCALAR"; force = 0; }; class ace_magazinerepack_TimePerMagazine { value = 2; typeName = "SCALAR"; force = 0; }; class ace_magazinerepack_TimePerBeltLink { value = 8; typeName = "SCALAR"; force = 0; }; class ace_map_BFT_Interval { value = 1; typeName = "SCALAR"; force = 0; }; class ace_map_BFT_Enabled { value = 0; typeName = "BOOL"; force = 0; }; class ace_map_BFT_HideAiGroups { value = 0; typeName = "BOOL"; force = 0; }; class ace_map_mapIllumination { value = 1; typeName = "BOOL"; force = 0; }; class ace_map_mapShake { value = 1; typeName = "BOOL"; force = 0; }; class ace_map_mapLimitZoom { value = 0; typeName = "BOOL"; force = 0; }; class ace_map_mapShowCursorCoordinates { value = 0; typeName = "BOOL"; force = 0; }; class ace_maptools_EveryoneCanDrawOnBriefing { value = 1; typeName = "BOOL"; force = 0; }; class ace_microdagr_MapDataAvailable { value = 1; typeName = "SCALAR"; force = 0; }; class ace_mk6mortar_airResistanceEnabled { value = 1; typeName = "BOOL"; force = 0; }; class ace_mk6mortar_allowComputerRangefinder { value = 0; typeName = "BOOL"; force = 0; }; class ace_mk6mortar_allowCompass { value = 1; typeName = "BOOL"; force = 0; }; class ace_nametags_showCursorTagForVehicles { value = 0; typeName = "BOOL"; force = 0; }; class ace_nametags_PlayerNamesViewDistance { value = 5; typeName = "SCALAR"; force = 0; }; class ace_nametags_PlayerNamesMaxAlpha { value = 0.8; typeName = "SCALAR"; force = 0; }; class ace_respawn_SavePreDeathGear { value = 0; typeName = "BOOL"; force = 0; }; class ace_respawn_RemoveDeadBodiesDisconnected { value = 1; typeName = "BOOL"; force = 0; }; class ace_switchunits_EnableSwitchUnits { value = 0; typeName = "BOOL"; force = 0; }; class ace_switchunits_SwitchToWest { value = 0; typeName = "BOOL"; force = 0; }; class ace_switchunits_SwitchToEast { value = 0; typeName = "BOOL"; force = 0; }; class ace_switchunits_SwitchToIndependent { value = 0; typeName = "BOOL"; force = 0; }; class ace_switchunits_SwitchToCivilian { value = 0; typeName = "BOOL"; force = 0; }; class ace_switchunits_EnableSafeZone { value = 1; typeName = "BOOL"; force = 0; }; class ace_switchunits_SafeZoneRadius { value = 100; typeName = "SCALAR"; force = 0; }; class ace_vehiclelock_DefaultLockpickStrength { value = 10; typeName = "SCALAR"; force = 0; }; class ace_vehiclelock_LockVehicleInventory { value = 0; typeName = "BOOL"; force = 0; }; class ace_vehiclelock_VehicleStartingLockState { value = -1; typeName = "SCALAR"; force = 0; }; class ace_weather_enableServerController { value = 1; typeName = "BOOL"; force = 0; }; class ace_weather_useACEWeather { value = 1; typeName = "BOOL"; force = 0; }; class ace_weather_syncRain { value = 1; typeName = "BOOL"; force = 0; }; class ace_weather_syncWind { value = 1; typeName = "BOOL"; force = 0; }; class ace_weather_syncMisc { value = 1; typeName = "BOOL"; force = 0; }; class ace_weather_serverUpdateInterval { value = 60; typeName = "SCALAR"; force = 0; }; class ace_winddeflection_enabled { value = 1; typeName = "BOOL"; force = 0; }; class ace_winddeflection_vehicleEnabled { value = 1; typeName = "BOOL"; force = 0; }; class ace_winddeflection_simulationInterval { value = 0.05; typeName = "SCALAR"; force = 0; }; class ace_winddeflection_simulationRadius { value = 3000; typeName = "SCALAR"; force = 0; }; class ace_explosives_RequireSpecialist { value = 0; typeName = "BOOL"; force = 0; }; class ace_explosives_PunishNonSpecialists { value = 1; typeName = "BOOL"; force = 0; }; class ace_hearing_EnableCombatDeafness { value = 1; typeName = "BOOL"; force = 0; }; class ace_hearing_EarplugsVolume { value = 0.5; typeName = "SCALAR"; force = 0; }; class ace_hearing_UnconsciousnessVolume { value = 0.2; typeName = "SCALAR"; force = 0; }; class ace_medical_level { value = 2; typeName = "SCALAR"; force = 0; }; class ace_medical_medicSetting { value = 2; typeName = "SCALAR"; force = 0; }; class ace_medical_enableFor { value = 1; typeName = "SCALAR"; force = 0; }; class ace_medical_enableOverdosing { value = 1; typeName = "BOOL"; force = 0; }; class ace_medical_bleedingCoefficient { value = 0.9; typeName = "SCALAR"; force = 0; }; class ace_medical_painCoefficient { value = 0.9; typeName = "SCALAR"; force = 0; }; class ace_medical_enableAirway { value = 0; typeName = "BOOL"; force = 0; }; class ace_medical_enableFractures { value = 0; typeName = "BOOL"; force = 0; }; class ace_medical_enableAdvancedWounds { value = 1; typeName = "BOOL"; force = 0; }; class ace_medical_enableVehicleCrashes { value = 1; typeName = "BOOL"; force = 0; }; class ace_medical_enableScreams { value = 1; typeName = "BOOL"; force = 0; }; class ace_medical_playerDamageThreshold { value = 0.9; typeName = "SCALAR"; force = 0; }; class ace_medical_AIDamageThreshold { value = 1; typeName = "SCALAR"; force = 0; }; class ace_medical_enableUnconsciousnessAI { value = 1; typeName = "SCALAR"; force = 0; }; class ace_medical_preventInstaDeath { value = 0; typeName = "BOOL"; force = 0; }; class ace_medical_enableRevive { value = 0; typeName = "SCALAR"; force = 0; }; class ace_medical_maxReviveTime { value = 120; typeName = "SCALAR"; force = 0; }; class ace_medical_amountOfReviveLives { value = -1; typeName = "SCALAR"; force = 0; }; class ace_medical_allowDeadBodyMovement { value = 0; typeName = "BOOL"; force = 0; }; class ace_medical_allowLitterCreation { value = 1; typeName = "BOOL"; force = 0; }; class ace_medical_litterCleanUpDelay { value = 1800; typeName = "SCALAR"; force = 0; }; class ace_medical_medicSetting_PAK { value = 1; typeName = "SCALAR"; force = 0; }; class ace_medical_medicSetting_SurgicalKit { value = 1; typeName = "SCALAR"; force = 0; }; class ace_medical_consumeItem_PAK { value = 1; typeName = "SCALAR"; force = 0; }; class ace_medical_consumeItem_SurgicalKit { value = 1; typeName = "SCALAR"; force = 0; }; class ace_medical_useLocation_PAK { value = 3; typeName = "SCALAR"; force = 0; }; class ace_medical_useLocation_SurgicalKit { value = 3; typeName = "SCALAR"; force = 0; }; class ace_medical_keepLocalSettingsSynced { value = 1; typeName = "BOOL"; force = 0; }; class ace_medical_healHitPointAfterAdvBandage { value = 1; typeName = "BOOL"; force = 0; }; class ace_medical_allowUnconsciousAnimationOnTreatment { value = 0; typeName = "BOOL"; force = 0; }; class ace_medical_moveUnitsFromGroupOnUnconscious { value = 0; typeName = "BOOL"; force = 0; }; class ace_medical_menuTypeStyle { value = 0; typeName = "SCALAR"; force = 0; }; class ace_advanced_ballistics_enabled { value = 1; typeName = "BOOL"; force = 0; }; class ace_advanced_ballistics_alwaysSimulateForSnipers { value = 1; typeName = "BOOL"; force = 0; }; class ace_advanced_ballistics_disabledInFullAutoMode { value = 0; typeName = "BOOL"; force = 0; }; class ace_advanced_ballistics_onlyActiveForLocalPlayers { value = 1; typeName = "BOOL"; force = 0; }; class ace_advanced_ballistics_ammoTemperatureEnabled { value = 1; typeName = "BOOL"; force = 0; }; class ace_advanced_ballistics_barrelLengthInfluenceEnabled { value = 1; typeName = "BOOL"; force = 0; }; class ace_advanced_ballistics_bulletTraceEnabled { value = 1; typeName = "BOOL"; force = 0; }; class ace_advanced_ballistics_simulationInterval { value = 0; typeName = "SCALAR"; force = 0; }; class ace_advanced_ballistics_simulationRadius { value = 3000; typeName = "SCALAR"; force = 0; }; ``` **Description:** *We had so far many situations, when player after being hit by bullets, or for example smashing on the ground on parachute were often hurt from both ACE3 Adv. medical and vanilla wounding system. Happened so far on both BWMod soldiers and RHS ones. I even tried gear up vanilla NATO to be 1:1 BWMod soldiers or RHS ones, didn't work. It's problematic cause players often complain that they are still hurt even when they are completely healed (ACE3 wounding). Only way to overcome this is to heal up by using vanilla actions near vanilla medevac vehicle. Seems to happen only on MP.* **Steps to reproduce:** * *Use RHS soldier or BWMod one, also advanced medical system* * *Play on MP* * *Get hit by AI or other player* * *Heal yourself up and observe that you are still yelling from pain, screen is blurry (A3 vanilla wounding)* **Where did the issue occur?** *"Multiplayer"* **RPT log file:** *Well nothing related to medical inside mine. It just happens...*
test
vanilla wounding system kicks in often version mods cba rhs leightopforpack bwmod placed modules class ace common forceallsettings value typename bool force class ace frag enabled value typename bool force class ace interaction enableteammanagement value typename bool force class ace magazinerepack timeperammo value typename scalar force class ace magazinerepack timepermagazine value typename scalar force class ace magazinerepack timeperbeltlink value typename scalar force class ace map bft interval value typename scalar force class ace map bft enabled value typename bool force class ace map bft hideaigroups value typename bool force class ace map mapillumination value typename bool force class ace map mapshake value typename bool force class ace map maplimitzoom value typename bool force class ace map mapshowcursorcoordinates value typename bool force class ace maptools everyonecandrawonbriefing value typename bool force class ace microdagr mapdataavailable value typename scalar force class ace airresistanceenabled value typename bool force class ace allowcomputerrangefinder value typename bool force class ace allowcompass value typename bool force class ace nametags showcursortagforvehicles value typename bool force class ace nametags playernamesviewdistance value typename scalar force class ace nametags playernamesmaxalpha value typename scalar force class ace respawn savepredeathgear value typename bool force class ace respawn removedeadbodiesdisconnected value typename bool force class ace switchunits enableswitchunits value typename bool force class ace switchunits switchtowest value typename bool force class ace switchunits switchtoeast value typename bool force class ace switchunits switchtoindependent value typename bool force class ace switchunits switchtocivilian value typename bool force class ace switchunits enablesafezone value typename bool force class ace switchunits safezoneradius value typename scalar force class ace vehiclelock defaultlockpickstrength value typename scalar force class ace vehiclelock lockvehicleinventory value typename bool force class ace vehiclelock vehiclestartinglockstate value typename scalar force class ace weather enableservercontroller value typename bool force class ace weather useaceweather value typename bool force class ace weather syncrain value typename bool force class ace weather syncwind value typename bool force class ace weather syncmisc value typename bool force class ace weather serverupdateinterval value typename scalar force class ace winddeflection enabled value typename bool force class ace winddeflection vehicleenabled value typename bool force class ace winddeflection simulationinterval value typename scalar force class ace winddeflection simulationradius value typename scalar force class ace explosives requirespecialist value typename bool force class ace explosives punishnonspecialists value typename bool force class ace hearing enablecombatdeafness value typename bool force class ace hearing earplugsvolume value typename scalar force class ace hearing unconsciousnessvolume value typename scalar force class ace medical level value typename scalar force class ace medical medicsetting value typename scalar force class ace medical enablefor value typename scalar force class ace medical enableoverdosing value typename bool force class ace medical bleedingcoefficient value typename scalar force class ace medical paincoefficient value typename scalar force class ace medical enableairway value typename bool force class ace medical enablefractures value typename bool force class ace medical enableadvancedwounds value typename bool force class ace medical enablevehiclecrashes value typename bool force class ace medical enablescreams value typename bool force class ace medical playerdamagethreshold value typename scalar force class ace medical aidamagethreshold value typename scalar force class ace medical enableunconsciousnessai value typename scalar force class ace medical preventinstadeath value typename bool force class ace medical enablerevive value typename scalar force class ace medical maxrevivetime value typename scalar force class ace medical amountofrevivelives value typename scalar force class ace medical allowdeadbodymovement value typename bool force class ace medical allowlittercreation value typename bool force class ace medical littercleanupdelay value typename scalar force class ace medical medicsetting pak value typename scalar force class ace medical medicsetting surgicalkit value typename scalar force class ace medical consumeitem pak value typename scalar force class ace medical consumeitem surgicalkit value typename scalar force class ace medical uselocation pak value typename scalar force class ace medical uselocation surgicalkit value typename scalar force class ace medical keeplocalsettingssynced value typename bool force class ace medical healhitpointafteradvbandage value typename bool force class ace medical allowunconsciousanimationontreatment value typename bool force class ace medical moveunitsfromgrouponunconscious value typename bool force class ace medical menutypestyle value typename scalar force class ace advanced ballistics enabled value typename bool force class ace advanced ballistics alwayssimulateforsnipers value typename bool force class ace advanced ballistics disabledinfullautomode value typename bool force class ace advanced ballistics onlyactiveforlocalplayers value typename bool force class ace advanced ballistics ammotemperatureenabled value typename bool force class ace advanced ballistics barrellengthinfluenceenabled value typename bool force class ace advanced ballistics bullettraceenabled value typename bool force class ace advanced ballistics simulationinterval value typename scalar force class ace advanced ballistics simulationradius value typename scalar force description we had so far many situations when player after being hit by bullets or for example smashing on the ground on parachute were often hurt from both adv medical and vanilla wounding system happened so far on both bwmod soldiers and rhs ones i even tried gear up vanilla nato to be bwmod soldiers or rhs ones didn t work it s problematic cause players often complain that they are still hurt even when they are completely healed wounding only way to overcome this is to heal up by using vanilla actions near vanilla medevac vehicle seems to happen only on mp steps to reproduce use rhs soldier or bwmod one also advanced medical system play on mp get hit by ai or other player heal yourself up and observe that you are still yelling from pain screen is blurry vanilla wounding where did the issue occur multiplayer rpt log file well nothing related to medical inside mine it just happens
1
70,918
7,203,735,656
IssuesEvent
2018-02-06 10:11:50
arescentral/antares
https://api.github.com/repos/arescentral/antares
opened
Unit test fragment shader(s)
Complexity:Low Priority:Low Project:OpenGL Testing Type:Enhancement
The Antares fragment shader has six modes: fill, dither, draw sprite, tint sprite, static sprite, and outline sprite. All of them are seen in one regression test or another, but it would be nice to have unit tests specifically for each of them. It would be particularly good to see a unit test for outlining a sprite with alpha. I think it should work fine for sprites with a hard/anti-aliased edge, but not for sprites with a soft edge. I’m not sure how you *can* outline something that doesn’t have a hard edge. (post-#170, there would be multiple fragment shaders instead of different uniform values; same idea applies in either case)
1.0
Unit test fragment shader(s) - The Antares fragment shader has six modes: fill, dither, draw sprite, tint sprite, static sprite, and outline sprite. All of them are seen in one regression test or another, but it would be nice to have unit tests specifically for each of them. It would be particularly good to see a unit test for outlining a sprite with alpha. I think it should work fine for sprites with a hard/anti-aliased edge, but not for sprites with a soft edge. I’m not sure how you *can* outline something that doesn’t have a hard edge. (post-#170, there would be multiple fragment shaders instead of different uniform values; same idea applies in either case)
test
unit test fragment shader s the antares fragment shader has six modes fill dither draw sprite tint sprite static sprite and outline sprite all of them are seen in one regression test or another but it would be nice to have unit tests specifically for each of them it would be particularly good to see a unit test for outlining a sprite with alpha i think it should work fine for sprites with a hard anti aliased edge but not for sprites with a soft edge i’m not sure how you can outline something that doesn’t have a hard edge post there would be multiple fragment shaders instead of different uniform values same idea applies in either case
1
716,677
24,644,314,429
IssuesEvent
2022-10-17 13:52:57
ramp4-pcar4/story-ramp
https://api.github.com/repos/ramp4-pcar4/story-ramp
closed
Create basic "Switch to French" function
StoryRAMP Editor Needs: estimate Priority: Urgent
_As a Storylines Editor, I need to be able to seamlessly switch between editing the French or English equivalent slide that I'm working on._ Dependant on #226 and #228. Add a tab to the editor that allows a user to flip between English and French products.
1.0
Create basic "Switch to French" function - _As a Storylines Editor, I need to be able to seamlessly switch between editing the French or English equivalent slide that I'm working on._ Dependant on #226 and #228. Add a tab to the editor that allows a user to flip between English and French products.
non_test
create basic switch to french function as a storylines editor i need to be able to seamlessly switch between editing the french or english equivalent slide that i m working on dependant on and add a tab to the editor that allows a user to flip between english and french products
0
7,959
3,641,496,634
IssuesEvent
2016-02-13 17:37:02
Gfurst/Jack-o-Nine-2
https://api.github.com/repos/Gfurst/Jack-o-Nine-2
opened
remake singleton
code improvement
here comes the need again for a unique singleton to handle some major stuff in game, for example asset tracking, scene changes, and some other stuff I forgot
1.0
remake singleton - here comes the need again for a unique singleton to handle some major stuff in game, for example asset tracking, scene changes, and some other stuff I forgot
non_test
remake singleton here comes the need again for a unique singleton to handle some major stuff in game for example asset tracking scene changes and some other stuff i forgot
0
88,115
8,131,975,010
IssuesEvent
2018-08-18 05:07:39
brave/browser-laptop
https://api.github.com/repos/brave/browser-laptop
closed
Transaction goes through even though it shows errors and not executed message
QA/test-plan-specified bug bug in mist feature/ETH-Wallet priority/P3 release-notes/exclude release/blocking reverted
<!-- Have you searched for similar issues? We have received a lot of feedback and bug reports that we have closed as duplicates. Before submitting this issue, please visit our community site for common ones: https://community.brave.com/c/common-issues --> ### Description Transaction goes through even though it shows errors and not executed message ### Steps to Reproduce 1. Build from beeff64 and start with `GETH_LOG=1 LEDGER_ENVIRONMENT=staging npm run start` 2. Enable ETH-Wallet and create a new ETH-Wallet 3. Fund newly created Wallet with ETH (I used 0.015 ETH) 4. Create Brave Wallet 5. Switch to ETH-Wallet and click on Fund Brave Wallet 6. Enter Amount as `0.015` and retain default fee value and click on Send 7. Enter password to complete transaction and wait for the transaction to get listed 8. Open transaction link which shows as success (Mine was https://etherscan.io/tx/0x4e6eeb8abccd84ce6adce333f0584d4ad4e05aa0af77b2015c9b236b05c3377e) 9. Click on latest transaction, shows error message about not sufficient gas to execute ![image](https://user-images.githubusercontent.com/17010094/43621399-3232ee7c-96a5-11e8-9dfb-c83a0719c951.png) 10. Wait for few mins, Transaction failed notifications show up in ETH-Wallet ![image](https://user-images.githubusercontent.com/17010094/43621421-55f9baca-96a5-11e8-9cb4-939d086c7f88.png) **Actual result:** Transaction goes through even though it shows errors and not executed message **Expected result:** Should properly show if the transaction is done successfully or not **Reproduces how often:** Just tried once since its on main net ### Brave Version **about:brave info:** beeff64 **Reproducible on current live release:** N/A ### Additional Information cc: @LaurenWags @kjozwiak @btlechowski
1.0
Transaction goes through even though it shows errors and not executed message - <!-- Have you searched for similar issues? We have received a lot of feedback and bug reports that we have closed as duplicates. Before submitting this issue, please visit our community site for common ones: https://community.brave.com/c/common-issues --> ### Description Transaction goes through even though it shows errors and not executed message ### Steps to Reproduce 1. Build from beeff64 and start with `GETH_LOG=1 LEDGER_ENVIRONMENT=staging npm run start` 2. Enable ETH-Wallet and create a new ETH-Wallet 3. Fund newly created Wallet with ETH (I used 0.015 ETH) 4. Create Brave Wallet 5. Switch to ETH-Wallet and click on Fund Brave Wallet 6. Enter Amount as `0.015` and retain default fee value and click on Send 7. Enter password to complete transaction and wait for the transaction to get listed 8. Open transaction link which shows as success (Mine was https://etherscan.io/tx/0x4e6eeb8abccd84ce6adce333f0584d4ad4e05aa0af77b2015c9b236b05c3377e) 9. Click on latest transaction, shows error message about not sufficient gas to execute ![image](https://user-images.githubusercontent.com/17010094/43621399-3232ee7c-96a5-11e8-9dfb-c83a0719c951.png) 10. Wait for few mins, Transaction failed notifications show up in ETH-Wallet ![image](https://user-images.githubusercontent.com/17010094/43621421-55f9baca-96a5-11e8-9cb4-939d086c7f88.png) **Actual result:** Transaction goes through even though it shows errors and not executed message **Expected result:** Should properly show if the transaction is done successfully or not **Reproduces how often:** Just tried once since its on main net ### Brave Version **about:brave info:** beeff64 **Reproducible on current live release:** N/A ### Additional Information cc: @LaurenWags @kjozwiak @btlechowski
test
transaction goes through even though it shows errors and not executed message have you searched for similar issues we have received a lot of feedback and bug reports that we have closed as duplicates before submitting this issue please visit our community site for common ones description transaction goes through even though it shows errors and not executed message steps to reproduce build from and start with geth log ledger environment staging npm run start enable eth wallet and create a new eth wallet fund newly created wallet with eth i used eth create brave wallet switch to eth wallet and click on fund brave wallet enter amount as and retain default fee value and click on send enter password to complete transaction and wait for the transaction to get listed open transaction link which shows as success mine was click on latest transaction shows error message about not sufficient gas to execute wait for few mins transaction failed notifications show up in eth wallet actual result transaction goes through even though it shows errors and not executed message expected result should properly show if the transaction is done successfully or not reproduces how often just tried once since its on main net brave version about brave info reproducible on current live release n a additional information cc laurenwags kjozwiak btlechowski
1
192,340
14,614,620,210
IssuesEvent
2020-12-22 10:11:24
github-vet/rangeloop-pointer-findings
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
closed
dedisuryadi/tile38: vendor/github.com/tidwall/geoindex/geoindex_test.go; 4 LoC
fresh test tiny vendored
Found a possible issue in [dedisuryadi/tile38](https://www.github.com/dedisuryadi/tile38) at [vendor/github.com/tidwall/geoindex/geoindex_test.go](https://github.com/dedisuryadi/tile38/blob/d84903ea4f21b75166220d1ed8ecfb935994593d/vendor/github.com/tidwall/geoindex/geoindex_test.go#L365-L368) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to city at line 367 may start a goroutine [Click here to see the code in its original context.](https://github.com/dedisuryadi/tile38/blob/d84903ea4f21b75166220d1ed8ecfb935994593d/vendor/github.com/tidwall/geoindex/geoindex_test.go#L365-L368) <details> <summary>Click here to show the 4 line(s) of Go which triggered the analyzer.</summary> ```go for _, city := range cities.Cities { p := [2]float64{city.Longitude, city.Latitude} tr.Insert(p, p, &city) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: d84903ea4f21b75166220d1ed8ecfb935994593d
1.0
dedisuryadi/tile38: vendor/github.com/tidwall/geoindex/geoindex_test.go; 4 LoC - Found a possible issue in [dedisuryadi/tile38](https://www.github.com/dedisuryadi/tile38) at [vendor/github.com/tidwall/geoindex/geoindex_test.go](https://github.com/dedisuryadi/tile38/blob/d84903ea4f21b75166220d1ed8ecfb935994593d/vendor/github.com/tidwall/geoindex/geoindex_test.go#L365-L368) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to city at line 367 may start a goroutine [Click here to see the code in its original context.](https://github.com/dedisuryadi/tile38/blob/d84903ea4f21b75166220d1ed8ecfb935994593d/vendor/github.com/tidwall/geoindex/geoindex_test.go#L365-L368) <details> <summary>Click here to show the 4 line(s) of Go which triggered the analyzer.</summary> ```go for _, city := range cities.Cities { p := [2]float64{city.Longitude, city.Latitude} tr.Insert(p, p, &city) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: d84903ea4f21b75166220d1ed8ecfb935994593d
test
dedisuryadi vendor github com tidwall geoindex geoindex test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to city at line may start a goroutine click here to show the line s of go which triggered the analyzer go for city range cities cities p city longitude city latitude tr insert p p city leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
182,277
14,112,590,796
IssuesEvent
2020-11-07 06:19:08
apache/incubator-mxnet
https://api.github.com/repos/apache/incubator-mxnet
closed
test_operator_gpu.test_ndarray_reshape fails
Flaky Test
test_operator_gpu.test_ndarray_reshape Pipeline - http://jenkins.mxnet-ci.amazon-ml.com/blue/organizations/jenkins/mxnet-validation%2Funix-gpu/detail/PR-15807/4/pipeline Error Log ``` test_operator_gpu.test_ndarray_reshape ... [18:55:15] src/operator/tensor/./.././../common/../operator/mxnet_op.h:845: Check failed: (err) == (cudaSuccess) Name: mxnet_generic_kernel ErrStr:an illegal memory access was encountered /work/runtime_functions.sh: line 889: 6 Aborted (core dumped) nosetests-3.4 $NOSE_COVERAGE_ARGUMENTS $NOSE_TIMER_ARGUMENTS --with-xunit --xunit-file nosetests_gpu.xml --verbose tests/python/gpu build.py: 2019-08-09 18:56:05,569Z INFO Waiting for status of container 97fd5051cbd5 for 600 s. build.py: 2019-08-09 18:56:05,747Z INFO Container exit status: {'Error': None, 'StatusCode': 134} build.py: 2019-08-09 18:56:05,747Z ERROR Container exited with an error 😞 build.py: 2019-08-09 18:56:05,747Z INFO Executed command for reproduction: ci/build.py --docker-registry mxnetci --nvidiadocker --platform ubuntu_gpu_cu101 --docker-build-retries 3 --shm-size 500m /work/runtime_functions.sh unittest_ubuntu_python3_gpu ``` PR #15807
1.0
test_operator_gpu.test_ndarray_reshape fails - test_operator_gpu.test_ndarray_reshape Pipeline - http://jenkins.mxnet-ci.amazon-ml.com/blue/organizations/jenkins/mxnet-validation%2Funix-gpu/detail/PR-15807/4/pipeline Error Log ``` test_operator_gpu.test_ndarray_reshape ... [18:55:15] src/operator/tensor/./.././../common/../operator/mxnet_op.h:845: Check failed: (err) == (cudaSuccess) Name: mxnet_generic_kernel ErrStr:an illegal memory access was encountered /work/runtime_functions.sh: line 889: 6 Aborted (core dumped) nosetests-3.4 $NOSE_COVERAGE_ARGUMENTS $NOSE_TIMER_ARGUMENTS --with-xunit --xunit-file nosetests_gpu.xml --verbose tests/python/gpu build.py: 2019-08-09 18:56:05,569Z INFO Waiting for status of container 97fd5051cbd5 for 600 s. build.py: 2019-08-09 18:56:05,747Z INFO Container exit status: {'Error': None, 'StatusCode': 134} build.py: 2019-08-09 18:56:05,747Z ERROR Container exited with an error 😞 build.py: 2019-08-09 18:56:05,747Z INFO Executed command for reproduction: ci/build.py --docker-registry mxnetci --nvidiadocker --platform ubuntu_gpu_cu101 --docker-build-retries 3 --shm-size 500m /work/runtime_functions.sh unittest_ubuntu_python3_gpu ``` PR #15807
test
test operator gpu test ndarray reshape fails test operator gpu test ndarray reshape pipeline error log test operator gpu test ndarray reshape src operator tensor common operator mxnet op h check failed err cudasuccess name mxnet generic kernel errstr an illegal memory access was encountered work runtime functions sh line aborted core dumped nosetests nose coverage arguments nose timer arguments with xunit xunit file nosetests gpu xml verbose tests python gpu build py info waiting for status of container for s build py info container exit status error none statuscode build py error container exited with an error 😞 build py info executed command for reproduction ci build py docker registry mxnetci nvidiadocker platform ubuntu gpu docker build retries shm size work runtime functions sh unittest ubuntu gpu pr
1
61,405
6,735,318,548
IssuesEvent
2017-10-18 21:18:45
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Not able to start LB containers with older images.
area/balancer area/loadbalancer kind/bug status/resolved status/to-test
**Rancher versions:** rancher/server: v1.6-development Steps to reproduce the problem: Scenario1: Create a new LB service using older image ```rancher/lb-service-haproxy:v0.7.9``` using rancher cli. Lb service fails to start with the following errors in logs: ``` 10/10/2017 2:21:11 PMtime="2017-10-10T21:21:11Z" level=error msg="Failed to initialize Kubernetes controller: KUBERNETES_URL is not set" 10/10/2017 2:21:11 PMtime="2017-10-10T21:21:11Z" level=info msg="CATTLE_ACCESS_KEY is not set, skipping init of Rancher LB provider" 10/10/2017 2:21:11 PMtime="2017-10-10T21:21:11Z" level=info msg="Starting Rancher LB service" 10/10/2017 2:21:11 PMtime="2017-10-10T21:21:11Z" level=fatal msg="CATTLE_ACCESS_KEY is not set, fail to init of Rancher LB provider" ```
1.0
Not able to start LB containers with older images. - **Rancher versions:** rancher/server: v1.6-development Steps to reproduce the problem: Scenario1: Create a new LB service using older image ```rancher/lb-service-haproxy:v0.7.9``` using rancher cli. Lb service fails to start with the following errors in logs: ``` 10/10/2017 2:21:11 PMtime="2017-10-10T21:21:11Z" level=error msg="Failed to initialize Kubernetes controller: KUBERNETES_URL is not set" 10/10/2017 2:21:11 PMtime="2017-10-10T21:21:11Z" level=info msg="CATTLE_ACCESS_KEY is not set, skipping init of Rancher LB provider" 10/10/2017 2:21:11 PMtime="2017-10-10T21:21:11Z" level=info msg="Starting Rancher LB service" 10/10/2017 2:21:11 PMtime="2017-10-10T21:21:11Z" level=fatal msg="CATTLE_ACCESS_KEY is not set, fail to init of Rancher LB provider" ```
test
not able to start lb containers with older images rancher versions rancher server development steps to reproduce the problem create a new lb service using older image rancher lb service haproxy using rancher cli lb service fails to start with the following errors in logs pmtime level error msg failed to initialize kubernetes controller kubernetes url is not set pmtime level info msg cattle access key is not set skipping init of rancher lb provider pmtime level info msg starting rancher lb service pmtime level fatal msg cattle access key is not set fail to init of rancher lb provider
1
141,829
18,990,634,908
IssuesEvent
2021-11-22 06:41:26
ignatandrei/BlocklyAutomation
https://api.github.com/repos/ignatandrei/BlocklyAutomation
opened
CVE-2021-3807 (High) detected in ansi-regex-4.1.0.tgz
security vulnerability
## CVE-2021-3807 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ansi-regex-4.1.0.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz</a></p> <p>Path to dependency file: BlocklyAutomation/src/package.json</p> <p>Path to vulnerable library: BlocklyAutomation/src/node_modules/cliui/node_modules/ansi-regex/package.json,BlocklyAutomation/src/node_modules/yargs/node_modules/ansi-regex/package.json,BlocklyAutomation/src/node_modules/wrap-ansi/node_modules/ansi-regex/package.json</p> <p> Dependency Hierarchy: - build-angular-12.2.13.tgz (Root Library) - webpack-dev-server-3.11.2.tgz - yargs-13.3.2.tgz - string-width-3.1.0.tgz - strip-ansi-5.2.0.tgz - :x: **ansi-regex-4.1.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ignatandrei/BlocklyAutomation/commit/9f77708e72b991e3017aa5d983141ab65f412ffd">9f77708e72b991e3017aa5d983141ab65f412ffd</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ansi-regex is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807>CVE-2021-3807</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/">https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/</a></p> <p>Release Date: 2021-09-17</p> <p>Fix Resolution: ansi-regex - 5.0.1,6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-3807 (High) detected in ansi-regex-4.1.0.tgz - ## CVE-2021-3807 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ansi-regex-4.1.0.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz</a></p> <p>Path to dependency file: BlocklyAutomation/src/package.json</p> <p>Path to vulnerable library: BlocklyAutomation/src/node_modules/cliui/node_modules/ansi-regex/package.json,BlocklyAutomation/src/node_modules/yargs/node_modules/ansi-regex/package.json,BlocklyAutomation/src/node_modules/wrap-ansi/node_modules/ansi-regex/package.json</p> <p> Dependency Hierarchy: - build-angular-12.2.13.tgz (Root Library) - webpack-dev-server-3.11.2.tgz - yargs-13.3.2.tgz - string-width-3.1.0.tgz - strip-ansi-5.2.0.tgz - :x: **ansi-regex-4.1.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ignatandrei/BlocklyAutomation/commit/9f77708e72b991e3017aa5d983141ab65f412ffd">9f77708e72b991e3017aa5d983141ab65f412ffd</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ansi-regex is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807>CVE-2021-3807</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/">https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/</a></p> <p>Release Date: 2021-09-17</p> <p>Fix Resolution: ansi-regex - 5.0.1,6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in ansi regex tgz cve high severity vulnerability vulnerable library ansi regex tgz regular expression for matching ansi escape codes library home page a href path to dependency file blocklyautomation src package json path to vulnerable library blocklyautomation src node modules cliui node modules ansi regex package json blocklyautomation src node modules yargs node modules ansi regex package json blocklyautomation src node modules wrap ansi node modules ansi regex package json dependency hierarchy build angular tgz root library webpack dev server tgz yargs tgz string width tgz strip ansi tgz x ansi regex tgz vulnerable library found in head commit a href found in base branch main vulnerability details ansi regex is vulnerable to inefficient regular expression complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ansi regex step up your open source security game with whitesource
0
287,346
21,651,038,811
IssuesEvent
2022-05-06 09:22:36
keploy/keploy
https://api.github.com/repos/keploy/keploy
closed
Add CITATION.cff
documentation good first issue
### What do you want to add to the docs? (please state reasons if any) The standard CITATION format is the [Citation File Format](https://citation-file-format.github.io/), proposed by GitHub: `CITATION.cff ` files are plain text files with human- and machine-readable citation information for software (and datasets). Code developers can include them in their repositories to let others know how to correctly cite their software. example : ```yaml cff-version: 1.2.0 title: Beautify GitHub sample repository message: If you really want to cite this repository, here's how you should cite it. type: software authors: - given-names: Nicolas family-names: Fränkel repository-code: 'https://github.com/ajavageek/beautifygithub' license: Unlicense ``` ### Where is this stated? The CITATION file allows you to answer the following questions: What is the name of the software? What label should I use to uniquely identify the version of the software I have used? What is the appropriate set of people that should be cited as authors?
1.0
Add CITATION.cff - ### What do you want to add to the docs? (please state reasons if any) The standard CITATION format is the [Citation File Format](https://citation-file-format.github.io/), proposed by GitHub: `CITATION.cff ` files are plain text files with human- and machine-readable citation information for software (and datasets). Code developers can include them in their repositories to let others know how to correctly cite their software. example : ```yaml cff-version: 1.2.0 title: Beautify GitHub sample repository message: If you really want to cite this repository, here's how you should cite it. type: software authors: - given-names: Nicolas family-names: Fränkel repository-code: 'https://github.com/ajavageek/beautifygithub' license: Unlicense ``` ### Where is this stated? The CITATION file allows you to answer the following questions: What is the name of the software? What label should I use to uniquely identify the version of the software I have used? What is the appropriate set of people that should be cited as authors?
non_test
add citation cff what do you want to add to the docs please state reasons if any the standard citation format is the proposed by github citation cff files are plain text files with human and machine readable citation information for software and datasets code developers can include them in their repositories to let others know how to correctly cite their software example yaml cff version title beautify github sample repository message if you really want to cite this repository here s how you should cite it type software authors given names nicolas family names fränkel repository code license unlicense where is this stated the citation file allows you to answer the following questions what is the name of the software what label should i use to uniquely identify the version of the software i have used what is the appropriate set of people that should be cited as authors
0
103,729
8,941,333,602
IssuesEvent
2019-01-24 03:53:10
phetsims/coulombs-law
https://api.github.com/repos/phetsims/coulombs-law
closed
CT running lint-all task
type:automated-testing
``` coulombs-law : build : phet Failure to grunt snapshot-1548281897096/coulombs-law: Running "lint-all" task Running "report-media" task >> missing-file: coulombs-law/images/image06.png Fatal error: There is an issue with the licenses for media types. Approximately 1/23/2019, 3:18:17 PM coulombs-law : build : phet-io Failure to grunt snapshot-1548281897096-phet-io/coulombs-law: Running "lint-all" task Running "report-media" task >> missing-file: coulombs-law/images/image06.png Fatal error: There is an issue with the licenses for media types. Approximately 1/23/2019, 3:18:17 PM ```
1.0
CT running lint-all task - ``` coulombs-law : build : phet Failure to grunt snapshot-1548281897096/coulombs-law: Running "lint-all" task Running "report-media" task >> missing-file: coulombs-law/images/image06.png Fatal error: There is an issue with the licenses for media types. Approximately 1/23/2019, 3:18:17 PM coulombs-law : build : phet-io Failure to grunt snapshot-1548281897096-phet-io/coulombs-law: Running "lint-all" task Running "report-media" task >> missing-file: coulombs-law/images/image06.png Fatal error: There is an issue with the licenses for media types. Approximately 1/23/2019, 3:18:17 PM ```
test
ct running lint all task coulombs law build phet failure to grunt snapshot coulombs law running lint all task running report media task missing file coulombs law images png fatal error there is an issue with the licenses for media types  approximately pm coulombs law build phet io failure to grunt snapshot phet io coulombs law running lint all task running report media task missing file coulombs law images png fatal error there is an issue with the licenses for media types  approximately pm
1
203,597
15,376,356,805
IssuesEvent
2021-03-02 15:53:41
Pack-Smart/pack-smart-fe
https://api.github.com/repos/Pack-Smart/pack-smart-fe
opened
Cypress
Frontend Phase 5 testing
As a developer, When I create an app, I want to see integration testing, So that I have a robust testing suite .
1.0
Cypress - As a developer, When I create an app, I want to see integration testing, So that I have a robust testing suite .
test
cypress as a developer when i create an app i want to see integration testing so that i have a robust testing suite
1
528,836
15,375,396,217
IssuesEvent
2021-03-02 14:53:24
woocommerce/woocommerce-gutenberg-products-block
https://api.github.com/repos/woocommerce/woocommerce-gutenberg-products-block
closed
Full site editor renders blank page with WooCommerce activated
priority: high type: bug
Full Site Editor renders blank when WooCommerce is enabled. This issue was reported by a user in GB repo (https://github.com/WordPress/gutenberg/issues/27938) and according to the discussion there, the cause might be the work-around introduced in this PR: https://github.com/woocommerce/woocommerce-gutenberg-products-block/pull/3219. ### Steps to reproduce 1. Install a FSE compatible theme (ie, [Armando](https://wordpress.org/themes/download/armando.1.0.0.zip?nostats=1)). 2. Enable Gutenberg and WooCommerce. 3. Click on the _Site Editor_ sidebar menu item. 4. Notice the page renders blank.
1.0
Full site editor renders blank page with WooCommerce activated - Full Site Editor renders blank when WooCommerce is enabled. This issue was reported by a user in GB repo (https://github.com/WordPress/gutenberg/issues/27938) and according to the discussion there, the cause might be the work-around introduced in this PR: https://github.com/woocommerce/woocommerce-gutenberg-products-block/pull/3219. ### Steps to reproduce 1. Install a FSE compatible theme (ie, [Armando](https://wordpress.org/themes/download/armando.1.0.0.zip?nostats=1)). 2. Enable Gutenberg and WooCommerce. 3. Click on the _Site Editor_ sidebar menu item. 4. Notice the page renders blank.
non_test
full site editor renders blank page with woocommerce activated full site editor renders blank when woocommerce is enabled this issue was reported by a user in gb repo and according to the discussion there the cause might be the work around introduced in this pr steps to reproduce install a fse compatible theme ie enable gutenberg and woocommerce click on the site editor sidebar menu item notice the page renders blank
0
235,392
7,737,953,788
IssuesEvent
2018-05-28 10:04:04
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
Controllers should define flags in their own codebase.
lifecycle/rotten priority/backlog sig/api-machinery
Currently all flags are defined in the binary which makes a big mess. We need some way to deal with duplicate names. cc @kubernetes/goog-control-plane
1.0
Controllers should define flags in their own codebase. - Currently all flags are defined in the binary which makes a big mess. We need some way to deal with duplicate names. cc @kubernetes/goog-control-plane
non_test
controllers should define flags in their own codebase currently all flags are defined in the binary which makes a big mess we need some way to deal with duplicate names cc kubernetes goog control plane
0
224,847
17,778,425,862
IssuesEvent
2021-08-30 22:54:53
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
[Test UI] Testresult are not shown if running tests over different test controller/profiles
bug unreleased testing
<!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ --> <!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ --> <!-- 🕮 Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions --> <!-- 🔎 Search existing issues to avoid creating duplicates. --> <!-- 🧪 Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ --> <!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. --> <!-- 🔧 Launch with `code --disable-extensions` to check. --> Does this issue occur when all extensions are disabled?: Yes/No <!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. --> <!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. --> - VS Code Version: 1.59 and 1.60-insider - OS Version: Windows 10/Arch Linux If you have more then one test controller registered, and if you want to run all tests with the "Run Tests" button, then the test run stops after the first controller ends the test run. Or better, the result of the other, longer running test runs are not shown in the UI. If you run only one test profile, everything works fine. See here ... ![sample-1629983529652](https://user-images.githubusercontent.com/7069968/130968950-1da69121-b38d-40d5-a944-74055a4e3813.gif) There is a message in debug console: Setting the state of test "file:///c%3A/tmp/vscodetest/test-provider-sample/sample/test.md/2 + 2 = 5" is a no-op after the run ends. Steps to Reproduce: To repoduce this, you need at least 2 test controllers registered. For testing purposes, I have prepared a github repo based on the test-provider-sample and the insider version of the python extension. https://github.com/d-biehl/test-provider-sample --- just one thougth: it seems, running tests over different test controllers, the test profiles are executed in parallel. I think, it should be better to run the test profiles sequentialy or it should be configurable. Because if there could be side effects running tests profiles in parallel, ie. using the same resources or testing the UI with different language or so...
1.0
[Test UI] Testresult are not shown if running tests over different test controller/profiles - <!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ --> <!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ --> <!-- 🕮 Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions --> <!-- 🔎 Search existing issues to avoid creating duplicates. --> <!-- 🧪 Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ --> <!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. --> <!-- 🔧 Launch with `code --disable-extensions` to check. --> Does this issue occur when all extensions are disabled?: Yes/No <!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. --> <!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. --> - VS Code Version: 1.59 and 1.60-insider - OS Version: Windows 10/Arch Linux If you have more then one test controller registered, and if you want to run all tests with the "Run Tests" button, then the test run stops after the first controller ends the test run. Or better, the result of the other, longer running test runs are not shown in the UI. If you run only one test profile, everything works fine. See here ... ![sample-1629983529652](https://user-images.githubusercontent.com/7069968/130968950-1da69121-b38d-40d5-a944-74055a4e3813.gif) There is a message in debug console: Setting the state of test "file:///c%3A/tmp/vscodetest/test-provider-sample/sample/test.md/2 + 2 = 5" is a no-op after the run ends. Steps to Reproduce: To repoduce this, you need at least 2 test controllers registered. For testing purposes, I have prepared a github repo based on the test-provider-sample and the insider version of the python extension. https://github.com/d-biehl/test-provider-sample --- just one thougth: it seems, running tests over different test controllers, the test profiles are executed in parallel. I think, it should be better to run the test profiles sequentialy or it should be configurable. Because if there could be side effects running tests profiles in parallel, ie. using the same resources or testing the UI with different language or so...
test
testresult are not shown if running tests over different test controller profiles does this issue occur when all extensions are disabled yes no report issue dialog can assist with this vs code version and insider os version windows arch linux if you have more then one test controller registered and if you want to run all tests with the run tests button then the test run stops after the first controller ends the test run or better the result of the other longer running test runs are not shown in the ui if you run only one test profile everything works fine see here there is a message in debug console setting the state of test file c tmp vscodetest test provider sample sample test md is a no op after the run ends steps to reproduce to repoduce this you need at least test controllers registered for testing purposes i have prepared a github repo based on the test provider sample and the insider version of the python extension just one thougth it seems running tests over different test controllers the test profiles are executed in parallel i think it should be better to run the test profiles sequentialy or it should be configurable because if there could be side effects running tests profiles in parallel ie using the same resources or testing the ui with different language or so
1
21,339
11,196,519,372
IssuesEvent
2020-01-03 10:21:03
mozilla-mobile/fenix
https://api.github.com/repos/mozilla-mobile/fenix
closed
Move unsetOpenLinksInAPrivateTabIfNecessary() to worker thread to gain some time on start up
P1 eng:performance eng:qa:needed
### Why/User Benefit/User Problem unsetOpenLinksInAPrivateTabIfNecessary() seems to be called on start of Fenix most of the time. It can take from 50ms to 150ms, although not too sure why, it varies with my starts. *There is a comment in the method that says : ` // Toggle off the open_link_in_private_tab pref if we are no longer set as the default browser` ` // We do this on a separate thread to alleviate performance issues` However we do `lifeCycleScope.launch {...}` without calling any other dispatchers to start the Job, which means it runs on the UI thread. ### Impact This is a low hanging fruit with a quick fix. It just has to be changed to a different dispatcher ### Acceptance Criteria (how do I know when I’m done?) unsetOpenLinksInAPrivateTabIfNecessary no longer runs on the Main thread when starting Fenix
True
Move unsetOpenLinksInAPrivateTabIfNecessary() to worker thread to gain some time on start up - ### Why/User Benefit/User Problem unsetOpenLinksInAPrivateTabIfNecessary() seems to be called on start of Fenix most of the time. It can take from 50ms to 150ms, although not too sure why, it varies with my starts. *There is a comment in the method that says : ` // Toggle off the open_link_in_private_tab pref if we are no longer set as the default browser` ` // We do this on a separate thread to alleviate performance issues` However we do `lifeCycleScope.launch {...}` without calling any other dispatchers to start the Job, which means it runs on the UI thread. ### Impact This is a low hanging fruit with a quick fix. It just has to be changed to a different dispatcher ### Acceptance Criteria (how do I know when I’m done?) unsetOpenLinksInAPrivateTabIfNecessary no longer runs on the Main thread when starting Fenix
non_test
move unsetopenlinksinaprivatetabifnecessary to worker thread to gain some time on start up why user benefit user problem unsetopenlinksinaprivatetabifnecessary seems to be called on start of fenix most of the time it can take from to although not too sure why it varies with my starts there is a comment in the method that says toggle off the open link in private tab pref if we are no longer set as the default browser we do this on a separate thread to alleviate performance issues however we do lifecyclescope launch without calling any other dispatchers to start the job which means it runs on the ui thread impact this is a low hanging fruit with a quick fix it just has to be changed to a different dispatcher acceptance criteria how do i know when i’m done unsetopenlinksinaprivatetabifnecessary no longer runs on the main thread when starting fenix
0
278,975
24,189,161,212
IssuesEvent
2022-09-23 15:47:05
junit-team/junit5
https://api.github.com/repos/junit-team/junit5
reopened
ParameterizedTestExtension disallows acceptable case of zero invocations
status: team discussion component: Jupiter theme: parameterized tests status: stale
## Overview Using: JUnit: 5.2.0 Great job on the new `ParameterizedTest` support in v.5. The replacement of the static, one-per-class Parameters annotation with more flexible `MethodSource`, etc. has been like a breath of fresh air and allowed me to remove thousands (!!) of lines of supporting code from my system. I'm really loving it. However, someone decided to disallow zero parameters with this precondition check in `ParameterizedTestExtension`. ```java .onClose(() -> Preconditions.condition(invocationCount.get() > 0, "Configuration error: You must provide at least one argument for this @ParameterizedTest")); ``` The problem with this is that we have some testing situations where parameterized tests with zero parameters are not exceptional. For example, we run tests against thousands of a certain type of class generically against a database of past production failures, and many of these classes have never experienced a production failure. When the tests are run, we now get failures due the above precondition check. JUnit 4 handled this cleanly: it would simply not run any tests on those classes. ## Workaround I can get around this by adding a `null` to the method creating the collection of parameters if it is empty, and then return from the beginning of the `@ParameterizedTest` method code if the passed parameter is `null`. That lets us continue to run the parameterized tests against all of the classes, but comes with some disadvantages: - We must add specific code to the front and back of every parameterized pair just to avoid a failure that doesn't matter to the tests. - The handling of the `null` causes the test count inflation for these "phantom" tests. - `null` is now reserved as a signal for no parameters, rather than something wrong in the parameter creation. ## Proposal If nobody has any strong feelings about disallowing the no-parameter case, can we just have this precondition removed from a future version? Thanks.
1.0
ParameterizedTestExtension disallows acceptable case of zero invocations - ## Overview Using: JUnit: 5.2.0 Great job on the new `ParameterizedTest` support in v.5. The replacement of the static, one-per-class Parameters annotation with more flexible `MethodSource`, etc. has been like a breath of fresh air and allowed me to remove thousands (!!) of lines of supporting code from my system. I'm really loving it. However, someone decided to disallow zero parameters with this precondition check in `ParameterizedTestExtension`. ```java .onClose(() -> Preconditions.condition(invocationCount.get() > 0, "Configuration error: You must provide at least one argument for this @ParameterizedTest")); ``` The problem with this is that we have some testing situations where parameterized tests with zero parameters are not exceptional. For example, we run tests against thousands of a certain type of class generically against a database of past production failures, and many of these classes have never experienced a production failure. When the tests are run, we now get failures due the above precondition check. JUnit 4 handled this cleanly: it would simply not run any tests on those classes. ## Workaround I can get around this by adding a `null` to the method creating the collection of parameters if it is empty, and then return from the beginning of the `@ParameterizedTest` method code if the passed parameter is `null`. That lets us continue to run the parameterized tests against all of the classes, but comes with some disadvantages: - We must add specific code to the front and back of every parameterized pair just to avoid a failure that doesn't matter to the tests. - The handling of the `null` causes the test count inflation for these "phantom" tests. - `null` is now reserved as a signal for no parameters, rather than something wrong in the parameter creation. ## Proposal If nobody has any strong feelings about disallowing the no-parameter case, can we just have this precondition removed from a future version? Thanks.
test
parameterizedtestextension disallows acceptable case of zero invocations overview using junit great job on the new parameterizedtest support in v the replacement of the static one per class parameters annotation with more flexible methodsource etc has been like a breath of fresh air and allowed me to remove thousands of lines of supporting code from my system i m really loving it however someone decided to disallow zero parameters with this precondition check in parameterizedtestextension java onclose preconditions condition invocationcount get configuration error you must provide at least one argument for this parameterizedtest the problem with this is that we have some testing situations where parameterized tests with zero parameters are not exceptional for example we run tests against thousands of a certain type of class generically against a database of past production failures and many of these classes have never experienced a production failure when the tests are run we now get failures due the above precondition check junit handled this cleanly it would simply not run any tests on those classes workaround i can get around this by adding a null to the method creating the collection of parameters if it is empty and then return from the beginning of the parameterizedtest method code if the passed parameter is null that lets us continue to run the parameterized tests against all of the classes but comes with some disadvantages we must add specific code to the front and back of every parameterized pair just to avoid a failure that doesn t matter to the tests the handling of the null causes the test count inflation for these phantom tests null is now reserved as a signal for no parameters rather than something wrong in the parameter creation proposal if nobody has any strong feelings about disallowing the no parameter case can we just have this precondition removed from a future version thanks
1
701,540
24,100,143,304
IssuesEvent
2022-09-19 23:14:31
googleapis/nodejs-media-translation
https://api.github.com/repos/googleapis/nodejs-media-translation
closed
MediaTranslation: should translate from a streamed file failed
type: bug priority: p1 api: mediatranslation flakybot: issue
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 8257ade7fd05bd65161d7a8c156e6c04fc09bfbe buildURL: [Build Status](https://source.cloud.google.com/results/invocations/2e29d795-c2f5-4ba6-a160-b1fb7a6e9692), [Sponge](http://sponge2/2e29d795-c2f5-4ba6-a160-b1fb7a6e9692) status: failed <details><summary>Test output</summary><br><pre>Command failed: node translate_from_file.js /workspace/samples/resources/audio.raw linear16 en-US es-ES events.js:291 throw er; // Unhandled 'error' event ^ Error: 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. at Object.callErrorFromStatus (/workspace/node_modules/@grpc/grpc-js/build/src/call.js:31:19) at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client.js:413:49) at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:328:181) at /workspace/node_modules/@grpc/grpc-js/build/src/call-stream.js:188:78 at processTicksAndRejections (internal/process/task_queues.js:79:11) for call at at ServiceClientImpl.makeBidiStreamRequest (/workspace/node_modules/@grpc/grpc-js/build/src/client.js:398:30) at ServiceClientImpl.<anonymous> (/workspace/node_modules/@grpc/grpc-js/build/src/make-client.js:105:19) at /workspace/build/src/v1beta1/speech_translation_service_client.js:190:29 at /workspace/node_modules/google-gax/build/src/streamingCalls/streamingApiCaller.js:46:28 at /workspace/node_modules/google-gax/build/src/normalCalls/timeout.js:44:16 at StreamProxy.setStream (/workspace/node_modules/google-gax/build/src/streamingCalls/streaming.js:144:24) at StreamingApiCaller.call (/workspace/node_modules/google-gax/build/src/streamingCalls/streamingApiCaller.js:54:16) at /workspace/node_modules/google-gax/build/src/createApiCall.js:84:30 at processTicksAndRejections (internal/process/task_queues.js:97:5) Emitted 'error' event on StreamProxy instance at: at errorOrDestroy (/workspace/node_modules/readable-stream/lib/internal/streams/destroy.js:98:101) at onwriteError (/workspace/node_modules/readable-stream/lib/_stream_writable.js:430:5) at onwrite (/workspace/node_modules/readable-stream/lib/_stream_writable.js:450:11) at WritableState.onwrite (/workspace/node_modules/readable-stream/lib/_stream_writable.js:160:5) at StreamProxy.Duplexify._destroy (/workspace/node_modules/duplexify/index.js:194:18) at /workspace/node_modules/duplexify/index.js:185:10 at processTicksAndRejections (internal/process/task_queues.js:79:11) { code: 16, details: 'Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.', metadata: Metadata { internalRepr: Map { 'grpc-server-stats-bin' => [ Buffer(10) [Uint8Array] [ 0, 0, 217, 248, 68, 0, 0, 0, 0, 0 ] ] }, options: {} } } Error: Command failed: node translate_from_file.js /workspace/samples/resources/audio.raw linear16 en-US es-ES events.js:291 throw er; // Unhandled 'error' event ^ Error: 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. at Object.callErrorFromStatus (/workspace/node_modules/@grpc/grpc-js/build/src/call.js:31:19) at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client.js:413:49) at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:328:181) at /workspace/node_modules/@grpc/grpc-js/build/src/call-stream.js:188:78 at processTicksAndRejections (internal/process/task_queues.js:79:11) for call at at ServiceClientImpl.makeBidiStreamRequest (/workspace/node_modules/@grpc/grpc-js/build/src/client.js:398:30) at ServiceClientImpl.<anonymous> (/workspace/node_modules/@grpc/grpc-js/build/src/make-client.js:105:19) at /workspace/build/src/v1beta1/speech_translation_service_client.js:190:29 at /workspace/node_modules/google-gax/build/src/streamingCalls/streamingApiCaller.js:46:28 at /workspace/node_modules/google-gax/build/src/normalCalls/timeout.js:44:16 at StreamProxy.setStream (/workspace/node_modules/google-gax/build/src/streamingCalls/streaming.js:144:24) at StreamingApiCaller.call (/workspace/node_modules/google-gax/build/src/streamingCalls/streamingApiCaller.js:54:16) at /workspace/node_modules/google-gax/build/src/createApiCall.js:84:30 at processTicksAndRejections (internal/process/task_queues.js:97:5) Emitted 'error' event on StreamProxy instance at: at errorOrDestroy (/workspace/node_modules/readable-stream/lib/internal/streams/destroy.js:98:101) at onwriteError (/workspace/node_modules/readable-stream/lib/_stream_writable.js:430:5) at onwrite (/workspace/node_modules/readable-stream/lib/_stream_writable.js:450:11) at WritableState.onwrite (/workspace/node_modules/readable-stream/lib/_stream_writable.js:160:5) at StreamProxy.Duplexify._destroy (/workspace/node_modules/duplexify/index.js:194:18) at /workspace/node_modules/duplexify/index.js:185:10 at processTicksAndRejections (internal/process/task_queues.js:79:11) { code: 16, details: 'Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.', metadata: Metadata { internalRepr: Map { 'grpc-server-stats-bin' => [ Buffer(10) [Uint8Array] [ 0, 0, 217, 248, 68, 0, 0, 0, 0, 0 ] ] }, options: {} } } at checkExecSyncError (child_process.js:635:11) at execSync (child_process.js:671:15) at exec (test/translate_from_file.test.js:24:21) at Context.<anonymous> (test/translate_from_file.test.js:28:20) at processImmediate (internal/timers.js:461:21)</pre></details>
1.0
MediaTranslation: should translate from a streamed file failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 8257ade7fd05bd65161d7a8c156e6c04fc09bfbe buildURL: [Build Status](https://source.cloud.google.com/results/invocations/2e29d795-c2f5-4ba6-a160-b1fb7a6e9692), [Sponge](http://sponge2/2e29d795-c2f5-4ba6-a160-b1fb7a6e9692) status: failed <details><summary>Test output</summary><br><pre>Command failed: node translate_from_file.js /workspace/samples/resources/audio.raw linear16 en-US es-ES events.js:291 throw er; // Unhandled 'error' event ^ Error: 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. at Object.callErrorFromStatus (/workspace/node_modules/@grpc/grpc-js/build/src/call.js:31:19) at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client.js:413:49) at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:328:181) at /workspace/node_modules/@grpc/grpc-js/build/src/call-stream.js:188:78 at processTicksAndRejections (internal/process/task_queues.js:79:11) for call at at ServiceClientImpl.makeBidiStreamRequest (/workspace/node_modules/@grpc/grpc-js/build/src/client.js:398:30) at ServiceClientImpl.<anonymous> (/workspace/node_modules/@grpc/grpc-js/build/src/make-client.js:105:19) at /workspace/build/src/v1beta1/speech_translation_service_client.js:190:29 at /workspace/node_modules/google-gax/build/src/streamingCalls/streamingApiCaller.js:46:28 at /workspace/node_modules/google-gax/build/src/normalCalls/timeout.js:44:16 at StreamProxy.setStream (/workspace/node_modules/google-gax/build/src/streamingCalls/streaming.js:144:24) at StreamingApiCaller.call (/workspace/node_modules/google-gax/build/src/streamingCalls/streamingApiCaller.js:54:16) at /workspace/node_modules/google-gax/build/src/createApiCall.js:84:30 at processTicksAndRejections (internal/process/task_queues.js:97:5) Emitted 'error' event on StreamProxy instance at: at errorOrDestroy (/workspace/node_modules/readable-stream/lib/internal/streams/destroy.js:98:101) at onwriteError (/workspace/node_modules/readable-stream/lib/_stream_writable.js:430:5) at onwrite (/workspace/node_modules/readable-stream/lib/_stream_writable.js:450:11) at WritableState.onwrite (/workspace/node_modules/readable-stream/lib/_stream_writable.js:160:5) at StreamProxy.Duplexify._destroy (/workspace/node_modules/duplexify/index.js:194:18) at /workspace/node_modules/duplexify/index.js:185:10 at processTicksAndRejections (internal/process/task_queues.js:79:11) { code: 16, details: 'Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.', metadata: Metadata { internalRepr: Map { 'grpc-server-stats-bin' => [ Buffer(10) [Uint8Array] [ 0, 0, 217, 248, 68, 0, 0, 0, 0, 0 ] ] }, options: {} } } Error: Command failed: node translate_from_file.js /workspace/samples/resources/audio.raw linear16 en-US es-ES events.js:291 throw er; // Unhandled 'error' event ^ Error: 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. at Object.callErrorFromStatus (/workspace/node_modules/@grpc/grpc-js/build/src/call.js:31:19) at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client.js:413:49) at Object.onReceiveStatus (/workspace/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:328:181) at /workspace/node_modules/@grpc/grpc-js/build/src/call-stream.js:188:78 at processTicksAndRejections (internal/process/task_queues.js:79:11) for call at at ServiceClientImpl.makeBidiStreamRequest (/workspace/node_modules/@grpc/grpc-js/build/src/client.js:398:30) at ServiceClientImpl.<anonymous> (/workspace/node_modules/@grpc/grpc-js/build/src/make-client.js:105:19) at /workspace/build/src/v1beta1/speech_translation_service_client.js:190:29 at /workspace/node_modules/google-gax/build/src/streamingCalls/streamingApiCaller.js:46:28 at /workspace/node_modules/google-gax/build/src/normalCalls/timeout.js:44:16 at StreamProxy.setStream (/workspace/node_modules/google-gax/build/src/streamingCalls/streaming.js:144:24) at StreamingApiCaller.call (/workspace/node_modules/google-gax/build/src/streamingCalls/streamingApiCaller.js:54:16) at /workspace/node_modules/google-gax/build/src/createApiCall.js:84:30 at processTicksAndRejections (internal/process/task_queues.js:97:5) Emitted 'error' event on StreamProxy instance at: at errorOrDestroy (/workspace/node_modules/readable-stream/lib/internal/streams/destroy.js:98:101) at onwriteError (/workspace/node_modules/readable-stream/lib/_stream_writable.js:430:5) at onwrite (/workspace/node_modules/readable-stream/lib/_stream_writable.js:450:11) at WritableState.onwrite (/workspace/node_modules/readable-stream/lib/_stream_writable.js:160:5) at StreamProxy.Duplexify._destroy (/workspace/node_modules/duplexify/index.js:194:18) at /workspace/node_modules/duplexify/index.js:185:10 at processTicksAndRejections (internal/process/task_queues.js:79:11) { code: 16, details: 'Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.', metadata: Metadata { internalRepr: Map { 'grpc-server-stats-bin' => [ Buffer(10) [Uint8Array] [ 0, 0, 217, 248, 68, 0, 0, 0, 0, 0 ] ] }, options: {} } } at checkExecSyncError (child_process.js:635:11) at execSync (child_process.js:671:15) at exec (test/translate_from_file.test.js:24:21) at Context.<anonymous> (test/translate_from_file.test.js:28:20) at processImmediate (internal/timers.js:461:21)</pre></details>
non_test
mediatranslation should translate from a streamed file failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output command failed node translate from file js workspace samples resources audio raw en us es es events js throw er unhandled error event error unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see at object callerrorfromstatus workspace node modules grpc grpc js build src call js at object onreceivestatus workspace node modules grpc grpc js build src client js at object onreceivestatus workspace node modules grpc grpc js build src client interceptors js at workspace node modules grpc grpc js build src call stream js at processticksandrejections internal process task queues js for call at at serviceclientimpl makebidistreamrequest workspace node modules grpc grpc js build src client js at serviceclientimpl workspace node modules grpc grpc js build src make client js at workspace build src speech translation service client js at workspace node modules google gax build src streamingcalls streamingapicaller js at workspace node modules google gax build src normalcalls timeout js at streamproxy setstream workspace node modules google gax build src streamingcalls streaming js at streamingapicaller call workspace node modules google gax build src streamingcalls streamingapicaller js at workspace node modules google gax build src createapicall js at processticksandrejections internal process task queues js emitted error event on streamproxy instance at at errorordestroy workspace node modules readable stream lib internal streams destroy js at onwriteerror workspace node modules readable stream lib stream writable js at onwrite workspace node modules readable stream lib stream writable js at writablestate onwrite workspace node modules readable stream lib stream writable js at streamproxy duplexify destroy workspace node modules duplexify index js at workspace node modules duplexify index js at processticksandrejections internal process task queues js code details request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see metadata metadata internalrepr map grpc server stats bin buffer options error command failed node translate from file js workspace samples resources audio raw en us es es events js throw er unhandled error event error unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see at object callerrorfromstatus workspace node modules grpc grpc js build src call js at object onreceivestatus workspace node modules grpc grpc js build src client js at object onreceivestatus workspace node modules grpc grpc js build src client interceptors js at workspace node modules grpc grpc js build src call stream js at processticksandrejections internal process task queues js for call at at serviceclientimpl makebidistreamrequest workspace node modules grpc grpc js build src client js at serviceclientimpl workspace node modules grpc grpc js build src make client js at workspace build src speech translation service client js at workspace node modules google gax build src streamingcalls streamingapicaller js at workspace node modules google gax build src normalcalls timeout js at streamproxy setstream workspace node modules google gax build src streamingcalls streaming js at streamingapicaller call workspace node modules google gax build src streamingcalls streamingapicaller js at workspace node modules google gax build src createapicall js at processticksandrejections internal process task queues js emitted error event on streamproxy instance at at errorordestroy workspace node modules readable stream lib internal streams destroy js at onwriteerror workspace node modules readable stream lib stream writable js at onwrite workspace node modules readable stream lib stream writable js at writablestate onwrite workspace node modules readable stream lib stream writable js at streamproxy duplexify destroy workspace node modules duplexify index js at workspace node modules duplexify index js at processticksandrejections internal process task queues js code details request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see metadata metadata internalrepr map grpc server stats bin buffer options at checkexecsyncerror child process js at execsync child process js at exec test translate from file test js at context test translate from file test js at processimmediate internal timers js
0
157,357
5,997,422,972
IssuesEvent
2017-06-04 00:06:10
mreishman/Log-Hog
https://api.github.com/repos/mreishman/Log-Hog
opened
Setting to enable / disable right click menu
enhancement Priority - 4 - Low
- [ ] Setting to enable / disable right click menu
1.0
Setting to enable / disable right click menu - - [ ] Setting to enable / disable right click menu
non_test
setting to enable disable right click menu setting to enable disable right click menu
0
263,650
23,072,686,943
IssuesEvent
2022-07-25 19:42:52
strangelove-ventures/sommelier
https://api.github.com/repos/strangelove-ventures/sommelier
closed
Non-critical UX bugs
testing
Let's collect here all non-critical UX bugs 1. by @AdrianDobrican Staging: ![2022-06-30 17 12 36](https://user-images.githubusercontent.com/26877917/176700192-d9062753-cb11-4c66-9b68-c489d98d93fb.jpg) Figma: ![2022-06-30 17 12 29](https://user-images.githubusercontent.com/26877917/176700258-300d768d-5732-4880-9529-74cb5ca8fe62.jpg) 2. by @AdrianDobrican Staging: ![2022-06-30 17 16 05](https://user-images.githubusercontent.com/26877917/176700469-0c7902a8-55c3-4284-a45b-3c017f89b967.jpg) Figma: ![2022-06-30 17 16 09](https://user-images.githubusercontent.com/26877917/176700438-1a7d1d89-401b-4956-ab57-fe758941da96.jpg) 3. by @AdrianDobrican Staging: ![2022-06-30 17 17 14](https://user-images.githubusercontent.com/26877917/176700751-e3c1059b-e4c7-4317-9d7a-7282d0a82ad9.jpg) Figma: ![2022-06-30 17 17 20](https://user-images.githubusercontent.com/26877917/176700816-eb4dd599-fc8f-4828-9576-c825243b8700.jpg) 4. Staging: ![Image](https://user-images.githubusercontent.com/26877917/178059608-b9af0b08-bf12-459e-b1da-0ee382e9fe52.jpeg) Figma: ![Image](https://user-images.githubusercontent.com/26877917/178059643-d23956c1-5178-4c4b-86f6-6fb5e4c18c3c.jpeg)
1.0
Non-critical UX bugs - Let's collect here all non-critical UX bugs 1. by @AdrianDobrican Staging: ![2022-06-30 17 12 36](https://user-images.githubusercontent.com/26877917/176700192-d9062753-cb11-4c66-9b68-c489d98d93fb.jpg) Figma: ![2022-06-30 17 12 29](https://user-images.githubusercontent.com/26877917/176700258-300d768d-5732-4880-9529-74cb5ca8fe62.jpg) 2. by @AdrianDobrican Staging: ![2022-06-30 17 16 05](https://user-images.githubusercontent.com/26877917/176700469-0c7902a8-55c3-4284-a45b-3c017f89b967.jpg) Figma: ![2022-06-30 17 16 09](https://user-images.githubusercontent.com/26877917/176700438-1a7d1d89-401b-4956-ab57-fe758941da96.jpg) 3. by @AdrianDobrican Staging: ![2022-06-30 17 17 14](https://user-images.githubusercontent.com/26877917/176700751-e3c1059b-e4c7-4317-9d7a-7282d0a82ad9.jpg) Figma: ![2022-06-30 17 17 20](https://user-images.githubusercontent.com/26877917/176700816-eb4dd599-fc8f-4828-9576-c825243b8700.jpg) 4. Staging: ![Image](https://user-images.githubusercontent.com/26877917/178059608-b9af0b08-bf12-459e-b1da-0ee382e9fe52.jpeg) Figma: ![Image](https://user-images.githubusercontent.com/26877917/178059643-d23956c1-5178-4c4b-86f6-6fb5e4c18c3c.jpeg)
test
non critical ux bugs let s collect here all non critical ux bugs by adriandobrican staging figma by adriandobrican staging figma by adriandobrican staging figma staging figma
1
288,143
24,882,783,278
IssuesEvent
2022-10-28 03:48:33
MPMG-DCC-UFMG/F01
https://api.github.com/repos/MPMG-DCC-UFMG/F01
closed
Teste de generalizacao para a tag Orçamento - Execução - Riachinho
generalization test development template - Memory (66) tag - Orçamento subtag - Execução
DoD: Realizar o teste de Generalização do validador da tag Orçamento - Execução para o Município de Riachinho.
1.0
Teste de generalizacao para a tag Orçamento - Execução - Riachinho - DoD: Realizar o teste de Generalização do validador da tag Orçamento - Execução para o Município de Riachinho.
test
teste de generalizacao para a tag orçamento execução riachinho dod realizar o teste de generalização do validador da tag orçamento execução para o município de riachinho
1
37,239
4,789,321,666
IssuesEvent
2016-10-31 00:20:31
arknano/Icarus
https://api.github.com/repos/arknano/Icarus
opened
Diagetic UI Design
Art Design Feature
This includes showing the score/time to the player at the end and potentially some sort of mid-level display.
1.0
Diagetic UI Design - This includes showing the score/time to the player at the end and potentially some sort of mid-level display.
non_test
diagetic ui design this includes showing the score time to the player at the end and potentially some sort of mid level display
0
269,500
23,446,612,397
IssuesEvent
2022-08-15 20:21:01
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
DISABLED test_aot_autograd_exhaustive_min_binary_cpu_float32 (__main__.TestEagerFusionOpInfoCPU)
module: flaky-tests skipped module: functorch
Platforms: mac, macos This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_aot_autograd_exhaustive_min_binary_cpu_float32&suite=TestEagerFusionOpInfoCPU&file=/Users/runner/work/pytorch/pytorch/functorch/test/test_pythonkey.py) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/7840162739). Over the past 3 hours, it has been determined flaky in 3 workflow(s) with 3 failures and 3 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT BE ALARMED THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_aot_autograd_exhaustive_min_binary_cpu_float32` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. cc @zou3519 @Chillee @samdow
1.0
DISABLED test_aot_autograd_exhaustive_min_binary_cpu_float32 (__main__.TestEagerFusionOpInfoCPU) - Platforms: mac, macos This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_aot_autograd_exhaustive_min_binary_cpu_float32&suite=TestEagerFusionOpInfoCPU&file=/Users/runner/work/pytorch/pytorch/functorch/test/test_pythonkey.py) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/7840162739). Over the past 3 hours, it has been determined flaky in 3 workflow(s) with 3 failures and 3 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT BE ALARMED THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_aot_autograd_exhaustive_min_binary_cpu_float32` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. cc @zou3519 @Chillee @samdow
test
disabled test aot autograd exhaustive min binary cpu main testeagerfusionopinfocpu platforms mac macos this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not be alarmed the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test aot autograd exhaustive min binary cpu there should be several instances run as flaky tests are rerun in ci from which you can study the logs cc chillee samdow
1
94,458
8,492,033,958
IssuesEvent
2018-10-27 18:44:44
MSCMP/MSCMP
https://api.github.com/repos/MSCMP/MSCMP
closed
Crash when other players enter vehicles
retest-needed
The game crashes immediately when another player enters a vehicle. Here's a screenshot of the error: https://i.imgur.com/kEaPcF7.png
1.0
Crash when other players enter vehicles - The game crashes immediately when another player enters a vehicle. Here's a screenshot of the error: https://i.imgur.com/kEaPcF7.png
test
crash when other players enter vehicles the game crashes immediately when another player enters a vehicle here s a screenshot of the error
1
247,466
20,981,203,656
IssuesEvent
2022-03-28 20:08:15
ECP-WarpX/WarpX
https://api.github.com/repos/ECP-WarpX/WarpX
opened
maxLevel -> finestLevel
bug bug: affects latest release component: core
There are a few locations in WarpX where `maxLevel` is used but maybe `finestLevel` would be the better choice, similar to #2970. Some of it is init and resampling code that I am not too familiar with. This needs to be checked (not much, like <40 lines in the code or so).
1.0
maxLevel -> finestLevel - There are a few locations in WarpX where `maxLevel` is used but maybe `finestLevel` would be the better choice, similar to #2970. Some of it is init and resampling code that I am not too familiar with. This needs to be checked (not much, like <40 lines in the code or so).
test
maxlevel finestlevel there are a few locations in warpx where maxlevel is used but maybe finestlevel would be the better choice similar to some of it is init and resampling code that i am not too familiar with this needs to be checked not much like lines in the code or so
1
325,671
24,056,460,412
IssuesEvent
2022-09-16 17:25:08
snowmang1/nvim
https://api.github.com/repos/snowmang1/nvim
closed
fix Readme
documentation
deprecated hop shortcut n is not for nvim-tree new plugins have been added all plugins need links
1.0
fix Readme - deprecated hop shortcut n is not for nvim-tree new plugins have been added all plugins need links
non_test
fix readme deprecated hop shortcut n is not for nvim tree new plugins have been added all plugins need links
0
320,103
27,420,145,173
IssuesEvent
2023-03-01 16:11:20
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix modules.test_module_training
Sub Task Ivy Stateful API Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4301928602/jobs/7499775516" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4305405432/jobs/7507808142" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4301928602/jobs/7499798022" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> <details> <summary>Not found</summary> Not found </details>
1.0
Fix modules.test_module_training - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4301928602/jobs/7499775516" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4305405432/jobs/7507808142" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4301928602/jobs/7499798022" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> <details> <summary>Not found</summary> Not found </details>
test
fix modules test module training tensorflow img src torch img src numpy img src jax img src not found not found
1
64,608
6,912,664,993
IssuesEvent
2017-11-28 12:51:03
curationexperts/epigaea
https://api.github.com/repos/curationexperts/epigaea
closed
Inability to save edits on object updated by template
acceptance testing bug ready
If you update an object via template, and the template completes, and then you open the object and attempt to edit it, you cannot save the object. We have gotten two different error message in testing, appended.
1.0
Inability to save edits on object updated by template - If you update an object via template, and the template completes, and then you open the object and attempt to edit it, you cannot save the object. We have gotten two different error message in testing, appended.
test
inability to save edits on object updated by template if you update an object via template and the template completes and then you open the object and attempt to edit it you cannot save the object we have gotten two different error message in testing appended
1
163,123
13,911,562,908
IssuesEvent
2020-10-20 17:34:31
vicdoja/MetaCountdown
https://api.github.com/repos/vicdoja/MetaCountdown
closed
Write a useful README
documentation
The README file is one of the most important files in a project. This is like this because the README file is the first content that users or other developers see when they first enter the project page. For our early development stage, it should include the following sections: - [ ] **Description**: define the problem, it complexity and the used emetaheuristics. - [ ] **Origin**: this is a Master's subject project. - [ ] **Contributing** - [ ] **References**: include links to revelant materials to the project - [ ] **License**
1.0
Write a useful README - The README file is one of the most important files in a project. This is like this because the README file is the first content that users or other developers see when they first enter the project page. For our early development stage, it should include the following sections: - [ ] **Description**: define the problem, it complexity and the used emetaheuristics. - [ ] **Origin**: this is a Master's subject project. - [ ] **Contributing** - [ ] **References**: include links to revelant materials to the project - [ ] **License**
non_test
write a useful readme the readme file is one of the most important files in a project this is like this because the readme file is the first content that users or other developers see when they first enter the project page for our early development stage it should include the following sections description define the problem it complexity and the used emetaheuristics origin this is a master s subject project contributing references include links to revelant materials to the project license
0
295,902
25,514,105,355
IssuesEvent
2022-11-28 15:08:41
ntop/ntopng
https://api.github.com/repos/ntop/ntopng
closed
Broken Exclusions on Domains
Bug Ready to Test
**Environment**: * ntopng version/revision: v.5.4.220915] **What happened**: Exclusions about domains, implemented both via UI (starting directly from alert exclusion popup) and via ndpi protos file on ntopng and via ndpi file on nprobe have no effect. Here's an example: We are flooded by "Binary Application Transfer" alerts on windows updates downloads ![ev_01](https://user-images.githubusercontent.com/108343272/193069700-6a23a60c-4cae-44ce-8c6e-bc3f5264bb9c.png) So we tried to exclude it via UI ![ev_02](https://user-images.githubusercontent.com/108343272/193070789-d7a74cf4-255b-4a50-b50d-03e7179cd838.png) And in fact it appears as excluded: ![ev_03](https://user-images.githubusercontent.com/108343272/193071438-4521a07f-b714-44b1-b5ee-965fdfb03699.PNG) Anyway it had no effect, alert kept on generating. So as stated [here](https://www.ntop.org/ndpi/howto-define-ndpi-risk-exceptions-for-networks-and-domains/) we created a new .protos file on ntopng instance, specifying the string: host_risk_mask:".windowsupdate.com"=0xFFFFFFFFFFFFFFEF and pointed at it via conf file (-p=/etc/ntopng/myfile.protos). It had no effect, even restarting ntopng. Finally we put the same string on nprobe instance (specifying --ndpi-custom-protos=/etc/nprobe/myfile.protos - anyway on the guide asks for PATH, not file, I believe it's a typo). Alerts keep popping up, even after restarting nprobe. If a pcap is needed to solve the issue, please let me know how to execute tcpdump to achieve this. This issue is meant to be an extension of a previous one [#6886 ](https://github.com/ntop/ntopng/issues/6886), please feel free to close it. Thank you
1.0
Broken Exclusions on Domains - **Environment**: * ntopng version/revision: v.5.4.220915] **What happened**: Exclusions about domains, implemented both via UI (starting directly from alert exclusion popup) and via ndpi protos file on ntopng and via ndpi file on nprobe have no effect. Here's an example: We are flooded by "Binary Application Transfer" alerts on windows updates downloads ![ev_01](https://user-images.githubusercontent.com/108343272/193069700-6a23a60c-4cae-44ce-8c6e-bc3f5264bb9c.png) So we tried to exclude it via UI ![ev_02](https://user-images.githubusercontent.com/108343272/193070789-d7a74cf4-255b-4a50-b50d-03e7179cd838.png) And in fact it appears as excluded: ![ev_03](https://user-images.githubusercontent.com/108343272/193071438-4521a07f-b714-44b1-b5ee-965fdfb03699.PNG) Anyway it had no effect, alert kept on generating. So as stated [here](https://www.ntop.org/ndpi/howto-define-ndpi-risk-exceptions-for-networks-and-domains/) we created a new .protos file on ntopng instance, specifying the string: host_risk_mask:".windowsupdate.com"=0xFFFFFFFFFFFFFFEF and pointed at it via conf file (-p=/etc/ntopng/myfile.protos). It had no effect, even restarting ntopng. Finally we put the same string on nprobe instance (specifying --ndpi-custom-protos=/etc/nprobe/myfile.protos - anyway on the guide asks for PATH, not file, I believe it's a typo). Alerts keep popping up, even after restarting nprobe. If a pcap is needed to solve the issue, please let me know how to execute tcpdump to achieve this. This issue is meant to be an extension of a previous one [#6886 ](https://github.com/ntop/ntopng/issues/6886), please feel free to close it. Thank you
test
broken exclusions on domains environment ntopng version revision v what happened exclusions about domains implemented both via ui starting directly from alert exclusion popup and via ndpi protos file on ntopng and via ndpi file on nprobe have no effect here s an example we are flooded by binary application transfer alerts on windows updates downloads so we tried to exclude it via ui and in fact it appears as excluded anyway it had no effect alert kept on generating so as stated we created a new protos file on ntopng instance specifying the string host risk mask windowsupdate com and pointed at it via conf file p etc ntopng myfile protos it had no effect even restarting ntopng finally we put the same string on nprobe instance specifying ndpi custom protos etc nprobe myfile protos anyway on the guide asks for path not file i believe it s a typo alerts keep popping up even after restarting nprobe if a pcap is needed to solve the issue please let me know how to execute tcpdump to achieve this this issue is meant to be an extension of a previous one please feel free to close it thank you
1
104,964
9,013,271,754
IssuesEvent
2019-02-05 19:03:22
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
teamcity: failed test: TestImportCSVStmt
C-test-failure O-robot
The following tests appear to have failed on master (testrace): TestImportCSVStmt/schema-in-file-implicit-gzip, TestImportCSVStmt/schema-in-file-no-decompress, TestImportCSVStmt/empty-file, TestImportCSVStmt/schema-in-query-opts, TestImportCSVStmt/schema-in-file-sstsize, TestImportCSVStmt/schema-in-file-auto-decompress, TestImportCSVStmt/schema-in-file-auto-gzip, TestImportCSVStmt, TestImportCSVStmt/schema-in-query-transform-only, TestImportCSVStmt/schema-in-file-explicit-gzip, TestImportCSVStmt/empty-with-files You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+TestImportCSVStmt). [#1125945](https://teamcity.cockroachdb.com/viewLog.html?buildId=1125945): ``` TestImportCSVStmt/schema-in-file-implicit-gzip ...002/"Y…-834/"Y"}] initiating a split of this range at key /Table/76/3/4224/"M" [r402] (manual) I190205 18:04:39.087848 9224 storage/store.go:2669 [n2,s2,r366/2:/Table/74/2/"{N"/794-Q"/3241}] removing replica r368/2 I190205 18:04:39.094016 9541 storage/store.go:2669 [n3,s3,r366/3:/Table/74/2/"{N"/794-Q"/3241}] removing replica r368/3 I190205 18:04:39.094483 8905 storage/store.go:2669 [n1,s1,r366/1:/Table/74/2/"{N"/794-Q"/3241}] removing replica r368/1 I190205 18:04:39.172897 9214 storage/store.go:2669 [n2,s2,r341/2:/Table/74/1/{3266-4842}] removing replica r289/2 I190205 18:04:39.174095 8938 storage/store.go:2669 [n1,s1,r341/1:/Table/74/1/{3266-4842}] removing replica r289/1 I190205 18:04:39.189718 9486 storage/store.go:2669 [n3,s3,r341/3:/Table/74/1/{3266-4842}] removing replica r289/3 I190205 18:04:39.195738 42996 storage/replica_command.go:383 [n2,merge,s2,r289/2:/Table/74/{1/4842-2/"C"/3…}] initiating a merge of r361:/Table/74/2/"{C"/3539-G"/917} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=18 KiB+18 KiB qps=2.69+0.00 --> 2.69qps) below threshold (size=37 KiB, qps=2.69)) I190205 18:04:39.674788 42929 storage/replica_command.go:383 [n3,merge,s3,r343/3:/Table/74/1/{741-2632}] initiating a merge of r346:/Table/74/1/{2632-3266} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=0] into this range (lhs+rhs has (size=50 KiB+17 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=67 KiB, qps=0.00)) I190205 18:04:40.045754 42443 sql/event_log.go:135 [n1] Event: "create_statistics", target: 74, info: {StatisticName:__auto__ Statement:CREATE STATISTICS __auto__ FROM [74] AS OF SYSTEM TIME '-30s'} I190205 18:04:40.174598 9531 storage/store.go:2669 [n3,s3,r343/3:/Table/74/1/{741-2632}] removing replica r346/3 I190205 18:04:40.175909 8910 storage/store.go:2669 [n1,s1,r343/1:/Table/74/1/{741-2632}] removing replica r346/1 I190205 18:04:40.187748 9195 storage/store.go:2669 [n2,s2,r343/2:/Table/74/1/{741-2632}] removing replica r346/2 I190205 18:04:40.247398 43054 storage/replica_command.go:383 [n2,merge,s2,r366/2:/Table/74/2/"{N"/794-U"/723}] initiating a merge of r369:/Table/74/2/"{U"/723-X"/3170} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=37 KiB+18 KiB qps=0.97+0.00 --> 0.97qps) below threshold (size=55 KiB, qps=0.97)) I190205 18:04:40.343699 43081 storage/replica_command.go:383 [n3,merge,s3,r341/3:/Table/74/{1/3266-2/"C"/3…}] initiating a merge of r361:/Table/74/2/"{C"/3539-G"/917} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=60 KiB+18 KiB qps=5.63+1.70 --> 7.33qps) below threshold (size=78 KiB, qps=7.33)) I190205 18:04:40.570936 9199 storage/store.go:2669 [n2,s2,r366/2:/Table/74/2/"{N"/794-U"/723}] removing replica r369/2 I190205 18:04:40.576583 9531 storage/store.go:2669 [n3,s3,r366/3:/Table/74/2/"{N"/794-U"/723}] removing replica r369/3 I190205 18:04:40.598783 8939 storage/store.go:2669 [n1,s1,r366/1:/Table/74/2/"{N"/794-U"/723}] removing replica r369/1 I190205 18:04:40.827940 9481 storage/store.go:2669 [n3,s3,r341/3:/Table/74/{1/3266-2/"C"/3…}] removing replica r361/3 I190205 18:04:40.830389 9209 storage/store.go:2669 [n2,s2,r341/2:/Table/74/{1/3266-2/"C"/3…}] removing replica r361/2 I190205 18:04:40.834986 8918 storage/store.go:2669 [n1,s1,r341/1:/Table/74/{1/3266-2/"C"/3…}] removing replica r361/1 import_stmt_test.go:1180: job 12 did not match: Description: "IMPORT TABLE csv12.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:////csv/data-0.gz', 'nodelocal:////csv/data-1.gz', 'nodelocal:////csv/data-2.gz', 'nodelocal:////csv/data-3.gz', 'nodelocal:////csv/data-4.gz')" != "IMPORT TABLE csv8.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///csv/data-0', 'nodelocal:///csv/data-1', 'nodelocal:///csv/data-2', 'nodelocal:///csv/data-3', 'nodelocal:///csv/data-4') WITH decompress = 'auto'" TestImportCSVStmt/schema-in-file-auto-decompress ...2:34.891226 29904 storage/replica_command.go:244 [n2,s2,r273/2:/Table/68/3/{1187/"…-2549/"…}] initiating a split of this range at key /Table/68/3/1854/"I"/PrefixEnd [r274] (manual) I190205 18:02:35.120817 30188 storage/replica_command.go:383 [n3,merge,s3,r208/3:/Table/66/2/"{B"/4968-F"/2346}] initiating a merge of r222:/Table/66/2/"{F"/2346-H"/1541} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=18 KiB+9.7 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=28 KiB, qps=0.00)) I190205 18:02:35.328598 30142 storage/replica_command.go:383 [n1,merge,s1,r208/1:/Table/66/2/"{B"/4968-F"/2346}] initiating a merge of r222:/Table/66/2/"{F"/2346-H"/1541} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=18 KiB+9.7 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=28 KiB, qps=0.00)) I190205 18:02:35.661168 9255 server/status/runtime.go:464 [n2] runtime stats: 1.6 GiB RSS, 676 goroutines, 95 MiB/66 MiB/204 MiB GO alloc/idle/total, 190 MiB/234 MiB CGO alloc/total, 4992.3 CGO/sec, 142.7/12.2 %(u/s)time, 0.6 %gc (5x), 3.3 MiB/3.3 MiB (r/w)net I190205 18:02:35.902427 30207 storage/replica_command.go:244 [n1,s1,r274/1:/Table/68/3/{1854/"…-2549/"…}] initiating a split of this range at key /Table/68/3/1883/"L" [r142] (manual) I190205 18:02:36.562791 8950 storage/store.go:2669 [n1,s1,r208/1:/Table/66/2/"{B"/4968-F"/2346}] removing replica r222/1 I190205 18:02:36.568458 9535 storage/store.go:2669 [n3,s3,r208/3:/Table/66/2/"{B"/4968-F"/2346}] removing replica r222/3 I190205 18:02:36.638714 9579 server/status/runtime.go:464 [n3] runtime stats: 1.6 GiB RSS, 670 goroutines, 142 MiB/26 MiB/204 MiB GO alloc/idle/total, 190 MiB/234 MiB CGO alloc/total, 4787.4 CGO/sec, 140.5/12.0 %(u/s)time, 0.6 %gc (5x), 3.3 MiB/3.3 MiB (r/w)net I190205 18:02:36.639448 9219 storage/store.go:2669 [n2,s2,r208/2:/Table/66/2/"{B"/4968-F"/2346}] removing replica r222/2 I190205 18:02:36.837577 30349 storage/replica_command.go:383 [n1,merge,s1,r208/1:/Table/66/2/"{B"/4968-H"/1541}] initiating a merge of r226:/Table/66/2/"{H"/1541-K"/3937} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=2] into this range (lhs+rhs has (size=28 KiB+18 KiB qps=19.86+0.00 --> 19.86qps) below threshold (size=46 KiB, qps=19.86)) I190205 18:02:37.509519 30376 storage/replica_command.go:383 [n3,merge,s3,r226/3:/Table/66/2/"{H"/1541-K"/3937}] initiating a merge of r233:/Table/66/2/"{K"/3937-O"/1419} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=18 KiB+18 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=37 KiB, qps=0.00)) I190205 18:02:37.602138 8908 storage/store.go:2669 [n1,s1,r208/1:/Table/66/2/"{B"/4968-H"/1541}] removing replica r226/1 I190205 18:02:37.603497 9201 storage/store.go:2669 [n2,s2,r208/2:/Table/66/2/"{B"/4968-H"/1541}] removing replica r226/2 I190205 18:02:37.629172 9496 storage/store.go:2669 [n3,s3,r208/3:/Table/66/2/"{B"/4968-H"/1541}] removing replica r226/3 I190205 18:02:37.840870 30462 storage/replica_command.go:383 [n1,merge,s1,r249/1:/Table/6{6/3/4431…-7}] initiating a merge of r248:/Table/6{7-8/1/741} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=3] into this range (lhs+rhs has (size=16 KiB+19 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=35 KiB, qps=0.00)) import_stmt_test.go:1180: job 8 did not match: Description: "IMPORT TABLE csv8.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///csv/data-0', 'nodelocal:///csv/data-1', 'nodelocal:///csv/data-2', 'nodelocal:///csv/data-3', 'nodelocal:///csv/data-4') WITH decompress = 'auto'" != "IMPORT TABLE \"\".\"\".t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///csv/data-0-opts', 'nodelocal:///csv/data-1-opts', 'nodelocal:///csv/data-2-opts', 'nodelocal:///csv/data-3-opts', 'nodelocal:///csv/data-4-opts') WITH comment = '#', delimiter = '|', \"nullif\" = '', skip = '2', transform = 'nodelocal:///5'" TestImportCSVStmt/schema-in-query-opts ...n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=135 KiB+19 KiB qps=2.98+2.69 --> 5.67qps) below threshold (size=154 KiB, qps=5.67)) I190205 18:00:52.078010 19718 storage/replica_command.go:244 [n3,s3,r171/3:/{Table/59/3/2…-Max}] initiating a split of this range at key /Table/59/3/3329/"B"/PrefixEnd [r172] (manual) I190205 18:00:52.278113 9537 storage/store.go:2669 [n3,s3,r117/3:/Table/57/{1/4612-2/"X"/3…}] removing replica r126/3 I190205 18:00:52.279167 8719 storage/store.go:2669 [n1,s1,r117/1:/Table/57/{1/4612-2/"X"/3…}] removing replica r126/1 I190205 18:00:52.279897 9204 storage/store.go:2669 [n2,s2,r117/2:/Table/57/{1/4612-2/"X"/3…}] removing replica r126/2 I190205 18:00:52.345961 19791 storage/replica_command.go:383 [n3,merge,s3,r56/3:/Table/57/1/{741-1456}] initiating a merge of r59:/Table/57/1/1{456-752} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=19 KiB+7.8 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=27 KiB, qps=0.00)) I190205 18:00:52.807849 19809 storage/replica_command.go:244 [n3,s3,r172/3:/{Table/59/3/3…-Max}] initiating a split of this range at key /Table/59/3/4031/"B"/PrefixEnd [r173] (manual) I190205 18:00:52.825113 8953 storage/store.go:2669 [n1,s1,r56/1:/Table/57/1/{741-1456}] removing replica r59/1 I190205 18:00:52.839472 9225 storage/store.go:2669 [n2,s2,r56/2:/Table/57/1/{741-1456}] removing replica r59/2 I190205 18:00:52.871815 9543 storage/store.go:2669 [n3,s3,r56/3:/Table/57/1/{741-1456}] removing replica r59/3 I190205 18:00:53.204322 8835 server/status/runtime.go:464 [n1] runtime stats: 1.3 GiB RSS, 671 goroutines, 76 MiB/85 MiB/204 MiB GO alloc/idle/total, 140 MiB/184 MiB CGO alloc/total, 4762.7 CGO/sec, 146.4/14.1 %(u/s)time, 1.1 %gc (6x), 3.1 MiB/3.1 MiB (r/w)net I190205 18:00:53.327528 19863 storage/replica_command.go:383 [n3,merge,s3,r56/3:/Table/57/1/{741-1752}] initiating a merge of r72:/Table/57/1/{1752-2467} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=27 KiB+19 KiB qps=4.58+0.00 --> 4.58qps) below threshold (size=46 KiB, qps=4.58)) I190205 18:00:53.449266 19893 storage/replica_command.go:244 [n3,s3,r173/3:/{Table/59/3/4…-Max}] initiating a split of this range at key /Table/59/3/4733/"B"/PrefixEnd [r174] (manual); delayed split for 0.2s to avoid Raft snapshot I190205 18:00:53.530908 9488 storage/store.go:2669 [n3,s3,r56/3:/Table/57/1/{741-1752}] removing replica r72/3 I190205 18:00:53.535078 9232 storage/store.go:2669 [n2,s2,r56/2:/Table/57/1/{741-1752}] removing replica r72/2 I190205 18:00:53.537881 8949 storage/store.go:2669 [n1,s1,r56/1:/Table/57/1/{741-1752}] removing replica r72/1 I190205 18:00:53.760796 19825 storage/replica_command.go:244 [n3,s3,r174/3:/{Table/59/3/4…-Max}] initiating a split of this range at key /Table/60 [r175] (manual) I190205 18:00:54.389805 19946 storage/replica_command.go:383 [n3,merge,s3,r56/3:/Table/57/1/{741-2467}] initiating a merge of r75:/Table/57/1/{2467-3182} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=2] into this range (lhs+rhs has (size=46 KiB+19 KiB qps=5.66+0.00 --> 5.66qps) below threshold (size=64 KiB, qps=5.66)) I190205 18:00:54.812161 9543 storage/store.go:2669 [n3,s3,r56/3:/Table/57/1/{741-2467}] removing replica r75/3 I190205 18:00:54.817986 9228 storage/store.go:2669 [n2,s2,r56/2:/Table/57/1/{741-2467}] removing replica r75/2 I190205 18:00:54.849856 8953 storage/store.go:2669 [n1,s1,r56/1:/Table/57/1/{741-2467}] removing replica r75/1 import_stmt_test.go:1180: job 3 did not match: Description: "IMPORT TABLE csv3.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///csv/data-0-opts', 'nodelocal:///csv/data-1-opts', 'nodelocal:///csv/data-2-opts', 'nodelocal:///csv/data-3-opts', 'nodelocal:///csv/data-4-opts') WITH comment = '#', delimiter = '|', \"nullif\" = '', skip = '2'" != "CREATE STATISTICS __auto__ FROM [53] AS OF SYSTEM TIME '-30s'" TestImportCSVStmt/schema-in-file-explicit-gzip ...iB CGO alloc/total, 8636.8 CGO/sec, 143.8/12.7 %(u/s)time, 0.9 %gc (5x), 3.7 MiB/3.7 MiB (r/w)net I190205 18:03:36.967954 36306 storage/replica_command.go:244 [n2,s2,r339/2:/{Table/72/3/2…-Max}] initiating a split of this range at key /Table/72/3/3614/"A"/PrefixEnd [r275] (manual) I190205 18:03:37.568594 9573 storage/compactor/compactor.go:329 [n3,s3,compactor] purging suggested compaction for range /Table/68/2/"H"/3232 - /Table/68/2/"L"/688 that contains live data I190205 18:03:37.569024 9573 storage/compactor/compactor.go:329 [n3,s3,compactor] purging suggested compaction for range /Table/68/2/"L"/688 - /Table/68/2/"S"/617 that contains live data I190205 18:03:37.569337 9573 storage/compactor/compactor.go:329 [n3,s3,compactor] purging suggested compaction for range /Table/68/2/"S"/617 - /Table/68/2/"Z"/546 that contains live data I190205 18:03:37.569589 9573 storage/compactor/compactor.go:329 [n3,s3,compactor] purging suggested compaction for range /Table/68/2/"V"/3064 - /Table/68/2/"Z"/546 that contains live data I190205 18:03:37.585786 9573 storage/compactor/compactor.go:329 [n3,s3,compactor] purging suggested compaction for range /Table/68/3/520/"A"/PrefixEnd - /Table/68/3/1187/"R"/PrefixEnd that contains live data I190205 18:03:37.586234 9573 storage/compactor/compactor.go:329 [n3,s3,compactor] purging suggested compaction for range /Table/68/3/1187/"R"/PrefixEnd - /Table/70 that contains live data I190205 18:03:37.586560 9573 storage/compactor/compactor.go:329 [n3,s3,compactor] purging suggested compaction for range /Table/68/3/3815/"T"/PrefixEnd - /Table/70 that contains live data I190205 18:03:37.599331 9249 storage/compactor/compactor.go:329 [n2,s2,compactor] purging suggested compaction for range /Table/68/2/"H"/3232 - /Table/68/2/"L"/688 that contains live data I190205 18:03:37.599682 9249 storage/compactor/compactor.go:329 [n2,s2,compactor] purging suggested compaction for range /Table/68/2/"L"/688 - /Table/68/2/"S"/617 that contains live data I190205 18:03:37.600004 9249 storage/compactor/compactor.go:329 [n2,s2,compactor] purging suggested compaction for range /Table/68/2/"S"/617 - /Table/68/2/"Z"/546 that contains live data I190205 18:03:37.600286 9249 storage/compactor/compactor.go:329 [n2,s2,compactor] purging suggested compaction for range /Table/68/2/"V"/3064 - /Table/68/2/"Z"/546 that contains live data I190205 18:03:37.600723 9249 storage/compactor/compactor.go:329 [n2,s2,compactor] purging suggested compaction for range /Table/68/3/520/"A"/PrefixEnd - /Table/68/3/1187/"R"/PrefixEnd that contains live data I190205 18:03:37.600980 9249 storage/compactor/compactor.go:329 [n2,s2,compactor] purging suggested compaction for range /Table/68/3/1187/"R"/PrefixEnd - /Table/70 that contains live data I190205 18:03:37.601267 9249 storage/compactor/compactor.go:329 [n2,s2,compactor] purging suggested compaction for range /Table/68/3/3815/"T"/PrefixEnd - /Table/70 that contains live data I190205 18:03:38.022540 36444 storage/replica_command.go:244 [n2,s2,r275/2:/{Table/72/3/3…-Max}] initiating a split of this range at key /Table/72/3/4281/"R"/PrefixEnd [r276] (manual) I190205 18:03:38.550538 36473 storage/replica_command.go:244 [n2,s2,r276/2:/{Table/72/3/4…-Max}] initiating a split of this range at key /Table/72/3/4948/"I"/PrefixEnd [r277] (manual) I190205 18:03:39.021688 36533 storage/replica_command.go:244 [n2,s2,r277/2:/{Table/72/3/4…-Max}] initiating a split of this range at key /Table/73 [r278] (manual) import_stmt_test.go:1180: job 10 did not match: Description: "IMPORT TABLE csv10.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:////csv/data-0.gz', 'nodelocal:////csv/data-1.gz', 'nodelocal:////csv/data-2.gz', 'nodelocal:////csv/data-3.gz', 'nodelocal:////csv/data-4.gz') WITH decompress = 'gzip'" != "IMPORT TABLE csv6.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///empty.csv')" TestImportCSVStmt/empty-with-files ...storage/replica_command.go:244 [n3,s3,r233/3:/Table/66/{2/"K"/3…-3/2872/…}] initiating a split of this range at key /Table/66/2/"O"/1419 [r235] (manual) I190205 18:02:11.967595 28068 storage/replica_command.go:244 [n3,s3,r234/3:/{Table/66/3/3…-Max}] initiating a split of this range at key /Table/66/3/4093/"L" [r236] (manual) I190205 18:02:12.771569 28084 storage/replica_command.go:244 [n3,s3,r235/3:/Table/66/{2/"O"/1…-3/2872/…}] initiating a split of this range at key /Table/66/3/1471/"P"/PrefixEnd [r237] (manual) I190205 18:02:12.780075 28190 storage/replica_command.go:244 [n3,s3,r235/3:/Table/66/{2/"O"/1…-3/2872/…}] initiating a split of this range at key /Table/66/2/"R"/3866 [r238] (manual) I190205 18:02:13.153221 28130 storage/replica_command.go:244 [n3,s3,r238/3:/Table/66/{2/"R"/3…-3/2872/…}] initiating a split of this range at key /Table/66/2/"V"/1348 [r239] (manual) I190205 18:02:13.163599 28084 storage/replica_command.go:244 [n3,s3,r238/3:/Table/66/{2/"R"/3…-3/2872/…}] initiating a split of this range at key /Table/66/3/1471/"P"/PrefixEnd [r240] (manual) I190205 18:02:13.252688 8835 server/status/runtime.go:464 [n1] runtime stats: 1.5 GiB RSS, 673 goroutines, 131 MiB/33 MiB/204 MiB GO alloc/idle/total, 197 MiB/245 MiB CGO alloc/total, 5826.6 CGO/sec, 136.2/14.2 %(u/s)time, 0.6 %gc (4x), 4.1 MiB/4.1 MiB (r/w)net I190205 18:02:13.597025 28084 storage/replica_command.go:244 [n3,s3,r239/3:/Table/66/{2/"V"/1…-3/2872/…}] initiating a split of this range at key /Table/66/3/1471/"P"/PrefixEnd [r241] (manual) I190205 18:02:13.683187 28222 storage/replica_command.go:244 [n3,s3,r239/3:/Table/66/{2/"V"/1…-3/2872/…}] initiating a split of this range at key /Table/66/2/"Y"/3795 [r242] (manual) I190205 18:02:13.951157 28084 storage/replica_command.go:244 [n3,s3,r241/3:/Table/66/3/{1471/"…-2872/"…}] initiating a split of this range at key /Table/66/3/2138/"G"/PrefixEnd [r243] (manual) I190205 18:02:13.966378 28222 storage/replica_command.go:244 [n3,s3,r239/3:/Table/66/{2/"V"/1…-3/1471/…}] initiating a split of this range at key /Table/66/2/"Y"/3795 [r244] (manual) I190205 18:02:14.398051 28390 storage/replica_command.go:244 [n3,s3,r244/3:/Table/66/{2/"Y"/3…-3/1471/…}] initiating a split of this range at key /Table/66/3/453/"L"/PrefixEnd [r245] (manual) I190205 18:02:14.461788 28084 storage/replica_command.go:244 [n3,s3,r243/3:/Table/66/3/2{138/"G…-872/"M…}] initiating a split of this range at key /Table/66/3/2206/"W" [r246] (manual) I190205 18:02:15.385777 28409 storage/replica_command.go:244 [n3,s3,r245/3:/Table/66/3/{453/"L…-1471/"…}] initiating a split of this range at key /Table/66/3/805/"Z" [r247] (manual); delayed split for 0.2s to avoid Raft snapshot I190205 18:02:15.537922 28385 storage/replica_command.go:244 [n3,s3,r236/3:/{Table/66/3/4…-Max}] initiating a split of this range at key /Table/67 [r248] (manual) I190205 18:02:15.643209 9255 server/status/runtime.go:464 [n2] runtime stats: 1.5 GiB RSS, 670 goroutines, 100 MiB/60 MiB/204 MiB GO alloc/idle/total, 196 MiB/245 MiB CGO alloc/total, 4912.6 CGO/sec, 133.0/13.9 %(u/s)time, 0.5 %gc (5x), 4.2 MiB/4.2 MiB (r/w)net I190205 18:02:16.456933 28569 storage/replica_command.go:244 [n3,s3,r236/3:/Table/6{6/3/4093…-7}] initiating a split of this range at key /Table/66/3/4431/"L" [r249] (manual) I190205 18:02:16.603408 9579 server/status/runtime.go:464 [n3] runtime stats: 1.5 GiB RSS, 670 goroutines, 144 MiB/24 MiB/204 MiB GO alloc/idle/total, 196 MiB/246 MiB CGO alloc/total, 4739.5 CGO/sec, 133.3/13.6 %(u/s)time, 0.5 %gc (5x), 4.1 MiB/4.1 MiB (r/w)net import_stmt_test.go:1180: job 7 did not match: Description: "IMPORT TABLE csv7.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///empty.csv', 'nodelocal:///csv/data-0', 'nodelocal:///csv/data-1', 'nodelocal:///csv/data-2', 'nodelocal:///csv/data-3', 'nodelocal:///csv/data-4')" != "CREATE STATISTICS __auto__ FROM [59] AS OF SYSTEM TIME '-30s'" TestImportCSVStmt/empty-file ...g replica r198/2 I190205 18:01:49.802452 8904 storage/store.go:2669 [n1,s1,r195/1:/Table/61/3/{607/"J…-774/"U"}] removing replica r198/1 I190205 18:01:49.815062 9500 storage/store.go:2669 [n3,s3,r195/3:/Table/61/3/{607/"J…-774/"U"}] removing replica r198/3 I190205 18:01:50.015819 25828 storage/replica_command.go:383 [n3,merge,s3,r179/3:/Table/61/1/{741-992}] initiating a merge of r184:/Table/61/1/{992-1707} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=0] into this range (lhs+rhs has (size=6.6 KiB+19 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=26 KiB, qps=0.00)) I190205 18:01:50.149924 25918 storage/replica_command.go:383 [n1,merge,s1,r201/1:/Table/61/3/3{145/"Z"-811/"P…}] initiating a merge of r109:/Table/6{1/3/3811/"P"/PrefixEnd-2} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=3] into this range (lhs+rhs has (size=18 KiB+32 KiB qps=0.00+6.78 --> 6.78qps) below threshold (size=51 KiB, qps=6.78)) I190205 18:01:50.388960 25875 ccl/importccl/read_import_proc.go:75 [n1,import-distsql] could not fetch file size; falling back to per-file progress: <nil> I190205 18:01:50.619253 8913 storage/store.go:2669 [n1,s1,r201/1:/Table/61/3/3{145/"Z"-811/"P…}] removing replica r109/1 I190205 18:01:50.625826 9499 storage/store.go:2669 [n3,s3,r201/3:/Table/61/3/3{145/"Z"-811/"P…}] removing replica r109/3 I190205 18:01:50.637114 9190 storage/store.go:2669 [n2,s2,r201/2:/Table/61/3/3{145/"Z"-811/"P…}] removing replica r109/2 I190205 18:01:50.787783 9525 storage/store.go:2669 [n3,s3,r179/3:/Table/61/1/{741-992}] removing replica r184/3 I190205 18:01:50.794165 8921 storage/store.go:2669 [n1,s1,r179/1:/Table/61/1/{741-992}] removing replica r184/1 I190205 18:01:50.826763 9231 storage/store.go:2669 [n2,s2,r179/2:/Table/61/1/{741-992}] removing replica r184/2 I190205 18:01:50.974613 26017 storage/replica_command.go:383 [n3,merge,s3,r186/3:/Table/61/2/"{I"/1438-K"/2792}] initiating a merge of r197:/Table/61/2/"{K"/2792-O"/249} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=12 KiB+18 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=30 KiB, qps=0.00)) I190205 18:01:51.083425 25998 storage/replica_command.go:383 [n2,merge,s2,r200/2:/Table/61/2/"{O"/249-Q"/3994}] initiating a merge of r161:/Table/61/2/"{Q"/3994-T"/2957} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=0] into this range (lhs+rhs has (size=14 KiB+15 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=29 KiB, qps=0.00)) I190205 18:01:51.515378 9503 storage/store.go:2669 [n3,s3,r186/3:/Table/61/2/"{I"/1438-K"/2792}] removing replica r197/3 I190205 18:01:51.523438 9226 storage/store.go:2669 [n2,s2,r186/2:/Table/61/2/"{I"/1438-K"/2792}] removing replica r197/2 I190205 18:01:51.550941 8912 storage/store.go:2669 [n1,s1,r186/1:/Table/61/2/"{I"/1438-K"/2792}] removing replica r197/1 I190205 18:01:51.626980 9535 storage/store.go:2669 [n3,s3,r200/3:/Table/61/2/"{O"/249-Q"/3994}] removing replica r161/3 I190205 18:01:51.630187 9205 storage/store.go:2669 [n2,s2,r200/2:/Table/61/2/"{O"/249-Q"/3994}] removing replica r161/2 I190205 18:01:51.637129 8722 storage/store.go:2669 [n1,s1,r200/1:/Table/61/2/"{O"/249-Q"/3994}] removing replica r161/1 I190205 18:01:51.709839 26101 storage/replica_command.go:383 [n1,merge,s1,r197/1:/Table/61/2/"{K"/2792-O"/249}] initiating a merge of r200:/Table/61/2/"{O"/249-T"/2957} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=2] into this range (lhs+rhs has (size=18 KiB+14 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=33 KiB, qps=0.00)) import_stmt_test.go:1180: job 6 did not match: Description: "IMPORT TABLE csv6.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///empty.csv')" != "IMPORT TABLE csv4.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///csv/data-0', 'nodelocal:///csv/data-1', 'nodelocal:///csv/data-2', 'nodelocal:///csv/data-3', 'nodelocal:///csv/data-4') WITH sstsize = '10K'" TestImportCSVStmt ...sMax-tsd} [(n1,s1):1, (n2,s2):2, next=3, gen=0] I190205 17:59:11.751245 8675 storage/replica_raft.go:372 [n1,s1,r3/1:/System/{NodeLive…-tsd}] proposing ADD_REPLICA((n3,s3):3): updated=[(n1,s1):1 (n2,s2):2 (n3,s3):3] next=4 I190205 17:59:11.774953 8675 storage/store_snapshot.go:762 [n1,replicate,s1,r15/1:/Table/{19-20}] sending preemptive snapshot e4be3f02 at applied index 16 I190205 17:59:11.775984 8675 storage/store_snapshot.go:805 [n1,replicate,s1,r15/1:/Table/{19-20}] streamed snapshot to (n2,s2):?: kv pairs: 7, log entries: 6, rate-limit: 8.0 MiB/sec, 0.01s I190205 17:59:11.777909 10086 storage/replica_raftstorage.go:805 [n2,s2,r15/?:{-}] applying preemptive snapshot at index 16 (id=e4be3f02, encoded size=1245, 1 rocksdb batches, 6 log entries) I190205 17:59:11.780196 10086 storage/replica_raftstorage.go:811 [n2,s2,r15/?:/Table/{19-20}] applied preemptive snapshot in 2ms [clear=0ms batch=0ms entries=1ms commit=0ms] I190205 17:59:11.783115 8675 storage/replica_command.go:798 [n1,replicate,s1,r15/1:/Table/{19-20}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r15:/Table/{19-20} [(n1,s1):1, (n3,s3):2, next=3, gen=0] I190205 17:59:11.880721 8675 storage/replica_raft.go:372 [n1,s1,r15/1:/Table/{19-20}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3] next=4 I190205 17:59:11.908664 8675 storage/store_snapshot.go:762 [n1,replicate,s1,r13/1:/Table/1{7-8}] sending preemptive snapshot 0d086ece at applied index 16 I190205 17:59:11.909887 8675 storage/store_snapshot.go:805 [n1,replicate,s1,r13/1:/Table/1{7-8}] streamed snapshot to (n3,s3):?: kv pairs: 7, log entries: 6, rate-limit: 8.0 MiB/sec, 0.01s I190205 17:59:11.912876 10105 storage/replica_raftstorage.go:805 [n3,s3,r13/?:{-}] applying preemptive snapshot at index 16 (id=0d086ece, encoded size=1245, 1 rocksdb batches, 6 log entries) I190205 17:59:11.916074 10105 storage/replica_raftstorage.go:811 [n3,s3,r13/?:/Table/1{7-8}] applied preemptive snapshot in 3ms [clear=0ms batch=0ms entries=2ms commit=0ms] I190205 17:59:11.918849 8675 storage/replica_command.go:798 [n1,replicate,s1,r13/1:/Table/1{7-8}] change replicas (ADD_REPLICA (n3,s3):3): read existing descriptor r13:/Table/1{7-8} [(n1,s1):1, (n2,s2):2, next=3, gen=0] I190205 17:59:11.971325 8675 storage/replica_raft.go:372 [n1,s1,r13/1:/Table/1{7-8}] proposing ADD_REPLICA((n3,s3):3): updated=[(n1,s1):1 (n2,s2):2 (n3,s3):3] next=4 I190205 17:59:12.034294 8675 storage/store_snapshot.go:762 [n1,replicate,s1,r9/1:/Table/1{3-4}] sending preemptive snapshot 23e47208 at applied index 93 I190205 17:59:12.057123 8675 storage/store_snapshot.go:805 [n1,replicate,s1,r9/1:/Table/1{3-4}] streamed snapshot to (n2,s2):?: kv pairs: 207, log entries: 83, rate-limit: 8.0 MiB/sec, 0.05s I190205 17:59:12.107757 10066 storage/replica_raftstorage.go:805 [n2,s2,r9/?:{-}] applying preemptive snapshot at index 93 (id=23e47208, encoded size=78231, 1 rocksdb batches, 83 log entries) I190205 17:59:12.226984 10066 storage/replica_raftstorage.go:811 [n2,s2,r9/?:/Table/1{3-4}] applied preemptive snapshot in 108ms [clear=0ms batch=0ms entries=106ms commit=1ms] I190205 17:59:12.244846 8675 storage/replica_command.go:798 [n1,replicate,s1,r9/1:/Table/1{3-4}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r9:/Table/1{3-4} [(n1,s1):1, (n3,s3):2, next=3, gen=0] I190205 17:59:12.344852 8675 storage/replica_raft.go:372 [n1,s1,r9/1:/Table/1{3-4}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3] next=4 I190205 17:59:12.666934 10046 sql/event_log.go:135 [n1,client=127.0.0.1:58322,user=root] Event: "set_cluster_setting", target: 0, info: {SettingName:kv.import.batch_size Value:10KB User:root} I190205 17:59:13.024772 8835 server/status/runtime.go:464 [n1] runtime stats: 945 MiB RSS, 627 goroutines, 108 MiB/61 MiB/203 MiB GO alloc/idle/total, 82 MiB/124 MiB CGO alloc/total, 0.0 CGO/sec, 0.0/0.0 %(u/s)time, 0.0 %gc (76x), 4.1 MiB/4.1 MiB (r/w)net TestImportCSVStmt/schema-in-file-sstsize ...action for range /Table/57/1/4612 - /Table/57/3/263/"D"/PrefixEnd that contains live data I190205 18:01:17.867897 9249 storage/compactor/compactor.go:329 [n2,s2,compactor] purging suggested compaction for range /Table/57/3/263/"D"/PrefixEnd - /Table/59 that contains live data I190205 18:01:17.870473 9249 storage/compactor/compactor.go:329 [n2,s2,compactor] purging suggested compaction for range /Table/57/3/1979/"D"/PrefixEnd - /Table/57/3/3313/"L"/PrefixEnd that contains live data I190205 18:01:17.870911 9249 storage/compactor/compactor.go:329 [n2,s2,compactor] purging suggested compaction for range /Table/57/3/3313/"L"/PrefixEnd - /Table/59 that contains live data I190205 18:01:18.732636 22570 storage/replica_command.go:244 [n2,s2,r162/2:/Table/61/3/{1440/"…-3145/"…}] initiating a split of this range at key /Table/61/3/2107/"B"/PrefixEnd [r163] (manual) I190205 18:01:18.795606 22599 storage/replica_command.go:244 [n3,s3,r201/3:/{Table/61/3/3…-Max}] initiating a split of this range at key /Table/61/3/4807/"X" [r203] (manual) I190205 18:01:19.607903 22661 storage/replica_command.go:244 [n2,s2,r163/2:/Table/61/3/{2107/"…-3145/"…}] initiating a split of this range at key /Table/61/3/2188/"E" [r164] (manual) I190205 18:01:20.593575 22790 storage/replica_command.go:244 [n3,s3,r199/3:/Table/61/{2/"W"/3…-3/607/"…}] initiating a split of this range at key /Table/61/2/"Z"/2833 [r204] (manual) I190205 18:01:21.023431 22810 storage/replica_command.go:244 [n1,s1,r201/1:/Table/61/3/{3145/"…-4807/"…}] initiating a split of this range at key /Table/61/3/3811/"P"/PrefixEnd [r109] (manual) I190205 18:01:21.374816 22810 storage/replica_command.go:244 [n1,s1,r109/1:/Table/61/3/{3811/"…-4807/"…}] initiating a split of this range at key /Table/61/3/3873/"Z" [r110] (manual) I190205 18:01:21.482793 22740 storage/replica_command.go:244 [n2,s2,r164/2:/Table/61/3/{2188/"…-3145/"…}] initiating a split of this range at key /Table/61/3/2752/"W" [r165] (manual) I190205 18:01:22.148452 22936 storage/replica_command.go:244 [n3,s3,r203/3:/{Table/61/3/4…-Max}] initiating a split of this range at key /Table/62 [r205] (manual) I190205 18:01:22.765115 23049 storage/replica_command.go:244 [n1,s1,r110/1:/Table/61/3/{3873/"…-4807/"…}] initiating a split of this range at key /Table/61/3/4385/"R" [r141] (manual) I190205 18:01:23.131342 23052 storage/replica_command.go:383 [n3,merge,s3,r174/3:/Table/{59/3/4733…-60}] initiating a merge of r175:/Table/6{0-1/1/741} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=2] into this range (lhs+rhs has (size=7.1 KiB+19 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=26 KiB, qps=0.00)) I190205 18:01:23.214735 8835 server/status/runtime.go:464 [n1] runtime stats: 1.4 GiB RSS, 678 goroutines, 119 MiB/45 MiB/204 MiB GO alloc/idle/total, 188 MiB/231 MiB CGO alloc/total, 4737.6 CGO/sec, 143.1/13.8 %(u/s)time, 0.8 %gc (5x), 3.2 MiB/3.2 MiB (r/w)net I190205 18:01:23.781051 8922 storage/store.go:2669 [n1,s1,r174/1:/Table/{59/3/4733…-60}] removing replica r175/1 I190205 18:01:23.786445 9199 storage/store.go:2669 [n2,s2,r174/2:/Table/{59/3/4733…-60}] removing replica r175/2 I190205 18:01:23.786540 9505 storage/store.go:2669 [n3,s3,r174/3:/Table/{59/3/4733…-60}] removing replica r175/3 I190205 18:01:24.219326 23158 storage/replica_command.go:383 [n3,merge,s3,r131/3:/Table/59/{1/4485-2/NULL/…}] initiating a merge of r134:/Table/59/2/NULL/{1089-4425} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=20 KiB+20 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=40 KiB, qps=0.00)) import_stmt_test.go:1180: job 4 did not match: Description: "IMPORT TABLE csv4.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///csv/data-0', 'nodelocal:///csv/data-1', 'nodelocal:///csv/data-2', 'nodelocal:///csv/data-3', 'nodelocal:///csv/data-4') WITH sstsize = '10K'" != "CREATE STATISTICS __auto__ FROM [55] AS OF SYSTEM TIME '-30s'" TestImportCSVStmt/schema-in-file-auto-gzip ... (manual); delayed split for 0.2s to avoid Raft snapshot I190205 18:04:12.747814 8951 storage/store.go:2669 [n1,s1,r337/1:/Table/72/3/{2280/"…-4948/"…}] removing replica r277/1 I190205 18:04:12.749147 9230 storage/store.go:2669 [n2,s2,r337/2:/Table/72/3/{2280/"…-4948/"…}] removing replica r277/2 I190205 18:04:12.752104 9526 storage/store.go:2669 [n3,s3,r337/3:/Table/72/3/{2280/"…-4948/"…}] removing replica r277/3 I190205 18:04:13.044405 39469 storage/replica_command.go:244 [n2,s2,r369/2:/Table/74/{2/"U"/7…-3/756/"…}] initiating a split of this range at key /Table/74/2/"X"/3170 [r370] (manual) I190205 18:04:13.369867 39469 storage/replica_command.go:244 [n2,s2,r370/2:/Table/74/{2/"X"/3…-3/756/"…}] initiating a split of this range at key /Table/74/3/76/"Y" [r371] (manual) I190205 18:04:13.409700 8835 server/status/runtime.go:464 [n1] runtime stats: 1.8 GiB RSS, 671 goroutines, 101 MiB/122 MiB/271 MiB GO alloc/idle/total, 277 MiB/320 MiB CGO alloc/total, 4790.3 CGO/sec, 140.2/11.3 %(u/s)time, 0.7 %gc (5x), 2.9 MiB/2.9 MiB (r/w)net I190205 18:04:13.581569 40251 storage/replica_command.go:383 [n1,merge,s1,r144/1:/Table/72/{1/741-2/"V"/4…}] initiating a merge of r320:/Table/72/2/"{V"/4337-Z"/1794} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=227 KiB+18 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=245 KiB, qps=0.00)) I190205 18:04:13.790872 8937 storage/store.go:2669 [n1,s1,r144/1:/Table/72/{1/741-2/"V"/4…}] removing replica r320/1 I190205 18:04:13.804087 9507 storage/store.go:2669 [n3,s3,r144/3:/Table/72/{1/741-2/"V"/4…}] removing replica r320/3 I190205 18:04:13.806617 9210 storage/store.go:2669 [n2,s2,r144/2:/Table/72/{1/741-2/"V"/4…}] removing replica r320/2 I190205 18:04:13.890425 40227 storage/replica_command.go:383 [n1,merge,s1,r144/1:/Table/72/{1/741-2/"Z"/1…}] initiating a merge of r331:/Table/72/{2/"Z"/1794-3/1614/"C"} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=3] into this range (lhs+rhs has (size=245 KiB+47 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=292 KiB, qps=0.00)) I190205 18:04:14.063040 9505 storage/store.go:2669 [n3,s3,r144/3:/Table/72/{1/741-2/"Z"/1…}] removing replica r331/3 I190205 18:04:14.065216 8928 storage/store.go:2669 [n1,s1,r144/1:/Table/72/{1/741-2/"Z"/1…}] removing replica r331/1 I190205 18:04:14.088413 9185 storage/store.go:2669 [n2,s2,r144/2:/Table/72/{1/741-2/"Z"/1…}] removing replica r331/2 I190205 18:04:14.239558 40271 storage/replica_command.go:383 [n1,merge,s1,r144/1:/Table/72/{1/741-3/1614/…}] initiating a merge of r340:/Table/72/3/{1614/"C"-2280/"S"/PrefixEnd} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=0] into this range (lhs+rhs has (size=292 KiB+18 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=311 KiB, qps=0.00)) I190205 18:04:14.584782 40379 storage/replica_command.go:383 [n3,merge,s3,r340/3:/Table/72/3/{1614/"…-2280/"…}] initiating a merge of r337:/Table/7{2/3/2280/"S"/PrefixEnd-4/1/741} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=5] into this range (lhs+rhs has (size=18 KiB+94 KiB qps=0.00+3.62 --> 3.62qps) below threshold (size=112 KiB, qps=3.62)) I190205 18:04:14.590946 8941 storage/store.go:2669 [n1,s1,r144/1:/Table/72/{1/741-3/1614/…}] removing replica r340/1 I190205 18:04:14.604285 9206 storage/store.go:2669 [n2,s2,r144/2:/Table/72/{1/741-3/1614/…}] removing replica r340/2 import_stmt_test.go:1180: job 11 did not match: Description: "IMPORT TABLE csv11.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:////csv/data-0.gz', 'nodelocal:////csv/data-1.gz', 'nodelocal:////csv/data-2.gz', 'nodelocal:////csv/data-3.gz', 'nodelocal:////csv/data-4.gz') WITH decompress = 'auto'" != "IMPORT TABLE csv7.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///empty.csv', 'nodelocal:///csv/data-0', 'nodelocal:///csv/data-1', 'nodelocal:///csv/data-2', 'nodelocal:///csv/data-3', 'nodelocal:///csv/data-4')" TestImportCSVStmt/schema-in-file-no-decompress ...:/Table/68/1/{741-1456}] removing replica r170/3 I190205 18:03:09.588628 8910 storage/store.go:2669 [n1,s1,r168/1:/Table/68/1/{741-1456}] removing replica r170/1 I190205 18:03:09.728167 33570 storage/replica_command.go:383 [n2,merge,s2,r168/2:/Table/68/1/{741-1689}] initiating a merge of r212:/Table/68/1/{1689-2404} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=0] into this range (lhs+rhs has (size=25 KiB+19 KiB qps=3.99+0.00 --> 3.99qps) below threshold (size=44 KiB, qps=3.99)) I190205 18:03:09.735905 33683 storage/replica_command.go:244 [n3,s3,r301/3:/Table/70/{2/"O"/4…-3/360/"…}] initiating a split of this range at key /Table/70/2/"R"/2852 [r302] (manual) I190205 18:03:10.116941 9509 storage/store.go:2669 [n3,s3,r168/3:/Table/68/1/{741-1689}] removing replica r212/3 I190205 18:03:10.128932 9223 storage/store.go:2669 [n2,s2,r168/2:/Table/68/1/{741-1689}] removing replica r212/2 I190205 18:03:10.142650 8909 storage/store.go:2669 [n1,s1,r168/1:/Table/68/1/{741-1689}] removing replica r212/1 I190205 18:03:10.233154 33783 storage/replica_command.go:244 [n3,s3,r302/3:/Table/70/{2/"R"/2…-3/360/"…}] initiating a split of this range at key /Table/70/2/"V"/334 [r303] (manual) I190205 18:03:10.538610 33769 storage/replica_command.go:244 [n3,s3,r303/3:/Table/70/{2/"V"/3…-3/360/"…}] initiating a split of this range at key /Table/70/2/"Y"/1402 [r304] (manual) I190205 18:03:10.613954 33746 storage/replica_command.go:383 [n2,merge,s2,r168/2:/Table/68/1/{741-2404}] initiating a merge of r166:/Table/68/{1/2404-2/"B"/3824} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=5] into this range (lhs+rhs has (size=44 KiB+78 KiB qps=4.84+6.21 --> 11.05qps) below threshold (size=122 KiB, qps=11.05)) I190205 18:03:10.971673 9225 storage/store.go:2669 [n2,s2,r168/2:/Table/68/1/{741-2404}] removing replica r166/2 I190205 18:03:11.007038 8717 storage/store.go:2669 [n1,s1,r168/1:/Table/68/1/{741-2404}] removing replica r166/1 I190205 18:03:11.058169 9529 storage/store.go:2669 [n3,s3,r168/3:/Table/68/1/{741-2404}] removing replica r166/3 I190205 18:03:11.330578 33848 storage/replica_command.go:244 [n3,s3,r294/3:/{Table/70/3/4…-Max}] initiating a split of this range at key /Table/71 [r305] (manual) I190205 18:03:11.730055 33855 storage/replica_command.go:383 [n2,merge,s2,r168/2:/Table/68/{1/741-2/"B"/3…}] initiating a merge of r254:/Table/68/2/"{B"/3824-E"/888} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=122 KiB+13 KiB qps=7.99+0.00 --> 7.99qps) below threshold (size=134 KiB, qps=7.99)) I190205 18:03:12.299917 9200 storage/store.go:2669 [n2,s2,r168/2:/Table/68/{1/741-2/"B"/3…}] removing replica r254/2 I190205 18:03:12.314674 9513 storage/store.go:2669 [n3,s3,r168/3:/Table/68/{1/741-2/"B"/3…}] removing replica r254/3 I190205 18:03:12.350290 8929 storage/store.go:2669 [n1,s1,r168/1:/Table/68/{1/741-2/"B"/3…}] removing replica r254/1 I190205 18:03:12.385265 33899 storage/replica_command.go:383 [n1,merge,s1,r254/1:/Table/68/2/"{B"/3824-E"/888}] initiating a merge of r255:/Table/68/2/"{E"/888-H"/3232} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=3] into this range (lhs+rhs has (size=13 KiB+18 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=31 KiB, qps=0.00)) I190205 18:03:12.477914 33965 storage/replica_command.go:383 [n2,merge,s2,r168/2:/Table/68/{1/741-2/"E"/8…}] initiating a merge of r255:/Table/68/2/"{E"/888-H"/3232} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=3] into this range (lhs+rhs has (size=134 KiB+18 KiB qps=4.59+0.00 --> 4.59qps) below threshold (size=152 KiB, qps=4.59)) import_stmt_test.go:1180: job 9 did not match: Description: "IMPORT TABLE csv9.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///csv/data-0', 'nodelocal:///csv/data-1', 'nodelocal:///csv/data-2', 'nodelocal:///csv/data-3', 'nodelocal:///csv/data-4') WITH decompress = 'none'" != "CREATE STATISTICS __auto__ FROM [57] AS OF SYSTEM TIME '-30s'" TestImportCSVStmt/schema-in-query-transform-only ... 9192 storage/store.go:2669 [n2,s2,r100/2:/Table/59{-/2/"R"/4…}] removing replica r152/2 I190205 18:01:43.938554 9528 storage/store.go:2669 [n3,s3,r100/3:/Table/59{-/2/"R"/4…}] removing replica r152/3 I190205 18:01:43.949112 8936 storage/store.go:2669 [n1,s1,r100/1:/Table/59{-/2/"R"/4…}] removing replica r152/1 I190205 18:01:44.442921 9535 storage/store.go:2669 [n3,s3,r185/3:/Table/61/1/{2267-4455}] removing replica r191/3 I190205 18:01:44.462897 8914 storage/store.go:2669 [n1,s1,r185/1:/Table/61/1/{2267-4455}] removing replica r191/1 I190205 18:01:44.527437 9219 storage/store.go:2669 [n2,s2,r185/2:/Table/61/1/{2267-4455}] removing replica r191/2 I190205 18:01:44.648428 25343 storage/replica_command.go:383 [n3,merge,s3,r182/3:/Table/61/1/{1707-2267}] initiating a merge of r185:/Table/61/{1/2267-2/"A"/3432} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=5] into this range (lhs+rhs has (size=15 KiB+76 KiB qps=1.77+5.59 --> 7.37qps) below threshold (size=90 KiB, qps=7.37)) I190205 18:01:45.018810 9503 storage/store.go:2669 [n3,s3,r182/3:/Table/61/1/{1707-2267}] removing replica r185/3 I190205 18:01:45.029332 9209 storage/store.go:2669 [n2,s2,r182/2:/Table/61/1/{1707-2267}] removing replica r185/2 I190205 18:01:45.033788 8903 storage/store.go:2669 [n1,s1,r182/1:/Table/61/1/{1707-2267}] removing replica r185/1 I190205 18:01:45.527526 25347 storage/replica_command.go:383 [n3,merge,s3,r182/3:/Table/61/{1/1707-2/"A"/3…}] initiating a merge of r192:/Table/61/2/"{A"/3432-E"/785} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=90 KiB+18 KiB qps=5.99+0.00 --> 5.99qps) below threshold (size=109 KiB, qps=5.99)) I190205 18:01:45.626120 9255 server/status/runtime.go:464 [n2] runtime stats: 1.5 GiB RSS, 674 goroutines, 106 MiB/56 MiB/204 MiB GO alloc/idle/total, 189 MiB/233 MiB CGO alloc/total, 5393.0 CGO/sec, 145.7/11.6 %(u/s)time, 1.4 %gc (5x), 2.7 MiB/2.7 MiB (r/w)net I190205 18:01:46.164665 9500 storage/store.go:2669 [n3,s3,r182/3:/Table/61/{1/1707-2/"A"/3…}] removing replica r192/3 I190205 18:01:46.186703 8939 storage/store.go:2669 [n1,s1,r182/1:/Table/61/{1/1707-2/"A"/3…}] removing replica r192/1 I190205 18:01:46.247077 9178 storage/store.go:2669 [n2,s2,r182/2:/Table/61/{1/1707-2/"A"/3…}] removing replica r192/2 I190205 18:01:46.291058 25558 storage/replica_command.go:383 [n3,merge,s3,r182/3:/Table/61/{1/1707-2/"E"/7…}] initiating a merge of r194:/Table/61/2/"{E"/785-F"/3489} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=109 KiB+8.1 KiB qps=5.99+0.00 --> 5.99qps) below threshold (size=117 KiB, qps=5.99)) I190205 18:01:46.529935 9579 server/status/runtime.go:464 [n3] runtime stats: 1.5 GiB RSS, 663 goroutines, 127 MiB/42 MiB/204 MiB GO alloc/idle/total, 180 MiB/224 MiB CGO alloc/total, 9305.5 CGO/sec, 148.9/13.8 %(u/s)time, 1.6 %gc (6x), 2.8 MiB/2.8 MiB (r/w)net I190205 18:01:46.537164 25578 storage/replica_command.go:383 [n1,merge,s1,r109/1:/Table/61/3/38{11/"P…-73/"Z"}] initiating a merge of r110:/Table/61/3/{3873/"Z"-4385/"R"} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=1.7 KiB+14 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=16 KiB, qps=0.00)) import_stmt_test.go:1180: job 5 did not match: Description: "IMPORT TABLE \"\".\"\".t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///csv/data-0-opts', 'nodelocal:///csv/data-1-opts', 'nodelocal:///csv/data-2-opts', 'nodelocal:///csv/data-3-opts', 'nodelocal:///csv/data-4-opts') WITH comment = '#', delimiter = '|', \"nullif\" = '', skip = '2', transform = 'nodelocal:///5'" != "IMPORT TABLE csv3.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///csv/data-0-opts', 'nodelocal:///csv/data-1-opts', 'nodelocal:///csv/data-2-opts', 'nodelocal:///csv/data-3-opts', 'nodelocal:///csv/data-4-opts') WITH comment = '#', delimiter = '|', \"nullif\" = '', skip = '2'" ``` Please assign, take a look and update the issue accordingly.
1.0
teamcity: failed test: TestImportCSVStmt - The following tests appear to have failed on master (testrace): TestImportCSVStmt/schema-in-file-implicit-gzip, TestImportCSVStmt/schema-in-file-no-decompress, TestImportCSVStmt/empty-file, TestImportCSVStmt/schema-in-query-opts, TestImportCSVStmt/schema-in-file-sstsize, TestImportCSVStmt/schema-in-file-auto-decompress, TestImportCSVStmt/schema-in-file-auto-gzip, TestImportCSVStmt, TestImportCSVStmt/schema-in-query-transform-only, TestImportCSVStmt/schema-in-file-explicit-gzip, TestImportCSVStmt/empty-with-files You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+TestImportCSVStmt). [#1125945](https://teamcity.cockroachdb.com/viewLog.html?buildId=1125945): ``` TestImportCSVStmt/schema-in-file-implicit-gzip ...002/"Y…-834/"Y"}] initiating a split of this range at key /Table/76/3/4224/"M" [r402] (manual) I190205 18:04:39.087848 9224 storage/store.go:2669 [n2,s2,r366/2:/Table/74/2/"{N"/794-Q"/3241}] removing replica r368/2 I190205 18:04:39.094016 9541 storage/store.go:2669 [n3,s3,r366/3:/Table/74/2/"{N"/794-Q"/3241}] removing replica r368/3 I190205 18:04:39.094483 8905 storage/store.go:2669 [n1,s1,r366/1:/Table/74/2/"{N"/794-Q"/3241}] removing replica r368/1 I190205 18:04:39.172897 9214 storage/store.go:2669 [n2,s2,r341/2:/Table/74/1/{3266-4842}] removing replica r289/2 I190205 18:04:39.174095 8938 storage/store.go:2669 [n1,s1,r341/1:/Table/74/1/{3266-4842}] removing replica r289/1 I190205 18:04:39.189718 9486 storage/store.go:2669 [n3,s3,r341/3:/Table/74/1/{3266-4842}] removing replica r289/3 I190205 18:04:39.195738 42996 storage/replica_command.go:383 [n2,merge,s2,r289/2:/Table/74/{1/4842-2/"C"/3…}] initiating a merge of r361:/Table/74/2/"{C"/3539-G"/917} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=18 KiB+18 KiB qps=2.69+0.00 --> 2.69qps) below threshold (size=37 KiB, qps=2.69)) I190205 18:04:39.674788 42929 storage/replica_command.go:383 [n3,merge,s3,r343/3:/Table/74/1/{741-2632}] initiating a merge of r346:/Table/74/1/{2632-3266} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=0] into this range (lhs+rhs has (size=50 KiB+17 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=67 KiB, qps=0.00)) I190205 18:04:40.045754 42443 sql/event_log.go:135 [n1] Event: "create_statistics", target: 74, info: {StatisticName:__auto__ Statement:CREATE STATISTICS __auto__ FROM [74] AS OF SYSTEM TIME '-30s'} I190205 18:04:40.174598 9531 storage/store.go:2669 [n3,s3,r343/3:/Table/74/1/{741-2632}] removing replica r346/3 I190205 18:04:40.175909 8910 storage/store.go:2669 [n1,s1,r343/1:/Table/74/1/{741-2632}] removing replica r346/1 I190205 18:04:40.187748 9195 storage/store.go:2669 [n2,s2,r343/2:/Table/74/1/{741-2632}] removing replica r346/2 I190205 18:04:40.247398 43054 storage/replica_command.go:383 [n2,merge,s2,r366/2:/Table/74/2/"{N"/794-U"/723}] initiating a merge of r369:/Table/74/2/"{U"/723-X"/3170} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=37 KiB+18 KiB qps=0.97+0.00 --> 0.97qps) below threshold (size=55 KiB, qps=0.97)) I190205 18:04:40.343699 43081 storage/replica_command.go:383 [n3,merge,s3,r341/3:/Table/74/{1/3266-2/"C"/3…}] initiating a merge of r361:/Table/74/2/"{C"/3539-G"/917} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=60 KiB+18 KiB qps=5.63+1.70 --> 7.33qps) below threshold (size=78 KiB, qps=7.33)) I190205 18:04:40.570936 9199 storage/store.go:2669 [n2,s2,r366/2:/Table/74/2/"{N"/794-U"/723}] removing replica r369/2 I190205 18:04:40.576583 9531 storage/store.go:2669 [n3,s3,r366/3:/Table/74/2/"{N"/794-U"/723}] removing replica r369/3 I190205 18:04:40.598783 8939 storage/store.go:2669 [n1,s1,r366/1:/Table/74/2/"{N"/794-U"/723}] removing replica r369/1 I190205 18:04:40.827940 9481 storage/store.go:2669 [n3,s3,r341/3:/Table/74/{1/3266-2/"C"/3…}] removing replica r361/3 I190205 18:04:40.830389 9209 storage/store.go:2669 [n2,s2,r341/2:/Table/74/{1/3266-2/"C"/3…}] removing replica r361/2 I190205 18:04:40.834986 8918 storage/store.go:2669 [n1,s1,r341/1:/Table/74/{1/3266-2/"C"/3…}] removing replica r361/1 import_stmt_test.go:1180: job 12 did not match: Description: "IMPORT TABLE csv12.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:////csv/data-0.gz', 'nodelocal:////csv/data-1.gz', 'nodelocal:////csv/data-2.gz', 'nodelocal:////csv/data-3.gz', 'nodelocal:////csv/data-4.gz')" != "IMPORT TABLE csv8.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///csv/data-0', 'nodelocal:///csv/data-1', 'nodelocal:///csv/data-2', 'nodelocal:///csv/data-3', 'nodelocal:///csv/data-4') WITH decompress = 'auto'" TestImportCSVStmt/schema-in-file-auto-decompress ...2:34.891226 29904 storage/replica_command.go:244 [n2,s2,r273/2:/Table/68/3/{1187/"…-2549/"…}] initiating a split of this range at key /Table/68/3/1854/"I"/PrefixEnd [r274] (manual) I190205 18:02:35.120817 30188 storage/replica_command.go:383 [n3,merge,s3,r208/3:/Table/66/2/"{B"/4968-F"/2346}] initiating a merge of r222:/Table/66/2/"{F"/2346-H"/1541} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=18 KiB+9.7 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=28 KiB, qps=0.00)) I190205 18:02:35.328598 30142 storage/replica_command.go:383 [n1,merge,s1,r208/1:/Table/66/2/"{B"/4968-F"/2346}] initiating a merge of r222:/Table/66/2/"{F"/2346-H"/1541} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=18 KiB+9.7 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=28 KiB, qps=0.00)) I190205 18:02:35.661168 9255 server/status/runtime.go:464 [n2] runtime stats: 1.6 GiB RSS, 676 goroutines, 95 MiB/66 MiB/204 MiB GO alloc/idle/total, 190 MiB/234 MiB CGO alloc/total, 4992.3 CGO/sec, 142.7/12.2 %(u/s)time, 0.6 %gc (5x), 3.3 MiB/3.3 MiB (r/w)net I190205 18:02:35.902427 30207 storage/replica_command.go:244 [n1,s1,r274/1:/Table/68/3/{1854/"…-2549/"…}] initiating a split of this range at key /Table/68/3/1883/"L" [r142] (manual) I190205 18:02:36.562791 8950 storage/store.go:2669 [n1,s1,r208/1:/Table/66/2/"{B"/4968-F"/2346}] removing replica r222/1 I190205 18:02:36.568458 9535 storage/store.go:2669 [n3,s3,r208/3:/Table/66/2/"{B"/4968-F"/2346}] removing replica r222/3 I190205 18:02:36.638714 9579 server/status/runtime.go:464 [n3] runtime stats: 1.6 GiB RSS, 670 goroutines, 142 MiB/26 MiB/204 MiB GO alloc/idle/total, 190 MiB/234 MiB CGO alloc/total, 4787.4 CGO/sec, 140.5/12.0 %(u/s)time, 0.6 %gc (5x), 3.3 MiB/3.3 MiB (r/w)net I190205 18:02:36.639448 9219 storage/store.go:2669 [n2,s2,r208/2:/Table/66/2/"{B"/4968-F"/2346}] removing replica r222/2 I190205 18:02:36.837577 30349 storage/replica_command.go:383 [n1,merge,s1,r208/1:/Table/66/2/"{B"/4968-H"/1541}] initiating a merge of r226:/Table/66/2/"{H"/1541-K"/3937} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=2] into this range (lhs+rhs has (size=28 KiB+18 KiB qps=19.86+0.00 --> 19.86qps) below threshold (size=46 KiB, qps=19.86)) I190205 18:02:37.509519 30376 storage/replica_command.go:383 [n3,merge,s3,r226/3:/Table/66/2/"{H"/1541-K"/3937}] initiating a merge of r233:/Table/66/2/"{K"/3937-O"/1419} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=18 KiB+18 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=37 KiB, qps=0.00)) I190205 18:02:37.602138 8908 storage/store.go:2669 [n1,s1,r208/1:/Table/66/2/"{B"/4968-H"/1541}] removing replica r226/1 I190205 18:02:37.603497 9201 storage/store.go:2669 [n2,s2,r208/2:/Table/66/2/"{B"/4968-H"/1541}] removing replica r226/2 I190205 18:02:37.629172 9496 storage/store.go:2669 [n3,s3,r208/3:/Table/66/2/"{B"/4968-H"/1541}] removing replica r226/3 I190205 18:02:37.840870 30462 storage/replica_command.go:383 [n1,merge,s1,r249/1:/Table/6{6/3/4431…-7}] initiating a merge of r248:/Table/6{7-8/1/741} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=3] into this range (lhs+rhs has (size=16 KiB+19 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=35 KiB, qps=0.00)) import_stmt_test.go:1180: job 8 did not match: Description: "IMPORT TABLE csv8.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///csv/data-0', 'nodelocal:///csv/data-1', 'nodelocal:///csv/data-2', 'nodelocal:///csv/data-3', 'nodelocal:///csv/data-4') WITH decompress = 'auto'" != "IMPORT TABLE \"\".\"\".t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///csv/data-0-opts', 'nodelocal:///csv/data-1-opts', 'nodelocal:///csv/data-2-opts', 'nodelocal:///csv/data-3-opts', 'nodelocal:///csv/data-4-opts') WITH comment = '#', delimiter = '|', \"nullif\" = '', skip = '2', transform = 'nodelocal:///5'" TestImportCSVStmt/schema-in-query-opts ...n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=135 KiB+19 KiB qps=2.98+2.69 --> 5.67qps) below threshold (size=154 KiB, qps=5.67)) I190205 18:00:52.078010 19718 storage/replica_command.go:244 [n3,s3,r171/3:/{Table/59/3/2…-Max}] initiating a split of this range at key /Table/59/3/3329/"B"/PrefixEnd [r172] (manual) I190205 18:00:52.278113 9537 storage/store.go:2669 [n3,s3,r117/3:/Table/57/{1/4612-2/"X"/3…}] removing replica r126/3 I190205 18:00:52.279167 8719 storage/store.go:2669 [n1,s1,r117/1:/Table/57/{1/4612-2/"X"/3…}] removing replica r126/1 I190205 18:00:52.279897 9204 storage/store.go:2669 [n2,s2,r117/2:/Table/57/{1/4612-2/"X"/3…}] removing replica r126/2 I190205 18:00:52.345961 19791 storage/replica_command.go:383 [n3,merge,s3,r56/3:/Table/57/1/{741-1456}] initiating a merge of r59:/Table/57/1/1{456-752} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=19 KiB+7.8 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=27 KiB, qps=0.00)) I190205 18:00:52.807849 19809 storage/replica_command.go:244 [n3,s3,r172/3:/{Table/59/3/3…-Max}] initiating a split of this range at key /Table/59/3/4031/"B"/PrefixEnd [r173] (manual) I190205 18:00:52.825113 8953 storage/store.go:2669 [n1,s1,r56/1:/Table/57/1/{741-1456}] removing replica r59/1 I190205 18:00:52.839472 9225 storage/store.go:2669 [n2,s2,r56/2:/Table/57/1/{741-1456}] removing replica r59/2 I190205 18:00:52.871815 9543 storage/store.go:2669 [n3,s3,r56/3:/Table/57/1/{741-1456}] removing replica r59/3 I190205 18:00:53.204322 8835 server/status/runtime.go:464 [n1] runtime stats: 1.3 GiB RSS, 671 goroutines, 76 MiB/85 MiB/204 MiB GO alloc/idle/total, 140 MiB/184 MiB CGO alloc/total, 4762.7 CGO/sec, 146.4/14.1 %(u/s)time, 1.1 %gc (6x), 3.1 MiB/3.1 MiB (r/w)net I190205 18:00:53.327528 19863 storage/replica_command.go:383 [n3,merge,s3,r56/3:/Table/57/1/{741-1752}] initiating a merge of r72:/Table/57/1/{1752-2467} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=27 KiB+19 KiB qps=4.58+0.00 --> 4.58qps) below threshold (size=46 KiB, qps=4.58)) I190205 18:00:53.449266 19893 storage/replica_command.go:244 [n3,s3,r173/3:/{Table/59/3/4…-Max}] initiating a split of this range at key /Table/59/3/4733/"B"/PrefixEnd [r174] (manual); delayed split for 0.2s to avoid Raft snapshot I190205 18:00:53.530908 9488 storage/store.go:2669 [n3,s3,r56/3:/Table/57/1/{741-1752}] removing replica r72/3 I190205 18:00:53.535078 9232 storage/store.go:2669 [n2,s2,r56/2:/Table/57/1/{741-1752}] removing replica r72/2 I190205 18:00:53.537881 8949 storage/store.go:2669 [n1,s1,r56/1:/Table/57/1/{741-1752}] removing replica r72/1 I190205 18:00:53.760796 19825 storage/replica_command.go:244 [n3,s3,r174/3:/{Table/59/3/4…-Max}] initiating a split of this range at key /Table/60 [r175] (manual) I190205 18:00:54.389805 19946 storage/replica_command.go:383 [n3,merge,s3,r56/3:/Table/57/1/{741-2467}] initiating a merge of r75:/Table/57/1/{2467-3182} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=2] into this range (lhs+rhs has (size=46 KiB+19 KiB qps=5.66+0.00 --> 5.66qps) below threshold (size=64 KiB, qps=5.66)) I190205 18:00:54.812161 9543 storage/store.go:2669 [n3,s3,r56/3:/Table/57/1/{741-2467}] removing replica r75/3 I190205 18:00:54.817986 9228 storage/store.go:2669 [n2,s2,r56/2:/Table/57/1/{741-2467}] removing replica r75/2 I190205 18:00:54.849856 8953 storage/store.go:2669 [n1,s1,r56/1:/Table/57/1/{741-2467}] removing replica r75/1 import_stmt_test.go:1180: job 3 did not match: Description: "IMPORT TABLE csv3.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///csv/data-0-opts', 'nodelocal:///csv/data-1-opts', 'nodelocal:///csv/data-2-opts', 'nodelocal:///csv/data-3-opts', 'nodelocal:///csv/data-4-opts') WITH comment = '#', delimiter = '|', \"nullif\" = '', skip = '2'" != "CREATE STATISTICS __auto__ FROM [53] AS OF SYSTEM TIME '-30s'" TestImportCSVStmt/schema-in-file-explicit-gzip ...iB CGO alloc/total, 8636.8 CGO/sec, 143.8/12.7 %(u/s)time, 0.9 %gc (5x), 3.7 MiB/3.7 MiB (r/w)net I190205 18:03:36.967954 36306 storage/replica_command.go:244 [n2,s2,r339/2:/{Table/72/3/2…-Max}] initiating a split of this range at key /Table/72/3/3614/"A"/PrefixEnd [r275] (manual) I190205 18:03:37.568594 9573 storage/compactor/compactor.go:329 [n3,s3,compactor] purging suggested compaction for range /Table/68/2/"H"/3232 - /Table/68/2/"L"/688 that contains live data I190205 18:03:37.569024 9573 storage/compactor/compactor.go:329 [n3,s3,compactor] purging suggested compaction for range /Table/68/2/"L"/688 - /Table/68/2/"S"/617 that contains live data I190205 18:03:37.569337 9573 storage/compactor/compactor.go:329 [n3,s3,compactor] purging suggested compaction for range /Table/68/2/"S"/617 - /Table/68/2/"Z"/546 that contains live data I190205 18:03:37.569589 9573 storage/compactor/compactor.go:329 [n3,s3,compactor] purging suggested compaction for range /Table/68/2/"V"/3064 - /Table/68/2/"Z"/546 that contains live data I190205 18:03:37.585786 9573 storage/compactor/compactor.go:329 [n3,s3,compactor] purging suggested compaction for range /Table/68/3/520/"A"/PrefixEnd - /Table/68/3/1187/"R"/PrefixEnd that contains live data I190205 18:03:37.586234 9573 storage/compactor/compactor.go:329 [n3,s3,compactor] purging suggested compaction for range /Table/68/3/1187/"R"/PrefixEnd - /Table/70 that contains live data I190205 18:03:37.586560 9573 storage/compactor/compactor.go:329 [n3,s3,compactor] purging suggested compaction for range /Table/68/3/3815/"T"/PrefixEnd - /Table/70 that contains live data I190205 18:03:37.599331 9249 storage/compactor/compactor.go:329 [n2,s2,compactor] purging suggested compaction for range /Table/68/2/"H"/3232 - /Table/68/2/"L"/688 that contains live data I190205 18:03:37.599682 9249 storage/compactor/compactor.go:329 [n2,s2,compactor] purging suggested compaction for range /Table/68/2/"L"/688 - /Table/68/2/"S"/617 that contains live data I190205 18:03:37.600004 9249 storage/compactor/compactor.go:329 [n2,s2,compactor] purging suggested compaction for range /Table/68/2/"S"/617 - /Table/68/2/"Z"/546 that contains live data I190205 18:03:37.600286 9249 storage/compactor/compactor.go:329 [n2,s2,compactor] purging suggested compaction for range /Table/68/2/"V"/3064 - /Table/68/2/"Z"/546 that contains live data I190205 18:03:37.600723 9249 storage/compactor/compactor.go:329 [n2,s2,compactor] purging suggested compaction for range /Table/68/3/520/"A"/PrefixEnd - /Table/68/3/1187/"R"/PrefixEnd that contains live data I190205 18:03:37.600980 9249 storage/compactor/compactor.go:329 [n2,s2,compactor] purging suggested compaction for range /Table/68/3/1187/"R"/PrefixEnd - /Table/70 that contains live data I190205 18:03:37.601267 9249 storage/compactor/compactor.go:329 [n2,s2,compactor] purging suggested compaction for range /Table/68/3/3815/"T"/PrefixEnd - /Table/70 that contains live data I190205 18:03:38.022540 36444 storage/replica_command.go:244 [n2,s2,r275/2:/{Table/72/3/3…-Max}] initiating a split of this range at key /Table/72/3/4281/"R"/PrefixEnd [r276] (manual) I190205 18:03:38.550538 36473 storage/replica_command.go:244 [n2,s2,r276/2:/{Table/72/3/4…-Max}] initiating a split of this range at key /Table/72/3/4948/"I"/PrefixEnd [r277] (manual) I190205 18:03:39.021688 36533 storage/replica_command.go:244 [n2,s2,r277/2:/{Table/72/3/4…-Max}] initiating a split of this range at key /Table/73 [r278] (manual) import_stmt_test.go:1180: job 10 did not match: Description: "IMPORT TABLE csv10.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:////csv/data-0.gz', 'nodelocal:////csv/data-1.gz', 'nodelocal:////csv/data-2.gz', 'nodelocal:////csv/data-3.gz', 'nodelocal:////csv/data-4.gz') WITH decompress = 'gzip'" != "IMPORT TABLE csv6.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///empty.csv')" TestImportCSVStmt/empty-with-files ...storage/replica_command.go:244 [n3,s3,r233/3:/Table/66/{2/"K"/3…-3/2872/…}] initiating a split of this range at key /Table/66/2/"O"/1419 [r235] (manual) I190205 18:02:11.967595 28068 storage/replica_command.go:244 [n3,s3,r234/3:/{Table/66/3/3…-Max}] initiating a split of this range at key /Table/66/3/4093/"L" [r236] (manual) I190205 18:02:12.771569 28084 storage/replica_command.go:244 [n3,s3,r235/3:/Table/66/{2/"O"/1…-3/2872/…}] initiating a split of this range at key /Table/66/3/1471/"P"/PrefixEnd [r237] (manual) I190205 18:02:12.780075 28190 storage/replica_command.go:244 [n3,s3,r235/3:/Table/66/{2/"O"/1…-3/2872/…}] initiating a split of this range at key /Table/66/2/"R"/3866 [r238] (manual) I190205 18:02:13.153221 28130 storage/replica_command.go:244 [n3,s3,r238/3:/Table/66/{2/"R"/3…-3/2872/…}] initiating a split of this range at key /Table/66/2/"V"/1348 [r239] (manual) I190205 18:02:13.163599 28084 storage/replica_command.go:244 [n3,s3,r238/3:/Table/66/{2/"R"/3…-3/2872/…}] initiating a split of this range at key /Table/66/3/1471/"P"/PrefixEnd [r240] (manual) I190205 18:02:13.252688 8835 server/status/runtime.go:464 [n1] runtime stats: 1.5 GiB RSS, 673 goroutines, 131 MiB/33 MiB/204 MiB GO alloc/idle/total, 197 MiB/245 MiB CGO alloc/total, 5826.6 CGO/sec, 136.2/14.2 %(u/s)time, 0.6 %gc (4x), 4.1 MiB/4.1 MiB (r/w)net I190205 18:02:13.597025 28084 storage/replica_command.go:244 [n3,s3,r239/3:/Table/66/{2/"V"/1…-3/2872/…}] initiating a split of this range at key /Table/66/3/1471/"P"/PrefixEnd [r241] (manual) I190205 18:02:13.683187 28222 storage/replica_command.go:244 [n3,s3,r239/3:/Table/66/{2/"V"/1…-3/2872/…}] initiating a split of this range at key /Table/66/2/"Y"/3795 [r242] (manual) I190205 18:02:13.951157 28084 storage/replica_command.go:244 [n3,s3,r241/3:/Table/66/3/{1471/"…-2872/"…}] initiating a split of this range at key /Table/66/3/2138/"G"/PrefixEnd [r243] (manual) I190205 18:02:13.966378 28222 storage/replica_command.go:244 [n3,s3,r239/3:/Table/66/{2/"V"/1…-3/1471/…}] initiating a split of this range at key /Table/66/2/"Y"/3795 [r244] (manual) I190205 18:02:14.398051 28390 storage/replica_command.go:244 [n3,s3,r244/3:/Table/66/{2/"Y"/3…-3/1471/…}] initiating a split of this range at key /Table/66/3/453/"L"/PrefixEnd [r245] (manual) I190205 18:02:14.461788 28084 storage/replica_command.go:244 [n3,s3,r243/3:/Table/66/3/2{138/"G…-872/"M…}] initiating a split of this range at key /Table/66/3/2206/"W" [r246] (manual) I190205 18:02:15.385777 28409 storage/replica_command.go:244 [n3,s3,r245/3:/Table/66/3/{453/"L…-1471/"…}] initiating a split of this range at key /Table/66/3/805/"Z" [r247] (manual); delayed split for 0.2s to avoid Raft snapshot I190205 18:02:15.537922 28385 storage/replica_command.go:244 [n3,s3,r236/3:/{Table/66/3/4…-Max}] initiating a split of this range at key /Table/67 [r248] (manual) I190205 18:02:15.643209 9255 server/status/runtime.go:464 [n2] runtime stats: 1.5 GiB RSS, 670 goroutines, 100 MiB/60 MiB/204 MiB GO alloc/idle/total, 196 MiB/245 MiB CGO alloc/total, 4912.6 CGO/sec, 133.0/13.9 %(u/s)time, 0.5 %gc (5x), 4.2 MiB/4.2 MiB (r/w)net I190205 18:02:16.456933 28569 storage/replica_command.go:244 [n3,s3,r236/3:/Table/6{6/3/4093…-7}] initiating a split of this range at key /Table/66/3/4431/"L" [r249] (manual) I190205 18:02:16.603408 9579 server/status/runtime.go:464 [n3] runtime stats: 1.5 GiB RSS, 670 goroutines, 144 MiB/24 MiB/204 MiB GO alloc/idle/total, 196 MiB/246 MiB CGO alloc/total, 4739.5 CGO/sec, 133.3/13.6 %(u/s)time, 0.5 %gc (5x), 4.1 MiB/4.1 MiB (r/w)net import_stmt_test.go:1180: job 7 did not match: Description: "IMPORT TABLE csv7.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///empty.csv', 'nodelocal:///csv/data-0', 'nodelocal:///csv/data-1', 'nodelocal:///csv/data-2', 'nodelocal:///csv/data-3', 'nodelocal:///csv/data-4')" != "CREATE STATISTICS __auto__ FROM [59] AS OF SYSTEM TIME '-30s'" TestImportCSVStmt/empty-file ...g replica r198/2 I190205 18:01:49.802452 8904 storage/store.go:2669 [n1,s1,r195/1:/Table/61/3/{607/"J…-774/"U"}] removing replica r198/1 I190205 18:01:49.815062 9500 storage/store.go:2669 [n3,s3,r195/3:/Table/61/3/{607/"J…-774/"U"}] removing replica r198/3 I190205 18:01:50.015819 25828 storage/replica_command.go:383 [n3,merge,s3,r179/3:/Table/61/1/{741-992}] initiating a merge of r184:/Table/61/1/{992-1707} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=0] into this range (lhs+rhs has (size=6.6 KiB+19 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=26 KiB, qps=0.00)) I190205 18:01:50.149924 25918 storage/replica_command.go:383 [n1,merge,s1,r201/1:/Table/61/3/3{145/"Z"-811/"P…}] initiating a merge of r109:/Table/6{1/3/3811/"P"/PrefixEnd-2} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=3] into this range (lhs+rhs has (size=18 KiB+32 KiB qps=0.00+6.78 --> 6.78qps) below threshold (size=51 KiB, qps=6.78)) I190205 18:01:50.388960 25875 ccl/importccl/read_import_proc.go:75 [n1,import-distsql] could not fetch file size; falling back to per-file progress: <nil> I190205 18:01:50.619253 8913 storage/store.go:2669 [n1,s1,r201/1:/Table/61/3/3{145/"Z"-811/"P…}] removing replica r109/1 I190205 18:01:50.625826 9499 storage/store.go:2669 [n3,s3,r201/3:/Table/61/3/3{145/"Z"-811/"P…}] removing replica r109/3 I190205 18:01:50.637114 9190 storage/store.go:2669 [n2,s2,r201/2:/Table/61/3/3{145/"Z"-811/"P…}] removing replica r109/2 I190205 18:01:50.787783 9525 storage/store.go:2669 [n3,s3,r179/3:/Table/61/1/{741-992}] removing replica r184/3 I190205 18:01:50.794165 8921 storage/store.go:2669 [n1,s1,r179/1:/Table/61/1/{741-992}] removing replica r184/1 I190205 18:01:50.826763 9231 storage/store.go:2669 [n2,s2,r179/2:/Table/61/1/{741-992}] removing replica r184/2 I190205 18:01:50.974613 26017 storage/replica_command.go:383 [n3,merge,s3,r186/3:/Table/61/2/"{I"/1438-K"/2792}] initiating a merge of r197:/Table/61/2/"{K"/2792-O"/249} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=12 KiB+18 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=30 KiB, qps=0.00)) I190205 18:01:51.083425 25998 storage/replica_command.go:383 [n2,merge,s2,r200/2:/Table/61/2/"{O"/249-Q"/3994}] initiating a merge of r161:/Table/61/2/"{Q"/3994-T"/2957} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=0] into this range (lhs+rhs has (size=14 KiB+15 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=29 KiB, qps=0.00)) I190205 18:01:51.515378 9503 storage/store.go:2669 [n3,s3,r186/3:/Table/61/2/"{I"/1438-K"/2792}] removing replica r197/3 I190205 18:01:51.523438 9226 storage/store.go:2669 [n2,s2,r186/2:/Table/61/2/"{I"/1438-K"/2792}] removing replica r197/2 I190205 18:01:51.550941 8912 storage/store.go:2669 [n1,s1,r186/1:/Table/61/2/"{I"/1438-K"/2792}] removing replica r197/1 I190205 18:01:51.626980 9535 storage/store.go:2669 [n3,s3,r200/3:/Table/61/2/"{O"/249-Q"/3994}] removing replica r161/3 I190205 18:01:51.630187 9205 storage/store.go:2669 [n2,s2,r200/2:/Table/61/2/"{O"/249-Q"/3994}] removing replica r161/2 I190205 18:01:51.637129 8722 storage/store.go:2669 [n1,s1,r200/1:/Table/61/2/"{O"/249-Q"/3994}] removing replica r161/1 I190205 18:01:51.709839 26101 storage/replica_command.go:383 [n1,merge,s1,r197/1:/Table/61/2/"{K"/2792-O"/249}] initiating a merge of r200:/Table/61/2/"{O"/249-T"/2957} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=2] into this range (lhs+rhs has (size=18 KiB+14 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=33 KiB, qps=0.00)) import_stmt_test.go:1180: job 6 did not match: Description: "IMPORT TABLE csv6.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///empty.csv')" != "IMPORT TABLE csv4.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///csv/data-0', 'nodelocal:///csv/data-1', 'nodelocal:///csv/data-2', 'nodelocal:///csv/data-3', 'nodelocal:///csv/data-4') WITH sstsize = '10K'" TestImportCSVStmt ...sMax-tsd} [(n1,s1):1, (n2,s2):2, next=3, gen=0] I190205 17:59:11.751245 8675 storage/replica_raft.go:372 [n1,s1,r3/1:/System/{NodeLive…-tsd}] proposing ADD_REPLICA((n3,s3):3): updated=[(n1,s1):1 (n2,s2):2 (n3,s3):3] next=4 I190205 17:59:11.774953 8675 storage/store_snapshot.go:762 [n1,replicate,s1,r15/1:/Table/{19-20}] sending preemptive snapshot e4be3f02 at applied index 16 I190205 17:59:11.775984 8675 storage/store_snapshot.go:805 [n1,replicate,s1,r15/1:/Table/{19-20}] streamed snapshot to (n2,s2):?: kv pairs: 7, log entries: 6, rate-limit: 8.0 MiB/sec, 0.01s I190205 17:59:11.777909 10086 storage/replica_raftstorage.go:805 [n2,s2,r15/?:{-}] applying preemptive snapshot at index 16 (id=e4be3f02, encoded size=1245, 1 rocksdb batches, 6 log entries) I190205 17:59:11.780196 10086 storage/replica_raftstorage.go:811 [n2,s2,r15/?:/Table/{19-20}] applied preemptive snapshot in 2ms [clear=0ms batch=0ms entries=1ms commit=0ms] I190205 17:59:11.783115 8675 storage/replica_command.go:798 [n1,replicate,s1,r15/1:/Table/{19-20}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r15:/Table/{19-20} [(n1,s1):1, (n3,s3):2, next=3, gen=0] I190205 17:59:11.880721 8675 storage/replica_raft.go:372 [n1,s1,r15/1:/Table/{19-20}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3] next=4 I190205 17:59:11.908664 8675 storage/store_snapshot.go:762 [n1,replicate,s1,r13/1:/Table/1{7-8}] sending preemptive snapshot 0d086ece at applied index 16 I190205 17:59:11.909887 8675 storage/store_snapshot.go:805 [n1,replicate,s1,r13/1:/Table/1{7-8}] streamed snapshot to (n3,s3):?: kv pairs: 7, log entries: 6, rate-limit: 8.0 MiB/sec, 0.01s I190205 17:59:11.912876 10105 storage/replica_raftstorage.go:805 [n3,s3,r13/?:{-}] applying preemptive snapshot at index 16 (id=0d086ece, encoded size=1245, 1 rocksdb batches, 6 log entries) I190205 17:59:11.916074 10105 storage/replica_raftstorage.go:811 [n3,s3,r13/?:/Table/1{7-8}] applied preemptive snapshot in 3ms [clear=0ms batch=0ms entries=2ms commit=0ms] I190205 17:59:11.918849 8675 storage/replica_command.go:798 [n1,replicate,s1,r13/1:/Table/1{7-8}] change replicas (ADD_REPLICA (n3,s3):3): read existing descriptor r13:/Table/1{7-8} [(n1,s1):1, (n2,s2):2, next=3, gen=0] I190205 17:59:11.971325 8675 storage/replica_raft.go:372 [n1,s1,r13/1:/Table/1{7-8}] proposing ADD_REPLICA((n3,s3):3): updated=[(n1,s1):1 (n2,s2):2 (n3,s3):3] next=4 I190205 17:59:12.034294 8675 storage/store_snapshot.go:762 [n1,replicate,s1,r9/1:/Table/1{3-4}] sending preemptive snapshot 23e47208 at applied index 93 I190205 17:59:12.057123 8675 storage/store_snapshot.go:805 [n1,replicate,s1,r9/1:/Table/1{3-4}] streamed snapshot to (n2,s2):?: kv pairs: 207, log entries: 83, rate-limit: 8.0 MiB/sec, 0.05s I190205 17:59:12.107757 10066 storage/replica_raftstorage.go:805 [n2,s2,r9/?:{-}] applying preemptive snapshot at index 93 (id=23e47208, encoded size=78231, 1 rocksdb batches, 83 log entries) I190205 17:59:12.226984 10066 storage/replica_raftstorage.go:811 [n2,s2,r9/?:/Table/1{3-4}] applied preemptive snapshot in 108ms [clear=0ms batch=0ms entries=106ms commit=1ms] I190205 17:59:12.244846 8675 storage/replica_command.go:798 [n1,replicate,s1,r9/1:/Table/1{3-4}] change replicas (ADD_REPLICA (n2,s2):3): read existing descriptor r9:/Table/1{3-4} [(n1,s1):1, (n3,s3):2, next=3, gen=0] I190205 17:59:12.344852 8675 storage/replica_raft.go:372 [n1,s1,r9/1:/Table/1{3-4}] proposing ADD_REPLICA((n2,s2):3): updated=[(n1,s1):1 (n3,s3):2 (n2,s2):3] next=4 I190205 17:59:12.666934 10046 sql/event_log.go:135 [n1,client=127.0.0.1:58322,user=root] Event: "set_cluster_setting", target: 0, info: {SettingName:kv.import.batch_size Value:10KB User:root} I190205 17:59:13.024772 8835 server/status/runtime.go:464 [n1] runtime stats: 945 MiB RSS, 627 goroutines, 108 MiB/61 MiB/203 MiB GO alloc/idle/total, 82 MiB/124 MiB CGO alloc/total, 0.0 CGO/sec, 0.0/0.0 %(u/s)time, 0.0 %gc (76x), 4.1 MiB/4.1 MiB (r/w)net TestImportCSVStmt/schema-in-file-sstsize ...action for range /Table/57/1/4612 - /Table/57/3/263/"D"/PrefixEnd that contains live data I190205 18:01:17.867897 9249 storage/compactor/compactor.go:329 [n2,s2,compactor] purging suggested compaction for range /Table/57/3/263/"D"/PrefixEnd - /Table/59 that contains live data I190205 18:01:17.870473 9249 storage/compactor/compactor.go:329 [n2,s2,compactor] purging suggested compaction for range /Table/57/3/1979/"D"/PrefixEnd - /Table/57/3/3313/"L"/PrefixEnd that contains live data I190205 18:01:17.870911 9249 storage/compactor/compactor.go:329 [n2,s2,compactor] purging suggested compaction for range /Table/57/3/3313/"L"/PrefixEnd - /Table/59 that contains live data I190205 18:01:18.732636 22570 storage/replica_command.go:244 [n2,s2,r162/2:/Table/61/3/{1440/"…-3145/"…}] initiating a split of this range at key /Table/61/3/2107/"B"/PrefixEnd [r163] (manual) I190205 18:01:18.795606 22599 storage/replica_command.go:244 [n3,s3,r201/3:/{Table/61/3/3…-Max}] initiating a split of this range at key /Table/61/3/4807/"X" [r203] (manual) I190205 18:01:19.607903 22661 storage/replica_command.go:244 [n2,s2,r163/2:/Table/61/3/{2107/"…-3145/"…}] initiating a split of this range at key /Table/61/3/2188/"E" [r164] (manual) I190205 18:01:20.593575 22790 storage/replica_command.go:244 [n3,s3,r199/3:/Table/61/{2/"W"/3…-3/607/"…}] initiating a split of this range at key /Table/61/2/"Z"/2833 [r204] (manual) I190205 18:01:21.023431 22810 storage/replica_command.go:244 [n1,s1,r201/1:/Table/61/3/{3145/"…-4807/"…}] initiating a split of this range at key /Table/61/3/3811/"P"/PrefixEnd [r109] (manual) I190205 18:01:21.374816 22810 storage/replica_command.go:244 [n1,s1,r109/1:/Table/61/3/{3811/"…-4807/"…}] initiating a split of this range at key /Table/61/3/3873/"Z" [r110] (manual) I190205 18:01:21.482793 22740 storage/replica_command.go:244 [n2,s2,r164/2:/Table/61/3/{2188/"…-3145/"…}] initiating a split of this range at key /Table/61/3/2752/"W" [r165] (manual) I190205 18:01:22.148452 22936 storage/replica_command.go:244 [n3,s3,r203/3:/{Table/61/3/4…-Max}] initiating a split of this range at key /Table/62 [r205] (manual) I190205 18:01:22.765115 23049 storage/replica_command.go:244 [n1,s1,r110/1:/Table/61/3/{3873/"…-4807/"…}] initiating a split of this range at key /Table/61/3/4385/"R" [r141] (manual) I190205 18:01:23.131342 23052 storage/replica_command.go:383 [n3,merge,s3,r174/3:/Table/{59/3/4733…-60}] initiating a merge of r175:/Table/6{0-1/1/741} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=2] into this range (lhs+rhs has (size=7.1 KiB+19 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=26 KiB, qps=0.00)) I190205 18:01:23.214735 8835 server/status/runtime.go:464 [n1] runtime stats: 1.4 GiB RSS, 678 goroutines, 119 MiB/45 MiB/204 MiB GO alloc/idle/total, 188 MiB/231 MiB CGO alloc/total, 4737.6 CGO/sec, 143.1/13.8 %(u/s)time, 0.8 %gc (5x), 3.2 MiB/3.2 MiB (r/w)net I190205 18:01:23.781051 8922 storage/store.go:2669 [n1,s1,r174/1:/Table/{59/3/4733…-60}] removing replica r175/1 I190205 18:01:23.786445 9199 storage/store.go:2669 [n2,s2,r174/2:/Table/{59/3/4733…-60}] removing replica r175/2 I190205 18:01:23.786540 9505 storage/store.go:2669 [n3,s3,r174/3:/Table/{59/3/4733…-60}] removing replica r175/3 I190205 18:01:24.219326 23158 storage/replica_command.go:383 [n3,merge,s3,r131/3:/Table/59/{1/4485-2/NULL/…}] initiating a merge of r134:/Table/59/2/NULL/{1089-4425} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=20 KiB+20 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=40 KiB, qps=0.00)) import_stmt_test.go:1180: job 4 did not match: Description: "IMPORT TABLE csv4.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///csv/data-0', 'nodelocal:///csv/data-1', 'nodelocal:///csv/data-2', 'nodelocal:///csv/data-3', 'nodelocal:///csv/data-4') WITH sstsize = '10K'" != "CREATE STATISTICS __auto__ FROM [55] AS OF SYSTEM TIME '-30s'" TestImportCSVStmt/schema-in-file-auto-gzip ... (manual); delayed split for 0.2s to avoid Raft snapshot I190205 18:04:12.747814 8951 storage/store.go:2669 [n1,s1,r337/1:/Table/72/3/{2280/"…-4948/"…}] removing replica r277/1 I190205 18:04:12.749147 9230 storage/store.go:2669 [n2,s2,r337/2:/Table/72/3/{2280/"…-4948/"…}] removing replica r277/2 I190205 18:04:12.752104 9526 storage/store.go:2669 [n3,s3,r337/3:/Table/72/3/{2280/"…-4948/"…}] removing replica r277/3 I190205 18:04:13.044405 39469 storage/replica_command.go:244 [n2,s2,r369/2:/Table/74/{2/"U"/7…-3/756/"…}] initiating a split of this range at key /Table/74/2/"X"/3170 [r370] (manual) I190205 18:04:13.369867 39469 storage/replica_command.go:244 [n2,s2,r370/2:/Table/74/{2/"X"/3…-3/756/"…}] initiating a split of this range at key /Table/74/3/76/"Y" [r371] (manual) I190205 18:04:13.409700 8835 server/status/runtime.go:464 [n1] runtime stats: 1.8 GiB RSS, 671 goroutines, 101 MiB/122 MiB/271 MiB GO alloc/idle/total, 277 MiB/320 MiB CGO alloc/total, 4790.3 CGO/sec, 140.2/11.3 %(u/s)time, 0.7 %gc (5x), 2.9 MiB/2.9 MiB (r/w)net I190205 18:04:13.581569 40251 storage/replica_command.go:383 [n1,merge,s1,r144/1:/Table/72/{1/741-2/"V"/4…}] initiating a merge of r320:/Table/72/2/"{V"/4337-Z"/1794} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=227 KiB+18 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=245 KiB, qps=0.00)) I190205 18:04:13.790872 8937 storage/store.go:2669 [n1,s1,r144/1:/Table/72/{1/741-2/"V"/4…}] removing replica r320/1 I190205 18:04:13.804087 9507 storage/store.go:2669 [n3,s3,r144/3:/Table/72/{1/741-2/"V"/4…}] removing replica r320/3 I190205 18:04:13.806617 9210 storage/store.go:2669 [n2,s2,r144/2:/Table/72/{1/741-2/"V"/4…}] removing replica r320/2 I190205 18:04:13.890425 40227 storage/replica_command.go:383 [n1,merge,s1,r144/1:/Table/72/{1/741-2/"Z"/1…}] initiating a merge of r331:/Table/72/{2/"Z"/1794-3/1614/"C"} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=3] into this range (lhs+rhs has (size=245 KiB+47 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=292 KiB, qps=0.00)) I190205 18:04:14.063040 9505 storage/store.go:2669 [n3,s3,r144/3:/Table/72/{1/741-2/"Z"/1…}] removing replica r331/3 I190205 18:04:14.065216 8928 storage/store.go:2669 [n1,s1,r144/1:/Table/72/{1/741-2/"Z"/1…}] removing replica r331/1 I190205 18:04:14.088413 9185 storage/store.go:2669 [n2,s2,r144/2:/Table/72/{1/741-2/"Z"/1…}] removing replica r331/2 I190205 18:04:14.239558 40271 storage/replica_command.go:383 [n1,merge,s1,r144/1:/Table/72/{1/741-3/1614/…}] initiating a merge of r340:/Table/72/3/{1614/"C"-2280/"S"/PrefixEnd} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=0] into this range (lhs+rhs has (size=292 KiB+18 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=311 KiB, qps=0.00)) I190205 18:04:14.584782 40379 storage/replica_command.go:383 [n3,merge,s3,r340/3:/Table/72/3/{1614/"…-2280/"…}] initiating a merge of r337:/Table/7{2/3/2280/"S"/PrefixEnd-4/1/741} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=5] into this range (lhs+rhs has (size=18 KiB+94 KiB qps=0.00+3.62 --> 3.62qps) below threshold (size=112 KiB, qps=3.62)) I190205 18:04:14.590946 8941 storage/store.go:2669 [n1,s1,r144/1:/Table/72/{1/741-3/1614/…}] removing replica r340/1 I190205 18:04:14.604285 9206 storage/store.go:2669 [n2,s2,r144/2:/Table/72/{1/741-3/1614/…}] removing replica r340/2 import_stmt_test.go:1180: job 11 did not match: Description: "IMPORT TABLE csv11.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:////csv/data-0.gz', 'nodelocal:////csv/data-1.gz', 'nodelocal:////csv/data-2.gz', 'nodelocal:////csv/data-3.gz', 'nodelocal:////csv/data-4.gz') WITH decompress = 'auto'" != "IMPORT TABLE csv7.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///empty.csv', 'nodelocal:///csv/data-0', 'nodelocal:///csv/data-1', 'nodelocal:///csv/data-2', 'nodelocal:///csv/data-3', 'nodelocal:///csv/data-4')" TestImportCSVStmt/schema-in-file-no-decompress ...:/Table/68/1/{741-1456}] removing replica r170/3 I190205 18:03:09.588628 8910 storage/store.go:2669 [n1,s1,r168/1:/Table/68/1/{741-1456}] removing replica r170/1 I190205 18:03:09.728167 33570 storage/replica_command.go:383 [n2,merge,s2,r168/2:/Table/68/1/{741-1689}] initiating a merge of r212:/Table/68/1/{1689-2404} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=0] into this range (lhs+rhs has (size=25 KiB+19 KiB qps=3.99+0.00 --> 3.99qps) below threshold (size=44 KiB, qps=3.99)) I190205 18:03:09.735905 33683 storage/replica_command.go:244 [n3,s3,r301/3:/Table/70/{2/"O"/4…-3/360/"…}] initiating a split of this range at key /Table/70/2/"R"/2852 [r302] (manual) I190205 18:03:10.116941 9509 storage/store.go:2669 [n3,s3,r168/3:/Table/68/1/{741-1689}] removing replica r212/3 I190205 18:03:10.128932 9223 storage/store.go:2669 [n2,s2,r168/2:/Table/68/1/{741-1689}] removing replica r212/2 I190205 18:03:10.142650 8909 storage/store.go:2669 [n1,s1,r168/1:/Table/68/1/{741-1689}] removing replica r212/1 I190205 18:03:10.233154 33783 storage/replica_command.go:244 [n3,s3,r302/3:/Table/70/{2/"R"/2…-3/360/"…}] initiating a split of this range at key /Table/70/2/"V"/334 [r303] (manual) I190205 18:03:10.538610 33769 storage/replica_command.go:244 [n3,s3,r303/3:/Table/70/{2/"V"/3…-3/360/"…}] initiating a split of this range at key /Table/70/2/"Y"/1402 [r304] (manual) I190205 18:03:10.613954 33746 storage/replica_command.go:383 [n2,merge,s2,r168/2:/Table/68/1/{741-2404}] initiating a merge of r166:/Table/68/{1/2404-2/"B"/3824} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=5] into this range (lhs+rhs has (size=44 KiB+78 KiB qps=4.84+6.21 --> 11.05qps) below threshold (size=122 KiB, qps=11.05)) I190205 18:03:10.971673 9225 storage/store.go:2669 [n2,s2,r168/2:/Table/68/1/{741-2404}] removing replica r166/2 I190205 18:03:11.007038 8717 storage/store.go:2669 [n1,s1,r168/1:/Table/68/1/{741-2404}] removing replica r166/1 I190205 18:03:11.058169 9529 storage/store.go:2669 [n3,s3,r168/3:/Table/68/1/{741-2404}] removing replica r166/3 I190205 18:03:11.330578 33848 storage/replica_command.go:244 [n3,s3,r294/3:/{Table/70/3/4…-Max}] initiating a split of this range at key /Table/71 [r305] (manual) I190205 18:03:11.730055 33855 storage/replica_command.go:383 [n2,merge,s2,r168/2:/Table/68/{1/741-2/"B"/3…}] initiating a merge of r254:/Table/68/2/"{B"/3824-E"/888} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=122 KiB+13 KiB qps=7.99+0.00 --> 7.99qps) below threshold (size=134 KiB, qps=7.99)) I190205 18:03:12.299917 9200 storage/store.go:2669 [n2,s2,r168/2:/Table/68/{1/741-2/"B"/3…}] removing replica r254/2 I190205 18:03:12.314674 9513 storage/store.go:2669 [n3,s3,r168/3:/Table/68/{1/741-2/"B"/3…}] removing replica r254/3 I190205 18:03:12.350290 8929 storage/store.go:2669 [n1,s1,r168/1:/Table/68/{1/741-2/"B"/3…}] removing replica r254/1 I190205 18:03:12.385265 33899 storage/replica_command.go:383 [n1,merge,s1,r254/1:/Table/68/2/"{B"/3824-E"/888}] initiating a merge of r255:/Table/68/2/"{E"/888-H"/3232} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=3] into this range (lhs+rhs has (size=13 KiB+18 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=31 KiB, qps=0.00)) I190205 18:03:12.477914 33965 storage/replica_command.go:383 [n2,merge,s2,r168/2:/Table/68/{1/741-2/"E"/8…}] initiating a merge of r255:/Table/68/2/"{E"/888-H"/3232} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=3] into this range (lhs+rhs has (size=134 KiB+18 KiB qps=4.59+0.00 --> 4.59qps) below threshold (size=152 KiB, qps=4.59)) import_stmt_test.go:1180: job 9 did not match: Description: "IMPORT TABLE csv9.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///csv/data-0', 'nodelocal:///csv/data-1', 'nodelocal:///csv/data-2', 'nodelocal:///csv/data-3', 'nodelocal:///csv/data-4') WITH decompress = 'none'" != "CREATE STATISTICS __auto__ FROM [57] AS OF SYSTEM TIME '-30s'" TestImportCSVStmt/schema-in-query-transform-only ... 9192 storage/store.go:2669 [n2,s2,r100/2:/Table/59{-/2/"R"/4…}] removing replica r152/2 I190205 18:01:43.938554 9528 storage/store.go:2669 [n3,s3,r100/3:/Table/59{-/2/"R"/4…}] removing replica r152/3 I190205 18:01:43.949112 8936 storage/store.go:2669 [n1,s1,r100/1:/Table/59{-/2/"R"/4…}] removing replica r152/1 I190205 18:01:44.442921 9535 storage/store.go:2669 [n3,s3,r185/3:/Table/61/1/{2267-4455}] removing replica r191/3 I190205 18:01:44.462897 8914 storage/store.go:2669 [n1,s1,r185/1:/Table/61/1/{2267-4455}] removing replica r191/1 I190205 18:01:44.527437 9219 storage/store.go:2669 [n2,s2,r185/2:/Table/61/1/{2267-4455}] removing replica r191/2 I190205 18:01:44.648428 25343 storage/replica_command.go:383 [n3,merge,s3,r182/3:/Table/61/1/{1707-2267}] initiating a merge of r185:/Table/61/{1/2267-2/"A"/3432} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=5] into this range (lhs+rhs has (size=15 KiB+76 KiB qps=1.77+5.59 --> 7.37qps) below threshold (size=90 KiB, qps=7.37)) I190205 18:01:45.018810 9503 storage/store.go:2669 [n3,s3,r182/3:/Table/61/1/{1707-2267}] removing replica r185/3 I190205 18:01:45.029332 9209 storage/store.go:2669 [n2,s2,r182/2:/Table/61/1/{1707-2267}] removing replica r185/2 I190205 18:01:45.033788 8903 storage/store.go:2669 [n1,s1,r182/1:/Table/61/1/{1707-2267}] removing replica r185/1 I190205 18:01:45.527526 25347 storage/replica_command.go:383 [n3,merge,s3,r182/3:/Table/61/{1/1707-2/"A"/3…}] initiating a merge of r192:/Table/61/2/"{A"/3432-E"/785} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=90 KiB+18 KiB qps=5.99+0.00 --> 5.99qps) below threshold (size=109 KiB, qps=5.99)) I190205 18:01:45.626120 9255 server/status/runtime.go:464 [n2] runtime stats: 1.5 GiB RSS, 674 goroutines, 106 MiB/56 MiB/204 MiB GO alloc/idle/total, 189 MiB/233 MiB CGO alloc/total, 5393.0 CGO/sec, 145.7/11.6 %(u/s)time, 1.4 %gc (5x), 2.7 MiB/2.7 MiB (r/w)net I190205 18:01:46.164665 9500 storage/store.go:2669 [n3,s3,r182/3:/Table/61/{1/1707-2/"A"/3…}] removing replica r192/3 I190205 18:01:46.186703 8939 storage/store.go:2669 [n1,s1,r182/1:/Table/61/{1/1707-2/"A"/3…}] removing replica r192/1 I190205 18:01:46.247077 9178 storage/store.go:2669 [n2,s2,r182/2:/Table/61/{1/1707-2/"A"/3…}] removing replica r192/2 I190205 18:01:46.291058 25558 storage/replica_command.go:383 [n3,merge,s3,r182/3:/Table/61/{1/1707-2/"E"/7…}] initiating a merge of r194:/Table/61/2/"{E"/785-F"/3489} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=109 KiB+8.1 KiB qps=5.99+0.00 --> 5.99qps) below threshold (size=117 KiB, qps=5.99)) I190205 18:01:46.529935 9579 server/status/runtime.go:464 [n3] runtime stats: 1.5 GiB RSS, 663 goroutines, 127 MiB/42 MiB/204 MiB GO alloc/idle/total, 180 MiB/224 MiB CGO alloc/total, 9305.5 CGO/sec, 148.9/13.8 %(u/s)time, 1.6 %gc (6x), 2.8 MiB/2.8 MiB (r/w)net I190205 18:01:46.537164 25578 storage/replica_command.go:383 [n1,merge,s1,r109/1:/Table/61/3/38{11/"P…-73/"Z"}] initiating a merge of r110:/Table/61/3/{3873/"Z"-4385/"R"} [(n1,s1):1, (n2,s2):2, (n3,s3):3, next=4, gen=1] into this range (lhs+rhs has (size=1.7 KiB+14 KiB qps=0.00+0.00 --> 0.00qps) below threshold (size=16 KiB, qps=0.00)) import_stmt_test.go:1180: job 5 did not match: Description: "IMPORT TABLE \"\".\"\".t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///csv/data-0-opts', 'nodelocal:///csv/data-1-opts', 'nodelocal:///csv/data-2-opts', 'nodelocal:///csv/data-3-opts', 'nodelocal:///csv/data-4-opts') WITH comment = '#', delimiter = '|', \"nullif\" = '', skip = '2', transform = 'nodelocal:///5'" != "IMPORT TABLE csv3.public.t (a INT8 PRIMARY KEY, b STRING, INDEX (b), INDEX (a, b)) CSV DATA ('nodelocal:///csv/data-0-opts', 'nodelocal:///csv/data-1-opts', 'nodelocal:///csv/data-2-opts', 'nodelocal:///csv/data-3-opts', 'nodelocal:///csv/data-4-opts') WITH comment = '#', delimiter = '|', \"nullif\" = '', skip = '2'" ``` Please assign, take a look and update the issue accordingly.
test
teamcity failed test testimportcsvstmt the following tests appear to have failed on master testrace testimportcsvstmt schema in file implicit gzip testimportcsvstmt schema in file no decompress testimportcsvstmt empty file testimportcsvstmt schema in query opts testimportcsvstmt schema in file sstsize testimportcsvstmt schema in file auto decompress testimportcsvstmt schema in file auto gzip testimportcsvstmt testimportcsvstmt schema in query transform only testimportcsvstmt schema in file explicit gzip testimportcsvstmt empty with files you may want to check testimportcsvstmt schema in file implicit gzip y… y initiating a split of this range at key table m manual storage store go removing replica storage store go removing replica storage store go removing replica storage store go removing replica storage store go removing replica storage store go removing replica storage replica command go initiating a merge of table c g into this range lhs rhs has size kib kib qps below threshold size kib qps storage replica command go initiating a merge of table into this range lhs rhs has size kib kib qps below threshold size kib qps sql event log go event create statistics target info statisticname auto statement create statistics auto from as of system time storage store go removing replica storage store go removing replica storage store go removing replica storage replica command go initiating a merge of table u x into this range lhs rhs has size kib kib qps below threshold size kib qps storage replica command go initiating a merge of table c g into this range lhs rhs has size kib kib qps below threshold size kib qps storage store go removing replica storage store go removing replica storage store go removing replica storage store go removing replica storage store go removing replica storage store go removing replica import stmt test go job did not match description import table public t a primary key b string index b index a b csv data nodelocal csv data gz nodelocal csv data gz nodelocal csv data gz nodelocal csv data gz nodelocal csv data gz import table public t a primary key b string index b index a b csv data nodelocal csv data nodelocal csv data nodelocal csv data nodelocal csv data nodelocal csv data with decompress auto testimportcsvstmt schema in file auto decompress storage replica command go initiating a split of this range at key table i prefixend manual storage replica command go initiating a merge of table f h into this range lhs rhs has size kib kib qps below threshold size kib qps storage replica command go initiating a merge of table f h into this range lhs rhs has size kib kib qps below threshold size kib qps server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total cgo sec u s time gc mib mib r w net storage replica command go initiating a split of this range at key table l manual storage store go removing replica storage store go removing replica server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total cgo sec u s time gc mib mib r w net storage store go removing replica storage replica command go initiating a merge of table h k into this range lhs rhs has size kib kib qps below threshold size kib qps storage replica command go initiating a merge of table k o into this range lhs rhs has size kib kib qps below threshold size kib qps storage store go removing replica storage store go removing replica storage store go removing replica storage replica command go initiating a merge of table into this range lhs rhs has size kib kib qps below threshold size kib qps import stmt test go job did not match description import table public t a primary key b string index b index a b csv data nodelocal csv data nodelocal csv data nodelocal csv data nodelocal csv data nodelocal csv data with decompress auto import table t a primary key b string index b index a b csv data nodelocal csv data opts nodelocal csv data opts nodelocal csv data opts nodelocal csv data opts nodelocal csv data opts with comment delimiter nullif skip transform nodelocal testimportcsvstmt schema in query opts next gen into this range lhs rhs has size kib kib qps below threshold size kib qps storage replica command go initiating a split of this range at key table b prefixend manual storage store go removing replica storage store go removing replica storage store go removing replica storage replica command go initiating a merge of table into this range lhs rhs has size kib kib qps below threshold size kib qps storage replica command go initiating a split of this range at key table b prefixend manual storage store go removing replica storage store go removing replica storage store go removing replica server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total cgo sec u s time gc mib mib r w net storage replica command go initiating a merge of table into this range lhs rhs has size kib kib qps below threshold size kib qps storage replica command go initiating a split of this range at key table b prefixend manual delayed split for to avoid raft snapshot storage store go removing replica storage store go removing replica storage store go removing replica storage replica command go initiating a split of this range at key table manual storage replica command go initiating a merge of table into this range lhs rhs has size kib kib qps below threshold size kib qps storage store go removing replica storage store go removing replica storage store go removing replica import stmt test go job did not match description import table public t a primary key b string index b index a b csv data nodelocal csv data opts nodelocal csv data opts nodelocal csv data opts nodelocal csv data opts nodelocal csv data opts with comment delimiter nullif skip create statistics auto from as of system time testimportcsvstmt schema in file explicit gzip ib cgo alloc total cgo sec u s time gc mib mib r w net storage replica command go initiating a split of this range at key table a prefixend manual storage compactor compactor go purging suggested compaction for range table h table l that contains live data storage compactor compactor go purging suggested compaction for range table l table s that contains live data storage compactor compactor go purging suggested compaction for range table s table z that contains live data storage compactor compactor go purging suggested compaction for range table v table z that contains live data storage compactor compactor go purging suggested compaction for range table a prefixend table r prefixend that contains live data storage compactor compactor go purging suggested compaction for range table r prefixend table that contains live data storage compactor compactor go purging suggested compaction for range table t prefixend table that contains live data storage compactor compactor go purging suggested compaction for range table h table l that contains live data storage compactor compactor go purging suggested compaction for range table l table s that contains live data storage compactor compactor go purging suggested compaction for range table s table z that contains live data storage compactor compactor go purging suggested compaction for range table v table z that contains live data storage compactor compactor go purging suggested compaction for range table a prefixend table r prefixend that contains live data storage compactor compactor go purging suggested compaction for range table r prefixend table that contains live data storage compactor compactor go purging suggested compaction for range table t prefixend table that contains live data storage replica command go initiating a split of this range at key table r prefixend manual storage replica command go initiating a split of this range at key table i prefixend manual storage replica command go initiating a split of this range at key table manual import stmt test go job did not match description import table public t a primary key b string index b index a b csv data nodelocal csv data gz nodelocal csv data gz nodelocal csv data gz nodelocal csv data gz nodelocal csv data gz with decompress gzip import table public t a primary key b string index b index a b csv data nodelocal empty csv testimportcsvstmt empty with files storage replica command go initiating a split of this range at key table o manual storage replica command go initiating a split of this range at key table l manual storage replica command go initiating a split of this range at key table p prefixend manual storage replica command go initiating a split of this range at key table r manual storage replica command go initiating a split of this range at key table v manual storage replica command go initiating a split of this range at key table p prefixend manual server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total cgo sec u s time gc mib mib r w net storage replica command go initiating a split of this range at key table p prefixend manual storage replica command go initiating a split of this range at key table y manual storage replica command go initiating a split of this range at key table g prefixend manual storage replica command go initiating a split of this range at key table y manual storage replica command go initiating a split of this range at key table l prefixend manual storage replica command go initiating a split of this range at key table w manual storage replica command go initiating a split of this range at key table z manual delayed split for to avoid raft snapshot storage replica command go initiating a split of this range at key table manual server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total cgo sec u s time gc mib mib r w net storage replica command go initiating a split of this range at key table l manual server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total cgo sec u s time gc mib mib r w net import stmt test go job did not match description import table public t a primary key b string index b index a b csv data nodelocal empty csv nodelocal csv data nodelocal csv data nodelocal csv data nodelocal csv data nodelocal csv data create statistics auto from as of system time testimportcsvstmt empty file g replica storage store go removing replica storage store go removing replica storage replica command go initiating a merge of table into this range lhs rhs has size kib kib qps below threshold size kib qps storage replica command go initiating a merge of table p prefixend into this range lhs rhs has size kib kib qps below threshold size kib qps ccl importccl read import proc go could not fetch file size falling back to per file progress storage store go removing replica storage store go removing replica storage store go removing replica storage store go removing replica storage store go removing replica storage store go removing replica storage replica command go initiating a merge of table k o into this range lhs rhs has size kib kib qps below threshold size kib qps storage replica command go initiating a merge of table q t into this range lhs rhs has size kib kib qps below threshold size kib qps storage store go removing replica storage store go removing replica storage store go removing replica storage store go removing replica storage store go removing replica storage store go removing replica storage replica command go initiating a merge of table o t into this range lhs rhs has size kib kib qps below threshold size kib qps import stmt test go job did not match description import table public t a primary key b string index b index a b csv data nodelocal empty csv import table public t a primary key b string index b index a b csv data nodelocal csv data nodelocal csv data nodelocal csv data nodelocal csv data nodelocal csv data with sstsize testimportcsvstmt smax tsd storage replica raft go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica raft go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica raft go proposing add replica updated next storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table storage replica raft go proposing add replica updated next sql event log go event set cluster setting target info settingname kv import batch size value user root server status runtime go runtime stats mib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total cgo sec u s time gc mib mib r w net testimportcsvstmt schema in file sstsize action for range table table d prefixend that contains live data storage compactor compactor go purging suggested compaction for range table d prefixend table that contains live data storage compactor compactor go purging suggested compaction for range table d prefixend table l prefixend that contains live data storage compactor compactor go purging suggested compaction for range table l prefixend table that contains live data storage replica command go initiating a split of this range at key table b prefixend manual storage replica command go initiating a split of this range at key table x manual storage replica command go initiating a split of this range at key table e manual storage replica command go initiating a split of this range at key table z manual storage replica command go initiating a split of this range at key table p prefixend manual storage replica command go initiating a split of this range at key table z manual storage replica command go initiating a split of this range at key table w manual storage replica command go initiating a split of this range at key table manual storage replica command go initiating a split of this range at key table r manual storage replica command go initiating a merge of table into this range lhs rhs has size kib kib qps below threshold size kib qps server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total cgo sec u s time gc mib mib r w net storage store go removing replica storage store go removing replica storage store go removing replica storage replica command go initiating a merge of table null into this range lhs rhs has size kib kib qps below threshold size kib qps import stmt test go job did not match description import table public t a primary key b string index b index a b csv data nodelocal csv data nodelocal csv data nodelocal csv data nodelocal csv data nodelocal csv data with sstsize create statistics auto from as of system time testimportcsvstmt schema in file auto gzip manual delayed split for to avoid raft snapshot storage store go removing replica storage store go removing replica storage store go removing replica storage replica command go initiating a split of this range at key table x manual storage replica command go initiating a split of this range at key table y manual server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total cgo sec u s time gc mib mib r w net storage replica command go initiating a merge of table v z into this range lhs rhs has size kib kib qps below threshold size kib qps storage store go removing replica storage store go removing replica storage store go removing replica storage replica command go initiating a merge of table z c into this range lhs rhs has size kib kib qps below threshold size kib qps storage store go removing replica storage store go removing replica storage store go removing replica storage replica command go initiating a merge of table c s prefixend into this range lhs rhs has size kib kib qps below threshold size kib qps storage replica command go initiating a merge of table s prefixend into this range lhs rhs has size kib kib qps below threshold size kib qps storage store go removing replica storage store go removing replica import stmt test go job did not match description import table public t a primary key b string index b index a b csv data nodelocal csv data gz nodelocal csv data gz nodelocal csv data gz nodelocal csv data gz nodelocal csv data gz with decompress auto import table public t a primary key b string index b index a b csv data nodelocal empty csv nodelocal csv data nodelocal csv data nodelocal csv data nodelocal csv data nodelocal csv data testimportcsvstmt schema in file no decompress table removing replica storage store go removing replica storage replica command go initiating a merge of table into this range lhs rhs has size kib kib qps below threshold size kib qps storage replica command go initiating a split of this range at key table r manual storage store go removing replica storage store go removing replica storage store go removing replica storage replica command go initiating a split of this range at key table v manual storage replica command go initiating a split of this range at key table y manual storage replica command go initiating a merge of table b into this range lhs rhs has size kib kib qps below threshold size kib qps storage store go removing replica storage store go removing replica storage store go removing replica storage replica command go initiating a split of this range at key table manual storage replica command go initiating a merge of table b e into this range lhs rhs has size kib kib qps below threshold size kib qps storage store go removing replica storage store go removing replica storage store go removing replica storage replica command go initiating a merge of table e h into this range lhs rhs has size kib kib qps below threshold size kib qps storage replica command go initiating a merge of table e h into this range lhs rhs has size kib kib qps below threshold size kib qps import stmt test go job did not match description import table public t a primary key b string index b index a b csv data nodelocal csv data nodelocal csv data nodelocal csv data nodelocal csv data nodelocal csv data with decompress none create statistics auto from as of system time testimportcsvstmt schema in query transform only storage store go removing replica storage store go removing replica storage store go removing replica storage store go removing replica storage store go removing replica storage store go removing replica storage replica command go initiating a merge of table a into this range lhs rhs has size kib kib qps below threshold size kib qps storage store go removing replica storage store go removing replica storage store go removing replica storage replica command go initiating a merge of table a e into this range lhs rhs has size kib kib qps below threshold size kib qps server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total cgo sec u s time gc mib mib r w net storage store go removing replica storage store go removing replica storage store go removing replica storage replica command go initiating a merge of table e f into this range lhs rhs has size kib kib qps below threshold size kib qps server status runtime go runtime stats gib rss goroutines mib mib mib go alloc idle total mib mib cgo alloc total cgo sec u s time gc mib mib r w net storage replica command go initiating a merge of table z r into this range lhs rhs has size kib kib qps below threshold size kib qps import stmt test go job did not match description import table t a primary key b string index b index a b csv data nodelocal csv data opts nodelocal csv data opts nodelocal csv data opts nodelocal csv data opts nodelocal csv data opts with comment delimiter nullif skip transform nodelocal import table public t a primary key b string index b index a b csv data nodelocal csv data opts nodelocal csv data opts nodelocal csv data opts nodelocal csv data opts nodelocal csv data opts with comment delimiter nullif skip please assign take a look and update the issue accordingly
1
7,280
5,959,324,176
IssuesEvent
2017-05-29 10:39:53
swagger-api/swagger-codegen
https://api.github.com/repos/swagger-api/swagger-codegen
closed
[Ruby] Lowering memory consumption when writing large file
Client: Ruby Enhancement: Performance
##### Description Currently the generated Ruby client buffers the response body before writing the file in one go. I tested this by using the Ruby client to download a file where response body's size is ~2Gb, and I monitored the ruby process spiked to ~2Gb memory consumption. This obviously won't scale well with larger files, and is a problem when the machine doesn't have large memory capacity, plus there's going to be a limit of how large a string can be, and how much memory the ruby process can use. Is there any particular reason why the whole response body for a file type has to be received as a whole, deserialised, and then written to a file in one go? If the approach further below in 'Suggests a Fix' section looks reasonable, I'm happy to take a stab at it. ##### Swagger-codegen version master branch. ##### Swagger declaration file content or url Sample spec segment: ```yaml /some/path/to/endpoint: get: operationId: getFile produces: - application/octet-stream parameters: responses: default: schema: type: file ``` ##### Command line used for generation <!-- including the language, libraries and various options --> ##### Steps to reproduce 1. Create an endpoint that returns a large file as part of its response body. 2. Create spec with response schema `type: File` . 3. Generate Ruby client. 4. Call the API and monitor the process' memory usage. ##### Related issues Couldn't find anything obvious. ##### Suggest a Fix When Typhoeus::Request is created https://github.com/swagger-api/swagger-codegen/blob/35d3fb82ce82e86370ca24c21b0c066e8aea2f90/modules/swagger-codegen/src/main/resources/ruby/api_client.mustache#L120 , it should be passed `on_headers`, `on_body`, and `on_complete` callbacks if `return_type` is `File`.
True
[Ruby] Lowering memory consumption when writing large file - ##### Description Currently the generated Ruby client buffers the response body before writing the file in one go. I tested this by using the Ruby client to download a file where response body's size is ~2Gb, and I monitored the ruby process spiked to ~2Gb memory consumption. This obviously won't scale well with larger files, and is a problem when the machine doesn't have large memory capacity, plus there's going to be a limit of how large a string can be, and how much memory the ruby process can use. Is there any particular reason why the whole response body for a file type has to be received as a whole, deserialised, and then written to a file in one go? If the approach further below in 'Suggests a Fix' section looks reasonable, I'm happy to take a stab at it. ##### Swagger-codegen version master branch. ##### Swagger declaration file content or url Sample spec segment: ```yaml /some/path/to/endpoint: get: operationId: getFile produces: - application/octet-stream parameters: responses: default: schema: type: file ``` ##### Command line used for generation <!-- including the language, libraries and various options --> ##### Steps to reproduce 1. Create an endpoint that returns a large file as part of its response body. 2. Create spec with response schema `type: File` . 3. Generate Ruby client. 4. Call the API and monitor the process' memory usage. ##### Related issues Couldn't find anything obvious. ##### Suggest a Fix When Typhoeus::Request is created https://github.com/swagger-api/swagger-codegen/blob/35d3fb82ce82e86370ca24c21b0c066e8aea2f90/modules/swagger-codegen/src/main/resources/ruby/api_client.mustache#L120 , it should be passed `on_headers`, `on_body`, and `on_complete` callbacks if `return_type` is `File`.
non_test
lowering memory consumption when writing large file description currently the generated ruby client buffers the response body before writing the file in one go i tested this by using the ruby client to download a file where response body s size is and i monitored the ruby process spiked to memory consumption this obviously won t scale well with larger files and is a problem when the machine doesn t have large memory capacity plus there s going to be a limit of how large a string can be and how much memory the ruby process can use is there any particular reason why the whole response body for a file type has to be received as a whole deserialised and then written to a file in one go if the approach further below in suggests a fix section looks reasonable i m happy to take a stab at it swagger codegen version master branch swagger declaration file content or url sample spec segment yaml some path to endpoint get operationid getfile produces application octet stream parameters responses default schema type file command line used for generation steps to reproduce create an endpoint that returns a large file as part of its response body create spec with response schema type file generate ruby client call the api and monitor the process memory usage related issues couldn t find anything obvious suggest a fix when typhoeus request is created it should be passed on headers on body and on complete callbacks if return type is file
0
140,007
11,301,406,581
IssuesEvent
2020-01-17 15:31:57
stevenschader/kabanero-foundation
https://api.github.com/repos/stevenschader/kabanero-foundation
closed
SVT: TER: Kabanero Automation test execution master branch - Setup: svtcrc-786520-1.fyre.ibm.com
SVT Kabanero Test Execution Test Execution Record
"msg": "******************** crc_fyresetup_start ********************" } TASK [load_secrets : Load all secrets] **************************************************************************************************************** Thursday 16 January 2020 07:40:35 -0500 (0:00:00.034) 0:00:00.347 ****** fatal: [localhost]: FAILED! => {"msg": "Syntax Error while loading YAML.\n found unexpected end of stream\n\nThe error appears to be in '/etc/ansible/svtinfo/svtvars.yml': line 22, column 1, but may\nbe elsewhere in the file depending on the exact syntax problem.\n\n(specified line no longer in file, maybe it changed?)"} PLAY RECAP ******************************************************************************************************************************************** localhost : ok=2 changed=0 unreachable=0 failed=1 skipped=0 rescued=0 ignored=0 Thursday 16 January 2020 07:40:35 -0500 (0:00:00.035) 0:00:00.383 ****** =============================================================================== log : include_tasks ---------------------------------------------------------------------------------------------------------------------------- 0.06s load_secrets : Load all secrets ---------------------------------------------------------------------------------------------------------------- 0.04s log : debug ------------------------------------------------------------------------------------------------------------------------------------ 0.03s
2.0
SVT: TER: Kabanero Automation test execution master branch - Setup: svtcrc-786520-1.fyre.ibm.com - "msg": "******************** crc_fyresetup_start ********************" } TASK [load_secrets : Load all secrets] **************************************************************************************************************** Thursday 16 January 2020 07:40:35 -0500 (0:00:00.034) 0:00:00.347 ****** fatal: [localhost]: FAILED! => {"msg": "Syntax Error while loading YAML.\n found unexpected end of stream\n\nThe error appears to be in '/etc/ansible/svtinfo/svtvars.yml': line 22, column 1, but may\nbe elsewhere in the file depending on the exact syntax problem.\n\n(specified line no longer in file, maybe it changed?)"} PLAY RECAP ******************************************************************************************************************************************** localhost : ok=2 changed=0 unreachable=0 failed=1 skipped=0 rescued=0 ignored=0 Thursday 16 January 2020 07:40:35 -0500 (0:00:00.035) 0:00:00.383 ****** =============================================================================== log : include_tasks ---------------------------------------------------------------------------------------------------------------------------- 0.06s load_secrets : Load all secrets ---------------------------------------------------------------------------------------------------------------- 0.04s log : debug ------------------------------------------------------------------------------------------------------------------------------------ 0.03s
test
svt ter kabanero automation test execution master branch setup svtcrc fyre ibm com msg crc fyresetup start task thursday january fatal failed msg syntax error while loading yaml n found unexpected end of stream n nthe error appears to be in etc ansible svtinfo svtvars yml line column but may nbe elsewhere in the file depending on the exact syntax problem n n specified line no longer in file maybe it changed play recap localhost ok changed unreachable failed skipped rescued ignored thursday january log include tasks load secrets load all secrets log debug
1
156,306
5,967,045,331
IssuesEvent
2017-05-30 15:12:19
openshift/origin
https://api.github.com/repos/openshift/origin
closed
API server panic due to garbage collection requiring delete strategy
component/restapi kind/bug priority/P1
Edits by @enj 1. The following resources in origin have a `nil` `store.DeleteStrategy`: ``` clusterpolicies.authorization.openshift.io clusterpolicybindings.authorization.openshift.io policies.authorization.openshift.io policybindings.authorization.openshift.io images.image.openshift.io imagestreams.image.openshift.io clusternetworks.network.openshift.io egressnetworkpolicies.network.openshift.io hostsubnets.network.openshift.io netnamespaces.network.openshift.io oauthaccesstokens.oauth.openshift.io oauthauthorizetokens.oauth.openshift.io oauthclientauthorizations.oauth.openshift.io oauthclients.oauth.openshift.io routes.route.openshift.io templates.template.openshift.io groups.user.openshift.io identities.user.openshift.io users.user.openshift.io ``` 2. `store.EnableGarbageCollection` is `true` for all resources now. 1 and 2 combined with any of the following `DeleteOptions`: - `OrphanDependents` = `true` - `PropagationPolicy` = `DeletePropagationOrphan` - `PropagationPolicy` = `DeletePropagationForeground` Lead to a `panic` in the API server because it tries to a do a graceful deletion and finalization which dereferences `DeleteStrategy`. We have some protection from https://github.com/openshift/origin/commit/54d84e6a8db4c07f78fb2823508fed7751ebf1bd but we need to reconsider. Upstream does not carry this patch at all. --- I'm seeing a panic trying to delete a route in the web console. We're passing `propagationPolicy: "Foreground"` (which we've switched to passing for all requests). ``` $ oc version oc v3.6.0-alpha.1+b581bfe-566 kubernetes v1.6.1+5115d708d7 features: Basic-Auth Server https://127.0.0.1:8443 openshift v3.6.0-alpha.1+b581bf ``` HTTP Request: ``` DELETE /oapi/v1/namespaces/node/routes/asdfasdf Content-Type: application/json ... {"kind":"DeleteOptions","apiVersion":"v1","propagationPolicy":"Foreground"} ``` Response: ``` Status: 500 ... This request caused apisever to panic. Look in log for details. ``` Logs: ``` E0515 16:23:22.567079 26140 runtime.go:66] Observed a panic: "invalid memory address or nil pointer dereference" (runtime error: invalid memory address or nil pointer dereference) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:72 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:65 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:51 /usr/local/go/src/runtime/asm_amd64.s:479 /usr/local/go/src/runtime/panic.go:458 /usr/local/go/src/runtime/panic.go:62 /usr/local/go/src/runtime/sigpanic_unix.go:24 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/registry/rest/create.go:128 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/registry/rest/delete.go:66 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/registry/generic/registry/store.go:761 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/storage/util.go:38 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/storage/etcd3/store.go:485 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/storage/etcd3/store.go:292 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/storage/cacher.go:529 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/registry/generic/registry/store.go:799 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/registry/generic/registry/store.go:879 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/handlers/rest.go:936 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/handlers/rest.go:1102 /usr/local/go/src/runtime/asm_amd64.s:2086 E0515 16:23:22.567424 26140 runtime.go:66] Observed a panic: "invalid memory address or nil pointer dereference" (runtime error: invalid memory address or nil pointer dereference) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:72 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:65 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:51 /usr/local/go/src/runtime/asm_amd64.s:479 /usr/local/go/src/runtime/panic.go:458 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/handlers/rest.go:1118 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/handlers/rest.go:939 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/metrics/metrics.go:102 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/github.com/emicklei/go-restful/container.go:272 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/github.com/emicklei/go-restful/container.go:120 /usr/local/go/src/net/http/server.go:1726 /usr/local/go/src/net/http/server.go:2022 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/handlers/authorization.go:64 /usr/local/go/src/net/http/server.go:1726 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/handlers/impersonation.go:30 /usr/local/go/src/net/http/server.go:1726 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/handlers/authentication.go:32 /usr/local/go/src/net/http/server.go:1726 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/origin/handlers.go:97 /usr/local/go/src/net/http/server.go:1726 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/origin/handlers.go:65 /usr/local/go/src/net/http/server.go:1726 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/github.com/gorilla/context/context.go:141 /usr/local/go/src/net/http/server.go:1726 /usr/local/go/src/net/http/server.go:2022 /usr/local/go/src/net/http/server.go:2022 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/origin/master.go:1109 /usr/local/go/src/net/http/server.go:1726 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/origin/handlers.go:211 /usr/local/go/src/net/http/server.go:1726 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters/cors.go:75 /usr/local/go/src/net/http/server.go:1726 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters/wrap.go:41 /usr/local/go/src/net/http/server.go:1726 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters/timeout.go:91 /usr/local/go/src/runtime/asm_amd64.s:2086 E0515 16:23:22.567807 26140 wrap.go:34] APIServer panic'd on DELETE /oapi/v1/namespaces/node/routes/asdfasdf: runtime error: invalid memory address or nil pointer dereference goroutine 49727 [running]: runtime/debug.Stack(0xa78f1e0, 0xc4314185b0, 0x55689ed) /usr/local/go/src/runtime/debug/stack.go:24 +0x79 github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters.WithPanicRecovery.func1.1(0x4be09a0, 0xc420012050) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters/wrap.go:34 +0x74 github.com/openshift/origin/vendor/k8s.io/apimachinery/pkg/util/runtime.HandleCrash(0xc42bd83ec8, 0x1, 0x1) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:54 +0xe5 panic(0x4be09a0, 0xc420012050) /usr/local/go/src/runtime/panic.go:458 +0x243 github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/handlers.finishRequest(0x6fc23ac00, 0xc431d860a0, 0x24, 0x0, 0x0, 0xa75bba0) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/handlers/rest.go:1118 +0x3aa github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/handlers.DeleteResource.func1(0xc42e215680, 0xc430297860) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/handlers/rest.go:939 +0x420 github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/metrics.InstrumentRouteFunc.func1(0xc42e215680, 0xc430297860) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/metrics/metrics.go:102 +0x1f8 github.com/openshift/origin/vendor/github.com/emicklei/go-restful.(*Container).dispatch(0xc422eeb830, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/github.com/emicklei/go-restful/container.go:272 +0xba0 github.com/openshift/origin/vendor/github.com/emicklei/go-restful.(*Container).(github.com/openshift/origin/vendor/github.com/emicklei/go-restful.dispatch)-fm(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/github.com/emicklei/go-restful/container.go:120 +0x48 net/http.HandlerFunc.ServeHTTP(0xc4279d1780, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 net/http.(*ServeMux).ServeHTTP(0xc4237a7530, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:2022 +0x7f github.com/openshift/origin/pkg/cmd/server/handlers.AuthorizationFilter.func1(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/handlers/authorization.go:64 +0x113 net/http.HandlerFunc.ServeHTTP(0xc4237b6ec0, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 github.com/openshift/origin/pkg/cmd/server/handlers.ImpersonationFilter.func1(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/handlers/impersonation.go:30 +0x3142 net/http.HandlerFunc.ServeHTTP(0xc4237cb2c0, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 github.com/openshift/origin/pkg/cmd/server/handlers.AuthenticationHandlerFilter.func1(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/handlers/authentication.go:32 +0x2ba net/http.HandlerFunc.ServeHTTP(0xc4237b6f00, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 github.com/openshift/origin/pkg/cmd/server/origin.namespacingFilter.func1(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/origin/handlers.go:97 +0xd2 net/http.HandlerFunc.ServeHTTP(0xc423820420, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 github.com/openshift/origin/pkg/cmd/server/origin.cacheControlFilter.func1(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/origin/handlers.go:65 +0xc2 net/http.HandlerFunc.ServeHTTP(0xc423820450, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 github.com/openshift/origin/vendor/github.com/gorilla/context.ClearHandler.func1(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/github.com/gorilla/context/context.go:141 +0x8b net/http.HandlerFunc.ServeHTTP(0xc4214d37c0, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 net/http.(*ServeMux).ServeHTTP(0xc423820e40, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:2022 +0x7f net/http.(*ServeMux).ServeHTTP(0xc4212488d0, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:2022 +0x7f github.com/openshift/origin/pkg/cmd/server/origin.WithPatternsHandler.func1(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/origin/master.go:1109 +0xcd net/http.HandlerFunc.ServeHTTP(0xc422ed45c0, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 github.com/openshift/origin/pkg/cmd/server/origin.WithAssetServerRedirect.func1(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/origin/handlers.go:211 +0x7f net/http.HandlerFunc.ServeHTTP(0xc422012e10, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters.WithCORS.func1(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters/cors.go:75 +0x1a2 net/http.HandlerFunc.ServeHTTP(0xc4203e4540, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters.WithPanicRecovery.func1(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters/wrap.go:41 +0x11b net/http.HandlerFunc.ServeHTTP(0xc4234b3480, 0x7f93fd7ab688, 0xc429f4ef20, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc4234b34c0, 0xa7a55e0, 0xc429f4ef20, 0xc42f507590, 0xc430296780) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters/timeout.go:91 +0x8d created by github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters/timeout.go:93 +0x1d1 ``` Route YAML: ```yaml apiVersion: v1 kind: Route metadata: name: asdfasdf namespace: node selfLink: /oapi/v1/namespaces/node/routes/asdfasdf uid: 2a70d1d2-398a-11e7-8813-12fcae64f2c0 resourceVersion: '2821' creationTimestamp: '2017-05-15T16:18:43Z' annotations: openshift.io/host.generated: 'true' spec: host: asdfasdf-node.127.0.0.1.nip.io to: kind: Service name: mongodb weight: 100 port: targetPort: mongodb wildcardPolicy: None status: ingress: - host: asdfasdf-node.127.0.0.1.nip.io routerName: router conditions: - type: Admitted status: 'True' lastTransitionTime: '2017-05-15T16:18:43Z' wildcardPolicy: None ``` cc @deads2k @jwforres @knobunc
1.0
API server panic due to garbage collection requiring delete strategy - Edits by @enj 1. The following resources in origin have a `nil` `store.DeleteStrategy`: ``` clusterpolicies.authorization.openshift.io clusterpolicybindings.authorization.openshift.io policies.authorization.openshift.io policybindings.authorization.openshift.io images.image.openshift.io imagestreams.image.openshift.io clusternetworks.network.openshift.io egressnetworkpolicies.network.openshift.io hostsubnets.network.openshift.io netnamespaces.network.openshift.io oauthaccesstokens.oauth.openshift.io oauthauthorizetokens.oauth.openshift.io oauthclientauthorizations.oauth.openshift.io oauthclients.oauth.openshift.io routes.route.openshift.io templates.template.openshift.io groups.user.openshift.io identities.user.openshift.io users.user.openshift.io ``` 2. `store.EnableGarbageCollection` is `true` for all resources now. 1 and 2 combined with any of the following `DeleteOptions`: - `OrphanDependents` = `true` - `PropagationPolicy` = `DeletePropagationOrphan` - `PropagationPolicy` = `DeletePropagationForeground` Lead to a `panic` in the API server because it tries to a do a graceful deletion and finalization which dereferences `DeleteStrategy`. We have some protection from https://github.com/openshift/origin/commit/54d84e6a8db4c07f78fb2823508fed7751ebf1bd but we need to reconsider. Upstream does not carry this patch at all. --- I'm seeing a panic trying to delete a route in the web console. We're passing `propagationPolicy: "Foreground"` (which we've switched to passing for all requests). ``` $ oc version oc v3.6.0-alpha.1+b581bfe-566 kubernetes v1.6.1+5115d708d7 features: Basic-Auth Server https://127.0.0.1:8443 openshift v3.6.0-alpha.1+b581bf ``` HTTP Request: ``` DELETE /oapi/v1/namespaces/node/routes/asdfasdf Content-Type: application/json ... {"kind":"DeleteOptions","apiVersion":"v1","propagationPolicy":"Foreground"} ``` Response: ``` Status: 500 ... This request caused apisever to panic. Look in log for details. ``` Logs: ``` E0515 16:23:22.567079 26140 runtime.go:66] Observed a panic: "invalid memory address or nil pointer dereference" (runtime error: invalid memory address or nil pointer dereference) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:72 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:65 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:51 /usr/local/go/src/runtime/asm_amd64.s:479 /usr/local/go/src/runtime/panic.go:458 /usr/local/go/src/runtime/panic.go:62 /usr/local/go/src/runtime/sigpanic_unix.go:24 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/registry/rest/create.go:128 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/registry/rest/delete.go:66 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/registry/generic/registry/store.go:761 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/storage/util.go:38 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/storage/etcd3/store.go:485 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/storage/etcd3/store.go:292 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/storage/cacher.go:529 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/registry/generic/registry/store.go:799 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/registry/generic/registry/store.go:879 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/handlers/rest.go:936 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/handlers/rest.go:1102 /usr/local/go/src/runtime/asm_amd64.s:2086 E0515 16:23:22.567424 26140 runtime.go:66] Observed a panic: "invalid memory address or nil pointer dereference" (runtime error: invalid memory address or nil pointer dereference) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:72 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:65 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:51 /usr/local/go/src/runtime/asm_amd64.s:479 /usr/local/go/src/runtime/panic.go:458 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/handlers/rest.go:1118 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/handlers/rest.go:939 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/metrics/metrics.go:102 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/github.com/emicklei/go-restful/container.go:272 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/github.com/emicklei/go-restful/container.go:120 /usr/local/go/src/net/http/server.go:1726 /usr/local/go/src/net/http/server.go:2022 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/handlers/authorization.go:64 /usr/local/go/src/net/http/server.go:1726 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/handlers/impersonation.go:30 /usr/local/go/src/net/http/server.go:1726 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/handlers/authentication.go:32 /usr/local/go/src/net/http/server.go:1726 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/origin/handlers.go:97 /usr/local/go/src/net/http/server.go:1726 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/origin/handlers.go:65 /usr/local/go/src/net/http/server.go:1726 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/github.com/gorilla/context/context.go:141 /usr/local/go/src/net/http/server.go:1726 /usr/local/go/src/net/http/server.go:2022 /usr/local/go/src/net/http/server.go:2022 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/origin/master.go:1109 /usr/local/go/src/net/http/server.go:1726 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/origin/handlers.go:211 /usr/local/go/src/net/http/server.go:1726 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters/cors.go:75 /usr/local/go/src/net/http/server.go:1726 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters/wrap.go:41 /usr/local/go/src/net/http/server.go:1726 /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters/timeout.go:91 /usr/local/go/src/runtime/asm_amd64.s:2086 E0515 16:23:22.567807 26140 wrap.go:34] APIServer panic'd on DELETE /oapi/v1/namespaces/node/routes/asdfasdf: runtime error: invalid memory address or nil pointer dereference goroutine 49727 [running]: runtime/debug.Stack(0xa78f1e0, 0xc4314185b0, 0x55689ed) /usr/local/go/src/runtime/debug/stack.go:24 +0x79 github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters.WithPanicRecovery.func1.1(0x4be09a0, 0xc420012050) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters/wrap.go:34 +0x74 github.com/openshift/origin/vendor/k8s.io/apimachinery/pkg/util/runtime.HandleCrash(0xc42bd83ec8, 0x1, 0x1) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:54 +0xe5 panic(0x4be09a0, 0xc420012050) /usr/local/go/src/runtime/panic.go:458 +0x243 github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/handlers.finishRequest(0x6fc23ac00, 0xc431d860a0, 0x24, 0x0, 0x0, 0xa75bba0) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/handlers/rest.go:1118 +0x3aa github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/handlers.DeleteResource.func1(0xc42e215680, 0xc430297860) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/handlers/rest.go:939 +0x420 github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/metrics.InstrumentRouteFunc.func1(0xc42e215680, 0xc430297860) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/endpoints/metrics/metrics.go:102 +0x1f8 github.com/openshift/origin/vendor/github.com/emicklei/go-restful.(*Container).dispatch(0xc422eeb830, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/github.com/emicklei/go-restful/container.go:272 +0xba0 github.com/openshift/origin/vendor/github.com/emicklei/go-restful.(*Container).(github.com/openshift/origin/vendor/github.com/emicklei/go-restful.dispatch)-fm(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/github.com/emicklei/go-restful/container.go:120 +0x48 net/http.HandlerFunc.ServeHTTP(0xc4279d1780, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 net/http.(*ServeMux).ServeHTTP(0xc4237a7530, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:2022 +0x7f github.com/openshift/origin/pkg/cmd/server/handlers.AuthorizationFilter.func1(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/handlers/authorization.go:64 +0x113 net/http.HandlerFunc.ServeHTTP(0xc4237b6ec0, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 github.com/openshift/origin/pkg/cmd/server/handlers.ImpersonationFilter.func1(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/handlers/impersonation.go:30 +0x3142 net/http.HandlerFunc.ServeHTTP(0xc4237cb2c0, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 github.com/openshift/origin/pkg/cmd/server/handlers.AuthenticationHandlerFilter.func1(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/handlers/authentication.go:32 +0x2ba net/http.HandlerFunc.ServeHTTP(0xc4237b6f00, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 github.com/openshift/origin/pkg/cmd/server/origin.namespacingFilter.func1(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/origin/handlers.go:97 +0xd2 net/http.HandlerFunc.ServeHTTP(0xc423820420, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 github.com/openshift/origin/pkg/cmd/server/origin.cacheControlFilter.func1(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/origin/handlers.go:65 +0xc2 net/http.HandlerFunc.ServeHTTP(0xc423820450, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 github.com/openshift/origin/vendor/github.com/gorilla/context.ClearHandler.func1(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/github.com/gorilla/context/context.go:141 +0x8b net/http.HandlerFunc.ServeHTTP(0xc4214d37c0, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 net/http.(*ServeMux).ServeHTTP(0xc423820e40, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:2022 +0x7f net/http.(*ServeMux).ServeHTTP(0xc4212488d0, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:2022 +0x7f github.com/openshift/origin/pkg/cmd/server/origin.WithPatternsHandler.func1(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/origin/master.go:1109 +0xcd net/http.HandlerFunc.ServeHTTP(0xc422ed45c0, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 github.com/openshift/origin/pkg/cmd/server/origin.WithAssetServerRedirect.func1(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/pkg/cmd/server/origin/handlers.go:211 +0x7f net/http.HandlerFunc.ServeHTTP(0xc422012e10, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters.WithCORS.func1(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters/cors.go:75 +0x1a2 net/http.HandlerFunc.ServeHTTP(0xc4203e4540, 0xa78f1e0, 0xc4314185b0, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters.WithPanicRecovery.func1(0xa78f1e0, 0xc4314185b0, 0xc42f507590) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters/wrap.go:41 +0x11b net/http.HandlerFunc.ServeHTTP(0xc4234b3480, 0x7f93fd7ab688, 0xc429f4ef20, 0xc42f507590) /usr/local/go/src/net/http/server.go:1726 +0x44 github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc4234b34c0, 0xa7a55e0, 0xc429f4ef20, 0xc42f507590, 0xc430296780) /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters/timeout.go:91 +0x8d created by github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP /go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/filters/timeout.go:93 +0x1d1 ``` Route YAML: ```yaml apiVersion: v1 kind: Route metadata: name: asdfasdf namespace: node selfLink: /oapi/v1/namespaces/node/routes/asdfasdf uid: 2a70d1d2-398a-11e7-8813-12fcae64f2c0 resourceVersion: '2821' creationTimestamp: '2017-05-15T16:18:43Z' annotations: openshift.io/host.generated: 'true' spec: host: asdfasdf-node.127.0.0.1.nip.io to: kind: Service name: mongodb weight: 100 port: targetPort: mongodb wildcardPolicy: None status: ingress: - host: asdfasdf-node.127.0.0.1.nip.io routerName: router conditions: - type: Admitted status: 'True' lastTransitionTime: '2017-05-15T16:18:43Z' wildcardPolicy: None ``` cc @deads2k @jwforres @knobunc
non_test
api server panic due to garbage collection requiring delete strategy edits by enj the following resources in origin have a nil store deletestrategy clusterpolicies authorization openshift io clusterpolicybindings authorization openshift io policies authorization openshift io policybindings authorization openshift io images image openshift io imagestreams image openshift io clusternetworks network openshift io egressnetworkpolicies network openshift io hostsubnets network openshift io netnamespaces network openshift io oauthaccesstokens oauth openshift io oauthauthorizetokens oauth openshift io oauthclientauthorizations oauth openshift io oauthclients oauth openshift io routes route openshift io templates template openshift io groups user openshift io identities user openshift io users user openshift io store enablegarbagecollection is true for all resources now and combined with any of the following deleteoptions orphandependents true propagationpolicy deletepropagationorphan propagationpolicy deletepropagationforeground lead to a panic in the api server because it tries to a do a graceful deletion and finalization which dereferences deletestrategy we have some protection from but we need to reconsider upstream does not carry this patch at all i m seeing a panic trying to delete a route in the web console we re passing propagationpolicy foreground which we ve switched to passing for all requests oc version oc alpha kubernetes features basic auth server openshift alpha http request delete oapi namespaces node routes asdfasdf content type application json kind deleteoptions apiversion propagationpolicy foreground response status this request caused apisever to panic look in log for details logs runtime go observed a panic invalid memory address or nil pointer dereference runtime error invalid memory address or nil pointer dereference go src github com openshift origin output local go src github com openshift origin vendor io apimachinery pkg util runtime runtime go go src github com openshift origin output local go src github com openshift origin vendor io apimachinery pkg util runtime runtime go go src github com openshift origin output local go src github com openshift origin vendor io apimachinery pkg util runtime runtime go usr local go src runtime asm s usr local go src runtime panic go usr local go src runtime panic go usr local go src runtime sigpanic unix go go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg registry rest create go go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg registry rest delete go go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg registry generic registry store go go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg storage util go go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg storage store go go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg storage store go go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg storage cacher go go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg registry generic registry store go go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg registry generic registry store go go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg endpoints handlers rest go go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg endpoints handlers rest go usr local go src runtime asm s runtime go observed a panic invalid memory address or nil pointer dereference runtime error invalid memory address or nil pointer dereference go src github com openshift origin output local go src github com openshift origin vendor io apimachinery pkg util runtime runtime go go src github com openshift origin output local go src github com openshift origin vendor io apimachinery pkg util runtime runtime go go src github com openshift origin output local go src github com openshift origin vendor io apimachinery pkg util runtime runtime go usr local go src runtime asm s usr local go src runtime panic go go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg endpoints handlers rest go go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg endpoints handlers rest go go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg endpoints metrics metrics go go src github com openshift origin output local go src github com openshift origin vendor github com emicklei go restful container go go src github com openshift origin output local go src github com openshift origin vendor github com emicklei go restful container go usr local go src net http server go usr local go src net http server go go src github com openshift origin output local go src github com openshift origin pkg cmd server handlers authorization go usr local go src net http server go go src github com openshift origin output local go src github com openshift origin pkg cmd server handlers impersonation go usr local go src net http server go go src github com openshift origin output local go src github com openshift origin pkg cmd server handlers authentication go usr local go src net http server go go src github com openshift origin output local go src github com openshift origin pkg cmd server origin handlers go usr local go src net http server go go src github com openshift origin output local go src github com openshift origin pkg cmd server origin handlers go usr local go src net http server go go src github com openshift origin output local go src github com openshift origin vendor github com gorilla context context go usr local go src net http server go usr local go src net http server go usr local go src net http server go go src github com openshift origin output local go src github com openshift origin pkg cmd server origin master go usr local go src net http server go go src github com openshift origin output local go src github com openshift origin pkg cmd server origin handlers go usr local go src net http server go go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg server filters cors go usr local go src net http server go go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg server filters wrap go usr local go src net http server go go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg server filters timeout go usr local go src runtime asm s wrap go apiserver panic d on delete oapi namespaces node routes asdfasdf runtime error invalid memory address or nil pointer dereference goroutine runtime debug stack usr local go src runtime debug stack go github com openshift origin vendor io apiserver pkg server filters withpanicrecovery go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg server filters wrap go github com openshift origin vendor io apimachinery pkg util runtime handlecrash go src github com openshift origin output local go src github com openshift origin vendor io apimachinery pkg util runtime runtime go panic usr local go src runtime panic go github com openshift origin vendor io apiserver pkg endpoints handlers finishrequest go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg endpoints handlers rest go github com openshift origin vendor io apiserver pkg endpoints handlers deleteresource go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg endpoints handlers rest go github com openshift origin vendor io apiserver pkg endpoints metrics instrumentroutefunc go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg endpoints metrics metrics go github com openshift origin vendor github com emicklei go restful container dispatch go src github com openshift origin output local go src github com openshift origin vendor github com emicklei go restful container go github com openshift origin vendor github com emicklei go restful container github com openshift origin vendor github com emicklei go restful dispatch fm go src github com openshift origin output local go src github com openshift origin vendor github com emicklei go restful container go net http handlerfunc servehttp usr local go src net http server go net http servemux servehttp usr local go src net http server go github com openshift origin pkg cmd server handlers authorizationfilter go src github com openshift origin output local go src github com openshift origin pkg cmd server handlers authorization go net http handlerfunc servehttp usr local go src net http server go github com openshift origin pkg cmd server handlers impersonationfilter go src github com openshift origin output local go src github com openshift origin pkg cmd server handlers impersonation go net http handlerfunc servehttp usr local go src net http server go github com openshift origin pkg cmd server handlers authenticationhandlerfilter go src github com openshift origin output local go src github com openshift origin pkg cmd server handlers authentication go net http handlerfunc servehttp usr local go src net http server go github com openshift origin pkg cmd server origin namespacingfilter go src github com openshift origin output local go src github com openshift origin pkg cmd server origin handlers go net http handlerfunc servehttp usr local go src net http server go github com openshift origin pkg cmd server origin cachecontrolfilter go src github com openshift origin output local go src github com openshift origin pkg cmd server origin handlers go net http handlerfunc servehttp usr local go src net http server go github com openshift origin vendor github com gorilla context clearhandler go src github com openshift origin output local go src github com openshift origin vendor github com gorilla context context go net http handlerfunc servehttp usr local go src net http server go net http servemux servehttp usr local go src net http server go net http servemux servehttp usr local go src net http server go github com openshift origin pkg cmd server origin withpatternshandler go src github com openshift origin output local go src github com openshift origin pkg cmd server origin master go net http handlerfunc servehttp usr local go src net http server go github com openshift origin pkg cmd server origin withassetserverredirect go src github com openshift origin output local go src github com openshift origin pkg cmd server origin handlers go net http handlerfunc servehttp usr local go src net http server go github com openshift origin vendor io apiserver pkg server filters withcors go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg server filters cors go net http handlerfunc servehttp usr local go src net http server go github com openshift origin vendor io apiserver pkg server filters withpanicrecovery go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg server filters wrap go net http handlerfunc servehttp usr local go src net http server go github com openshift origin vendor io apiserver pkg server filters timeouthandler servehttp go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg server filters timeout go created by github com openshift origin vendor io apiserver pkg server filters timeouthandler servehttp go src github com openshift origin output local go src github com openshift origin vendor io apiserver pkg server filters timeout go route yaml yaml apiversion kind route metadata name asdfasdf namespace node selflink oapi namespaces node routes asdfasdf uid resourceversion creationtimestamp annotations openshift io host generated true spec host asdfasdf node nip io to kind service name mongodb weight port targetport mongodb wildcardpolicy none status ingress host asdfasdf node nip io routername router conditions type admitted status true lasttransitiontime wildcardpolicy none cc jwforres knobunc
0
184,762
14,289,859,197
IssuesEvent
2020-11-23 19:56:33
galaxyproject/galaxy
https://api.github.com/repos/galaxyproject/galaxy
closed
Fix transiently failing workflow API tests
area/testing/api area/workflows kind/bug
I don't have an exact date or a commit, but the API tests have sharply increased their number of transient failures within the last month - maybe 2 or 3 weeks? An example is: - 1 error (test_recover_mapping_in_subworkflow) https://jenkins.galaxyproject.org/job/docker-api/10881/consoleFull Doesn't seem like work invocation scheduling was hanging on older workflows invocations in other parts of tests. Timed out waiting on workflow invocation. - 1 errror (test_extract_with_mapped_output_collections) https://jenkins.galaxyproject.org/job/docker-api/10880/consoleFull Timed out waiting on history jobs. - 1 error (test_workflow_resume_from_failed_step) https://jenkins.galaxyproject.org/job/docker-api/10908/testReport/junit/api.test_workflows/WorkflowsApiTestCase/test_workflow_resume_from_failed_step/ Waiting on history to become ok, job is in queued state at timeout it seems. ```test_workflow_resume_from_failed_step (api.test_workflows.WorkflowsApiTestCase) ... galaxy.web.framework.webapp INFO 2018-03-15 16:51:41,518 Session authenticated using Galaxy master api key galaxy.jobs DEBUG 2018-03-15 16:51:41,525 job 643 ended (finish() executed in (548.140 ms)) galaxy.model.metadata DEBUG 2018-03-15 16:51:41,539 Cleaning up external metadata files galaxy.web.framework.webapp INFO 2018-03-15 16:51:41,571 Session authenticated using Galaxy master api key galaxy.workflow.modules DEBUG 2018-03-15 16:51:41,771 job_properties: using version '1.0.0' instead of version 'None' specified in this workflow. galaxy.workflow.modules DEBUG 2018-03-15 16:51:41,772 identifier_multiple_in_conditional: using version '1.0.0' instead of version 'None' specified in this workflow. galaxy.workflow.run_request INFO 2018-03-15 16:51:42,019 Creating a step_state for step.id 350 galaxy.workflow.run_request INFO 2018-03-15 16:51:42,019 Creating a step_state for step.id 351 galaxy.workflow.scheduling_manager INFO 2018-03-15 16:51:42,020 Queueing workflow invocation for handler [main] galaxy.tools.actions INFO 2018-03-15 16:51:42,823 [{'__object__': <galaxy.model.HistoryDatasetAssociation object at 0x7fc7c9947290>, 'name': 'one'}] galaxy.tools.actions INFO 2018-03-15 16:51:42,857 [{'__object__': <galaxy.model.HistoryDatasetAssociation object at 0x7fc7c9947290>, 'name': 'one'}, {'__object__': <galaxy.model.HistoryDatasetAssociation object at 0x7fc7c9851c90>, 'name': 'two'}] galaxy.managers.collections DEBUG 2018-03-15 16:51:42,874 Created collection with 2 elements galaxy.tools.actions INFO 2018-03-15 16:51:42,901 Handled collection output named list_output for tool job_properties (123.307 ms) galaxy.tools.actions INFO 2018-03-15 16:51:42,927 Handled output named out_file1 for tool job_properties (25.492 ms) galaxy.tools.actions INFO 2018-03-15 16:51:42,934 Added output datasets to history (7.067 ms) galaxy.tools.actions INFO 2018-03-15 16:51:42,940 Verified access to datasets for Job[unflushed,tool_id=job_properties] (0.050 ms) galaxy.tools.actions INFO 2018-03-15 16:51:42,942 Setup for job Job[unflushed,tool_id=job_properties] complete, ready to flush (8.065 ms) galaxy.tools.actions INFO 2018-03-15 16:51:43,009 Flushed transaction for job Job[id=645,tool_id=job_properties] (66.192 ms) galaxy.tools.execute DEBUG 2018-03-15 16:51:43,011 Tool [job_properties] created job [645] (245.102 ms) galaxy.tools.execute DEBUG 2018-03-15 16:51:43,068 Executed 1 job(s) for tool job_properties request: (316.923 ms) galaxy.workflow.run DEBUG 2018-03-15 16:51:43,100 Workflow step 350 of invocation 79 invoked (349.264 ms) galaxy.tools.actions INFO 2018-03-15 16:51:43,223 Handled output named output1 for tool identifier_multiple_in_conditional (55.311 ms) galaxy.tools.actions INFO 2018-03-15 16:51:43,232 Added output datasets to history (8.528 ms) galaxy.tools.actions INFO 2018-03-15 16:51:43,269 Verified access to datasets for Job[unflushed,tool_id=identifier_multiple_in_conditional] (22.335 ms) galaxy.tools.actions INFO 2018-03-15 16:51:43,272 Setup for job Job[unflushed,tool_id=identifier_multiple_in_conditional] complete, ready to flush (39.652 ms) galaxy.tools.actions INFO 2018-03-15 16:51:43,324 Flushed transaction for job Job[id=646,tool_id=identifier_multiple_in_conditional] (51.780 ms) galaxy.tools.execute DEBUG 2018-03-15 16:51:43,325 Tool [identifier_multiple_in_conditional] created job [646] (205.437 ms) galaxy.tools.execute DEBUG 2018-03-15 16:51:43,351 Executed 1 job(s) for tool identifier_multiple_in_conditional request: (231.105 ms) galaxy.workflow.run DEBUG 2018-03-15 16:51:43,374 Workflow step 351 of invocation 79 invoked (271.494 ms) galaxy.jobs DEBUG 2018-03-15 16:51:43,486 (645) Working directory for job is: /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645 galaxy.jobs.handler DEBUG 2018-03-15 16:51:43,498 (645) Dispatching to local runner galaxy.jobs DEBUG 2018-03-15 16:51:43,511 (645) Persisting job destination (destination id: local) galaxy.jobs.runners DEBUG 2018-03-15 16:51:43,530 Job [645] queued (30.918 ms) galaxy.jobs.handler INFO 2018-03-15 16:51:43,550 (645) Job dispatched galaxy.tools.evaluation INFO 2018-03-15 16:51:43,667 Updating param_dict for one with /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_708.dat galaxy.tools.evaluation INFO 2018-03-15 16:51:43,668 Updating param_dict for two with /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_709.dat galaxy.jobs.command_factory INFO 2018-03-15 16:51:43,751 Built script [/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/tool_script.sh] for tool command [echo "The bool is true" && echo "The bool is really true" 1>&2 && echo "This is a line of text." > '/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_710.dat' && cp '/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_710.dat' /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_708.dat && cp '/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_710.dat' /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_709.dat ; sh -c "exit 127"] galaxy.jobs.runners DEBUG 2018-03-15 16:51:44,090 (645) command is: rm -rf working; mkdir -p working; cd working; /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/tool_script.sh; return_code=$?; cd '/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645'; [ "$GALAXY_VIRTUAL_ENV" = "None" ] && GALAXY_VIRTUAL_ENV="$_GALAXY_VIRTUAL_ENV"; _galaxy_setup_environment True python "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/set_metadata_A1WiSn.py" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/registry.xml" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/working/galaxy.json" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_in_HistoryDatasetAssociation_739_nDdOtS,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_kwds_HistoryDatasetAssociation_739_gaxjl6,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_out_HistoryDatasetAssociation_739_NLco6P,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_results_HistoryDatasetAssociation_739_Lpbi1d,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_710.dat,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_override_HistoryDatasetAssociation_739_HGndQo" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_in_HistoryDatasetAssociation_738_aEML82,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_kwds_HistoryDatasetAssociation_738_nHuYeJ,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_out_HistoryDatasetAssociation_738_I6KvJW,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_results_HistoryDatasetAssociation_738_AlxO6S,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_709.dat,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_override_HistoryDatasetAssociation_738_jk13bP" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_in_HistoryDatasetAssociation_737_0Xxvtz,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_kwds_HistoryDatasetAssociation_737_5QTTMc,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_out_HistoryDatasetAssociation_737_D2xBNh,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_results_HistoryDatasetAssociation_737_NwbZS7,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_708.dat,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_override_HistoryDatasetAssociation_737_8qBnrj" 5242880; sh -c "exit $return_code" galaxy.jobs.runners.local DEBUG 2018-03-15 16:51:44,139 (645) executing job script: /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/galaxy_645.sh galaxy.jobs DEBUG 2018-03-15 16:51:44,184 (645) Persisting job destination (destination id: local) galaxy.jobs.runners.local DEBUG 2018-03-15 16:51:47,220 execution finished: /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/galaxy_645.sh galaxy.jobs.output_checker INFO 2018-03-15 16:51:47,233 Job 645: Fatal error: Exit code 127 (Failing exit code.) galaxy.jobs.output_checker DEBUG 2018-03-15 16:51:47,234 Tool exit code indicates an error, failing job. galaxy.jobs DEBUG 2018-03-15 16:51:47,314 (645) setting dataset 710 state to ERROR galaxy.jobs DEBUG 2018-03-15 16:51:47,375 Pausing Job '646', Execution of this dataset's job is paused because its input datasets are in an error state. galaxy.jobs DEBUG 2018-03-15 16:51:47,396 (645) setting dataset 709 state to ERROR galaxy.jobs DEBUG 2018-03-15 16:51:47,410 (645) setting dataset 708 state to ERROR galaxy.jobs INFO 2018-03-15 16:51:47,524 Collecting metrics for Job 645 galaxy.jobs DEBUG 2018-03-15 16:51:47,539 job 645 ended (finish() executed in (297.480 ms)) galaxy.tools.error_reports DEBUG 2018-03-15 16:51:47,554 Bug report plugin <galaxy.tools.error_reports.plugins.sentry.SentryPlugin object at 0x7fc84161ea10> generated response None galaxy.tools.error_reports DEBUG 2018-03-15 16:51:47,555 Bug report plugin <galaxy.tools.error_reports.plugins.sentry.SentryPlugin object at 0x7fc84161ea10> generated response None galaxy.tools.error_reports DEBUG 2018-03-15 16:51:47,555 Bug report plugin <galaxy.tools.error_reports.plugins.sentry.SentryPlugin object at 0x7fc84161ea10> generated response None galaxy.tools DEBUG 2018-03-15 16:51:48,866 Validated and populated state for tool request (0.123 ms) galaxy.tools.actions INFO 2018-03-15 16:51:48,931 [{'__object__': <galaxy.model.HistoryDatasetAssociation object at 0x7fc7cad49d90>, 'name': 'one'}] galaxy.tools.actions INFO 2018-03-15 16:51:48,973 [{'__object__': <galaxy.model.HistoryDatasetAssociation object at 0x7fc7cad49d90>, 'name': 'one'}, {'__object__': <galaxy.model.HistoryDatasetAssociation object at 0x7fc7c9a7e290>, 'name': 'two'}] galaxy.managers.collections DEBUG 2018-03-15 16:51:48,985 Created collection with 2 elements galaxy.tools.actions INFO 2018-03-15 16:51:49,015 Handled collection output named list_output for tool job_properties (128.349 ms) galaxy.tools.actions INFO 2018-03-15 16:51:49,034 Handled output named out_file1 for tool job_properties (18.884 ms) galaxy.tools.actions INFO 2018-03-15 16:51:49,043 Added output datasets to history (8.459 ms) galaxy.tools.actions INFO 2018-03-15 16:51:49,048 Verified access to datasets for Job[unflushed,tool_id=job_properties] (0.039 ms) galaxy.tools.actions INFO 2018-03-15 16:51:49,186 Job 646 input HDA 739 remapped to new HDA 743 galaxy.tools.actions INFO 2018-03-15 16:51:49,198 Job 646 input HDA 739 remapped to new HDA 743 galaxy.tools.actions INFO 2018-03-15 16:51:49,265 Setup for job Job[id=647,tool_id=job_properties] complete, ready to flush (221.941 ms) galaxy.tools.actions INFO 2018-03-15 16:51:49,336 Flushed transaction for job Job[id=647,tool_id=job_properties] (69.740 ms) galaxy.tools.execute DEBUG 2018-03-15 16:51:49,337 Tool [job_properties] created job [647] (453.595 ms) galaxy.tools.execute DEBUG 2018-03-15 16:51:49,344 Executed 1 job(s) for tool job_properties request: (477.750 ms) galaxy.jobs DEBUG 2018-03-15 16:51:49,406 (647) Working directory for job is: /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647 galaxy.jobs.handler DEBUG 2018-03-15 16:51:49,420 (647) Dispatching to local runner galaxy.jobs DEBUG 2018-03-15 16:51:49,443 (647) Persisting job destination (destination id: local) galaxy.jobs.runners DEBUG 2018-03-15 16:51:49,456 Job [647] queued (27.106 ms) galaxy.jobs.handler INFO 2018-03-15 16:51:49,477 (647) Job dispatched galaxy.tools.evaluation INFO 2018-03-15 16:51:49,626 Updating param_dict for one with /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_712.dat galaxy.tools.evaluation INFO 2018-03-15 16:51:49,627 Updating param_dict for two with /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_713.dat galaxy.jobs.command_factory INFO 2018-03-15 16:51:49,728 Built script [/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/tool_script.sh] for tool command [echo "The bool is not true" && echo "The bool is very not true" 1>&2 && echo "This is a different line of text." > '/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_714.dat' && sh -c "exit 2"] galaxy.jobs.runners DEBUG 2018-03-15 16:51:50,017 (647) command is: rm -rf working; mkdir -p working; cd working; /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/tool_script.sh; return_code=$?; cd '/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647'; [ "$GALAXY_VIRTUAL_ENV" = "None" ] && GALAXY_VIRTUAL_ENV="$_GALAXY_VIRTUAL_ENV"; _galaxy_setup_environment True python "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/set_metadata_TyX6gz.py" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/registry.xml" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/working/galaxy.json" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_in_HistoryDatasetAssociation_743_YBOTUD,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_kwds_HistoryDatasetAssociation_743_oRP4wQ,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_out_HistoryDatasetAssociation_743_JUYn7V,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_results_HistoryDatasetAssociation_743_HQ2pp3,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_714.dat,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_override_HistoryDatasetAssociation_743_zw3eoH" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_in_HistoryDatasetAssociation_742_em53Vw,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_kwds_HistoryDatasetAssociation_742_88bKub,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_out_HistoryDatasetAssociation_742_hO8Sfz,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_results_HistoryDatasetAssociation_742_hanPoo,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_713.dat,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_override_HistoryDatasetAssociation_742_DtwgQI" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_in_HistoryDatasetAssociation_741_eX3aJY,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_kwds_HistoryDatasetAssociation_741_uUH0ud,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_out_HistoryDatasetAssociation_741_pW0zzZ,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_results_HistoryDatasetAssociation_741_pb4U0y,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_712.dat,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_override_HistoryDatasetAssociation_741_3saMk6" 5242880; sh -c "exit $return_code" galaxy.jobs.runners.local DEBUG 2018-03-15 16:51:50,070 (647) executing job script: /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/galaxy_647.sh galaxy.jobs DEBUG 2018-03-15 16:51:50,120 (647) Persisting job destination (destination id: local) galaxy.jobs.runners.local DEBUG 2018-03-15 16:51:53,144 execution finished: /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/galaxy_647.sh galaxy.model.metadata DEBUG 2018-03-15 16:51:53,264 loading metadata from file for: HistoryDatasetAssociation 743 galaxy.model.metadata DEBUG 2018-03-15 16:51:53,349 loading metadata from file for: HistoryDatasetAssociation 742 galaxy.model.metadata DEBUG 2018-03-15 16:51:53,533 loading metadata from file for: HistoryDatasetAssociation 741 galaxy.jobs DEBUG 2018-03-15 16:51:53,648 (646) Working directory for job is: /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646 galaxy.jobs.handler DEBUG 2018-03-15 16:51:53,661 (646) Dispatching to local runner galaxy.jobs DEBUG 2018-03-15 16:51:53,685 (646) Persisting job destination (destination id: local) galaxy.jobs.runners DEBUG 2018-03-15 16:51:53,703 Job [646] queued (41.243 ms) galaxy.jobs.handler INFO 2018-03-15 16:51:53,720 (646) Job dispatched galaxy.jobs INFO 2018-03-15 16:51:53,804 Collecting metrics for Job 647 galaxy.jobs DEBUG 2018-03-15 16:51:53,850 job 647 ended (finish() executed in (686.959 ms)) galaxy.model.metadata DEBUG 2018-03-15 16:51:53,866 Cleaning up external metadata files galaxy.jobs.command_factory INFO 2018-03-15 16:51:54,043 Built script [/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/tool_script.sh] for tool command [echo 'Test Job Properties' >> 'output1'] galaxy.jobs.runners DEBUG 2018-03-15 16:53:44,469 (646) command is: rm -rf working; mkdir -p working; cd working; /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/tool_script.sh; return_code=$?; if [ -f /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/working/output1 ] ; then cp /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/working/output1 /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_711.dat ; fi; cd '/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646'; [ "$GALAXY_VIRTUAL_ENV" = "None" ] && GALAXY_VIRTUAL_ENV="$_GALAXY_VIRTUAL_ENV"; _galaxy_setup_environment True python "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/set_metadata_963nPQ.py" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/registry.xml" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/working/galaxy.json" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/metadata_in_HistoryDatasetAssociation_740_hdwesK,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/metadata_kwds_HistoryDatasetAssociation_740_WvgeQ_,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/metadata_out_HistoryDatasetAssociation_740_ttNwMW,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/metadata_results_HistoryDatasetAssociation_740_QwZ3oD,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_711.dat,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/metadata_override_HistoryDatasetAssociation_740_V_17Ml" 5242880; sh -c "exit $return_code" galaxy.jobs.runners.local DEBUG 2018-03-15 16:53:44,522 (646) executing job script: /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/galaxy_646.sh galaxy.jobs DEBUG 2018-03-15 16:53:44,565 (646) Persisting job destination (destination id: local) FAIL ```
1.0
Fix transiently failing workflow API tests - I don't have an exact date or a commit, but the API tests have sharply increased their number of transient failures within the last month - maybe 2 or 3 weeks? An example is: - 1 error (test_recover_mapping_in_subworkflow) https://jenkins.galaxyproject.org/job/docker-api/10881/consoleFull Doesn't seem like work invocation scheduling was hanging on older workflows invocations in other parts of tests. Timed out waiting on workflow invocation. - 1 errror (test_extract_with_mapped_output_collections) https://jenkins.galaxyproject.org/job/docker-api/10880/consoleFull Timed out waiting on history jobs. - 1 error (test_workflow_resume_from_failed_step) https://jenkins.galaxyproject.org/job/docker-api/10908/testReport/junit/api.test_workflows/WorkflowsApiTestCase/test_workflow_resume_from_failed_step/ Waiting on history to become ok, job is in queued state at timeout it seems. ```test_workflow_resume_from_failed_step (api.test_workflows.WorkflowsApiTestCase) ... galaxy.web.framework.webapp INFO 2018-03-15 16:51:41,518 Session authenticated using Galaxy master api key galaxy.jobs DEBUG 2018-03-15 16:51:41,525 job 643 ended (finish() executed in (548.140 ms)) galaxy.model.metadata DEBUG 2018-03-15 16:51:41,539 Cleaning up external metadata files galaxy.web.framework.webapp INFO 2018-03-15 16:51:41,571 Session authenticated using Galaxy master api key galaxy.workflow.modules DEBUG 2018-03-15 16:51:41,771 job_properties: using version '1.0.0' instead of version 'None' specified in this workflow. galaxy.workflow.modules DEBUG 2018-03-15 16:51:41,772 identifier_multiple_in_conditional: using version '1.0.0' instead of version 'None' specified in this workflow. galaxy.workflow.run_request INFO 2018-03-15 16:51:42,019 Creating a step_state for step.id 350 galaxy.workflow.run_request INFO 2018-03-15 16:51:42,019 Creating a step_state for step.id 351 galaxy.workflow.scheduling_manager INFO 2018-03-15 16:51:42,020 Queueing workflow invocation for handler [main] galaxy.tools.actions INFO 2018-03-15 16:51:42,823 [{'__object__': <galaxy.model.HistoryDatasetAssociation object at 0x7fc7c9947290>, 'name': 'one'}] galaxy.tools.actions INFO 2018-03-15 16:51:42,857 [{'__object__': <galaxy.model.HistoryDatasetAssociation object at 0x7fc7c9947290>, 'name': 'one'}, {'__object__': <galaxy.model.HistoryDatasetAssociation object at 0x7fc7c9851c90>, 'name': 'two'}] galaxy.managers.collections DEBUG 2018-03-15 16:51:42,874 Created collection with 2 elements galaxy.tools.actions INFO 2018-03-15 16:51:42,901 Handled collection output named list_output for tool job_properties (123.307 ms) galaxy.tools.actions INFO 2018-03-15 16:51:42,927 Handled output named out_file1 for tool job_properties (25.492 ms) galaxy.tools.actions INFO 2018-03-15 16:51:42,934 Added output datasets to history (7.067 ms) galaxy.tools.actions INFO 2018-03-15 16:51:42,940 Verified access to datasets for Job[unflushed,tool_id=job_properties] (0.050 ms) galaxy.tools.actions INFO 2018-03-15 16:51:42,942 Setup for job Job[unflushed,tool_id=job_properties] complete, ready to flush (8.065 ms) galaxy.tools.actions INFO 2018-03-15 16:51:43,009 Flushed transaction for job Job[id=645,tool_id=job_properties] (66.192 ms) galaxy.tools.execute DEBUG 2018-03-15 16:51:43,011 Tool [job_properties] created job [645] (245.102 ms) galaxy.tools.execute DEBUG 2018-03-15 16:51:43,068 Executed 1 job(s) for tool job_properties request: (316.923 ms) galaxy.workflow.run DEBUG 2018-03-15 16:51:43,100 Workflow step 350 of invocation 79 invoked (349.264 ms) galaxy.tools.actions INFO 2018-03-15 16:51:43,223 Handled output named output1 for tool identifier_multiple_in_conditional (55.311 ms) galaxy.tools.actions INFO 2018-03-15 16:51:43,232 Added output datasets to history (8.528 ms) galaxy.tools.actions INFO 2018-03-15 16:51:43,269 Verified access to datasets for Job[unflushed,tool_id=identifier_multiple_in_conditional] (22.335 ms) galaxy.tools.actions INFO 2018-03-15 16:51:43,272 Setup for job Job[unflushed,tool_id=identifier_multiple_in_conditional] complete, ready to flush (39.652 ms) galaxy.tools.actions INFO 2018-03-15 16:51:43,324 Flushed transaction for job Job[id=646,tool_id=identifier_multiple_in_conditional] (51.780 ms) galaxy.tools.execute DEBUG 2018-03-15 16:51:43,325 Tool [identifier_multiple_in_conditional] created job [646] (205.437 ms) galaxy.tools.execute DEBUG 2018-03-15 16:51:43,351 Executed 1 job(s) for tool identifier_multiple_in_conditional request: (231.105 ms) galaxy.workflow.run DEBUG 2018-03-15 16:51:43,374 Workflow step 351 of invocation 79 invoked (271.494 ms) galaxy.jobs DEBUG 2018-03-15 16:51:43,486 (645) Working directory for job is: /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645 galaxy.jobs.handler DEBUG 2018-03-15 16:51:43,498 (645) Dispatching to local runner galaxy.jobs DEBUG 2018-03-15 16:51:43,511 (645) Persisting job destination (destination id: local) galaxy.jobs.runners DEBUG 2018-03-15 16:51:43,530 Job [645] queued (30.918 ms) galaxy.jobs.handler INFO 2018-03-15 16:51:43,550 (645) Job dispatched galaxy.tools.evaluation INFO 2018-03-15 16:51:43,667 Updating param_dict for one with /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_708.dat galaxy.tools.evaluation INFO 2018-03-15 16:51:43,668 Updating param_dict for two with /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_709.dat galaxy.jobs.command_factory INFO 2018-03-15 16:51:43,751 Built script [/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/tool_script.sh] for tool command [echo "The bool is true" && echo "The bool is really true" 1>&2 && echo "This is a line of text." > '/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_710.dat' && cp '/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_710.dat' /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_708.dat && cp '/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_710.dat' /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_709.dat ; sh -c "exit 127"] galaxy.jobs.runners DEBUG 2018-03-15 16:51:44,090 (645) command is: rm -rf working; mkdir -p working; cd working; /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/tool_script.sh; return_code=$?; cd '/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645'; [ "$GALAXY_VIRTUAL_ENV" = "None" ] && GALAXY_VIRTUAL_ENV="$_GALAXY_VIRTUAL_ENV"; _galaxy_setup_environment True python "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/set_metadata_A1WiSn.py" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/registry.xml" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/working/galaxy.json" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_in_HistoryDatasetAssociation_739_nDdOtS,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_kwds_HistoryDatasetAssociation_739_gaxjl6,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_out_HistoryDatasetAssociation_739_NLco6P,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_results_HistoryDatasetAssociation_739_Lpbi1d,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_710.dat,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_override_HistoryDatasetAssociation_739_HGndQo" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_in_HistoryDatasetAssociation_738_aEML82,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_kwds_HistoryDatasetAssociation_738_nHuYeJ,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_out_HistoryDatasetAssociation_738_I6KvJW,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_results_HistoryDatasetAssociation_738_AlxO6S,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_709.dat,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_override_HistoryDatasetAssociation_738_jk13bP" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_in_HistoryDatasetAssociation_737_0Xxvtz,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_kwds_HistoryDatasetAssociation_737_5QTTMc,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_out_HistoryDatasetAssociation_737_D2xBNh,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_results_HistoryDatasetAssociation_737_NwbZS7,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_708.dat,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/metadata_override_HistoryDatasetAssociation_737_8qBnrj" 5242880; sh -c "exit $return_code" galaxy.jobs.runners.local DEBUG 2018-03-15 16:51:44,139 (645) executing job script: /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/galaxy_645.sh galaxy.jobs DEBUG 2018-03-15 16:51:44,184 (645) Persisting job destination (destination id: local) galaxy.jobs.runners.local DEBUG 2018-03-15 16:51:47,220 execution finished: /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/645/galaxy_645.sh galaxy.jobs.output_checker INFO 2018-03-15 16:51:47,233 Job 645: Fatal error: Exit code 127 (Failing exit code.) galaxy.jobs.output_checker DEBUG 2018-03-15 16:51:47,234 Tool exit code indicates an error, failing job. galaxy.jobs DEBUG 2018-03-15 16:51:47,314 (645) setting dataset 710 state to ERROR galaxy.jobs DEBUG 2018-03-15 16:51:47,375 Pausing Job '646', Execution of this dataset's job is paused because its input datasets are in an error state. galaxy.jobs DEBUG 2018-03-15 16:51:47,396 (645) setting dataset 709 state to ERROR galaxy.jobs DEBUG 2018-03-15 16:51:47,410 (645) setting dataset 708 state to ERROR galaxy.jobs INFO 2018-03-15 16:51:47,524 Collecting metrics for Job 645 galaxy.jobs DEBUG 2018-03-15 16:51:47,539 job 645 ended (finish() executed in (297.480 ms)) galaxy.tools.error_reports DEBUG 2018-03-15 16:51:47,554 Bug report plugin <galaxy.tools.error_reports.plugins.sentry.SentryPlugin object at 0x7fc84161ea10> generated response None galaxy.tools.error_reports DEBUG 2018-03-15 16:51:47,555 Bug report plugin <galaxy.tools.error_reports.plugins.sentry.SentryPlugin object at 0x7fc84161ea10> generated response None galaxy.tools.error_reports DEBUG 2018-03-15 16:51:47,555 Bug report plugin <galaxy.tools.error_reports.plugins.sentry.SentryPlugin object at 0x7fc84161ea10> generated response None galaxy.tools DEBUG 2018-03-15 16:51:48,866 Validated and populated state for tool request (0.123 ms) galaxy.tools.actions INFO 2018-03-15 16:51:48,931 [{'__object__': <galaxy.model.HistoryDatasetAssociation object at 0x7fc7cad49d90>, 'name': 'one'}] galaxy.tools.actions INFO 2018-03-15 16:51:48,973 [{'__object__': <galaxy.model.HistoryDatasetAssociation object at 0x7fc7cad49d90>, 'name': 'one'}, {'__object__': <galaxy.model.HistoryDatasetAssociation object at 0x7fc7c9a7e290>, 'name': 'two'}] galaxy.managers.collections DEBUG 2018-03-15 16:51:48,985 Created collection with 2 elements galaxy.tools.actions INFO 2018-03-15 16:51:49,015 Handled collection output named list_output for tool job_properties (128.349 ms) galaxy.tools.actions INFO 2018-03-15 16:51:49,034 Handled output named out_file1 for tool job_properties (18.884 ms) galaxy.tools.actions INFO 2018-03-15 16:51:49,043 Added output datasets to history (8.459 ms) galaxy.tools.actions INFO 2018-03-15 16:51:49,048 Verified access to datasets for Job[unflushed,tool_id=job_properties] (0.039 ms) galaxy.tools.actions INFO 2018-03-15 16:51:49,186 Job 646 input HDA 739 remapped to new HDA 743 galaxy.tools.actions INFO 2018-03-15 16:51:49,198 Job 646 input HDA 739 remapped to new HDA 743 galaxy.tools.actions INFO 2018-03-15 16:51:49,265 Setup for job Job[id=647,tool_id=job_properties] complete, ready to flush (221.941 ms) galaxy.tools.actions INFO 2018-03-15 16:51:49,336 Flushed transaction for job Job[id=647,tool_id=job_properties] (69.740 ms) galaxy.tools.execute DEBUG 2018-03-15 16:51:49,337 Tool [job_properties] created job [647] (453.595 ms) galaxy.tools.execute DEBUG 2018-03-15 16:51:49,344 Executed 1 job(s) for tool job_properties request: (477.750 ms) galaxy.jobs DEBUG 2018-03-15 16:51:49,406 (647) Working directory for job is: /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647 galaxy.jobs.handler DEBUG 2018-03-15 16:51:49,420 (647) Dispatching to local runner galaxy.jobs DEBUG 2018-03-15 16:51:49,443 (647) Persisting job destination (destination id: local) galaxy.jobs.runners DEBUG 2018-03-15 16:51:49,456 Job [647] queued (27.106 ms) galaxy.jobs.handler INFO 2018-03-15 16:51:49,477 (647) Job dispatched galaxy.tools.evaluation INFO 2018-03-15 16:51:49,626 Updating param_dict for one with /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_712.dat galaxy.tools.evaluation INFO 2018-03-15 16:51:49,627 Updating param_dict for two with /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_713.dat galaxy.jobs.command_factory INFO 2018-03-15 16:51:49,728 Built script [/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/tool_script.sh] for tool command [echo "The bool is not true" && echo "The bool is very not true" 1>&2 && echo "This is a different line of text." > '/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_714.dat' && sh -c "exit 2"] galaxy.jobs.runners DEBUG 2018-03-15 16:51:50,017 (647) command is: rm -rf working; mkdir -p working; cd working; /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/tool_script.sh; return_code=$?; cd '/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647'; [ "$GALAXY_VIRTUAL_ENV" = "None" ] && GALAXY_VIRTUAL_ENV="$_GALAXY_VIRTUAL_ENV"; _galaxy_setup_environment True python "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/set_metadata_TyX6gz.py" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/registry.xml" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/working/galaxy.json" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_in_HistoryDatasetAssociation_743_YBOTUD,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_kwds_HistoryDatasetAssociation_743_oRP4wQ,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_out_HistoryDatasetAssociation_743_JUYn7V,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_results_HistoryDatasetAssociation_743_HQ2pp3,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_714.dat,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_override_HistoryDatasetAssociation_743_zw3eoH" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_in_HistoryDatasetAssociation_742_em53Vw,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_kwds_HistoryDatasetAssociation_742_88bKub,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_out_HistoryDatasetAssociation_742_hO8Sfz,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_results_HistoryDatasetAssociation_742_hanPoo,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_713.dat,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_override_HistoryDatasetAssociation_742_DtwgQI" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_in_HistoryDatasetAssociation_741_eX3aJY,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_kwds_HistoryDatasetAssociation_741_uUH0ud,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_out_HistoryDatasetAssociation_741_pW0zzZ,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_results_HistoryDatasetAssociation_741_pb4U0y,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_712.dat,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/metadata_override_HistoryDatasetAssociation_741_3saMk6" 5242880; sh -c "exit $return_code" galaxy.jobs.runners.local DEBUG 2018-03-15 16:51:50,070 (647) executing job script: /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/galaxy_647.sh galaxy.jobs DEBUG 2018-03-15 16:51:50,120 (647) Persisting job destination (destination id: local) galaxy.jobs.runners.local DEBUG 2018-03-15 16:51:53,144 execution finished: /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/647/galaxy_647.sh galaxy.model.metadata DEBUG 2018-03-15 16:51:53,264 loading metadata from file for: HistoryDatasetAssociation 743 galaxy.model.metadata DEBUG 2018-03-15 16:51:53,349 loading metadata from file for: HistoryDatasetAssociation 742 galaxy.model.metadata DEBUG 2018-03-15 16:51:53,533 loading metadata from file for: HistoryDatasetAssociation 741 galaxy.jobs DEBUG 2018-03-15 16:51:53,648 (646) Working directory for job is: /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646 galaxy.jobs.handler DEBUG 2018-03-15 16:51:53,661 (646) Dispatching to local runner galaxy.jobs DEBUG 2018-03-15 16:51:53,685 (646) Persisting job destination (destination id: local) galaxy.jobs.runners DEBUG 2018-03-15 16:51:53,703 Job [646] queued (41.243 ms) galaxy.jobs.handler INFO 2018-03-15 16:51:53,720 (646) Job dispatched galaxy.jobs INFO 2018-03-15 16:51:53,804 Collecting metrics for Job 647 galaxy.jobs DEBUG 2018-03-15 16:51:53,850 job 647 ended (finish() executed in (686.959 ms)) galaxy.model.metadata DEBUG 2018-03-15 16:51:53,866 Cleaning up external metadata files galaxy.jobs.command_factory INFO 2018-03-15 16:51:54,043 Built script [/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/tool_script.sh] for tool command [echo 'Test Job Properties' >> 'output1'] galaxy.jobs.runners DEBUG 2018-03-15 16:53:44,469 (646) command is: rm -rf working; mkdir -p working; cd working; /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/tool_script.sh; return_code=$?; if [ -f /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/working/output1 ] ; then cp /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/working/output1 /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_711.dat ; fi; cd '/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646'; [ "$GALAXY_VIRTUAL_ENV" = "None" ] && GALAXY_VIRTUAL_ENV="$_GALAXY_VIRTUAL_ENV"; _galaxy_setup_environment True python "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/set_metadata_963nPQ.py" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/registry.xml" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/working/galaxy.json" "/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/metadata_in_HistoryDatasetAssociation_740_hdwesK,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/metadata_kwds_HistoryDatasetAssociation_740_WvgeQ_,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/metadata_out_HistoryDatasetAssociation_740_ttNwMW,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/metadata_results_HistoryDatasetAssociation_740_QwZ3oD,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/files/000/dataset_711.dat,/tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/metadata_override_HistoryDatasetAssociation_740_V_17Ml" 5242880; sh -c "exit $return_code" galaxy.jobs.runners.local DEBUG 2018-03-15 16:53:44,522 (646) executing job script: /tmp/tmpDuZfEg/tmp5mSE2V/tmpyLJlv5/database/job_working_directory_8_W39B/000/646/galaxy_646.sh galaxy.jobs DEBUG 2018-03-15 16:53:44,565 (646) Persisting job destination (destination id: local) FAIL ```
test
fix transiently failing workflow api tests i don t have an exact date or a commit but the api tests have sharply increased their number of transient failures within the last month maybe or weeks an example is error test recover mapping in subworkflow doesn t seem like work invocation scheduling was hanging on older workflows invocations in other parts of tests timed out waiting on workflow invocation errror test extract with mapped output collections timed out waiting on history jobs error test workflow resume from failed step waiting on history to become ok job is in queued state at timeout it seems test workflow resume from failed step api test workflows workflowsapitestcase galaxy web framework webapp info session authenticated using galaxy master api key galaxy jobs debug job ended finish executed in ms galaxy model metadata debug cleaning up external metadata files galaxy web framework webapp info session authenticated using galaxy master api key galaxy workflow modules debug job properties using version instead of version none specified in this workflow galaxy workflow modules debug identifier multiple in conditional using version instead of version none specified in this workflow galaxy workflow run request info creating a step state for step id galaxy workflow run request info creating a step state for step id galaxy workflow scheduling manager info queueing workflow invocation for handler galaxy tools actions info galaxy tools actions info galaxy managers collections debug created collection with elements galaxy tools actions info handled collection output named list output for tool job properties ms galaxy tools actions info handled output named out for tool job properties ms galaxy tools actions info added output datasets to history ms galaxy tools actions info verified access to datasets for job ms galaxy tools actions info setup for job job complete ready to flush ms galaxy tools actions info flushed transaction for job job ms galaxy tools execute debug tool created job ms galaxy tools execute debug executed job s for tool job properties request ms galaxy workflow run debug workflow step of invocation invoked ms galaxy tools actions info handled output named for tool identifier multiple in conditional ms galaxy tools actions info added output datasets to history ms galaxy tools actions info verified access to datasets for job ms galaxy tools actions info setup for job job complete ready to flush ms galaxy tools actions info flushed transaction for job job ms galaxy tools execute debug tool created job ms galaxy tools execute debug executed job s for tool identifier multiple in conditional request ms galaxy workflow run debug workflow step of invocation invoked ms galaxy jobs debug working directory for job is tmp tmpduzfeg database job working directory galaxy jobs handler debug dispatching to local runner galaxy jobs debug persisting job destination destination id local galaxy jobs runners debug job queued ms galaxy jobs handler info job dispatched galaxy tools evaluation info updating param dict for one with tmp tmpduzfeg database files dataset dat galaxy tools evaluation info updating param dict for two with tmp tmpduzfeg database files dataset dat galaxy jobs command factory info built script for tool command galaxy jobs runners debug command is rm rf working mkdir p working cd working tmp tmpduzfeg database job working directory tool script sh return code cd tmp tmpduzfeg database job working directory galaxy virtual env galaxy virtual env galaxy setup environment true python tmp tmpduzfeg database job working directory set metadata py tmp tmpduzfeg database job working directory registry xml tmp tmpduzfeg database job working directory working galaxy json tmp tmpduzfeg database job working directory metadata in historydatasetassociation nddots tmp tmpduzfeg database job working directory metadata kwds historydatasetassociation tmp tmpduzfeg database job working directory metadata out historydatasetassociation tmp tmpduzfeg database job working directory metadata results historydatasetassociation tmp tmpduzfeg database files dataset dat tmp tmpduzfeg database job working directory metadata override historydatasetassociation hgndqo tmp tmpduzfeg database job working directory metadata in historydatasetassociation tmp tmpduzfeg database job working directory metadata kwds historydatasetassociation nhuyej tmp tmpduzfeg database job working directory metadata out historydatasetassociation tmp tmpduzfeg database job working directory metadata results historydatasetassociation tmp tmpduzfeg database files dataset dat tmp tmpduzfeg database job working directory metadata override historydatasetassociation tmp tmpduzfeg database job working directory metadata in historydatasetassociation tmp tmpduzfeg database job working directory metadata kwds historydatasetassociation tmp tmpduzfeg database job working directory metadata out historydatasetassociation tmp tmpduzfeg database job working directory metadata results historydatasetassociation tmp tmpduzfeg database files dataset dat tmp tmpduzfeg database job working directory metadata override historydatasetassociation sh c exit return code galaxy jobs runners local debug executing job script tmp tmpduzfeg database job working directory galaxy sh galaxy jobs debug persisting job destination destination id local galaxy jobs runners local debug execution finished tmp tmpduzfeg database job working directory galaxy sh galaxy jobs output checker info job fatal error exit code failing exit code galaxy jobs output checker debug tool exit code indicates an error failing job galaxy jobs debug setting dataset state to error galaxy jobs debug pausing job execution of this dataset s job is paused because its input datasets are in an error state galaxy jobs debug setting dataset state to error galaxy jobs debug setting dataset state to error galaxy jobs info collecting metrics for job galaxy jobs debug job ended finish executed in ms galaxy tools error reports debug bug report plugin generated response none galaxy tools error reports debug bug report plugin generated response none galaxy tools error reports debug bug report plugin generated response none galaxy tools debug validated and populated state for tool request ms galaxy tools actions info galaxy tools actions info galaxy managers collections debug created collection with elements galaxy tools actions info handled collection output named list output for tool job properties ms galaxy tools actions info handled output named out for tool job properties ms galaxy tools actions info added output datasets to history ms galaxy tools actions info verified access to datasets for job ms galaxy tools actions info job input hda remapped to new hda galaxy tools actions info job input hda remapped to new hda galaxy tools actions info setup for job job complete ready to flush ms galaxy tools actions info flushed transaction for job job ms galaxy tools execute debug tool created job ms galaxy tools execute debug executed job s for tool job properties request ms galaxy jobs debug working directory for job is tmp tmpduzfeg database job working directory galaxy jobs handler debug dispatching to local runner galaxy jobs debug persisting job destination destination id local galaxy jobs runners debug job queued ms galaxy jobs handler info job dispatched galaxy tools evaluation info updating param dict for one with tmp tmpduzfeg database files dataset dat galaxy tools evaluation info updating param dict for two with tmp tmpduzfeg database files dataset dat galaxy jobs command factory info built script for tool command galaxy jobs runners debug command is rm rf working mkdir p working cd working tmp tmpduzfeg database job working directory tool script sh return code cd tmp tmpduzfeg database job working directory galaxy virtual env galaxy virtual env galaxy setup environment true python tmp tmpduzfeg database job working directory set metadata py tmp tmpduzfeg database job working directory registry xml tmp tmpduzfeg database job working directory working galaxy json tmp tmpduzfeg database job working directory metadata in historydatasetassociation ybotud tmp tmpduzfeg database job working directory metadata kwds historydatasetassociation tmp tmpduzfeg database job working directory metadata out historydatasetassociation tmp tmpduzfeg database job working directory metadata results historydatasetassociation tmp tmpduzfeg database files dataset dat tmp tmpduzfeg database job working directory metadata override historydatasetassociation tmp tmpduzfeg database job working directory metadata in historydatasetassociation tmp tmpduzfeg database job working directory metadata kwds historydatasetassociation tmp tmpduzfeg database job working directory metadata out historydatasetassociation tmp tmpduzfeg database job working directory metadata results historydatasetassociation hanpoo tmp tmpduzfeg database files dataset dat tmp tmpduzfeg database job working directory metadata override historydatasetassociation dtwgqi tmp tmpduzfeg database job working directory metadata in historydatasetassociation tmp tmpduzfeg database job working directory metadata kwds historydatasetassociation tmp tmpduzfeg database job working directory metadata out historydatasetassociation tmp tmpduzfeg database job working directory metadata results historydatasetassociation tmp tmpduzfeg database files dataset dat tmp tmpduzfeg database job working directory metadata override historydatasetassociation sh c exit return code galaxy jobs runners local debug executing job script tmp tmpduzfeg database job working directory galaxy sh galaxy jobs debug persisting job destination destination id local galaxy jobs runners local debug execution finished tmp tmpduzfeg database job working directory galaxy sh galaxy model metadata debug loading metadata from file for historydatasetassociation galaxy model metadata debug loading metadata from file for historydatasetassociation galaxy model metadata debug loading metadata from file for historydatasetassociation galaxy jobs debug working directory for job is tmp tmpduzfeg database job working directory galaxy jobs handler debug dispatching to local runner galaxy jobs debug persisting job destination destination id local galaxy jobs runners debug job queued ms galaxy jobs handler info job dispatched galaxy jobs info collecting metrics for job galaxy jobs debug job ended finish executed in ms galaxy model metadata debug cleaning up external metadata files galaxy jobs command factory info built script for tool command galaxy jobs runners debug command is rm rf working mkdir p working cd working tmp tmpduzfeg database job working directory tool script sh return code if then cp tmp tmpduzfeg database job working directory working tmp tmpduzfeg database files dataset dat fi cd tmp tmpduzfeg database job working directory galaxy virtual env galaxy virtual env galaxy setup environment true python tmp tmpduzfeg database job working directory set metadata py tmp tmpduzfeg database job working directory registry xml tmp tmpduzfeg database job working directory working galaxy json tmp tmpduzfeg database job working directory metadata in historydatasetassociation hdwesk tmp tmpduzfeg database job working directory metadata kwds historydatasetassociation wvgeq tmp tmpduzfeg database job working directory metadata out historydatasetassociation ttnwmw tmp tmpduzfeg database job working directory metadata results historydatasetassociation tmp tmpduzfeg database files dataset dat tmp tmpduzfeg database job working directory metadata override historydatasetassociation v sh c exit return code galaxy jobs runners local debug executing job script tmp tmpduzfeg database job working directory galaxy sh galaxy jobs debug persisting job destination destination id local fail
1
44,396
12,130,738,698
IssuesEvent
2020-04-23 02:29:25
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
java.lang.IllegalStateException: Unknown protocol: CP2
Source: Community Team: Client Type: Defect
<!-- Thanks for reporting your issue. Please share with us the following information, to help us resolve your issue quickly and efficiently. --> **Describe the bug** Get error when using JHiptser with Hazelcast java.lang.IllegalStateException: Unknown protocol: CP2 **Expected behavior** **To Reproduce** **Additional context** <!-- Add any other context about the problem here. Common details that we're often interested in: - Detailed description of the steps to reproduce your issue - Logs and stack traces, if available - Hazelcast version that you use (e.g. 3.4, also specify whether it is a minor release or the latest snapshot) - If available, integration module versions (e.g. Tomcat, Jetty, Spring, Hibernate). Also, include their detailed configuration information such as web.xml, Hibernate configuration and `context.xml` for Spring - Cluster size, i.e. the number of Hazelcast cluster members - Number of the clients - Version of Java. It is also helpful to mention the JVM parameters - Operating system. If it is Linux, kernel version is helpful - Unit test with the `hazelcast.xml` file. If you could include a unit test which reproduces your issue, we would be grateful -->
1.0
java.lang.IllegalStateException: Unknown protocol: CP2 - <!-- Thanks for reporting your issue. Please share with us the following information, to help us resolve your issue quickly and efficiently. --> **Describe the bug** Get error when using JHiptser with Hazelcast java.lang.IllegalStateException: Unknown protocol: CP2 **Expected behavior** **To Reproduce** **Additional context** <!-- Add any other context about the problem here. Common details that we're often interested in: - Detailed description of the steps to reproduce your issue - Logs and stack traces, if available - Hazelcast version that you use (e.g. 3.4, also specify whether it is a minor release or the latest snapshot) - If available, integration module versions (e.g. Tomcat, Jetty, Spring, Hibernate). Also, include their detailed configuration information such as web.xml, Hibernate configuration and `context.xml` for Spring - Cluster size, i.e. the number of Hazelcast cluster members - Number of the clients - Version of Java. It is also helpful to mention the JVM parameters - Operating system. If it is Linux, kernel version is helpful - Unit test with the `hazelcast.xml` file. If you could include a unit test which reproduces your issue, we would be grateful -->
non_test
java lang illegalstateexception unknown protocol thanks for reporting your issue please share with us the following information to help us resolve your issue quickly and efficiently describe the bug get error when using jhiptser with hazelcast java lang illegalstateexception unknown protocol expected behavior to reproduce additional context add any other context about the problem here common details that we re often interested in detailed description of the steps to reproduce your issue logs and stack traces if available hazelcast version that you use e g also specify whether it is a minor release or the latest snapshot if available integration module versions e g tomcat jetty spring hibernate also include their detailed configuration information such as web xml hibernate configuration and context xml for spring cluster size i e the number of hazelcast cluster members number of the clients version of java it is also helpful to mention the jvm parameters operating system if it is linux kernel version is helpful unit test with the hazelcast xml file if you could include a unit test which reproduces your issue we would be grateful
0
264,944
23,145,082,067
IssuesEvent
2022-07-28 23:14:22
MPMG-DCC-UFMG/F01
https://api.github.com/repos/MPMG-DCC-UFMG/F01
closed
Teste de generalizacao para a tag Seridores - Registro por lotação - Divisa Alegre
generalization test development template-Síntese tecnologia informatica tag-Servidores subtag-Registro por lotação
DoD: Realizar o teste de Generalização do validador da tag Seridores - Registro por lotação para o Município de Divisa Alegre.
1.0
Teste de generalizacao para a tag Seridores - Registro por lotação - Divisa Alegre - DoD: Realizar o teste de Generalização do validador da tag Seridores - Registro por lotação para o Município de Divisa Alegre.
test
teste de generalizacao para a tag seridores registro por lotação divisa alegre dod realizar o teste de generalização do validador da tag seridores registro por lotação para o município de divisa alegre
1
296,823
25,578,155,836
IssuesEvent
2022-12-01 00:41:57
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[REMOTO]Analista de Desenvolvimento Front-end Pleno/Sênior na [D3Soft ]
FRONT-END BANCO DE DADOS MYSQL JAVASCRIPT CSS3 HTML SCRUM GIT REST SOAP BOOTSTRAP TESTE AUTOMATIZADO TESTE DE INTEGRAÇÃO NODE.JS POSTGRESQL UI/UX SASS REMOTO INGLÊS JASMINE SISTEMAS EM NUVEM Stale
<!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS! Use: "Desenvolvedor Front-end" ao invés de "Front-End Developer" \o/ Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]` ================================================== --> ## Analista de Desenvolvimento Front-end Pleno/Sênior -Estamos em busca de Analista de Desenvolvimento Front-end Pleno/Sênior ou Fullstack - Queremos um desenvolvedor que cumpra prazos de entregas, que saiba colaborar com a equipe, que possa assumir até se for o caso a posição de líder, seja atualizado e goste de aprender sempre. - Objetivo: Desenvolver Novos sistemas, Desenvolver e manter API em NodeJS, desenvolver sites, além de planejar, otimizar como também acrescentar novos recursos e muito mais. ## Local - Remoto ## Requisitos **Obrigatórios:** - Superior completo em Ciência da Computação, Eng da Computação, Sistema da Informação, Análise de Sistemas ou cursos similares (ou como tecnólogo) - Experiência com front-end, tendo trabalhado em aplicações client-side complexas e sistemas web responsivos; - Experiência com NodeJs, Javascript, ExpressJS, RequireJS, Sequelize, npm, HTML5, CSS3, SASS, Bootstrap 4 e GIT; Noções de UX/UI; - Automação de testes;Integração de serviços SOAP/REST; - Conhecer algum Banco de Dados preferencialmente MySQL ou PostGreeSQL - Ter atuado em times utilizando metodologias ágeis (SCRUM);Capacidade de trabalhar de forma independente (Ownership e proatividade); - Estabilidade nos últimos empregos; - Inglês técnico, com capacidade para leitura e interpretação de textos; **Diferenciais:** - Conhecimento em/na: Automação de testes utilizando o Jasmine;ExtJS; Sistemas reativos; Sistemas em nuvem. ## Contratação - PJ ## D3Soft - D3Soft é uma empresa que está crescendo a todo vapor, mas que precisa de novos colaboradores . Sempre buscamos dar ao nosso cliente soluções personalizadas e super criativas. ## Como se candidatar - LINK: https://www.recrutasimples.com.br/vaga/analista-de-desenvolvimento-front-end-plenosenior/natal-rn/-Lju_U8QXnIuRlWRO-r0?referrer=selfShareUrl
2.0
[REMOTO]Analista de Desenvolvimento Front-end Pleno/Sênior na [D3Soft ] - <!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS! Use: "Desenvolvedor Front-end" ao invés de "Front-End Developer" \o/ Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]` ================================================== --> ## Analista de Desenvolvimento Front-end Pleno/Sênior -Estamos em busca de Analista de Desenvolvimento Front-end Pleno/Sênior ou Fullstack - Queremos um desenvolvedor que cumpra prazos de entregas, que saiba colaborar com a equipe, que possa assumir até se for o caso a posição de líder, seja atualizado e goste de aprender sempre. - Objetivo: Desenvolver Novos sistemas, Desenvolver e manter API em NodeJS, desenvolver sites, além de planejar, otimizar como também acrescentar novos recursos e muito mais. ## Local - Remoto ## Requisitos **Obrigatórios:** - Superior completo em Ciência da Computação, Eng da Computação, Sistema da Informação, Análise de Sistemas ou cursos similares (ou como tecnólogo) - Experiência com front-end, tendo trabalhado em aplicações client-side complexas e sistemas web responsivos; - Experiência com NodeJs, Javascript, ExpressJS, RequireJS, Sequelize, npm, HTML5, CSS3, SASS, Bootstrap 4 e GIT; Noções de UX/UI; - Automação de testes;Integração de serviços SOAP/REST; - Conhecer algum Banco de Dados preferencialmente MySQL ou PostGreeSQL - Ter atuado em times utilizando metodologias ágeis (SCRUM);Capacidade de trabalhar de forma independente (Ownership e proatividade); - Estabilidade nos últimos empregos; - Inglês técnico, com capacidade para leitura e interpretação de textos; **Diferenciais:** - Conhecimento em/na: Automação de testes utilizando o Jasmine;ExtJS; Sistemas reativos; Sistemas em nuvem. ## Contratação - PJ ## D3Soft - D3Soft é uma empresa que está crescendo a todo vapor, mas que precisa de novos colaboradores . Sempre buscamos dar ao nosso cliente soluções personalizadas e super criativas. ## Como se candidatar - LINK: https://www.recrutasimples.com.br/vaga/analista-de-desenvolvimento-front-end-plenosenior/natal-rn/-Lju_U8QXnIuRlWRO-r0?referrer=selfShareUrl
test
analista de desenvolvimento front end pleno sênior na por favor só poste se a vaga for para salvador e cidades vizinhas use desenvolvedor front end ao invés de front end developer o exemplo desenvolvedor front end na analista de desenvolvimento front end pleno sênior estamos em busca de analista de desenvolvimento front end pleno sênior ou fullstack queremos um desenvolvedor que cumpra prazos de entregas que saiba colaborar com a equipe que possa assumir até se for o caso a posição de líder seja atualizado e goste de aprender sempre objetivo desenvolver novos sistemas desenvolver e manter api em nodejs desenvolver sites além de planejar otimizar como também acrescentar novos recursos e muito mais local remoto requisitos obrigatórios superior completo em ciência da computação eng da computação sistema da informação análise de sistemas ou cursos similares ou como tecnólogo experiência com front end tendo trabalhado em aplicações client side complexas e sistemas web responsivos experiência com nodejs javascript expressjs requirejs sequelize npm sass bootstrap e git noções de ux ui automação de testes integração de serviços soap rest conhecer algum banco de dados preferencialmente mysql ou postgreesql ter atuado em times utilizando metodologias ágeis scrum capacidade de trabalhar de forma independente ownership e proatividade estabilidade nos últimos empregos inglês técnico com capacidade para leitura e interpretação de textos diferenciais conhecimento em na automação de testes utilizando o jasmine extjs sistemas reativos sistemas em nuvem contratação pj é uma empresa que está crescendo a todo vapor mas que precisa de novos colaboradores sempre buscamos dar ao nosso cliente soluções personalizadas e super criativas como se candidatar link
1
163,633
12,737,302,877
IssuesEvent
2020-06-25 18:30:38
cornell-netlab/petr4
https://api.github.com/repos/cornell-netlab/petr4
opened
STF missing parameters for action in issue870_ebpf.p4
stf-testing
The P4 file has an action `pipe.Reject(IPv4Address add)` that updates the source IP of the packet before flagging it to be dropped. The STF file uses the action but doesn't pass any parameters to it, so the interpreter ends up crashing. ``` add pipe.Check_src_ip 0 key.field0:0x0a019846 pipe.Reject() ``` We should double check what the p4c/bmv2 does with this test... maybe I am misunderstanding the meaning of the STF here.
1.0
STF missing parameters for action in issue870_ebpf.p4 - The P4 file has an action `pipe.Reject(IPv4Address add)` that updates the source IP of the packet before flagging it to be dropped. The STF file uses the action but doesn't pass any parameters to it, so the interpreter ends up crashing. ``` add pipe.Check_src_ip 0 key.field0:0x0a019846 pipe.Reject() ``` We should double check what the p4c/bmv2 does with this test... maybe I am misunderstanding the meaning of the STF here.
test
stf missing parameters for action in ebpf the file has an action pipe reject add that updates the source ip of the packet before flagging it to be dropped the stf file uses the action but doesn t pass any parameters to it so the interpreter ends up crashing add pipe check src ip key pipe reject we should double check what the does with this test maybe i am misunderstanding the meaning of the stf here
1
22,434
11,733,005,558
IssuesEvent
2020-03-11 05:47:04
Azure/azure-cli
https://api.github.com/repos/Azure/azure-cli
closed
webapp: update_site_config() takes 2 positional arguments but 3 were given
Service Attention Web Apps
### **This is autogenerated. Please review and update as needed.** ## Describe the bug **Command Name** `az webapp create` **Errors:** ``` update_site_config() takes 2 positional arguments but 3 were given Traceback (most recent call last): python3.6/site-packages/knack/cli.py, ln 206, in invoke cmd_result = self.invocation.execute(args) cli/core/commands/__init__.py, ln 608, in execute raise ex cli/core/commands/__init__.py, ln 666, in _run_jobs_serially results.append(self._run_job(expanded_arg, cmd_copy)) cli/core/commands/__init__.py, ln 657, in _run_job cmd_copy.exception_handler(ex) ... cli/core/commands/__init__.py, ln 306, in __call__ return self.handler(*args, **kwargs) azure/cli/core/__init__.py, ln 493, in default_command_handler return op(**command_args) cli/command_modules/appservice/custom.py, ln 135, in create_webapp match['setter'](cmd, match, site_config) TypeError: update_site_config() takes 2 positional arguments but 3 were given ``` ## To Reproduce: Steps to reproduce the behavior. Note that argument values have been redacted, as they may contain sensitive information. - _Put any pre-requisite steps here..._ - `az webapp create --resource-group {} --plan {} --name {} --runtime {} --deployment-local-git` ## Expected Behavior ## Environment Summary ``` Linux-4.15.0-1071-azure-x86_64-with-debian-stretch-sid Python 3.6.5 azure-cli 2.1.0 ``` ## Additional Context <!--Please don't remove this:--> <!--auto-generated-->
1.0
webapp: update_site_config() takes 2 positional arguments but 3 were given - ### **This is autogenerated. Please review and update as needed.** ## Describe the bug **Command Name** `az webapp create` **Errors:** ``` update_site_config() takes 2 positional arguments but 3 were given Traceback (most recent call last): python3.6/site-packages/knack/cli.py, ln 206, in invoke cmd_result = self.invocation.execute(args) cli/core/commands/__init__.py, ln 608, in execute raise ex cli/core/commands/__init__.py, ln 666, in _run_jobs_serially results.append(self._run_job(expanded_arg, cmd_copy)) cli/core/commands/__init__.py, ln 657, in _run_job cmd_copy.exception_handler(ex) ... cli/core/commands/__init__.py, ln 306, in __call__ return self.handler(*args, **kwargs) azure/cli/core/__init__.py, ln 493, in default_command_handler return op(**command_args) cli/command_modules/appservice/custom.py, ln 135, in create_webapp match['setter'](cmd, match, site_config) TypeError: update_site_config() takes 2 positional arguments but 3 were given ``` ## To Reproduce: Steps to reproduce the behavior. Note that argument values have been redacted, as they may contain sensitive information. - _Put any pre-requisite steps here..._ - `az webapp create --resource-group {} --plan {} --name {} --runtime {} --deployment-local-git` ## Expected Behavior ## Environment Summary ``` Linux-4.15.0-1071-azure-x86_64-with-debian-stretch-sid Python 3.6.5 azure-cli 2.1.0 ``` ## Additional Context <!--Please don't remove this:--> <!--auto-generated-->
non_test
webapp update site config takes positional arguments but were given this is autogenerated please review and update as needed describe the bug command name az webapp create errors update site config takes positional arguments but were given traceback most recent call last site packages knack cli py ln in invoke cmd result self invocation execute args cli core commands init py ln in execute raise ex cli core commands init py ln in run jobs serially results append self run job expanded arg cmd copy cli core commands init py ln in run job cmd copy exception handler ex cli core commands init py ln in call return self handler args kwargs azure cli core init py ln in default command handler return op command args cli command modules appservice custom py ln in create webapp match cmd match site config typeerror update site config takes positional arguments but were given to reproduce steps to reproduce the behavior note that argument values have been redacted as they may contain sensitive information put any pre requisite steps here az webapp create resource group plan name runtime deployment local git expected behavior environment summary linux azure with debian stretch sid python azure cli additional context
0
248,819
26,840,051,363
IssuesEvent
2023-02-02 23:15:25
opensearch-project/documentation-website
https://api.github.com/repos/opensearch-project/documentation-website
closed
[DOC]Revise Security plugin Configuration TOC structure
security
**What do you want to do?** - [x] Request a change to existing documentation - [ ] Add new documentation - [ ] Report a technical problem with the documentation - [ ] Other **Tell us about your request.** Reorganize topics in the Security plugin Configuration section to improve logic and order. **What other resources are available?** Addresses #1036 for Security.
True
[DOC]Revise Security plugin Configuration TOC structure - **What do you want to do?** - [x] Request a change to existing documentation - [ ] Add new documentation - [ ] Report a technical problem with the documentation - [ ] Other **Tell us about your request.** Reorganize topics in the Security plugin Configuration section to improve logic and order. **What other resources are available?** Addresses #1036 for Security.
non_test
revise security plugin configuration toc structure what do you want to do request a change to existing documentation add new documentation report a technical problem with the documentation other tell us about your request reorganize topics in the security plugin configuration section to improve logic and order what other resources are available addresses for security
0
51,665
21,769,879,735
IssuesEvent
2022-05-13 08:00:53
Azure/azure-cli
https://api.github.com/repos/Azure/azure-cli
closed
WebApp:az webapp deployment (Max retries exceeded)
Web Apps Service Attention question customer-reported needs-author-feedback no-recent-activity
### **This is autogenerated. Please review and update as needed.** ## Describe the bug **Command Name** `az webapp deployment source config-zip` **Errors:** ``` The command failed with an unexpected error. Here is the traceback: HTTPSConnectionPool(host='imgapi100.scm.azurewebsites.net', port=443): Max retries exceeded with url: /api/zipdeploy?isAsync=true (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7fe2825b9588>: Failed to establish a new connection: [Errno -2] Name or service not known',)) Traceback (most recent call last): File "/opt/az/lib/python3.6/site-packages/urllib3/connection.py", line 170, in _new_conn (self._dns_host, self.port), self.timeout, **extra_kw File "/opt/az/lib/python3.6/site-packages/urllib3/util/connection.py", line 73, in create_connection for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM): File "/opt/az/lib/python3.6/socket.py", line 745, in getaddrinfo for res in _socket.getaddrinfo(host, port, family, type, proto, flags): socket.gaierror: [Errno -2] Name or service not known During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/opt/az/lib/python3.6/site-packages/urllib3/connectionpool.py", line 706, in urlopen chunked=chunked, File "/opt/az/lib/python3.6/site-packages/urllib3/connectionpool.py", line 382, in _make_request self._validate_conn(conn) File "/opt/az/lib/python3.6/site-packages/urllib3/connectionpool.py", line 1010, in _validate_conn conn.connect() File "/opt/az/lib/python3.6/site-packages/urllib3/connection.py", line 353, in connect conn = self._new_conn() File "/opt/az/lib/python3.6/site-packages/urllib3/connection.py", line 182, in _new_conn self, "Failed to establish a new connection: %s" % e urllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7fe2825b9588>: Failed to establish a new connection: [Errno -2] Name or service not known During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/opt/az/lib/python3.6/site-packages/requests/adapters.py", line 449, in send timeout=timeout File "/opt/az/lib/python3.6/si
1.0
WebApp:az webapp deployment (Max retries exceeded) - ### **This is autogenerated. Please review and update as needed.** ## Describe the bug **Command Name** `az webapp deployment source config-zip` **Errors:** ``` The command failed with an unexpected error. Here is the traceback: HTTPSConnectionPool(host='imgapi100.scm.azurewebsites.net', port=443): Max retries exceeded with url: /api/zipdeploy?isAsync=true (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7fe2825b9588>: Failed to establish a new connection: [Errno -2] Name or service not known',)) Traceback (most recent call last): File "/opt/az/lib/python3.6/site-packages/urllib3/connection.py", line 170, in _new_conn (self._dns_host, self.port), self.timeout, **extra_kw File "/opt/az/lib/python3.6/site-packages/urllib3/util/connection.py", line 73, in create_connection for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM): File "/opt/az/lib/python3.6/socket.py", line 745, in getaddrinfo for res in _socket.getaddrinfo(host, port, family, type, proto, flags): socket.gaierror: [Errno -2] Name or service not known During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/opt/az/lib/python3.6/site-packages/urllib3/connectionpool.py", line 706, in urlopen chunked=chunked, File "/opt/az/lib/python3.6/site-packages/urllib3/connectionpool.py", line 382, in _make_request self._validate_conn(conn) File "/opt/az/lib/python3.6/site-packages/urllib3/connectionpool.py", line 1010, in _validate_conn conn.connect() File "/opt/az/lib/python3.6/site-packages/urllib3/connection.py", line 353, in connect conn = self._new_conn() File "/opt/az/lib/python3.6/site-packages/urllib3/connection.py", line 182, in _new_conn self, "Failed to establish a new connection: %s" % e urllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7fe2825b9588>: Failed to establish a new connection: [Errno -2] Name or service not known During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/opt/az/lib/python3.6/site-packages/requests/adapters.py", line 449, in send timeout=timeout File "/opt/az/lib/python3.6/si
non_test
webapp az webapp deployment max retries exceeded this is autogenerated please review and update as needed describe the bug command name az webapp deployment source config zip errors the command failed with an unexpected error here is the traceback httpsconnectionpool host scm azurewebsites net port max retries exceeded with url api zipdeploy isasync true caused by newconnectionerror failed to establish a new connection name or service not known traceback most recent call last file opt az lib site packages connection py line in new conn self dns host self port self timeout extra kw file opt az lib site packages util connection py line in create connection for res in socket getaddrinfo host port family socket sock stream file opt az lib socket py line in getaddrinfo for res in socket getaddrinfo host port family type proto flags socket gaierror name or service not known during handling of the above exception another exception occurred traceback most recent call last file opt az lib site packages connectionpool py line in urlopen chunked chunked file opt az lib site packages connectionpool py line in make request self validate conn conn file opt az lib site packages connectionpool py line in validate conn conn connect file opt az lib site packages connection py line in connect conn self new conn file opt az lib site packages connection py line in new conn self failed to establish a new connection s e exceptions newconnectionerror failed to establish a new connection name or service not known during handling of the above exception another exception occurred traceback most recent call last file opt az lib site packages requests adapters py line in send timeout timeout file opt az lib si
0
126,357
10,419,505,277
IssuesEvent
2019-09-15 17:02:07
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Rewards panel shows white popup
QA/Test-Plan-Specified QA/Yes bug feature/rewards release/blocking
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> Discussed with @NejcZdovc over the call, this issue is slightly different from the issue #5336, The white popup issue gets resolved if we restart the browser/profile where we encountered the BR panel white popup issue. Hence opened the new issue ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> STR1: 1. Clean profile 0.68.132 with staging flags 2. Enable rewards, add a site to a-c table and complete the auto contribution 3. Upgrade to 0.68.135 4. Close the brave by selecting `Exit` from the hamburger menu 5. Restart the browser using staging flags 6. Click on BR panel shows the white popup 6. Add a site to a-c table and complete the auto contribution in 0.68.135 7. Click on BR panel unable to see a-c summary in BR panel due to white popup 8. Repeat steps 4 and 5 9. BR panel white popup issue resolves, Join rewards popup displayed STR2: 1. Clean profile 0.68.135 2. Join rewards from the panel but don't visit brave://rewards page. 3. Close Brave by selecting `Exit` from the hamburger menu 4. Restart the browser 5. Click on panel. 6. The panel shows white popup 7. Repeat the steps 3 and 4 8. BR panel white popup issue resolves, Join rewards popup displayed STR 3: 1. Clean profile 0.68.135 2. Enable rewards and restore a key which has the balance 3. Send a one time tip to verified pubs 4. Keep Rewards page open and in focus. Close the brave by selecting `Exit` from the hamburger menu 5. Restart the browser 6. Tip data reset to zero in rewards summary section 7. Click on panel. 8. The error message is shown over wallet balance on brave://rewards and white box for the panel. Error message goes away, but the white box on the panel is permanent and cannot be recovered from. 9. Close the browser by clicking on Exit in the hamburger menu 10. Restart the browser 11. Click on panel 12 BR panel white popup issue resolves, Join rewards popup displayed ## Actual result: <!--Please add screenshots if needed--> Rewards popup shows white popup, restart the browser resolves the issue and displayed Join rewards popup Recorded files: https://drive.google.com/open?id=1k6c-Dm9eSlTb50SvO4VN_otIfLMuwCu2 https://drive.google.com/open?id=1tz6CeS9Wxtt5OGWGix5HoucpKOzIOS0b https://drive.google.com/open?id=1jqc0kacHipgOE_MCAOShPMkWiZmJeZOi ## Expected result: White popup shouldn't be shown in the BR panel ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Always ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 0.68.135 Chromium: 77.0.3865.75 (Official Build) (64-bit) -- | -- Revision | 201e747d032611c5f2785cae06e894cf85be7f8a-refs/branch-heads/3865@{#776} OS | Windows 10 OS Version 1803 (Build 17134.523) ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? No - Can you reproduce this issue with the beta channel? No - Can you reproduce this issue with the dev channel? No - Can you reproduce this issue with the nightly channel? No ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? NA - Does the issue resolve itself when disabling Brave Rewards? NA - Is the issue reproducible on the latest version of Chrome? NA ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> @NejcZdovc @brave/legacy_qa
1.0
Rewards panel shows white popup - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> Discussed with @NejcZdovc over the call, this issue is slightly different from the issue #5336, The white popup issue gets resolved if we restart the browser/profile where we encountered the BR panel white popup issue. Hence opened the new issue ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> STR1: 1. Clean profile 0.68.132 with staging flags 2. Enable rewards, add a site to a-c table and complete the auto contribution 3. Upgrade to 0.68.135 4. Close the brave by selecting `Exit` from the hamburger menu 5. Restart the browser using staging flags 6. Click on BR panel shows the white popup 6. Add a site to a-c table and complete the auto contribution in 0.68.135 7. Click on BR panel unable to see a-c summary in BR panel due to white popup 8. Repeat steps 4 and 5 9. BR panel white popup issue resolves, Join rewards popup displayed STR2: 1. Clean profile 0.68.135 2. Join rewards from the panel but don't visit brave://rewards page. 3. Close Brave by selecting `Exit` from the hamburger menu 4. Restart the browser 5. Click on panel. 6. The panel shows white popup 7. Repeat the steps 3 and 4 8. BR panel white popup issue resolves, Join rewards popup displayed STR 3: 1. Clean profile 0.68.135 2. Enable rewards and restore a key which has the balance 3. Send a one time tip to verified pubs 4. Keep Rewards page open and in focus. Close the brave by selecting `Exit` from the hamburger menu 5. Restart the browser 6. Tip data reset to zero in rewards summary section 7. Click on panel. 8. The error message is shown over wallet balance on brave://rewards and white box for the panel. Error message goes away, but the white box on the panel is permanent and cannot be recovered from. 9. Close the browser by clicking on Exit in the hamburger menu 10. Restart the browser 11. Click on panel 12 BR panel white popup issue resolves, Join rewards popup displayed ## Actual result: <!--Please add screenshots if needed--> Rewards popup shows white popup, restart the browser resolves the issue and displayed Join rewards popup Recorded files: https://drive.google.com/open?id=1k6c-Dm9eSlTb50SvO4VN_otIfLMuwCu2 https://drive.google.com/open?id=1tz6CeS9Wxtt5OGWGix5HoucpKOzIOS0b https://drive.google.com/open?id=1jqc0kacHipgOE_MCAOShPMkWiZmJeZOi ## Expected result: White popup shouldn't be shown in the BR panel ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Always ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 0.68.135 Chromium: 77.0.3865.75 (Official Build) (64-bit) -- | -- Revision | 201e747d032611c5f2785cae06e894cf85be7f8a-refs/branch-heads/3865@{#776} OS | Windows 10 OS Version 1803 (Build 17134.523) ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? No - Can you reproduce this issue with the beta channel? No - Can you reproduce this issue with the dev channel? No - Can you reproduce this issue with the nightly channel? No ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? NA - Does the issue resolve itself when disabling Brave Rewards? NA - Is the issue reproducible on the latest version of Chrome? NA ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> @NejcZdovc @brave/legacy_qa
test
rewards panel shows white popup have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description discussed with nejczdovc over the call this issue is slightly different from the issue the white popup issue gets resolved if we restart the browser profile where we encountered the br panel white popup issue hence opened the new issue steps to reproduce clean profile with staging flags enable rewards add a site to a c table and complete the auto contribution upgrade to close the brave by selecting exit from the hamburger menu restart the browser using staging flags click on br panel shows the white popup add a site to a c table and complete the auto contribution in click on br panel unable to see a c summary in br panel due to white popup repeat steps and br panel white popup issue resolves join rewards popup displayed clean profile join rewards from the panel but don t visit brave rewards page close brave by selecting exit from the hamburger menu restart the browser click on panel the panel shows white popup repeat the steps and br panel white popup issue resolves join rewards popup displayed str clean profile enable rewards and restore a key which has the balance send a one time tip to verified pubs keep rewards page open and in focus close the brave by selecting exit from the hamburger menu restart the browser tip data reset to zero in rewards summary section click on panel the error message is shown over wallet balance on brave rewards and white box for the panel error message goes away but the white box on the panel is permanent and cannot be recovered from close the browser by clicking on exit in the hamburger menu restart the browser click on panel br panel white popup issue resolves join rewards popup displayed actual result rewards popup shows white popup restart the browser resolves the issue and displayed join rewards popup recorded files expected result white popup shouldn t be shown in the br panel reproduces how often always brave version brave version info brave chromium   official build   bit revision refs branch heads os windows  os version build version channel information can you reproduce this issue with the current release no can you reproduce this issue with the beta channel no can you reproduce this issue with the dev channel no can you reproduce this issue with the nightly channel no other additional information does the issue resolve itself when disabling brave shields na does the issue resolve itself when disabling brave rewards na is the issue reproducible on the latest version of chrome na miscellaneous information nejczdovc brave legacy qa
1
214,300
16,580,355,616
IssuesEvent
2021-05-31 10:54:25
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
opened
[CI] timeout in netty4.ESLoggingHandlerIT
:Distributed/Network >test-failure
`netty4.ESLoggingHandlerIT` times out, in a addition the log shows an OOM which might be the cause or a follow up error. Does not reproduce locally **Build scan**: https://gradle-enterprise.elastic.co/s/olvxkhfaegvww **Repro line**: `./gradlew ':modules:transport-netty4:internalClusterTest' --tests "org.elasticsearch.transport.netty4.ESLoggingHandlerIT.testLoggingHandler" -Dtests.seed=682BB49420EB9F5F -Dtests.security.manager=true` **Reproduces locally?**: no **Applicable branches**: 7.13 **Failure excerpt**: ``` 10:03:55 > Task :modules:transport-netty4:internalClusterTest 10:03:55 10:03:55 10:03:55 org.elasticsearch.transport.netty4.ESLoggingHandlerIT > testLoggingHandler FAILED 10:03:55 java.lang.Exception: Test abandoned because suite timeout was reached. 10:03:55 at __randomizedtesting.SeedInfo.seed([682BB49420EB9F5F]:0) 10:03:58 10:03:58 > Task :x-pack:plugin:fleet:javaRestTest 10:03:59 10:03:59 > Task :x-pack:plugin:repositories-metering-api:test 10:03:59 Picked up JAVA_TOOL_OPTIONS: -Dfile.encoding=UTF8 10:03:59 10:03:59 > Task :x-pack:plugin:ql:test 10:03:59 Picked up JAVA_TOOL_OPTIONS: -Dfile.encoding=UTF8 10:03:59 Picked up JAVA_TOOL_OPTIONS: -Dfile.encoding=UTF8 10:03:59 Picked up JAVA_TOOL_OPTIONS: -Dfile.encoding=UTF8 10:04:01 10:04:01 > Task :x-pack:plugin:monitoring:test 10:04:01 Picked up JAVA_TOOL_OPTIONS: -Dfile.encoding=UTF8 10:04:04 10:04:04 > Task :qa:translog-policy:v6.8.17#Step3NewClusterTest 10:04:04 > Task :x-pack:plugin:identity-provider:qa:idp-rest-tests:javaRestTest 10:04:11 10:04:11 Exception: java.lang.OutOfMemoryError thrown from the UncaughtExceptionHandler in thread "pool-1-thread-1" 10:04:20 10:04:20 Exception: java.lang.OutOfMemoryError thrown from the UncaughtExceptionHandler in thread "main" ```
1.0
[CI] timeout in netty4.ESLoggingHandlerIT - `netty4.ESLoggingHandlerIT` times out, in a addition the log shows an OOM which might be the cause or a follow up error. Does not reproduce locally **Build scan**: https://gradle-enterprise.elastic.co/s/olvxkhfaegvww **Repro line**: `./gradlew ':modules:transport-netty4:internalClusterTest' --tests "org.elasticsearch.transport.netty4.ESLoggingHandlerIT.testLoggingHandler" -Dtests.seed=682BB49420EB9F5F -Dtests.security.manager=true` **Reproduces locally?**: no **Applicable branches**: 7.13 **Failure excerpt**: ``` 10:03:55 > Task :modules:transport-netty4:internalClusterTest 10:03:55 10:03:55 10:03:55 org.elasticsearch.transport.netty4.ESLoggingHandlerIT > testLoggingHandler FAILED 10:03:55 java.lang.Exception: Test abandoned because suite timeout was reached. 10:03:55 at __randomizedtesting.SeedInfo.seed([682BB49420EB9F5F]:0) 10:03:58 10:03:58 > Task :x-pack:plugin:fleet:javaRestTest 10:03:59 10:03:59 > Task :x-pack:plugin:repositories-metering-api:test 10:03:59 Picked up JAVA_TOOL_OPTIONS: -Dfile.encoding=UTF8 10:03:59 10:03:59 > Task :x-pack:plugin:ql:test 10:03:59 Picked up JAVA_TOOL_OPTIONS: -Dfile.encoding=UTF8 10:03:59 Picked up JAVA_TOOL_OPTIONS: -Dfile.encoding=UTF8 10:03:59 Picked up JAVA_TOOL_OPTIONS: -Dfile.encoding=UTF8 10:04:01 10:04:01 > Task :x-pack:plugin:monitoring:test 10:04:01 Picked up JAVA_TOOL_OPTIONS: -Dfile.encoding=UTF8 10:04:04 10:04:04 > Task :qa:translog-policy:v6.8.17#Step3NewClusterTest 10:04:04 > Task :x-pack:plugin:identity-provider:qa:idp-rest-tests:javaRestTest 10:04:11 10:04:11 Exception: java.lang.OutOfMemoryError thrown from the UncaughtExceptionHandler in thread "pool-1-thread-1" 10:04:20 10:04:20 Exception: java.lang.OutOfMemoryError thrown from the UncaughtExceptionHandler in thread "main" ```
test
timeout in eslogginghandlerit eslogginghandlerit times out in a addition the log shows an oom which might be the cause or a follow up error does not reproduce locally build scan repro line gradlew modules transport internalclustertest tests org elasticsearch transport eslogginghandlerit testlogginghandler dtests seed dtests security manager true reproduces locally no applicable branches failure excerpt task modules transport internalclustertest org elasticsearch transport eslogginghandlerit testlogginghandler failed java lang exception test abandoned because suite timeout was reached at randomizedtesting seedinfo seed task x pack plugin fleet javaresttest task x pack plugin repositories metering api test picked up java tool options dfile encoding task x pack plugin ql test picked up java tool options dfile encoding picked up java tool options dfile encoding picked up java tool options dfile encoding task x pack plugin monitoring test picked up java tool options dfile encoding task qa translog policy task x pack plugin identity provider qa idp rest tests javaresttest exception java lang outofmemoryerror thrown from the uncaughtexceptionhandler in thread pool thread exception java lang outofmemoryerror thrown from the uncaughtexceptionhandler in thread main
1
201,652
15,806,369,412
IssuesEvent
2021-04-04 04:49:19
sqlalchemy/sqlalchemy
https://api.github.com/repos/sqlalchemy/sqlalchemy
closed
insert_primary_key_rows does not return primary keys of inserted rows.
documentation
**Describe the bug** After multiple insertions at same time, `insert_primary_key_rows` return `[(None,), (None,)]`. The primary key of each inserted row is expected. The code works as expected with on insertion. `insert_primary_key` has the pk of the inserted row. but it does work with multiples insertions. <!-- A clear and concise description of what the bug is. --> **Expected behavior** `[(1,), (2,)]` expected as it's an in memory database just created, and it's the first insertions. <!-- A clear and concise description of what you expected to happen. --> **To Reproduce** ```py from sqlalchemy import MetaData from sqlalchemy import create_engine from sqlalchemy import Table, Column, Integer, String from sqlalchemy import insert, select engine = create_engine("sqlite+pysqlite:///:memory:", echo=True, future=True) metadata = MetaData() user_table = Table( "user_account", metadata, Column('id', Integer, primary_key=True), Column('name', String(30)), Column('fullname', String) ) metadata.create_all(engine) # Insertion of 2 rows at same time. with engine.connect() as conn: result = conn.execute( insert(user_table), [ {"name": "sandy", "fullname": "sandy test"}, {"name": "tata", "fullname": "tata test"}, ] ) conn.commit() print(result.inserted_primary_key_rows) # Insertion of 1 row with engine.connect() as conn: result = conn.execute( insert(user_table), [ {"name": "ouba", "fullname": "ouba test"}, ] ) conn.commit() print(result.inserted_primary_key_rows) with engine.connect() as conn: result = conn.execute(select(user_table)) for row in result: print(row) ``` **Error** ``` 2021-04-03 21:53:50,895 INFO sqlalchemy.engine.Engine BEGIN (implicit) 2021-04-03 21:53:50,895 INFO sqlalchemy.engine.Engine PRAGMA main.table_info("user_account") 2021-04-03 21:53:50,895 INFO sqlalchemy.engine.Engine [raw sql] () 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine PRAGMA temp.table_info("user_account") 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine [raw sql] () 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine CREATE TABLE user_account ( id INTEGER NOT NULL, name VARCHAR(30), fullname VARCHAR, PRIMARY KEY (id) ) 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine [no key 0.00013s] () 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine COMMIT 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine BEGIN (implicit) 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine INSERT INTO user_account (name, fullname) VALUES (?, ?) 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine [generated in 0.00024s] (('sandy', 'sandy test'), ('tata', 'tata test')) 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine COMMIT [(None,), (None,)] 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine BEGIN (implicit) 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine INSERT INTO user_account (name, fullname) VALUES (?, ?) 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine [generated in 0.00020s] ('ouba', 'ouba test') 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine COMMIT [(3,)] 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine BEGIN (implicit) 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine SELECT user_account.id, user_account.name, user_account.fullname FROM user_account 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine [generated in 0.00023s] () (1, 'sandy', 'sandy test') (2, 'tata', 'tata test') (3, 'ouba', 'ouba test') 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine ROLLBACK ``` **Versions.** - OS: Windows-10-10.0.19041-SP0 - Python: Python 3.8.8 (tags/v3.8.8:024d805, Feb 19 2021, 13:18:16) [MSC v.1928 64 bit (AMD64)] on win32 - SQLAlchemy: 1.4.5 - Database: sqlite+pysqlite:///:memory - DBAPI: sqlite3 2.6.0 **Have a nice day!**
1.0
insert_primary_key_rows does not return primary keys of inserted rows. - **Describe the bug** After multiple insertions at same time, `insert_primary_key_rows` return `[(None,), (None,)]`. The primary key of each inserted row is expected. The code works as expected with on insertion. `insert_primary_key` has the pk of the inserted row. but it does work with multiples insertions. <!-- A clear and concise description of what the bug is. --> **Expected behavior** `[(1,), (2,)]` expected as it's an in memory database just created, and it's the first insertions. <!-- A clear and concise description of what you expected to happen. --> **To Reproduce** ```py from sqlalchemy import MetaData from sqlalchemy import create_engine from sqlalchemy import Table, Column, Integer, String from sqlalchemy import insert, select engine = create_engine("sqlite+pysqlite:///:memory:", echo=True, future=True) metadata = MetaData() user_table = Table( "user_account", metadata, Column('id', Integer, primary_key=True), Column('name', String(30)), Column('fullname', String) ) metadata.create_all(engine) # Insertion of 2 rows at same time. with engine.connect() as conn: result = conn.execute( insert(user_table), [ {"name": "sandy", "fullname": "sandy test"}, {"name": "tata", "fullname": "tata test"}, ] ) conn.commit() print(result.inserted_primary_key_rows) # Insertion of 1 row with engine.connect() as conn: result = conn.execute( insert(user_table), [ {"name": "ouba", "fullname": "ouba test"}, ] ) conn.commit() print(result.inserted_primary_key_rows) with engine.connect() as conn: result = conn.execute(select(user_table)) for row in result: print(row) ``` **Error** ``` 2021-04-03 21:53:50,895 INFO sqlalchemy.engine.Engine BEGIN (implicit) 2021-04-03 21:53:50,895 INFO sqlalchemy.engine.Engine PRAGMA main.table_info("user_account") 2021-04-03 21:53:50,895 INFO sqlalchemy.engine.Engine [raw sql] () 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine PRAGMA temp.table_info("user_account") 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine [raw sql] () 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine CREATE TABLE user_account ( id INTEGER NOT NULL, name VARCHAR(30), fullname VARCHAR, PRIMARY KEY (id) ) 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine [no key 0.00013s] () 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine COMMIT 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine BEGIN (implicit) 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine INSERT INTO user_account (name, fullname) VALUES (?, ?) 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine [generated in 0.00024s] (('sandy', 'sandy test'), ('tata', 'tata test')) 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine COMMIT [(None,), (None,)] 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine BEGIN (implicit) 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine INSERT INTO user_account (name, fullname) VALUES (?, ?) 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine [generated in 0.00020s] ('ouba', 'ouba test') 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine COMMIT [(3,)] 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine BEGIN (implicit) 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine SELECT user_account.id, user_account.name, user_account.fullname FROM user_account 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine [generated in 0.00023s] () (1, 'sandy', 'sandy test') (2, 'tata', 'tata test') (3, 'ouba', 'ouba test') 2021-04-03 21:53:50,908 INFO sqlalchemy.engine.Engine ROLLBACK ``` **Versions.** - OS: Windows-10-10.0.19041-SP0 - Python: Python 3.8.8 (tags/v3.8.8:024d805, Feb 19 2021, 13:18:16) [MSC v.1928 64 bit (AMD64)] on win32 - SQLAlchemy: 1.4.5 - Database: sqlite+pysqlite:///:memory - DBAPI: sqlite3 2.6.0 **Have a nice day!**
non_test
insert primary key rows does not return primary keys of inserted rows describe the bug after multiple insertions at same time insert primary key rows return the primary key of each inserted row is expected the code works as expected with on insertion insert primary key has the pk of the inserted row but it does work with multiples insertions expected behavior expected as it s an in memory database just created and it s the first insertions to reproduce py from sqlalchemy import metadata from sqlalchemy import create engine from sqlalchemy import table column integer string from sqlalchemy import insert select engine create engine sqlite pysqlite memory echo true future true metadata metadata user table table user account metadata column id integer primary key true column name string column fullname string metadata create all engine insertion of rows at same time with engine connect as conn result conn execute insert user table name sandy fullname sandy test name tata fullname tata test conn commit print result inserted primary key rows insertion of row with engine connect as conn result conn execute insert user table name ouba fullname ouba test conn commit print result inserted primary key rows with engine connect as conn result conn execute select user table for row in result print row error info sqlalchemy engine engine begin implicit info sqlalchemy engine engine pragma main table info user account info sqlalchemy engine engine info sqlalchemy engine engine pragma temp table info user account info sqlalchemy engine engine info sqlalchemy engine engine create table user account id integer not null name varchar fullname varchar primary key id info sqlalchemy engine engine info sqlalchemy engine engine commit info sqlalchemy engine engine begin implicit info sqlalchemy engine engine insert into user account name fullname values info sqlalchemy engine engine sandy sandy test tata tata test info sqlalchemy engine engine commit info sqlalchemy engine engine begin implicit info sqlalchemy engine engine insert into user account name fullname values info sqlalchemy engine engine ouba ouba test info sqlalchemy engine engine commit info sqlalchemy engine engine begin implicit info sqlalchemy engine engine select user account id user account name user account fullname from user account info sqlalchemy engine engine sandy sandy test tata tata test ouba ouba test info sqlalchemy engine engine rollback versions os windows python python tags feb on sqlalchemy database sqlite pysqlite memory dbapi have a nice day
0
193,241
14,644,106,111
IssuesEvent
2020-12-25 20:59:04
github-vet/rangeloop-pointer-findings
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
closed
hongyuanyang-uu/cfssl: bundler/bundle_from_remote_test.go; 14 LoC
fresh small test
Found a possible issue in [hongyuanyang-uu/cfssl](https://www.github.com/hongyuanyang-uu/cfssl) at [bundler/bundle_from_remote_test.go](https://github.com/hongyuanyang-uu/cfssl/blob/4b0e85c9fd6872261393aca42703b9002f78830d/bundler/bundle_from_remote_test.go#L167-L180) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to test at line 171 may start a goroutine [Click here to see the code in its original context.](https://github.com/hongyuanyang-uu/cfssl/blob/4b0e85c9fd6872261393aca42703b9002f78830d/bundler/bundle_from_remote_test.go#L167-L180) <details> <summary>Click here to show the 14 line(s) of Go which triggered the analyzer.</summary> ```go for _, test := range remoteSNITests { b := test.bundlerConstructor(t) bundle, err := b.BundleFromRemote(test.hostname, test.ip, bf) if test.errorCallback != nil { test.errorCallback(t, &test, err) } else { if err != nil { t.Errorf("expected no error. but an error occurred: %s", err.Error()) } if test.bundleCallback != nil { test.bundleCallback(t, &test, bundle) } } } ``` </details> <details> <summary>Click here to show extra information the analyzer produced.</summary> ``` The following dot graph describes paths through the callgraph that could lead to a function calling a goroutine: no paths found; call may have ended in third-party code; stay tuned for diagnostics ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 4b0e85c9fd6872261393aca42703b9002f78830d
1.0
hongyuanyang-uu/cfssl: bundler/bundle_from_remote_test.go; 14 LoC - Found a possible issue in [hongyuanyang-uu/cfssl](https://www.github.com/hongyuanyang-uu/cfssl) at [bundler/bundle_from_remote_test.go](https://github.com/hongyuanyang-uu/cfssl/blob/4b0e85c9fd6872261393aca42703b9002f78830d/bundler/bundle_from_remote_test.go#L167-L180) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to test at line 171 may start a goroutine [Click here to see the code in its original context.](https://github.com/hongyuanyang-uu/cfssl/blob/4b0e85c9fd6872261393aca42703b9002f78830d/bundler/bundle_from_remote_test.go#L167-L180) <details> <summary>Click here to show the 14 line(s) of Go which triggered the analyzer.</summary> ```go for _, test := range remoteSNITests { b := test.bundlerConstructor(t) bundle, err := b.BundleFromRemote(test.hostname, test.ip, bf) if test.errorCallback != nil { test.errorCallback(t, &test, err) } else { if err != nil { t.Errorf("expected no error. but an error occurred: %s", err.Error()) } if test.bundleCallback != nil { test.bundleCallback(t, &test, bundle) } } } ``` </details> <details> <summary>Click here to show extra information the analyzer produced.</summary> ``` The following dot graph describes paths through the callgraph that could lead to a function calling a goroutine: no paths found; call may have ended in third-party code; stay tuned for diagnostics ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 4b0e85c9fd6872261393aca42703b9002f78830d
test
hongyuanyang uu cfssl bundler bundle from remote test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to test at line may start a goroutine click here to show the line s of go which triggered the analyzer go for test range remotesnitests b test bundlerconstructor t bundle err b bundlefromremote test hostname test ip bf if test errorcallback nil test errorcallback t test err else if err nil t errorf expected no error but an error occurred s err error if test bundlecallback nil test bundlecallback t test bundle click here to show extra information the analyzer produced the following dot graph describes paths through the callgraph that could lead to a function calling a goroutine no paths found call may have ended in third party code stay tuned for diagnostics leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
299,193
25,889,500,253
IssuesEvent
2022-12-14 16:49:52
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: sqlsmith/setup=seed/setting=no-mutations failed
C-test-failure O-robot O-roachtest branch-master T-sql-queries
roachtest.sqlsmith/setup=seed/setting=no-mutations [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7952891?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7952891?buildTab=artifacts#/sqlsmith/setup=seed/setting=no-mutations) on master @ [052acc88ad9d7296ce6b8b441627fb469cc74d95](https://github.com/cockroachdb/cockroach/commits/052acc88ad9d7296ce6b8b441627fb469cc74d95): ``` test artifacts and logs in: /artifacts/sqlsmith/setup=seed/setting=no-mutations/run_1 (test_impl.go:297).Fatalf: error: pq: internal error: comparison overload not found (is, void, unknown) stmt: WITH with_733 (col_4358) AS (SELECT * FROM (VALUES ('':::VOID), (NULL), ('':::VOID)) AS tab_1801 (col_4358)), with_734 (col_4359) AS (SELECT * FROM (VALUES ('u':::STRING), ('8':::STRING)) AS tab_1802 (col_4359)), with_735 (col_4360) AS ( SELECT * FROM ( VALUES ('-63 years -5 mons -584 days -17:53:01.793462':::INTERVAL), ('-59 years -5 mons -454 days -08:42:58.54816':::INTERVAL), ('63 years 659 days 13:15:59.580449':::INTERVAL) ) AS tab_1803 (col_4360) ) SELECT NULL AS col_4361 FROM with_733 AS cte_ref_227 WHERE NULL ORDER BY cte_ref_227.col_4358 DESC, cte_ref_227.col_4358 DESC NULLS FIRST, cte_ref_227.col_4358 ASC, cte_ref_227.col_4358 DESC, cte_ref_227.col_4358 DESC NULLS FIRST LIMIT 3:::INT8; ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sqlsmith/setup=seed/setting=no-mutations.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-22413
2.0
roachtest: sqlsmith/setup=seed/setting=no-mutations failed - roachtest.sqlsmith/setup=seed/setting=no-mutations [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7952891?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7952891?buildTab=artifacts#/sqlsmith/setup=seed/setting=no-mutations) on master @ [052acc88ad9d7296ce6b8b441627fb469cc74d95](https://github.com/cockroachdb/cockroach/commits/052acc88ad9d7296ce6b8b441627fb469cc74d95): ``` test artifacts and logs in: /artifacts/sqlsmith/setup=seed/setting=no-mutations/run_1 (test_impl.go:297).Fatalf: error: pq: internal error: comparison overload not found (is, void, unknown) stmt: WITH with_733 (col_4358) AS (SELECT * FROM (VALUES ('':::VOID), (NULL), ('':::VOID)) AS tab_1801 (col_4358)), with_734 (col_4359) AS (SELECT * FROM (VALUES ('u':::STRING), ('8':::STRING)) AS tab_1802 (col_4359)), with_735 (col_4360) AS ( SELECT * FROM ( VALUES ('-63 years -5 mons -584 days -17:53:01.793462':::INTERVAL), ('-59 years -5 mons -454 days -08:42:58.54816':::INTERVAL), ('63 years 659 days 13:15:59.580449':::INTERVAL) ) AS tab_1803 (col_4360) ) SELECT NULL AS col_4361 FROM with_733 AS cte_ref_227 WHERE NULL ORDER BY cte_ref_227.col_4358 DESC, cte_ref_227.col_4358 DESC NULLS FIRST, cte_ref_227.col_4358 ASC, cte_ref_227.col_4358 DESC, cte_ref_227.col_4358 DESC NULLS FIRST LIMIT 3:::INT8; ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sqlsmith/setup=seed/setting=no-mutations.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-22413
test
roachtest sqlsmith setup seed setting no mutations failed roachtest sqlsmith setup seed setting no mutations with on master test artifacts and logs in artifacts sqlsmith setup seed setting no mutations run test impl go fatalf error pq internal error comparison overload not found is void unknown stmt with with col as select from values void null void as tab col with col as select from values u string string as tab col with col as select from values years mons days interval years mons days interval years days interval as tab col select null as col from with as cte ref where null order by cte ref col desc cte ref col desc nulls first cte ref col asc cte ref col desc cte ref col desc nulls first limit parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see cc cockroachdb sql queries jira issue crdb
1
55,889
14,751,104,154
IssuesEvent
2021-01-08 04:06:33
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
closed
Memory pressure may result in violated assertion in arc_wait_for_eviction()
Component: Memory Management Type: Defect
<!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Archlinux Distribution Version | rolling Linux Kernel | 5.9.10 Architecture | x64 ZFS Version | 8f158ae SPL Version | 8f158ae <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> ### Describe the problem you're observing Memory stressing a system with mprime results in a panic in `arc_wait_for_eviction()`. ### Describe how to reproduce the problem In a system with 128GB memory, of which 40GB were consumed by ARC running mprime panicked in the assertion of `arc_wait_for_eviction()`. mprime was run in torture-test mode, type of test was "4. Blend (tests all of the above)" (ie both large and small fast Fourrier transforms). In this test mode mprime tries to consume all available memory. I believe what happened is that the system was running low on memory so that the if clause at https://github.com/openzfs/zfs/blob/master/module/zfs/arc.c#L4166 was not true and arc_evict_waiters were not woken. However, arc_evict_count was still incremented at https://github.com/openzfs/zfs/blob/master/module/zfs/arc.c#L4164 which led to the panicked assertion in `arc_wait_for_eviction()`. Just a thought though. ### Include any warning/errors/backtraces from the system logs ``` VERIFY3(last->aew_count > arc_evict_count) failed (850903552 > 868262400) PANIC at arc.c:5255:arc_wait_for_eviction() Showing stack for process 1429448 CPU: 4 PID: 1429448 Comm: mprime Tainted: P OE 5.9.10-1-vfio #1 Call Trace: dump_stack+0x6b/0x83 spl_panic+0xef/0x117 [spl] ? sysvec_call_function+0x36/0x80 ? asm_sysvec_call_function+0x12/0x20 arc_wait_for_eviction+0x1db/0x1f0 [zfs] arc_shrinker_scan+0x36/0xd0 [zfs] do_shrink_slab+0x146/0x290 shrink_slab+0xd0/0x2f0 shrink_node+0x2c0/0x6e0 do_try_to_free_pages+0xda/0x4c0 try_to_free_pages+0xef/0x1c0 __alloc_pages_slowpath.constprop.0+0x384/0xce0 ? tick_nohz_next_event+0x8f/0x180 __alloc_pages_nodemask+0x2e6/0x310 alloc_pages_vma+0x80/0x250 handle_mm_fault+0xebd/0x1930 do_user_addr_fault+0x1b8/0x3f0 exc_page_fault+0x82/0x1a0 ? asm_exc_page_fault+0x8/0x30 asm_exc_page_fault+0x1e/0x30 ```
1.0
Memory pressure may result in violated assertion in arc_wait_for_eviction() - <!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Archlinux Distribution Version | rolling Linux Kernel | 5.9.10 Architecture | x64 ZFS Version | 8f158ae SPL Version | 8f158ae <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> ### Describe the problem you're observing Memory stressing a system with mprime results in a panic in `arc_wait_for_eviction()`. ### Describe how to reproduce the problem In a system with 128GB memory, of which 40GB were consumed by ARC running mprime panicked in the assertion of `arc_wait_for_eviction()`. mprime was run in torture-test mode, type of test was "4. Blend (tests all of the above)" (ie both large and small fast Fourrier transforms). In this test mode mprime tries to consume all available memory. I believe what happened is that the system was running low on memory so that the if clause at https://github.com/openzfs/zfs/blob/master/module/zfs/arc.c#L4166 was not true and arc_evict_waiters were not woken. However, arc_evict_count was still incremented at https://github.com/openzfs/zfs/blob/master/module/zfs/arc.c#L4164 which led to the panicked assertion in `arc_wait_for_eviction()`. Just a thought though. ### Include any warning/errors/backtraces from the system logs ``` VERIFY3(last->aew_count > arc_evict_count) failed (850903552 > 868262400) PANIC at arc.c:5255:arc_wait_for_eviction() Showing stack for process 1429448 CPU: 4 PID: 1429448 Comm: mprime Tainted: P OE 5.9.10-1-vfio #1 Call Trace: dump_stack+0x6b/0x83 spl_panic+0xef/0x117 [spl] ? sysvec_call_function+0x36/0x80 ? asm_sysvec_call_function+0x12/0x20 arc_wait_for_eviction+0x1db/0x1f0 [zfs] arc_shrinker_scan+0x36/0xd0 [zfs] do_shrink_slab+0x146/0x290 shrink_slab+0xd0/0x2f0 shrink_node+0x2c0/0x6e0 do_try_to_free_pages+0xda/0x4c0 try_to_free_pages+0xef/0x1c0 __alloc_pages_slowpath.constprop.0+0x384/0xce0 ? tick_nohz_next_event+0x8f/0x180 __alloc_pages_nodemask+0x2e6/0x310 alloc_pages_vma+0x80/0x250 handle_mm_fault+0xebd/0x1930 do_user_addr_fault+0x1b8/0x3f0 exc_page_fault+0x82/0x1a0 ? asm_exc_page_fault+0x8/0x30 asm_exc_page_fault+0x1e/0x30 ```
non_test
memory pressure may result in violated assertion in arc wait for eviction thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name archlinux distribution version rolling linux kernel architecture zfs version spl version commands to find zfs spl versions modinfo zfs grep iw version modinfo spl grep iw version describe the problem you re observing memory stressing a system with mprime results in a panic in arc wait for eviction describe how to reproduce the problem in a system with memory of which were consumed by arc running mprime panicked in the assertion of arc wait for eviction mprime was run in torture test mode type of test was blend tests all of the above ie both large and small fast fourrier transforms in this test mode mprime tries to consume all available memory i believe what happened is that the system was running low on memory so that the if clause at was not true and arc evict waiters were not woken however arc evict count was still incremented at which led to the panicked assertion in arc wait for eviction just a thought though include any warning errors backtraces from the system logs last aew count arc evict count failed panic at arc c arc wait for eviction showing stack for process cpu pid comm mprime tainted p oe vfio call trace dump stack spl panic sysvec call function asm sysvec call function arc wait for eviction arc shrinker scan do shrink slab shrink slab shrink node do try to free pages try to free pages alloc pages slowpath constprop tick nohz next event alloc pages nodemask alloc pages vma handle mm fault do user addr fault exc page fault asm exc page fault asm exc page fault
0
90,242
8,231,139,399
IssuesEvent
2018-09-07 15:01:10
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
workload/jsonload: (unknown) failed under stress
C-test-failure O-robot
SHA: https://github.com/cockroachdb/cockroach/commits/517663fa1489a2f36a593dc9722384a50bfcdbb5 Parameters: ``` TAGS= GOFLAGS= ``` Stress build found a failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=891244&tab=buildLog ``` build/werror.sh /go/native/x86_64-pc-linux-gnu/protobuf/protoc -I./pkg:./vendor/github.com/gogo/protobuf:./vendor/github.com/gogo/protobuf/protobuf --cpp_out=lite:/go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos ./pkg/roachpb/data.proto ./pkg/roachpb/internal.proto ./pkg/roachpb/metadata.proto ./pkg/storage/engine/enginepb/file_registry.proto ./pkg/storage/engine/enginepb/mvcc.proto ./pkg/storage/engine/enginepb/mvcc3.proto ./pkg/storage/engine/enginepb/rocksdb.proto ./pkg/util/hlc/legacy_timestamp.proto ./pkg/util/hlc/timestamp.proto ./pkg/util/unresolved_addr.proto set -e; for dir in ./pkg/acceptance/cluster/ ./pkg/build/ ./pkg/ccl/backupccl/ ./pkg/ccl/baseccl/ ./pkg/ccl/storageccl/engineccl/enginepbccl/ ./pkg/ccl/utilccl/licenseccl/ ./pkg/config/ ./pkg/gossip/ ./pkg/internal/client/ ./pkg/roachpb/ ./pkg/rpc/ ./pkg/server/diagnosticspb/ ./pkg/server/serverpb/ ./pkg/server/status/ ./pkg/settings/cluster/ ./pkg/sql/distsqlrun/ ./pkg/sql/jobs/ ./pkg/sql/pgwire/pgerror/ ./pkg/sql/sqlbase/ ./pkg/sql/stats/ ./pkg/storage/ ./pkg/storage/engine/enginepb/ ./pkg/storage/storagebase/ ./pkg/ts/tspb/ ./pkg/util/ ./pkg/util/hlc/ ./pkg/util/log/ ./pkg/util/protoutil/ ./pkg/util/tracing/; do \ build/werror.sh /go/native/x86_64-pc-linux-gnu/protobuf/protoc -I./pkg:./vendor/github.com/gogo/protobuf:./vendor/github.com/gogo/protobuf/protobuf:./vendor/github.com/coreos:./vendor/github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis --plugin=bin/protoc-gen-gogoroach --gogoroach_out=Mgoogle/api/annotations.proto=github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis/google/api,Mgoogle/protobuf/timestamp.proto=github.com/gogo/protobuf/types,,plugins=grpc,import_prefix=github.com/cockroachdb/cockroach/pkg/:./pkg $dir/*.proto; \ done sed -i -E '/gogoproto/d' /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protosccl/ccl/baseccl/encryption_options.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protosccl/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protosccl/ccl/baseccl/encryption_options.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protosccl/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.cc touch bin/.cpp_ccl_protobuf_sources sed -i -E '/gogoproto/d' /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/data.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/internal.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/metadata.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/file_registry.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/mvcc.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/mvcc3.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/rocksdb.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/hlc/legacy_timestamp.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/hlc/timestamp.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/unresolved_addr.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/data.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/internal.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/metadata.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/file_registry.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/mvcc.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/mvcc3.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/rocksdb.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/hlc/legacy_timestamp.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/hlc/timestamp.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/unresolved_addr.pb.cc github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/descriptor github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/suffixtree touch bin/.cpp_protobuf_sources github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/syntax yes checking for sys/types.h... github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/syntax/golang github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/output github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/generator yes github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/job checking for sys/stat.h... github.com/cockroachdb/cockroach/vendor/github.com/wadey/gocovmerge github.com/cockroachdb/cockroach/vendor/github.com/golang/dep github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/feedback github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/gengateway Scanning dependencies of target roach github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl [ 2%] Building CXX object CMakeFiles/roach.dir/batch.cc.o yes github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/internal/stats [ 5%] Building CXX object CMakeFiles/roach.dir/comparator.cc.o [ 8%] Building CXX object CMakeFiles/roach.dir/encoding.cc.o [ 11%] Building CXX object CMakeFiles/roach.dir/db.cc.o [ 14%] Building CXX object CMakeFiles/roach.dir/chunked_buffer.cc.o github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway checking for stdlib.h... github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/storage/benchfmt github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/imports github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/benchstat yes github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/base checking for string.h... github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/glide github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/glock yes github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/godep github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/govend github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/govendor checking for memory.h... github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/gvt github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/vndr github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/cmd/benchstat github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cmd/goyacc go/internal/gccgoimporter yes checking for strings.h... go/internal/gcimporter github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers go/internal/srcimporter yes github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/cmd/dep checking for inttypes.h... go/importer yes github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cmd/goimports github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cmd/stringer checking for stdint.h... yes checking for unistd.h... yes checking whether byte ordering is bigendian... no checking size of void *... 8 checking size of int... 4 checking size of long... 8 checking size of long long... 8 checking size of intmax_t... 8 checking build system type... x86_64-pc-linux-gnu checking host system type... x86_64-pc-linux-gnu checking whether pause instruction is compilable... yes checking for ar... ar checking malloc.h usability... yes checking malloc.h presence... [ 17%] Building CXX object CMakeFiles/roach.dir/engine.cc.o yes checking for malloc.h... yes checking whether malloc_usable_size definition can use const argument... no checking for library containing log... [ 20%] Building CXX object CMakeFiles/roach.dir/merge.cc.o touch bin/.bootstrap [ 23%] Building CXX object CMakeFiles/roach.dir/mvcc.cc.o -lm checking whether __attribute__ syntax is compilable... [ 26%] Building CXX object CMakeFiles/roach.dir/options.cc.o [ 29%] Building CXX object CMakeFiles/roach.dir/snapshot.cc.o yes checking whether compiler supports -fvisibility=hidden... yes checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether tls_model attribute is compilable... yes checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether alloc_size attribute is compilable... no checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether format(gnu_printf, ...) attribute is compilable... no checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether format(printf, ...) attribute is compilable... yes checking for a BSD-compatible install... /usr/bin/install -c checking for ranlib... ranlib checking for ld... /usr/bin/ld checking for autoconf... /usr/bin/autoconf checking for memalign... yes checking for valloc... yes checking whether compiler supports -O3... yes checking whether compiler supports -funroll-loops... yes checking unwind.h usability... yes checking unwind.h presence... yes checking for unwind.h... yes checking for _Unwind_Backtrace in -lgcc... yes checking configured backtracing method... libgcc checking for sbrk... yes checking whether utrace(2) is compilable... no checking whether valgrind is compilable... no checking whether a program using __builtin_unreachable is compilable... yes checking whether a program using __builtin_ffsl is compilable... yes checking LG_PAGE... 12 Missing VERSION file, and unable to generate it; creating bogus VERSION checking pthread.h usability... yes checking pthread.h presence... yes checking for pthread.h... yes checking for pthread_create in -lpthread... yes checking whether pthread_atfork(3) is compilable... [ 32%] Building CXX object CMakeFiles/roach.dir/protos/roachpb/data.pb.cc.o yes checking for library containing clock_gettime... none required checking whether clock_gettime(CLOCK_MONOTONIC_COARSE, ...) is compilable... yes checking whether clock_gettime(CLOCK_MONOTONIC, ...) is compilable... yes checking whether mach_absolute_time() is compilable... no checking whether compiler supports -Werror... yes checking whether syscall(2) is compilable... yes checking for secure_getenv... yes checking for issetugid... no checking for _malloc_thread_cleanup... no checking for _pthread_mutex_init_calloc_cb... [ 35%] Building CXX object CMakeFiles/roach.dir/protos/roachpb/internal.pb.cc.o no checking for TLS... yes checking whether C11 atomics is compilable... yes checking whether atomic(9) is compilable... no checking whether Darwin OSAtomic*() is compilable... [ 38%] Building CXX object CMakeFiles/roach.dir/protos/roachpb/metadata.pb.cc.o no checking whether madvise(2) is compilable... yes checking whether madvise(..., MADV_FREE) is compilable... no checking whether madvise(..., MADV_DONTNEED) is compilable... [ 41%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/mvcc.pb.cc.o [ 44%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/mvcc3.pb.cc.o yes checking whether madvise(..., MADV_[NO]HUGEPAGE) is compilable... yes checking whether to force 32-bit __sync_{add,sub}_and_fetch()... no checking whether to force 64-bit __sync_{add,sub}_and_fetch()... no checking for __builtin_clz... [ 47%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/file_registry.pb.cc.o yes checking whether Darwin os_unfair_lock_*() is compilable... no checking whether Darwin OSSpin*() is compilable... no checking whether glibc malloc hook is compilable... yes checking whether glibc memalign hook is compilable... yes checking whether pthreads adaptive mutexes is compilable... yes checking for stdbool.h that conforms to C99... yes checking for _Bool... yes configure: creating ./config.status config.status: creating Makefile config.status: creating jemalloc.pc config.status: creating doc/html.xsl config.status: creating doc/manpages.xsl config.status: creating doc/jemalloc.xml [ 50%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/rocksdb.pb.cc.o config.status: creating include/jemalloc/jemalloc_macros.h config.status: creating include/jemalloc/jemalloc_protos.h config.status: creating include/jemalloc/jemalloc_typedefs.h config.status: creating include/jemalloc/internal/jemalloc_internal.h config.status: creating test/test.sh config.status: creating test/include/test/jemalloc_test.h [ 52%] Building CXX object CMakeFiles/roach.dir/protos/util/hlc/legacy_timestamp.pb.cc.o config.status: creating config.stamp config.status: creating bin/jemalloc-config config.status: creating bin/jemalloc.sh config.status: creating bin/jeprof [ 55%] Building CXX object CMakeFiles/roach.dir/protos/util/hlc/timestamp.pb.cc.o config.status: creating include/jemalloc/jemalloc_defs.h config.status: creating include/jemalloc/internal/jemalloc_internal_defs.h config.status: creating test/include/test/jemalloc_test_defs.h config.status: executing include/jemalloc/internal/private_namespace.h commands config.status: executing include/jemalloc/internal/private_unnamespace.h commands config.status: executing include/jemalloc/internal/public_symbols.txt commands config.status: executing include/jemalloc/internal/public_namespace.h commands config.status: executing include/jemalloc/internal/public_unnamespace.h commands config.status: executing include/jemalloc/internal/size_classes.h commands [ 58%] Building CXX object CMakeFiles/roach.dir/protos/util/unresolved_addr.pb.cc.o config.status: executing include/jemalloc/jemalloc_protos_jet.h commands config.status: executing include/jemalloc/jemalloc_rename.h commands config.status: executing include/jemalloc/jemalloc_mangle.h commands config.status: executing include/jemalloc/jemalloc_mangle_jet.h commands set -euo pipefail; \ ret=$(cd ./pkg/sql/parser/gen && goyacc -p sql -o sql.go.tmp sql.y); \ if expr "$ret" : ".*conflicts" >/dev/null; then \ echo "$ret"; exit 1; \ fi config.status: executing include/jemalloc/jemalloc.h commands =============================================================================== jemalloc version : 0.0.0-0-g0000000000000000000000000000000000000000 library revision : 2 CONFIG : --enable-prof CC : cc CONFIGURE_CFLAGS : -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops SPECIFIED_CFLAGS : EXTRA_CFLAGS : CPPFLAGS : -D_GNU_SOURCE -D_REENTRANT LDFLAGS : EXTRA_LDFLAGS : LIBS : -lm -lgcc -lm -lpthread RPATH_EXTRA : XSLTPROC : false XSLROOT : PREFIX : /usr/local BINDIR : /usr/local/bin DATADIR : /usr/local/share INCLUDEDIR : /usr/local/include LIBDIR : /usr/local/lib MANDIR : /usr/local/share/man srcroot : /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/ abs_srcroot : /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/ objroot : abs_objroot : /go/native/x86_64-pc-linux-gnu/jemalloc/ JEMALLOC_PREFIX : JEMALLOC_PRIVATE_NAMESPACE : je_ install_suffix : malloc_conf : autogen : 0 cc-silence : 1 debug : 0 code-coverage : 0 stats : 1 prof : 1 prof-libunwind : 0 prof-libgcc : 1 prof-gcc : 0 tcache : 1 thp : 1 fill : 1 utrace : 0 valgrind : 0 xmalloc : 0 munmap : 0 lazy_lock : 0 tls : 1 cache-oblivious : 1 =============================================================================== find ./pkg -name node_modules -prune -o -type f -name '*.pb.gw.go' -exec rm {} + build/werror.sh /go/native/x86_64-pc-linux-gnu/protobuf/protoc -I./pkg:./vendor/github.com/gogo/protobuf:./vendor/github.com/gogo/protobuf/protobuf:./vendor/github.com/coreos:./vendor/github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis --grpc-gateway_out=logtostderr=true,request_context=true:./pkg ./pkg/server/serverpb/admin.proto ./pkg/server/serverpb/status.proto ./pkg/server/serverpb/authentication.proto build/werror.sh /go/native/x86_64-pc-linux-gnu/protobuf/protoc -I./pkg:./vendor/github.com/gogo/protobuf:./vendor/github.com/gogo/protobuf/protobuf:./vendor/github.com/coreos:./vendor/github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis --grpc-gateway_out=logtostderr=true,request_context=true:./pkg ./pkg/ts/tspb/timeseries.proto sed -i -E 's!golang.org/x/net/context!context!g' ./pkg/server/serverpb/admin.pb.gw.go ./pkg/server/serverpb/status.pb.gw.go ./pkg/server/serverpb/authentication.pb.gw.go ./pkg/ts/tspb/timeseries.pb.gw.go gofmt -s -w ./pkg/server/serverpb/admin.pb.gw.go ./pkg/server/serverpb/status.pb.gw.go ./pkg/server/serverpb/authentication.pb.gw.go ./pkg/ts/tspb/timeseries.pb.gw.go goimports -w ./pkg/server/serverpb/admin.pb.gw.go ./pkg/server/serverpb/status.pb.gw.go ./pkg/server/serverpb/authentication.pb.gw.go ./pkg/ts/tspb/timeseries.pb.gw.go sed -i '/import _/d' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go [ 61%] Linking CXX static library libroach.a sed -i -E 's!import (fmt|math) "github.com/cockroachdb/cockroach/pkg/(fmt|math)"! !g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go [ 82%] Built target roach touch bin/.gw_protobuf_sources Scanning dependencies of target roachccl [ 85%] Building CXX object CMakeFiles/roachccl.dir/ccl/db.cc.o sed -i -E 's!cockroachdb/cockroach/pkg/(etcd)!coreos/\1!g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go sed -i -E 's!github.com/cockroachdb/cockroach/pkg/(bytes|encoding/binary|errors|fmt|io|math|github\.com|(google\.)?golang\.org)!\1!g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go sed -i -E 's!golang.org/x/net/context!context!g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go gofmt -s -w ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/jemalloc.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/jemalloc.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/arena.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/arena.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/atomic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/atomic.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/base.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/base.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/bitmap.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/bitmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_dss.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_dss.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_mmap.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_mmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ckh.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ckh.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ctl.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ctl.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/extent.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/extent.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/hash.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/hash.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/huge.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/huge.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mb.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mb.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mutex.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mutex.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/nstime.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/nstime.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/pages.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/pages.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prng.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prng.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prof.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prof.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/quarantine.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/quarantine.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/rtree.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/rtree.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/stats.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/stats.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/spin.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/spin.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tcache.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tcache.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/t[ 88%] Building CXX object CMakeFiles/roachccl.dir/ccl/key_manager.cc.o touch bin/.go_protobuf_sources (echo "// Code generated by goyacc. DO NOT EDIT."; \ echo "// GENERATED FILE DO NOT EDIT"; \ cat pkg/sql/parser/gen/sql.go.tmp | \ sed -E 's/^const ([A-Z][_A-Z0-9]*) =.*$/const \1 = lex.\1/g') > pkg/sql/parser/sql.go (echo "// Code generated by make. DO NOT EDIT."; \ echo "// GENERATED FILE DO NOT EDIT"; \ echo; \ echo "package lex"; \ echo; \ grep '^const [A-Z][_A-Z0-9]* ' pkg/sql/parser/gen/sql.go.tmp) > pkg/sql/lex/tokens.go go run pkg/cmd/docgen/{main,funcs}.go functions docs/generated/sql --quiet [ 91%] Building CXX object CMakeFiles/roachccl.dir/protosccl/ccl/baseccl/encryption_options.pb.cc.o icker.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ticker.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tsd.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tsd.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/util.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/util.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/jemalloc.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/jemalloc.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/witness.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/witness.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/arena.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/arena.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/atomic.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/atomic.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/base.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/base.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/bitmap.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/bitmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_dss.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_dss.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_mmap.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_mmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemal[ 94%] Building CXX object CMakeFiles/roachccl.dir/protosccl/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.cc.o loc/include -Iinclude -o src/ckh.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ckh.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ctl.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ctl.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/extent.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/extent.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/hash.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/hash.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/huge.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/huge.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mb.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mb.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mutex.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mutex.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/nstime.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/nstime.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/pages.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/pages.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prng.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prng.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prof.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prof.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/quarantine.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/quarantine.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/sar: `u' modifier ignored since `D' is the default (see `U') ar: `u' modifier ignored since `D' is the default (see `U') rc/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/rtree.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/rtree.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/stats.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/stats.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/spin.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/spin.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tcache.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tcache.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ticker.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ticker.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tsd.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tsd.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/util.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/util.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/witness.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/witness.c ar crus lib/libjemalloc.a src/jemalloc.o src/arena.o src/atomic.o src/base.o src/bitmap.o src/chunk.o src/chunk_dss.o src/chunk_mmap.o src/ckh.o src/ctl.o src/extent.o src/hash.o src/huge.o src/mb.o src/mutex.o src/nstime.o src/pages.o src/prng.o src/prof.o src/quarantine.o src/rtree.o src/stats.o src/spin.o src/tcache.o src/ticker.o src/tsd.o src/util.o src/witness.o ar crus lib/libjemalloc_pic.a src/jemalloc.pic.o src/arena.pic.o src/atomic.pic.o src/base.pic.o src/bitmap.pic.o src/chunk.pic.o src/chunk_dss.pic.o src/chunk_mmap.pic.o src/ckh.pic.o src/ctl.pic.o src/extent.pic.o src/hash.pic.o src/huge.pic.o src/mb.pic.o src/mutex.pic.o src/nstime.pic.o src/pages.pic.o src/prng.pic.o src/prof.pic.o src/quarantine.pic.o src/rtree.pic.o src/stats.pic.o src/spin.pic.o src/tcache.pic.o src/ticker.pic.o src/tsd.pic.o src/util.pic.o src/witness.pic.o [ 0%] Built target build_version Scanning dependencies of target rocksdb [ 0%] Building CXX object CMakeFiles/rocksdb.dir/cache/sharded_cache.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/cache/clock_cache.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/cache/lru_cache.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/db/builder.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/db/c.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/db/column_family.cc.o touch bin/.docgen_functions [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compacted_db_impl.cc.o [ 97%] Linking CXX static library libroachccl.a [100%] Built target roachccl [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_iterator.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_job.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_picker.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_picker_universal.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/convenience.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_filesnapshot.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_write.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_compaction_flush.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_files.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_open.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_debug.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_experimental.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_readonly.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/db_info_dumper.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/db_iter.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/dbformat.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/event_helpers.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/experimental.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/external_sst_file_ingestion_job.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/file_indexer.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/flush_job.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/flush_scheduler.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/forward_iterator.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/internal_stats.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/log_reader.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/log_writer.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/malloc_stats.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/managed_iterator.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/memtable.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/memtable_list.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/merge_helper.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/range_del_aggregator.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/repair.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/table_cache.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/table_properties_collector.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/transaction_log_impl.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/version_builder.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/version_edit.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/version_set.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/db/wal_manager.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/db/write_batch.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/db/write_thread.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/env/env.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/env/env_encryption.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/env/mock_env.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/alloc_tracker.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/hash_cuckoo_rep.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/hash_linklist_rep.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/hash_skiplist_rep.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/skiplistrep.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/memtable/vectorrep.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/memtable/write_buffer_manager.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/histogram.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/instrumented_mutex.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/iostats_context.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/perf_context.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/perf_level.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/statistics.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/thread_status_updater.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/thread_status_util.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/thread_status_util_debug.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/options/cf_options.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/db_options.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/options.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/options_helper.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/options_parser.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/table/adaptive_table_factory.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/table/block.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_filter_block.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_table_builder.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_table_factory.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_table_reader.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_builder.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_prefix_index.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/bloom_block.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/cuckoo_table_builder.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/cuckoo_table_factory.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/cuckoo_table_reader.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/format.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/full_filter_block.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/get_context.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/index_builder.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/iterator.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/merging_iterator.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/meta_blocks.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/partitioned_filter_block.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/persistent_cache_helper.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_builder.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_factory.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_index.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_key_coding.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_reader.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/sst_file_writer.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/table_properties.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/two_level_iterator.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/tools/db_bench_tool.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/tools/dump/db_dump_tool.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/tools/ldb_cmd.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/tools/sst_dump_tool.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/util/auto_roll_logger.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/util/arena.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/util/bloom.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/coding.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/comparator.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/concurrent_arena.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/crc32c.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/delete_scheduler.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/dynamic_bloom.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/event_logger.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/file_reader_writer.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/file_util.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/filename.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/hash.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/log_buffer.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/random.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/rate_limiter.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/sst_file_manager_impl.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/status.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/sync_point.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/testutil.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/thread_local.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/threadpool_imp.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/transaction_test_util.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/utilities/backupable/backupable_db.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_db.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_db_impl.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_dump_tool.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_file.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_log_reader.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_log_writer.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_log_format.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/ttl_extractor.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/checkpoint/checkpoint_impl.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/col_buf_decoder.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/col_buf_encoder.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/column_aware_encoding_util.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/date_tiered/date_tiered_db_impl.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/debug.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/document/document_db.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/document/json_document.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/env_timed.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/geodb/geodb_impl.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/memory/memory_util.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/merge_operators/uint64add.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/options/options_util.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/block_cache_tier.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/block_cache_tier_file.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/block_cache_tier_metadata.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/volatile_tier_impl.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/redis/redis_lists.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/simulator_cache/sim_cache.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/spatialdb/spatial_db.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/optimistic_transaction_db_impl.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/optimistic_transaction.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/pessimistic_transaction.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/pessimistic_transaction_db.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/snapshot_checker.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/transaction_base.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/transaction_lock_mgr.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/transaction_util.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/write_prepared_txn.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/ttl/db_ttl_impl.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/write_batch_with_index/write_batch_with_index.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/write_batch_with_index/write_batch_with_index_internal.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/port/port_posix.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/env/env_posix.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/env/io_posix.cc.o [ 89%] Linking CXX static library librocksdb.a [100%] Built target rocksdb go test -v -tags ' make x86_64_pc_linux_gnu' -ldflags ' -X github.com/cockroachdb/cockroach/pkg/build.typ=development -X "github.com/cockroachdb/cockroach/pkg/build.tag=v2.0.5-34-g517663f-dirty" -X "github.com/cockroachdb/cockroach/pkg/build.utcTime=2018/09/07 08:52:18" -X "github.com/cockroachdb/cockroach/pkg/build.rev=517663fa1489a2f36a593dc9722384a50bfcdbb5" -X "github.com/cockroachdb/cockroach/pkg/build.cgoTargetTriple=x86_64-pc-linux-gnu" ' -i github.com/cockroachdb/cockroach/pkg/workload/jsonload can't load package: package github.com/cockroachdb/cockroach/pkg/workload/jsonload: cannot find package "github.com/cockroachdb/cockroach/pkg/workload/jsonload" in any of: /usr/local/go/src/github.com/cockroachdb/cockroach/pkg/workload/jsonload (from $GOROOT) /go/src/github.com/cockroachdb/cockroach/pkg/workload/jsonload (from $GOPATH) Makefile:756: recipe for target 'gotestdashi' failed make: *** [gotestdashi] Error 1 ```
1.0
workload/jsonload: (unknown) failed under stress - SHA: https://github.com/cockroachdb/cockroach/commits/517663fa1489a2f36a593dc9722384a50bfcdbb5 Parameters: ``` TAGS= GOFLAGS= ``` Stress build found a failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=891244&tab=buildLog ``` build/werror.sh /go/native/x86_64-pc-linux-gnu/protobuf/protoc -I./pkg:./vendor/github.com/gogo/protobuf:./vendor/github.com/gogo/protobuf/protobuf --cpp_out=lite:/go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos ./pkg/roachpb/data.proto ./pkg/roachpb/internal.proto ./pkg/roachpb/metadata.proto ./pkg/storage/engine/enginepb/file_registry.proto ./pkg/storage/engine/enginepb/mvcc.proto ./pkg/storage/engine/enginepb/mvcc3.proto ./pkg/storage/engine/enginepb/rocksdb.proto ./pkg/util/hlc/legacy_timestamp.proto ./pkg/util/hlc/timestamp.proto ./pkg/util/unresolved_addr.proto set -e; for dir in ./pkg/acceptance/cluster/ ./pkg/build/ ./pkg/ccl/backupccl/ ./pkg/ccl/baseccl/ ./pkg/ccl/storageccl/engineccl/enginepbccl/ ./pkg/ccl/utilccl/licenseccl/ ./pkg/config/ ./pkg/gossip/ ./pkg/internal/client/ ./pkg/roachpb/ ./pkg/rpc/ ./pkg/server/diagnosticspb/ ./pkg/server/serverpb/ ./pkg/server/status/ ./pkg/settings/cluster/ ./pkg/sql/distsqlrun/ ./pkg/sql/jobs/ ./pkg/sql/pgwire/pgerror/ ./pkg/sql/sqlbase/ ./pkg/sql/stats/ ./pkg/storage/ ./pkg/storage/engine/enginepb/ ./pkg/storage/storagebase/ ./pkg/ts/tspb/ ./pkg/util/ ./pkg/util/hlc/ ./pkg/util/log/ ./pkg/util/protoutil/ ./pkg/util/tracing/; do \ build/werror.sh /go/native/x86_64-pc-linux-gnu/protobuf/protoc -I./pkg:./vendor/github.com/gogo/protobuf:./vendor/github.com/gogo/protobuf/protobuf:./vendor/github.com/coreos:./vendor/github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis --plugin=bin/protoc-gen-gogoroach --gogoroach_out=Mgoogle/api/annotations.proto=github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis/google/api,Mgoogle/protobuf/timestamp.proto=github.com/gogo/protobuf/types,,plugins=grpc,import_prefix=github.com/cockroachdb/cockroach/pkg/:./pkg $dir/*.proto; \ done sed -i -E '/gogoproto/d' /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protosccl/ccl/baseccl/encryption_options.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protosccl/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protosccl/ccl/baseccl/encryption_options.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protosccl/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.cc touch bin/.cpp_ccl_protobuf_sources sed -i -E '/gogoproto/d' /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/data.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/internal.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/metadata.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/file_registry.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/mvcc.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/mvcc3.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/rocksdb.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/hlc/legacy_timestamp.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/hlc/timestamp.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/unresolved_addr.pb.h /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/data.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/internal.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/roachpb/metadata.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/file_registry.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/mvcc.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/mvcc3.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/storage/engine/enginepb/rocksdb.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/hlc/legacy_timestamp.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/hlc/timestamp.pb.cc /go/src/github.com/cockroachdb/cockroach/c-deps/libroach/protos/util/unresolved_addr.pb.cc github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/descriptor github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/suffixtree touch bin/.cpp_protobuf_sources github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/syntax yes checking for sys/types.h... github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/syntax/golang github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/output github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/generator yes github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl/job checking for sys/stat.h... github.com/cockroachdb/cockroach/vendor/github.com/wadey/gocovmerge github.com/cockroachdb/cockroach/vendor/github.com/golang/dep github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/feedback github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway/gengateway Scanning dependencies of target roach github.com/cockroachdb/cockroach/vendor/github.com/mibk/dupl [ 2%] Building CXX object CMakeFiles/roach.dir/batch.cc.o yes github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/internal/stats [ 5%] Building CXX object CMakeFiles/roach.dir/comparator.cc.o [ 8%] Building CXX object CMakeFiles/roach.dir/encoding.cc.o [ 11%] Building CXX object CMakeFiles/roach.dir/db.cc.o [ 14%] Building CXX object CMakeFiles/roach.dir/chunked_buffer.cc.o github.com/cockroachdb/cockroach/vendor/github.com/grpc-ecosystem/grpc-gateway/protoc-gen-grpc-gateway checking for stdlib.h... github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/storage/benchfmt github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/imports github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/benchstat yes github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/base checking for string.h... github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/glide github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/glock yes github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/godep github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/govend github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/govendor checking for memory.h... github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/gvt github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers/vndr github.com/cockroachdb/cockroach/vendor/golang.org/x/perf/cmd/benchstat github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cmd/goyacc go/internal/gccgoimporter yes checking for strings.h... go/internal/gcimporter github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/internal/importers go/internal/srcimporter yes github.com/cockroachdb/cockroach/vendor/github.com/golang/dep/cmd/dep checking for inttypes.h... go/importer yes github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cmd/goimports github.com/cockroachdb/cockroach/vendor/golang.org/x/tools/cmd/stringer checking for stdint.h... yes checking for unistd.h... yes checking whether byte ordering is bigendian... no checking size of void *... 8 checking size of int... 4 checking size of long... 8 checking size of long long... 8 checking size of intmax_t... 8 checking build system type... x86_64-pc-linux-gnu checking host system type... x86_64-pc-linux-gnu checking whether pause instruction is compilable... yes checking for ar... ar checking malloc.h usability... yes checking malloc.h presence... [ 17%] Building CXX object CMakeFiles/roach.dir/engine.cc.o yes checking for malloc.h... yes checking whether malloc_usable_size definition can use const argument... no checking for library containing log... [ 20%] Building CXX object CMakeFiles/roach.dir/merge.cc.o touch bin/.bootstrap [ 23%] Building CXX object CMakeFiles/roach.dir/mvcc.cc.o -lm checking whether __attribute__ syntax is compilable... [ 26%] Building CXX object CMakeFiles/roach.dir/options.cc.o [ 29%] Building CXX object CMakeFiles/roach.dir/snapshot.cc.o yes checking whether compiler supports -fvisibility=hidden... yes checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether tls_model attribute is compilable... yes checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether alloc_size attribute is compilable... no checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether format(gnu_printf, ...) attribute is compilable... no checking whether compiler supports -Werror... yes checking whether compiler supports -herror_on_warning... no checking whether format(printf, ...) attribute is compilable... yes checking for a BSD-compatible install... /usr/bin/install -c checking for ranlib... ranlib checking for ld... /usr/bin/ld checking for autoconf... /usr/bin/autoconf checking for memalign... yes checking for valloc... yes checking whether compiler supports -O3... yes checking whether compiler supports -funroll-loops... yes checking unwind.h usability... yes checking unwind.h presence... yes checking for unwind.h... yes checking for _Unwind_Backtrace in -lgcc... yes checking configured backtracing method... libgcc checking for sbrk... yes checking whether utrace(2) is compilable... no checking whether valgrind is compilable... no checking whether a program using __builtin_unreachable is compilable... yes checking whether a program using __builtin_ffsl is compilable... yes checking LG_PAGE... 12 Missing VERSION file, and unable to generate it; creating bogus VERSION checking pthread.h usability... yes checking pthread.h presence... yes checking for pthread.h... yes checking for pthread_create in -lpthread... yes checking whether pthread_atfork(3) is compilable... [ 32%] Building CXX object CMakeFiles/roach.dir/protos/roachpb/data.pb.cc.o yes checking for library containing clock_gettime... none required checking whether clock_gettime(CLOCK_MONOTONIC_COARSE, ...) is compilable... yes checking whether clock_gettime(CLOCK_MONOTONIC, ...) is compilable... yes checking whether mach_absolute_time() is compilable... no checking whether compiler supports -Werror... yes checking whether syscall(2) is compilable... yes checking for secure_getenv... yes checking for issetugid... no checking for _malloc_thread_cleanup... no checking for _pthread_mutex_init_calloc_cb... [ 35%] Building CXX object CMakeFiles/roach.dir/protos/roachpb/internal.pb.cc.o no checking for TLS... yes checking whether C11 atomics is compilable... yes checking whether atomic(9) is compilable... no checking whether Darwin OSAtomic*() is compilable... [ 38%] Building CXX object CMakeFiles/roach.dir/protos/roachpb/metadata.pb.cc.o no checking whether madvise(2) is compilable... yes checking whether madvise(..., MADV_FREE) is compilable... no checking whether madvise(..., MADV_DONTNEED) is compilable... [ 41%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/mvcc.pb.cc.o [ 44%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/mvcc3.pb.cc.o yes checking whether madvise(..., MADV_[NO]HUGEPAGE) is compilable... yes checking whether to force 32-bit __sync_{add,sub}_and_fetch()... no checking whether to force 64-bit __sync_{add,sub}_and_fetch()... no checking for __builtin_clz... [ 47%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/file_registry.pb.cc.o yes checking whether Darwin os_unfair_lock_*() is compilable... no checking whether Darwin OSSpin*() is compilable... no checking whether glibc malloc hook is compilable... yes checking whether glibc memalign hook is compilable... yes checking whether pthreads adaptive mutexes is compilable... yes checking for stdbool.h that conforms to C99... yes checking for _Bool... yes configure: creating ./config.status config.status: creating Makefile config.status: creating jemalloc.pc config.status: creating doc/html.xsl config.status: creating doc/manpages.xsl config.status: creating doc/jemalloc.xml [ 50%] Building CXX object CMakeFiles/roach.dir/protos/storage/engine/enginepb/rocksdb.pb.cc.o config.status: creating include/jemalloc/jemalloc_macros.h config.status: creating include/jemalloc/jemalloc_protos.h config.status: creating include/jemalloc/jemalloc_typedefs.h config.status: creating include/jemalloc/internal/jemalloc_internal.h config.status: creating test/test.sh config.status: creating test/include/test/jemalloc_test.h [ 52%] Building CXX object CMakeFiles/roach.dir/protos/util/hlc/legacy_timestamp.pb.cc.o config.status: creating config.stamp config.status: creating bin/jemalloc-config config.status: creating bin/jemalloc.sh config.status: creating bin/jeprof [ 55%] Building CXX object CMakeFiles/roach.dir/protos/util/hlc/timestamp.pb.cc.o config.status: creating include/jemalloc/jemalloc_defs.h config.status: creating include/jemalloc/internal/jemalloc_internal_defs.h config.status: creating test/include/test/jemalloc_test_defs.h config.status: executing include/jemalloc/internal/private_namespace.h commands config.status: executing include/jemalloc/internal/private_unnamespace.h commands config.status: executing include/jemalloc/internal/public_symbols.txt commands config.status: executing include/jemalloc/internal/public_namespace.h commands config.status: executing include/jemalloc/internal/public_unnamespace.h commands config.status: executing include/jemalloc/internal/size_classes.h commands [ 58%] Building CXX object CMakeFiles/roach.dir/protos/util/unresolved_addr.pb.cc.o config.status: executing include/jemalloc/jemalloc_protos_jet.h commands config.status: executing include/jemalloc/jemalloc_rename.h commands config.status: executing include/jemalloc/jemalloc_mangle.h commands config.status: executing include/jemalloc/jemalloc_mangle_jet.h commands set -euo pipefail; \ ret=$(cd ./pkg/sql/parser/gen && goyacc -p sql -o sql.go.tmp sql.y); \ if expr "$ret" : ".*conflicts" >/dev/null; then \ echo "$ret"; exit 1; \ fi config.status: executing include/jemalloc/jemalloc.h commands =============================================================================== jemalloc version : 0.0.0-0-g0000000000000000000000000000000000000000 library revision : 2 CONFIG : --enable-prof CC : cc CONFIGURE_CFLAGS : -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops SPECIFIED_CFLAGS : EXTRA_CFLAGS : CPPFLAGS : -D_GNU_SOURCE -D_REENTRANT LDFLAGS : EXTRA_LDFLAGS : LIBS : -lm -lgcc -lm -lpthread RPATH_EXTRA : XSLTPROC : false XSLROOT : PREFIX : /usr/local BINDIR : /usr/local/bin DATADIR : /usr/local/share INCLUDEDIR : /usr/local/include LIBDIR : /usr/local/lib MANDIR : /usr/local/share/man srcroot : /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/ abs_srcroot : /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/ objroot : abs_objroot : /go/native/x86_64-pc-linux-gnu/jemalloc/ JEMALLOC_PREFIX : JEMALLOC_PRIVATE_NAMESPACE : je_ install_suffix : malloc_conf : autogen : 0 cc-silence : 1 debug : 0 code-coverage : 0 stats : 1 prof : 1 prof-libunwind : 0 prof-libgcc : 1 prof-gcc : 0 tcache : 1 thp : 1 fill : 1 utrace : 0 valgrind : 0 xmalloc : 0 munmap : 0 lazy_lock : 0 tls : 1 cache-oblivious : 1 =============================================================================== find ./pkg -name node_modules -prune -o -type f -name '*.pb.gw.go' -exec rm {} + build/werror.sh /go/native/x86_64-pc-linux-gnu/protobuf/protoc -I./pkg:./vendor/github.com/gogo/protobuf:./vendor/github.com/gogo/protobuf/protobuf:./vendor/github.com/coreos:./vendor/github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis --grpc-gateway_out=logtostderr=true,request_context=true:./pkg ./pkg/server/serverpb/admin.proto ./pkg/server/serverpb/status.proto ./pkg/server/serverpb/authentication.proto build/werror.sh /go/native/x86_64-pc-linux-gnu/protobuf/protoc -I./pkg:./vendor/github.com/gogo/protobuf:./vendor/github.com/gogo/protobuf/protobuf:./vendor/github.com/coreos:./vendor/github.com/grpc-ecosystem/grpc-gateway/third_party/googleapis --grpc-gateway_out=logtostderr=true,request_context=true:./pkg ./pkg/ts/tspb/timeseries.proto sed -i -E 's!golang.org/x/net/context!context!g' ./pkg/server/serverpb/admin.pb.gw.go ./pkg/server/serverpb/status.pb.gw.go ./pkg/server/serverpb/authentication.pb.gw.go ./pkg/ts/tspb/timeseries.pb.gw.go gofmt -s -w ./pkg/server/serverpb/admin.pb.gw.go ./pkg/server/serverpb/status.pb.gw.go ./pkg/server/serverpb/authentication.pb.gw.go ./pkg/ts/tspb/timeseries.pb.gw.go goimports -w ./pkg/server/serverpb/admin.pb.gw.go ./pkg/server/serverpb/status.pb.gw.go ./pkg/server/serverpb/authentication.pb.gw.go ./pkg/ts/tspb/timeseries.pb.gw.go sed -i '/import _/d' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go [ 61%] Linking CXX static library libroach.a sed -i -E 's!import (fmt|math) "github.com/cockroachdb/cockroach/pkg/(fmt|math)"! !g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go [ 82%] Built target roach touch bin/.gw_protobuf_sources Scanning dependencies of target roachccl [ 85%] Building CXX object CMakeFiles/roachccl.dir/ccl/db.cc.o sed -i -E 's!cockroachdb/cockroach/pkg/(etcd)!coreos/\1!g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go sed -i -E 's!github.com/cockroachdb/cockroach/pkg/(bytes|encoding/binary|errors|fmt|io|math|github\.com|(google\.)?golang\.org)!\1!g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go sed -i -E 's!golang.org/x/net/context!context!g' ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go gofmt -s -w ./pkg/acceptance/cluster/testconfig.pb.go ./pkg/build/info.pb.go ./pkg/ccl/backupccl/backup.pb.go ./pkg/ccl/baseccl/encryption_options.pb.go ./pkg/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.go ./pkg/ccl/utilccl/licenseccl/license.pb.go ./pkg/config/system.pb.go ./pkg/config/zone.pb.go ./pkg/gossip/gossip.pb.go ./pkg/internal/client/lease.pb.go ./pkg/roachpb/api.pb.go ./pkg/roachpb/app_stats.pb.go ./pkg/roachpb/csv.pb.go ./pkg/roachpb/data.pb.go ./pkg/roachpb/errors.pb.go ./pkg/roachpb/internal.pb.go ./pkg/roachpb/internal_raft.pb.go ./pkg/roachpb/metadata.pb.go ./pkg/rpc/heartbeat.pb.go ./pkg/server/diagnosticspb/diagnostics.pb.go ./pkg/server/serverpb/admin.pb.go ./pkg/server/serverpb/authentication.pb.go ./pkg/server/serverpb/init.pb.go ./pkg/server/serverpb/status.pb.go ./pkg/server/status/status.pb.go ./pkg/settings/cluster/cluster_version.pb.go ./pkg/sql/distsqlrun/api.pb.go ./pkg/sql/distsqlrun/data.pb.go ./pkg/sql/distsqlrun/processors.pb.go ./pkg/sql/jobs/jobs.pb.go ./pkg/sql/pgwire/pgerror/errors.pb.go ./pkg/sql/sqlbase/encoded_datum.pb.go ./pkg/sql/sqlbase/join_type.pb.go ./pkg/sql/sqlbase/privilege.pb.go ./pkg/sql/sqlbase/structured.pb.go ./pkg/sql/stats/histogram.pb.go ./pkg/storage/api.pb.go ./pkg/storage/engine/enginepb/file_registry.pb.go ./pkg/storage/engine/enginepb/mvcc.pb.go ./pkg/storage/engine/enginepb/mvcc3.pb.go ./pkg/storage/engine/enginepb/rocksdb.pb.go ./pkg/storage/lease_status.pb.go ./pkg/storage/liveness.pb.go ./pkg/storage/log.pb.go ./pkg/storage/raft.pb.go ./pkg/storage/storagebase/proposer_kv.pb.go ./pkg/storage/storagebase/state.pb.go ./pkg/ts/tspb/timeseries.pb.go ./pkg/util/hlc/legacy_timestamp.pb.go ./pkg/util/hlc/timestamp.pb.go ./pkg/util/log/log.pb.go ./pkg/util/protoutil/clone.pb.go ./pkg/util/tracing/recorded_span.pb.go ./pkg/util/unresolved_addr.pb.go cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/jemalloc.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/jemalloc.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/arena.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/arena.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/atomic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/atomic.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/base.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/base.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/bitmap.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/bitmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_dss.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_dss.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_mmap.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_mmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ckh.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ckh.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ctl.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ctl.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/extent.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/extent.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/hash.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/hash.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/huge.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/huge.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mb.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mb.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mutex.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mutex.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/nstime.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/nstime.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/pages.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/pages.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prng.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prng.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prof.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prof.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/quarantine.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/quarantine.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/rtree.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/rtree.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/stats.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/stats.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/spin.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/spin.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tcache.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tcache.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/t[ 88%] Building CXX object CMakeFiles/roachccl.dir/ccl/key_manager.cc.o touch bin/.go_protobuf_sources (echo "// Code generated by goyacc. DO NOT EDIT."; \ echo "// GENERATED FILE DO NOT EDIT"; \ cat pkg/sql/parser/gen/sql.go.tmp | \ sed -E 's/^const ([A-Z][_A-Z0-9]*) =.*$/const \1 = lex.\1/g') > pkg/sql/parser/sql.go (echo "// Code generated by make. DO NOT EDIT."; \ echo "// GENERATED FILE DO NOT EDIT"; \ echo; \ echo "package lex"; \ echo; \ grep '^const [A-Z][_A-Z0-9]* ' pkg/sql/parser/gen/sql.go.tmp) > pkg/sql/lex/tokens.go go run pkg/cmd/docgen/{main,funcs}.go functions docs/generated/sql --quiet [ 91%] Building CXX object CMakeFiles/roachccl.dir/protosccl/ccl/baseccl/encryption_options.pb.cc.o icker.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ticker.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tsd.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tsd.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/util.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/util.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/jemalloc.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/jemalloc.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/witness.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/witness.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/arena.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/arena.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/atomic.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/atomic.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/base.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/base.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/bitmap.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/bitmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_dss.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_dss.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/chunk_mmap.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/chunk_mmap.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemal[ 94%] Building CXX object CMakeFiles/roachccl.dir/protosccl/ccl/storageccl/engineccl/enginepbccl/key_registry.pb.cc.o loc/include -Iinclude -o src/ckh.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ckh.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ctl.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ctl.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/extent.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/extent.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/hash.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/hash.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/huge.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/huge.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mb.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mb.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/mutex.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/mutex.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/nstime.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/nstime.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/pages.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/pages.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prng.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prng.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/prof.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/prof.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/quarantine.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/quarantine.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/sar: `u' modifier ignored since `D' is the default (see `U') ar: `u' modifier ignored since `D' is the default (see `U') rc/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/rtree.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/rtree.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/stats.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/stats.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/spin.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/spin.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tcache.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tcache.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/ticker.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/ticker.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/tsd.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/tsd.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/util.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/util.c cc -std=gnu11 -Wall -Werror=declaration-after-statement -Wshorten-64-to-32 -Wsign-compare -pipe -g3 -fvisibility=hidden -O3 -funroll-loops -fPIC -DPIC -c -D_GNU_SOURCE -D_REENTRANT -I/go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/include -Iinclude -o src/witness.pic.o /go/src/github.com/cockroachdb/cockroach/c-deps/jemalloc/src/witness.c ar crus lib/libjemalloc.a src/jemalloc.o src/arena.o src/atomic.o src/base.o src/bitmap.o src/chunk.o src/chunk_dss.o src/chunk_mmap.o src/ckh.o src/ctl.o src/extent.o src/hash.o src/huge.o src/mb.o src/mutex.o src/nstime.o src/pages.o src/prng.o src/prof.o src/quarantine.o src/rtree.o src/stats.o src/spin.o src/tcache.o src/ticker.o src/tsd.o src/util.o src/witness.o ar crus lib/libjemalloc_pic.a src/jemalloc.pic.o src/arena.pic.o src/atomic.pic.o src/base.pic.o src/bitmap.pic.o src/chunk.pic.o src/chunk_dss.pic.o src/chunk_mmap.pic.o src/ckh.pic.o src/ctl.pic.o src/extent.pic.o src/hash.pic.o src/huge.pic.o src/mb.pic.o src/mutex.pic.o src/nstime.pic.o src/pages.pic.o src/prng.pic.o src/prof.pic.o src/quarantine.pic.o src/rtree.pic.o src/stats.pic.o src/spin.pic.o src/tcache.pic.o src/ticker.pic.o src/tsd.pic.o src/util.pic.o src/witness.pic.o [ 0%] Built target build_version Scanning dependencies of target rocksdb [ 0%] Building CXX object CMakeFiles/rocksdb.dir/cache/sharded_cache.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/cache/clock_cache.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/cache/lru_cache.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/db/builder.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/db/c.cc.o [ 0%] Building CXX object CMakeFiles/rocksdb.dir/db/column_family.cc.o touch bin/.docgen_functions [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compacted_db_impl.cc.o [ 97%] Linking CXX static library libroachccl.a [100%] Built target roachccl [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_iterator.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_job.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_picker.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/compaction_picker_universal.cc.o [ 3%] Building CXX object CMakeFiles/rocksdb.dir/db/convenience.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_filesnapshot.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_write.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_compaction_flush.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_files.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_open.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_debug.cc.o [ 7%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_experimental.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/db_impl_readonly.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/db_info_dumper.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/db_iter.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/dbformat.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/event_helpers.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/experimental.cc.o [ 10%] Building CXX object CMakeFiles/rocksdb.dir/db/external_sst_file_ingestion_job.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/file_indexer.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/flush_job.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/flush_scheduler.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/forward_iterator.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/internal_stats.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/log_reader.cc.o [ 14%] Building CXX object CMakeFiles/rocksdb.dir/db/log_writer.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/malloc_stats.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/managed_iterator.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/memtable.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/memtable_list.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/merge_helper.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/range_del_aggregator.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/repair.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/table_cache.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/table_properties_collector.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/transaction_log_impl.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/version_builder.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/version_edit.cc.o [ 17%] Building CXX object CMakeFiles/rocksdb.dir/db/version_set.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/db/wal_manager.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/db/write_batch.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/db/write_thread.cc.o [ 21%] Building CXX object CMakeFiles/rocksdb.dir/env/env.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/env/env_encryption.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/env/mock_env.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/alloc_tracker.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/hash_cuckoo_rep.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/hash_linklist_rep.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/hash_skiplist_rep.cc.o [ 25%] Building CXX object CMakeFiles/rocksdb.dir/memtable/skiplistrep.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/memtable/vectorrep.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/memtable/write_buffer_manager.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/histogram.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/instrumented_mutex.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/iostats_context.cc.o [ 28%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/perf_context.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/perf_level.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/statistics.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/thread_status_updater.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/thread_status_util.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/monitoring/thread_status_util_debug.cc.o [ 32%] Building CXX object CMakeFiles/rocksdb.dir/options/cf_options.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/db_options.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/options.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/options_helper.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/options/options_parser.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/table/adaptive_table_factory.cc.o [ 35%] Building CXX object CMakeFiles/rocksdb.dir/table/block.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_filter_block.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_table_builder.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_table_factory.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_based_table_reader.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_builder.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/block_prefix_index.cc.o [ 39%] Building CXX object CMakeFiles/rocksdb.dir/table/bloom_block.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/cuckoo_table_builder.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/cuckoo_table_factory.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/cuckoo_table_reader.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/format.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/full_filter_block.cc.o [ 42%] Building CXX object CMakeFiles/rocksdb.dir/table/get_context.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/index_builder.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/iterator.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/merging_iterator.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/meta_blocks.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/partitioned_filter_block.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/persistent_cache_helper.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_builder.cc.o [ 46%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_factory.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_index.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_key_coding.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/plain_table_reader.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/sst_file_writer.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/table_properties.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/table/two_level_iterator.cc.o [ 50%] Building CXX object CMakeFiles/rocksdb.dir/tools/db_bench_tool.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/tools/dump/db_dump_tool.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/tools/ldb_cmd.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/tools/sst_dump_tool.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/util/auto_roll_logger.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/util/arena.cc.o [ 53%] Building CXX object CMakeFiles/rocksdb.dir/util/bloom.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/coding.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/comparator.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/concurrent_arena.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/crc32c.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/delete_scheduler.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/dynamic_bloom.cc.o [ 57%] Building CXX object CMakeFiles/rocksdb.dir/util/event_logger.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/file_reader_writer.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/file_util.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/filename.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/hash.cc.o [ 60%] Building CXX object CMakeFiles/rocksdb.dir/util/log_buffer.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/random.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/rate_limiter.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/sst_file_manager_impl.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/status.cc.o [ 64%] Building CXX object CMakeFiles/rocksdb.dir/util/sync_point.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/testutil.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/thread_local.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/threadpool_imp.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/util/transaction_test_util.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/utilities/backupable/backupable_db.cc.o [ 67%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_db.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_db_impl.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_dump_tool.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_file.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_log_reader.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_log_writer.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/blob_log_format.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/blob_db/ttl_extractor.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/checkpoint/checkpoint_impl.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/col_buf_decoder.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/col_buf_encoder.cc.o [ 71%] Building CXX object CMakeFiles/rocksdb.dir/utilities/column_aware_encoding_util.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/date_tiered/date_tiered_db_impl.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/debug.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/document/document_db.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/document/json_document.cc.o [ 75%] Building CXX object CMakeFiles/rocksdb.dir/utilities/env_timed.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/geodb/geodb_impl.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/memory/memory_util.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/merge_operators/uint64add.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/options/options_util.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/block_cache_tier.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/block_cache_tier_file.cc.o [ 78%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/block_cache_tier_metadata.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/persistent_cache/volatile_tier_impl.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/redis/redis_lists.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/simulator_cache/sim_cache.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/spatialdb/spatial_db.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/optimistic_transaction_db_impl.cc.o [ 82%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/optimistic_transaction.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/pessimistic_transaction.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/pessimistic_transaction_db.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/snapshot_checker.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/transaction_base.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/transaction_lock_mgr.cc.o [ 85%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/transaction_util.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/transactions/write_prepared_txn.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/ttl/db_ttl_impl.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/write_batch_with_index/write_batch_with_index.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/utilities/write_batch_with_index/write_batch_with_index_internal.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/port/port_posix.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/env/env_posix.cc.o [ 89%] Building CXX object CMakeFiles/rocksdb.dir/env/io_posix.cc.o [ 89%] Linking CXX static library librocksdb.a [100%] Built target rocksdb go test -v -tags ' make x86_64_pc_linux_gnu' -ldflags ' -X github.com/cockroachdb/cockroach/pkg/build.typ=development -X "github.com/cockroachdb/cockroach/pkg/build.tag=v2.0.5-34-g517663f-dirty" -X "github.com/cockroachdb/cockroach/pkg/build.utcTime=2018/09/07 08:52:18" -X "github.com/cockroachdb/cockroach/pkg/build.rev=517663fa1489a2f36a593dc9722384a50bfcdbb5" -X "github.com/cockroachdb/cockroach/pkg/build.cgoTargetTriple=x86_64-pc-linux-gnu" ' -i github.com/cockroachdb/cockroach/pkg/workload/jsonload can't load package: package github.com/cockroachdb/cockroach/pkg/workload/jsonload: cannot find package "github.com/cockroachdb/cockroach/pkg/workload/jsonload" in any of: /usr/local/go/src/github.com/cockroachdb/cockroach/pkg/workload/jsonload (from $GOROOT) /go/src/github.com/cockroachdb/cockroach/pkg/workload/jsonload (from $GOPATH) Makefile:756: recipe for target 'gotestdashi' failed make: *** [gotestdashi] Error 1 ```
test
workload jsonload unknown failed under stress sha parameters tags goflags stress build found a failed test build werror sh go native pc linux gnu protobuf protoc i pkg vendor github com gogo protobuf vendor github com gogo protobuf protobuf cpp out lite go src github com cockroachdb cockroach c deps libroach protos pkg roachpb data proto pkg roachpb internal proto pkg roachpb metadata proto pkg storage engine enginepb file registry proto pkg storage engine enginepb mvcc proto pkg storage engine enginepb proto pkg storage engine enginepb rocksdb proto pkg util hlc legacy timestamp proto pkg util hlc timestamp proto pkg util unresolved addr proto set e for dir in pkg acceptance cluster pkg build pkg ccl backupccl pkg ccl baseccl pkg ccl storageccl engineccl enginepbccl pkg ccl utilccl licenseccl pkg config pkg gossip pkg internal client pkg roachpb pkg rpc pkg server diagnosticspb pkg server serverpb pkg server status pkg settings cluster pkg sql distsqlrun pkg sql jobs pkg sql pgwire pgerror pkg sql sqlbase pkg sql stats pkg storage pkg storage engine enginepb pkg storage storagebase pkg ts tspb pkg util pkg util hlc pkg util log pkg util protoutil pkg util tracing do build werror sh go native pc linux gnu protobuf protoc i pkg vendor github com gogo protobuf vendor github com gogo protobuf protobuf vendor github com coreos vendor github com grpc ecosystem grpc gateway third party googleapis plugin bin protoc gen gogoroach gogoroach out mgoogle api annotations proto github com grpc ecosystem grpc gateway third party googleapis google api mgoogle protobuf timestamp proto github com gogo protobuf types plugins grpc import prefix github com cockroachdb cockroach pkg pkg dir proto done sed i e gogoproto d go src github com cockroachdb cockroach c deps libroach protosccl ccl baseccl encryption options pb h go src github com cockroachdb cockroach c deps libroach protosccl ccl storageccl engineccl enginepbccl key registry pb h go src github com cockroachdb cockroach c deps libroach protosccl ccl baseccl encryption options pb cc go src github com cockroachdb cockroach c deps libroach protosccl ccl storageccl engineccl enginepbccl key registry pb cc touch bin cpp ccl protobuf sources sed i e gogoproto d go src github com cockroachdb cockroach c deps libroach protos roachpb data pb h go src github com cockroachdb cockroach c deps libroach protos roachpb internal pb h go src github com cockroachdb cockroach c deps libroach protos roachpb metadata pb h go src github com cockroachdb cockroach c deps libroach protos storage engine enginepb file registry pb h go src github com cockroachdb cockroach c deps libroach protos storage engine enginepb mvcc pb h go src github com cockroachdb cockroach c deps libroach protos storage engine enginepb pb h go src github com cockroachdb cockroach c deps libroach protos storage engine enginepb rocksdb pb h go src github com cockroachdb cockroach c deps libroach protos util hlc legacy timestamp pb h go src github com cockroachdb cockroach c deps libroach protos util hlc timestamp pb h go src github com cockroachdb cockroach c deps libroach protos util unresolved addr pb h go src github com cockroachdb cockroach c deps libroach protos roachpb data pb cc go src github com cockroachdb cockroach c deps libroach protos roachpb internal pb cc go src github com cockroachdb cockroach c deps libroach protos roachpb metadata pb cc go src github com cockroachdb cockroach c deps libroach protos storage engine enginepb file registry pb cc go src github com cockroachdb cockroach c deps libroach protos storage engine enginepb mvcc pb cc go src github com cockroachdb cockroach c deps libroach protos storage engine enginepb pb cc go src github com cockroachdb cockroach c deps libroach protos storage engine enginepb rocksdb pb cc go src github com cockroachdb cockroach c deps libroach protos util hlc legacy timestamp pb cc go src github com cockroachdb cockroach c deps libroach protos util hlc timestamp pb cc go src github com cockroachdb cockroach c deps libroach protos util unresolved addr pb cc github com cockroachdb cockroach vendor github com grpc ecosystem grpc gateway protoc gen grpc gateway descriptor github com cockroachdb cockroach vendor github com mibk dupl suffixtree touch bin cpp protobuf sources github com cockroachdb cockroach vendor github com mibk dupl syntax yes checking for sys types h github com cockroachdb cockroach vendor github com mibk dupl syntax golang github com cockroachdb cockroach vendor github com mibk dupl output github com cockroachdb cockroach vendor github com grpc ecosystem grpc gateway protoc gen grpc gateway generator yes github com cockroachdb cockroach vendor github com mibk dupl job checking for sys stat h github com cockroachdb cockroach vendor github com wadey gocovmerge github com cockroachdb cockroach vendor github com golang dep github com cockroachdb cockroach vendor github com golang dep internal feedback github com cockroachdb cockroach vendor github com grpc ecosystem grpc gateway protoc gen grpc gateway gengateway scanning dependencies of target roach github com cockroachdb cockroach vendor github com mibk dupl building cxx object cmakefiles roach dir batch cc o yes github com cockroachdb cockroach vendor golang org x perf internal stats building cxx object cmakefiles roach dir comparator cc o building cxx object cmakefiles roach dir encoding cc o building cxx object cmakefiles roach dir db cc o building cxx object cmakefiles roach dir chunked buffer cc o github com cockroachdb cockroach vendor github com grpc ecosystem grpc gateway protoc gen grpc gateway checking for stdlib h github com cockroachdb cockroach vendor golang org x perf storage benchfmt github com cockroachdb cockroach vendor golang org x tools imports github com cockroachdb cockroach vendor golang org x perf benchstat yes github com cockroachdb cockroach vendor github com golang dep internal importers base checking for string h github com cockroachdb cockroach vendor github com golang dep internal importers glide github com cockroachdb cockroach vendor github com golang dep internal importers glock yes github com cockroachdb cockroach vendor github com golang dep internal importers godep github com cockroachdb cockroach vendor github com golang dep internal importers govend github com cockroachdb cockroach vendor github com golang dep internal importers govendor checking for memory h github com cockroachdb cockroach vendor github com golang dep internal importers gvt github com cockroachdb cockroach vendor github com golang dep internal importers vndr github com cockroachdb cockroach vendor golang org x perf cmd benchstat github com cockroachdb cockroach vendor golang org x tools cmd goyacc go internal gccgoimporter yes checking for strings h go internal gcimporter github com cockroachdb cockroach vendor github com golang dep internal importers go internal srcimporter yes github com cockroachdb cockroach vendor github com golang dep cmd dep checking for inttypes h go importer yes github com cockroachdb cockroach vendor golang org x tools cmd goimports github com cockroachdb cockroach vendor golang org x tools cmd stringer checking for stdint h yes checking for unistd h yes checking whether byte ordering is bigendian no checking size of void checking size of int checking size of long checking size of long long checking size of intmax t checking build system type pc linux gnu checking host system type pc linux gnu checking whether pause instruction is compilable yes checking for ar ar checking malloc h usability yes checking malloc h presence building cxx object cmakefiles roach dir engine cc o yes checking for malloc h yes checking whether malloc usable size definition can use const argument no checking for library containing log building cxx object cmakefiles roach dir merge cc o touch bin bootstrap building cxx object cmakefiles roach dir mvcc cc o lm checking whether attribute syntax is compilable building cxx object cmakefiles roach dir options cc o building cxx object cmakefiles roach dir snapshot cc o yes checking whether compiler supports fvisibility hidden yes checking whether compiler supports werror yes checking whether compiler supports herror on warning no checking whether tls model attribute is compilable yes checking whether compiler supports werror yes checking whether compiler supports herror on warning no checking whether alloc size attribute is compilable no checking whether compiler supports werror yes checking whether compiler supports herror on warning no checking whether format gnu printf attribute is compilable no checking whether compiler supports werror yes checking whether compiler supports herror on warning no checking whether format printf attribute is compilable yes checking for a bsd compatible install usr bin install c checking for ranlib ranlib checking for ld usr bin ld checking for autoconf usr bin autoconf checking for memalign yes checking for valloc yes checking whether compiler supports yes checking whether compiler supports funroll loops yes checking unwind h usability yes checking unwind h presence yes checking for unwind h yes checking for unwind backtrace in lgcc yes checking configured backtracing method libgcc checking for sbrk yes checking whether utrace is compilable no checking whether valgrind is compilable no checking whether a program using builtin unreachable is compilable yes checking whether a program using builtin ffsl is compilable yes checking lg page missing version file and unable to generate it creating bogus version checking pthread h usability yes checking pthread h presence yes checking for pthread h yes checking for pthread create in lpthread yes checking whether pthread atfork is compilable building cxx object cmakefiles roach dir protos roachpb data pb cc o yes checking for library containing clock gettime none required checking whether clock gettime clock monotonic coarse is compilable yes checking whether clock gettime clock monotonic is compilable yes checking whether mach absolute time is compilable no checking whether compiler supports werror yes checking whether syscall is compilable yes checking for secure getenv yes checking for issetugid no checking for malloc thread cleanup no checking for pthread mutex init calloc cb building cxx object cmakefiles roach dir protos roachpb internal pb cc o no checking for tls yes checking whether atomics is compilable yes checking whether atomic is compilable no checking whether darwin osatomic is compilable building cxx object cmakefiles roach dir protos roachpb metadata pb cc o no checking whether madvise is compilable yes checking whether madvise madv free is compilable no checking whether madvise madv dontneed is compilable building cxx object cmakefiles roach dir protos storage engine enginepb mvcc pb cc o building cxx object cmakefiles roach dir protos storage engine enginepb pb cc o yes checking whether madvise madv hugepage is compilable yes checking whether to force bit sync add sub and fetch no checking whether to force bit sync add sub and fetch no checking for builtin clz building cxx object cmakefiles roach dir protos storage engine enginepb file registry pb cc o yes checking whether darwin os unfair lock is compilable no checking whether darwin osspin is compilable no checking whether glibc malloc hook is compilable yes checking whether glibc memalign hook is compilable yes checking whether pthreads adaptive mutexes is compilable yes checking for stdbool h that conforms to yes checking for bool yes configure creating config status config status creating makefile config status creating jemalloc pc config status creating doc html xsl config status creating doc manpages xsl config status creating doc jemalloc xml building cxx object cmakefiles roach dir protos storage engine enginepb rocksdb pb cc o config status creating include jemalloc jemalloc macros h config status creating include jemalloc jemalloc protos h config status creating include jemalloc jemalloc typedefs h config status creating include jemalloc internal jemalloc internal h config status creating test test sh config status creating test include test jemalloc test h building cxx object cmakefiles roach dir protos util hlc legacy timestamp pb cc o config status creating config stamp config status creating bin jemalloc config config status creating bin jemalloc sh config status creating bin jeprof building cxx object cmakefiles roach dir protos util hlc timestamp pb cc o config status creating include jemalloc jemalloc defs h config status creating include jemalloc internal jemalloc internal defs h config status creating test include test jemalloc test defs h config status executing include jemalloc internal private namespace h commands config status executing include jemalloc internal private unnamespace h commands config status executing include jemalloc internal public symbols txt commands config status executing include jemalloc internal public namespace h commands config status executing include jemalloc internal public unnamespace h commands config status executing include jemalloc internal size classes h commands building cxx object cmakefiles roach dir protos util unresolved addr pb cc o config status executing include jemalloc jemalloc protos jet h commands config status executing include jemalloc jemalloc rename h commands config status executing include jemalloc jemalloc mangle h commands config status executing include jemalloc jemalloc mangle jet h commands set euo pipefail ret cd pkg sql parser gen goyacc p sql o sql go tmp sql y if expr ret conflicts dev null then echo ret exit fi config status executing include jemalloc jemalloc h commands jemalloc version library revision config enable prof cc cc configure cflags std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops specified cflags extra cflags cppflags d gnu source d reentrant ldflags extra ldflags libs lm lgcc lm lpthread rpath extra xsltproc false xslroot prefix usr local bindir usr local bin datadir usr local share includedir usr local include libdir usr local lib mandir usr local share man srcroot go src github com cockroachdb cockroach c deps jemalloc abs srcroot go src github com cockroachdb cockroach c deps jemalloc objroot abs objroot go native pc linux gnu jemalloc jemalloc prefix jemalloc private namespace je install suffix malloc conf autogen cc silence debug code coverage stats prof prof libunwind prof libgcc prof gcc tcache thp fill utrace valgrind xmalloc munmap lazy lock tls cache oblivious find pkg name node modules prune o type f name pb gw go exec rm build werror sh go native pc linux gnu protobuf protoc i pkg vendor github com gogo protobuf vendor github com gogo protobuf protobuf vendor github com coreos vendor github com grpc ecosystem grpc gateway third party googleapis grpc gateway out logtostderr true request context true pkg pkg server serverpb admin proto pkg server serverpb status proto pkg server serverpb authentication proto build werror sh go native pc linux gnu protobuf protoc i pkg vendor github com gogo protobuf vendor github com gogo protobuf protobuf vendor github com coreos vendor github com grpc ecosystem grpc gateway third party googleapis grpc gateway out logtostderr true request context true pkg pkg ts tspb timeseries proto sed i e s golang org x net context context g pkg server serverpb admin pb gw go pkg server serverpb status pb gw go pkg server serverpb authentication pb gw go pkg ts tspb timeseries pb gw go gofmt s w pkg server serverpb admin pb gw go pkg server serverpb status pb gw go pkg server serverpb authentication pb gw go pkg ts tspb timeseries pb gw go goimports w pkg server serverpb admin pb gw go pkg server serverpb status pb gw go pkg server serverpb authentication pb gw go pkg ts tspb timeseries pb gw go sed i import d pkg acceptance cluster testconfig pb go pkg build info pb go pkg ccl backupccl backup pb go pkg ccl baseccl encryption options pb go pkg ccl storageccl engineccl enginepbccl key registry pb go pkg ccl utilccl licenseccl license pb go pkg config system pb go pkg config zone pb go pkg gossip gossip pb go pkg internal client lease pb go pkg roachpb api pb go pkg roachpb app stats pb go pkg roachpb csv pb go pkg roachpb data pb go pkg roachpb errors pb go pkg roachpb internal pb go pkg roachpb internal raft pb go pkg roachpb metadata pb go pkg rpc heartbeat pb go pkg server diagnosticspb diagnostics pb go pkg server serverpb admin pb go pkg server serverpb authentication pb go pkg server serverpb init pb go pkg server serverpb status pb go pkg server status status pb go pkg settings cluster cluster version pb go pkg sql distsqlrun api pb go pkg sql distsqlrun data pb go pkg sql distsqlrun processors pb go pkg sql jobs jobs pb go pkg sql pgwire pgerror errors pb go pkg sql sqlbase encoded datum pb go pkg sql sqlbase join type pb go pkg sql sqlbase privilege pb go pkg sql sqlbase structured pb go pkg sql stats histogram pb go pkg storage api pb go pkg storage engine enginepb file registry pb go pkg storage engine enginepb mvcc pb go pkg storage engine enginepb pb go pkg storage engine enginepb rocksdb pb go pkg storage lease status pb go pkg storage liveness pb go pkg storage log pb go pkg storage raft pb go pkg storage storagebase proposer kv pb go pkg storage storagebase state pb go pkg ts tspb timeseries pb go pkg util hlc legacy timestamp pb go pkg util hlc timestamp pb go pkg util log log pb go pkg util protoutil clone pb go pkg util tracing recorded span pb go pkg util unresolved addr pb go linking cxx static library libroach a sed i e s import fmt math github com cockroachdb cockroach pkg fmt math g pkg acceptance cluster testconfig pb go pkg build info pb go pkg ccl backupccl backup pb go pkg ccl baseccl encryption options pb go pkg ccl storageccl engineccl enginepbccl key registry pb go pkg ccl utilccl licenseccl license pb go pkg config system pb go pkg config zone pb go pkg gossip gossip pb go pkg internal client lease pb go pkg roachpb api pb go pkg roachpb app stats pb go pkg roachpb csv pb go pkg roachpb data pb go pkg roachpb errors pb go pkg roachpb internal pb go pkg roachpb internal raft pb go pkg roachpb metadata pb go pkg rpc heartbeat pb go pkg server diagnosticspb diagnostics pb go pkg server serverpb admin pb go pkg server serverpb authentication pb go pkg server serverpb init pb go pkg server serverpb status pb go pkg server status status pb go pkg settings cluster cluster version pb go pkg sql distsqlrun api pb go pkg sql distsqlrun data pb go pkg sql distsqlrun processors pb go pkg sql jobs jobs pb go pkg sql pgwire pgerror errors pb go pkg sql sqlbase encoded datum pb go pkg sql sqlbase join type pb go pkg sql sqlbase privilege pb go pkg sql sqlbase structured pb go pkg sql stats histogram pb go pkg storage api pb go pkg storage engine enginepb file registry pb go pkg storage engine enginepb mvcc pb go pkg storage engine enginepb pb go pkg storage engine enginepb rocksdb pb go pkg storage lease status pb go pkg storage liveness pb go pkg storage log pb go pkg storage raft pb go pkg storage storagebase proposer kv pb go pkg storage storagebase state pb go pkg ts tspb timeseries pb go pkg util hlc legacy timestamp pb go pkg util hlc timestamp pb go pkg util log log pb go pkg util protoutil clone pb go pkg util tracing recorded span pb go pkg util unresolved addr pb go built target roach touch bin gw protobuf sources scanning dependencies of target roachccl building cxx object cmakefiles roachccl dir ccl db cc o sed i e s cockroachdb cockroach pkg etcd coreos g pkg acceptance cluster testconfig pb go pkg build info pb go pkg ccl backupccl backup pb go pkg ccl baseccl encryption options pb go pkg ccl storageccl engineccl enginepbccl key registry pb go pkg ccl utilccl licenseccl license pb go pkg config system pb go pkg config zone pb go pkg gossip gossip pb go pkg internal client lease pb go pkg roachpb api pb go pkg roachpb app stats pb go pkg roachpb csv pb go pkg roachpb data pb go pkg roachpb errors pb go pkg roachpb internal pb go pkg roachpb internal raft pb go pkg roachpb metadata pb go pkg rpc heartbeat pb go pkg server diagnosticspb diagnostics pb go pkg server serverpb admin pb go pkg server serverpb authentication pb go pkg server serverpb init pb go pkg server serverpb status pb go pkg server status status pb go pkg settings cluster cluster version pb go pkg sql distsqlrun api pb go pkg sql distsqlrun data pb go pkg sql distsqlrun processors pb go pkg sql jobs jobs pb go pkg sql pgwire pgerror errors pb go pkg sql sqlbase encoded datum pb go pkg sql sqlbase join type pb go pkg sql sqlbase privilege pb go pkg sql sqlbase structured pb go pkg sql stats histogram pb go pkg storage api pb go pkg storage engine enginepb file registry pb go pkg storage engine enginepb mvcc pb go pkg storage engine enginepb pb go pkg storage engine enginepb rocksdb pb go pkg storage lease status pb go pkg storage liveness pb go pkg storage log pb go pkg storage raft pb go pkg storage storagebase proposer kv pb go pkg storage storagebase state pb go pkg ts tspb timeseries pb go pkg util hlc legacy timestamp pb go pkg util hlc timestamp pb go pkg util log log pb go pkg util protoutil clone pb go pkg util tracing recorded span pb go pkg util unresolved addr pb go sed i e s github com cockroachdb cockroach pkg bytes encoding binary errors fmt io math github com google golang org g pkg acceptance cluster testconfig pb go pkg build info pb go pkg ccl backupccl backup pb go pkg ccl baseccl encryption options pb go pkg ccl storageccl engineccl enginepbccl key registry pb go pkg ccl utilccl licenseccl license pb go pkg config system pb go pkg config zone pb go pkg gossip gossip pb go pkg internal client lease pb go pkg roachpb api pb go pkg roachpb app stats pb go pkg roachpb csv pb go pkg roachpb data pb go pkg roachpb errors pb go pkg roachpb internal pb go pkg roachpb internal raft pb go pkg roachpb metadata pb go pkg rpc heartbeat pb go pkg server diagnosticspb diagnostics pb go pkg server serverpb admin pb go pkg server serverpb authentication pb go pkg server serverpb init pb go pkg server serverpb status pb go pkg server status status pb go pkg settings cluster cluster version pb go pkg sql distsqlrun api pb go pkg sql distsqlrun data pb go pkg sql distsqlrun processors pb go pkg sql jobs jobs pb go pkg sql pgwire pgerror errors pb go pkg sql sqlbase encoded datum pb go pkg sql sqlbase join type pb go pkg sql sqlbase privilege pb go pkg sql sqlbase structured pb go pkg sql stats histogram pb go pkg storage api pb go pkg storage engine enginepb file registry pb go pkg storage engine enginepb mvcc pb go pkg storage engine enginepb pb go pkg storage engine enginepb rocksdb pb go pkg storage lease status pb go pkg storage liveness pb go pkg storage log pb go pkg storage raft pb go pkg storage storagebase proposer kv pb go pkg storage storagebase state pb go pkg ts tspb timeseries pb go pkg util hlc legacy timestamp pb go pkg util hlc timestamp pb go pkg util log log pb go pkg util protoutil clone pb go pkg util tracing recorded span pb go pkg util unresolved addr pb go sed i e s golang org x net context context g pkg acceptance cluster testconfig pb go pkg build info pb go pkg ccl backupccl backup pb go pkg ccl baseccl encryption options pb go pkg ccl storageccl engineccl enginepbccl key registry pb go pkg ccl utilccl licenseccl license pb go pkg config system pb go pkg config zone pb go pkg gossip gossip pb go pkg internal client lease pb go pkg roachpb api pb go pkg roachpb app stats pb go pkg roachpb csv pb go pkg roachpb data pb go pkg roachpb errors pb go pkg roachpb internal pb go pkg roachpb internal raft pb go pkg roachpb metadata pb go pkg rpc heartbeat pb go pkg server diagnosticspb diagnostics pb go pkg server serverpb admin pb go pkg server serverpb authentication pb go pkg server serverpb init pb go pkg server serverpb status pb go pkg server status status pb go pkg settings cluster cluster version pb go pkg sql distsqlrun api pb go pkg sql distsqlrun data pb go pkg sql distsqlrun processors pb go pkg sql jobs jobs pb go pkg sql pgwire pgerror errors pb go pkg sql sqlbase encoded datum pb go pkg sql sqlbase join type pb go pkg sql sqlbase privilege pb go pkg sql sqlbase structured pb go pkg sql stats histogram pb go pkg storage api pb go pkg storage engine enginepb file registry pb go pkg storage engine enginepb mvcc pb go pkg storage engine enginepb pb go pkg storage engine enginepb rocksdb pb go pkg storage lease status pb go pkg storage liveness pb go pkg storage log pb go pkg storage raft pb go pkg storage storagebase proposer kv pb go pkg storage storagebase state pb go pkg ts tspb timeseries pb go pkg util hlc legacy timestamp pb go pkg util hlc timestamp pb go pkg util log log pb go pkg util protoutil clone pb go pkg util tracing recorded span pb go pkg util unresolved addr pb go gofmt s w pkg acceptance cluster testconfig pb go pkg build info pb go pkg ccl backupccl backup pb go pkg ccl baseccl encryption options pb go pkg ccl storageccl engineccl enginepbccl key registry pb go pkg ccl utilccl licenseccl license pb go pkg config system pb go pkg config zone pb go pkg gossip gossip pb go pkg internal client lease pb go pkg roachpb api pb go pkg roachpb app stats pb go pkg roachpb csv pb go pkg roachpb data pb go pkg roachpb errors pb go pkg roachpb internal pb go pkg roachpb internal raft pb go pkg roachpb metadata pb go pkg rpc heartbeat pb go pkg server diagnosticspb diagnostics pb go pkg server serverpb admin pb go pkg server serverpb authentication pb go pkg server serverpb init pb go pkg server serverpb status pb go pkg server status status pb go pkg settings cluster cluster version pb go pkg sql distsqlrun api pb go pkg sql distsqlrun data pb go pkg sql distsqlrun processors pb go pkg sql jobs jobs pb go pkg sql pgwire pgerror errors pb go pkg sql sqlbase encoded datum pb go pkg sql sqlbase join type pb go pkg sql sqlbase privilege pb go pkg sql sqlbase structured pb go pkg sql stats histogram pb go pkg storage api pb go pkg storage engine enginepb file registry pb go pkg storage engine enginepb mvcc pb go pkg storage engine enginepb pb go pkg storage engine enginepb rocksdb pb go pkg storage lease status pb go pkg storage liveness pb go pkg storage log pb go pkg storage raft pb go pkg storage storagebase proposer kv pb go pkg storage storagebase state pb go pkg ts tspb timeseries pb go pkg util hlc legacy timestamp pb go pkg util hlc timestamp pb go pkg util log log pb go pkg util protoutil clone pb go pkg util tracing recorded span pb go pkg util unresolved addr pb go cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src jemalloc o go src github com cockroachdb cockroach c deps jemalloc src jemalloc c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src arena o go src github com cockroachdb cockroach c deps jemalloc src arena c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src atomic o go src github com cockroachdb cockroach c deps jemalloc src atomic c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src base o go src github com cockroachdb cockroach c deps jemalloc src base c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src bitmap o go src github com cockroachdb cockroach c deps jemalloc src bitmap c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src chunk o go src github com cockroachdb cockroach c deps jemalloc src chunk c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src chunk dss o go src github com cockroachdb cockroach c deps jemalloc src chunk dss c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src chunk mmap o go src github com cockroachdb cockroach c deps jemalloc src chunk mmap c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src ckh o go src github com cockroachdb cockroach c deps jemalloc src ckh c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src ctl o go src github com cockroachdb cockroach c deps jemalloc src ctl c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src extent o go src github com cockroachdb cockroach c deps jemalloc src extent c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src hash o go src github com cockroachdb cockroach c deps jemalloc src hash c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src huge o go src github com cockroachdb cockroach c deps jemalloc src huge c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src mb o go src github com cockroachdb cockroach c deps jemalloc src mb c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src mutex o go src github com cockroachdb cockroach c deps jemalloc src mutex c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src nstime o go src github com cockroachdb cockroach c deps jemalloc src nstime c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src pages o go src github com cockroachdb cockroach c deps jemalloc src pages c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src prng o go src github com cockroachdb cockroach c deps jemalloc src prng c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src prof o go src github com cockroachdb cockroach c deps jemalloc src prof c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src quarantine o go src github com cockroachdb cockroach c deps jemalloc src quarantine c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src rtree o go src github com cockroachdb cockroach c deps jemalloc src rtree c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src stats o go src github com cockroachdb cockroach c deps jemalloc src stats c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src spin o go src github com cockroachdb cockroach c deps jemalloc src spin c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src tcache o go src github com cockroachdb cockroach c deps jemalloc src tcache c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src t building cxx object cmakefiles roachccl dir ccl key manager cc o touch bin go protobuf sources echo code generated by goyacc do not edit echo generated file do not edit cat pkg sql parser gen sql go tmp sed e s const const lex g pkg sql parser sql go echo code generated by make do not edit echo generated file do not edit echo echo package lex echo grep const pkg sql parser gen sql go tmp pkg sql lex tokens go go run pkg cmd docgen main funcs go functions docs generated sql quiet building cxx object cmakefiles roachccl dir protosccl ccl baseccl encryption options pb cc o icker o go src github com cockroachdb cockroach c deps jemalloc src ticker c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src tsd o go src github com cockroachdb cockroach c deps jemalloc src tsd c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src util o go src github com cockroachdb cockroach c deps jemalloc src util c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src jemalloc pic o go src github com cockroachdb cockroach c deps jemalloc src jemalloc c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src witness o go src github com cockroachdb cockroach c deps jemalloc src witness c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src arena pic o go src github com cockroachdb cockroach c deps jemalloc src arena c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src atomic pic o go src github com cockroachdb cockroach c deps jemalloc src atomic c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src base pic o go src github com cockroachdb cockroach c deps jemalloc src base c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src bitmap pic o go src github com cockroachdb cockroach c deps jemalloc src bitmap c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src chunk pic o go src github com cockroachdb cockroach c deps jemalloc src chunk c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src chunk dss pic o go src github com cockroachdb cockroach c deps jemalloc src chunk dss c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src chunk mmap pic o go src github com cockroachdb cockroach c deps jemalloc src chunk mmap c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemal building cxx object cmakefiles roachccl dir protosccl ccl storageccl engineccl enginepbccl key registry pb cc o loc include iinclude o src ckh pic o go src github com cockroachdb cockroach c deps jemalloc src ckh c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src ctl pic o go src github com cockroachdb cockroach c deps jemalloc src ctl c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src extent pic o go src github com cockroachdb cockroach c deps jemalloc src extent c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src hash pic o go src github com cockroachdb cockroach c deps jemalloc src hash c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src huge pic o go src github com cockroachdb cockroach c deps jemalloc src huge c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src mb pic o go src github com cockroachdb cockroach c deps jemalloc src mb c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src mutex pic o go src github com cockroachdb cockroach c deps jemalloc src mutex c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src nstime pic o go src github com cockroachdb cockroach c deps jemalloc src nstime c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src pages pic o go src github com cockroachdb cockroach c deps jemalloc src pages c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src prng pic o go src github com cockroachdb cockroach c deps jemalloc src prng c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src prof pic o go src github com cockroachdb cockroach c deps jemalloc src prof c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src quarantine pic o go src github com cockroachdb cockroach c deps jemalloc src quarantine c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go sar u modifier ignored since d is the default see u ar u modifier ignored since d is the default see u rc github com cockroachdb cockroach c deps jemalloc include iinclude o src rtree pic o go src github com cockroachdb cockroach c deps jemalloc src rtree c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src stats pic o go src github com cockroachdb cockroach c deps jemalloc src stats c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src spin pic o go src github com cockroachdb cockroach c deps jemalloc src spin c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src tcache pic o go src github com cockroachdb cockroach c deps jemalloc src tcache c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src ticker pic o go src github com cockroachdb cockroach c deps jemalloc src ticker c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src tsd pic o go src github com cockroachdb cockroach c deps jemalloc src tsd c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src util pic o go src github com cockroachdb cockroach c deps jemalloc src util c cc std wall werror declaration after statement wshorten to wsign compare pipe fvisibility hidden funroll loops fpic dpic c d gnu source d reentrant i go src github com cockroachdb cockroach c deps jemalloc include iinclude o src witness pic o go src github com cockroachdb cockroach c deps jemalloc src witness c ar crus lib libjemalloc a src jemalloc o src arena o src atomic o src base o src bitmap o src chunk o src chunk dss o src chunk mmap o src ckh o src ctl o src extent o src hash o src huge o src mb o src mutex o src nstime o src pages o src prng o src prof o src quarantine o src rtree o src stats o src spin o src tcache o src ticker o src tsd o src util o src witness o ar crus lib libjemalloc pic a src jemalloc pic o src arena pic o src atomic pic o src base pic o src bitmap pic o src chunk pic o src chunk dss pic o src chunk mmap pic o src ckh pic o src ctl pic o src extent pic o src hash pic o src huge pic o src mb pic o src mutex pic o src nstime pic o src pages pic o src prng pic o src prof pic o src quarantine pic o src rtree pic o src stats pic o src spin pic o src tcache pic o src ticker pic o src tsd pic o src util pic o src witness pic o built target build version scanning dependencies of target rocksdb building cxx object cmakefiles rocksdb dir cache sharded cache cc o building cxx object cmakefiles rocksdb dir cache clock cache cc o building cxx object cmakefiles rocksdb dir cache lru cache cc o building cxx object cmakefiles rocksdb dir db builder cc o building cxx object cmakefiles rocksdb dir db c cc o building cxx object cmakefiles rocksdb dir db column family cc o touch bin docgen functions building cxx object cmakefiles rocksdb dir db compacted db impl cc o linking cxx static library libroachccl a built target roachccl building cxx object cmakefiles rocksdb dir db compaction cc o building cxx object cmakefiles rocksdb dir db compaction iterator cc o building cxx object cmakefiles rocksdb dir db compaction job cc o building cxx object cmakefiles rocksdb dir db compaction picker cc o building cxx object cmakefiles rocksdb dir db compaction picker universal cc o building cxx object cmakefiles rocksdb dir db convenience cc o building cxx object cmakefiles rocksdb dir db db filesnapshot cc o building cxx object cmakefiles rocksdb dir db db impl cc o building cxx object cmakefiles rocksdb dir db db impl write cc o building cxx object cmakefiles rocksdb dir db db impl compaction flush cc o building cxx object cmakefiles rocksdb dir db db impl files cc o building cxx object cmakefiles rocksdb dir db db impl open cc o building cxx object cmakefiles rocksdb dir db db impl debug cc o building cxx object cmakefiles rocksdb dir db db impl experimental cc o building cxx object cmakefiles rocksdb dir db db impl readonly cc o building cxx object cmakefiles rocksdb dir db db info dumper cc o building cxx object cmakefiles rocksdb dir db db iter cc o building cxx object cmakefiles rocksdb dir db dbformat cc o building cxx object cmakefiles rocksdb dir db event helpers cc o building cxx object cmakefiles rocksdb dir db experimental cc o building cxx object cmakefiles rocksdb dir db external sst file ingestion job cc o building cxx object cmakefiles rocksdb dir db file indexer cc o building cxx object cmakefiles rocksdb dir db flush job cc o building cxx object cmakefiles rocksdb dir db flush scheduler cc o building cxx object cmakefiles rocksdb dir db forward iterator cc o building cxx object cmakefiles rocksdb dir db internal stats cc o building cxx object cmakefiles rocksdb dir db log reader cc o building cxx object cmakefiles rocksdb dir db log writer cc o building cxx object cmakefiles rocksdb dir db malloc stats cc o building cxx object cmakefiles rocksdb dir db managed iterator cc o building cxx object cmakefiles rocksdb dir db memtable cc o building cxx object cmakefiles rocksdb dir db memtable list cc o building cxx object cmakefiles rocksdb dir db merge helper cc o building cxx object cmakefiles rocksdb dir db range del aggregator cc o building cxx object cmakefiles rocksdb dir db repair cc o building cxx object cmakefiles rocksdb dir db table cache cc o building cxx object cmakefiles rocksdb dir db table properties collector cc o building cxx object cmakefiles rocksdb dir db transaction log impl cc o building cxx object cmakefiles rocksdb dir db version builder cc o building cxx object cmakefiles rocksdb dir db version edit cc o building cxx object cmakefiles rocksdb dir db version set cc o building cxx object cmakefiles rocksdb dir db wal manager cc o building cxx object cmakefiles rocksdb dir db write batch cc o building cxx object cmakefiles rocksdb dir db write thread cc o building cxx object cmakefiles rocksdb dir env env cc o building cxx object cmakefiles rocksdb dir env env encryption cc o building cxx object cmakefiles rocksdb dir env mock env cc o building cxx object cmakefiles rocksdb dir memtable alloc tracker cc o building cxx object cmakefiles rocksdb dir memtable hash cuckoo rep cc o building cxx object cmakefiles rocksdb dir memtable hash linklist rep cc o building cxx object cmakefiles rocksdb dir memtable hash skiplist rep cc o building cxx object cmakefiles rocksdb dir memtable skiplistrep cc o building cxx object cmakefiles rocksdb dir memtable vectorrep cc o building cxx object cmakefiles rocksdb dir memtable write buffer manager cc o building cxx object cmakefiles rocksdb dir monitoring histogram cc o building cxx object cmakefiles rocksdb dir monitoring instrumented mutex cc o building cxx object cmakefiles rocksdb dir monitoring iostats context cc o building cxx object cmakefiles rocksdb dir monitoring perf context cc o building cxx object cmakefiles rocksdb dir monitoring perf level cc o building cxx object cmakefiles rocksdb dir monitoring statistics cc o building cxx object cmakefiles rocksdb dir monitoring thread status updater cc o building cxx object cmakefiles rocksdb dir monitoring thread status util cc o building cxx object cmakefiles rocksdb dir monitoring thread status util debug cc o building cxx object cmakefiles rocksdb dir options cf options cc o building cxx object cmakefiles rocksdb dir options db options cc o building cxx object cmakefiles rocksdb dir options options cc o building cxx object cmakefiles rocksdb dir options options helper cc o building cxx object cmakefiles rocksdb dir options options parser cc o building cxx object cmakefiles rocksdb dir table adaptive table factory cc o building cxx object cmakefiles rocksdb dir table block cc o building cxx object cmakefiles rocksdb dir table block based filter block cc o building cxx object cmakefiles rocksdb dir table block based table builder cc o building cxx object cmakefiles rocksdb dir table block based table factory cc o building cxx object cmakefiles rocksdb dir table block based table reader cc o building cxx object cmakefiles rocksdb dir table block builder cc o building cxx object cmakefiles rocksdb dir table block prefix index cc o building cxx object cmakefiles rocksdb dir table bloom block cc o building cxx object cmakefiles rocksdb dir table cuckoo table builder cc o building cxx object cmakefiles rocksdb dir table cuckoo table factory cc o building cxx object cmakefiles rocksdb dir table cuckoo table reader cc o building cxx object cmakefiles rocksdb dir table format cc o building cxx object cmakefiles rocksdb dir table full filter block cc o building cxx object cmakefiles rocksdb dir table get context cc o building cxx object cmakefiles rocksdb dir table index builder cc o building cxx object cmakefiles rocksdb dir table iterator cc o building cxx object cmakefiles rocksdb dir table merging iterator cc o building cxx object cmakefiles rocksdb dir table meta blocks cc o building cxx object cmakefiles rocksdb dir table partitioned filter block cc o building cxx object cmakefiles rocksdb dir table persistent cache helper cc o building cxx object cmakefiles rocksdb dir table plain table builder cc o building cxx object cmakefiles rocksdb dir table plain table factory cc o building cxx object cmakefiles rocksdb dir table plain table index cc o building cxx object cmakefiles rocksdb dir table plain table key coding cc o building cxx object cmakefiles rocksdb dir table plain table reader cc o building cxx object cmakefiles rocksdb dir table sst file writer cc o building cxx object cmakefiles rocksdb dir table table properties cc o building cxx object cmakefiles rocksdb dir table two level iterator cc o building cxx object cmakefiles rocksdb dir tools db bench tool cc o building cxx object cmakefiles rocksdb dir tools dump db dump tool cc o building cxx object cmakefiles rocksdb dir tools ldb cmd cc o building cxx object cmakefiles rocksdb dir tools sst dump tool cc o building cxx object cmakefiles rocksdb dir util auto roll logger cc o building cxx object cmakefiles rocksdb dir util arena cc o building cxx object cmakefiles rocksdb dir util bloom cc o building cxx object cmakefiles rocksdb dir util coding cc o building cxx object cmakefiles rocksdb dir util comparator cc o building cxx object cmakefiles rocksdb dir util concurrent arena cc o building cxx object cmakefiles rocksdb dir util cc o building cxx object cmakefiles rocksdb dir util delete scheduler cc o building cxx object cmakefiles rocksdb dir util dynamic bloom cc o building cxx object cmakefiles rocksdb dir util event logger cc o building cxx object cmakefiles rocksdb dir util file reader writer cc o building cxx object cmakefiles rocksdb dir util file util cc o building cxx object cmakefiles rocksdb dir util filename cc o building cxx object cmakefiles rocksdb dir util hash cc o building cxx object cmakefiles rocksdb dir util log buffer cc o building cxx object cmakefiles rocksdb dir util random cc o building cxx object cmakefiles rocksdb dir util rate limiter cc o building cxx object cmakefiles rocksdb dir util sst file manager impl cc o building cxx object cmakefiles rocksdb dir util status cc o building cxx object cmakefiles rocksdb dir util sync point cc o building cxx object cmakefiles rocksdb dir util testutil cc o building cxx object cmakefiles rocksdb dir util thread local cc o building cxx object cmakefiles rocksdb dir util threadpool imp cc o building cxx object cmakefiles rocksdb dir util transaction test util cc o building cxx object cmakefiles rocksdb dir utilities backupable backupable db cc o building cxx object cmakefiles rocksdb dir utilities blob db blob db cc o building cxx object cmakefiles rocksdb dir utilities blob db blob db impl cc o building cxx object cmakefiles rocksdb dir utilities blob db blob dump tool cc o building cxx object cmakefiles rocksdb dir utilities blob db blob file cc o building cxx object cmakefiles rocksdb dir utilities blob db blob log reader cc o building cxx object cmakefiles rocksdb dir utilities blob db blob log writer cc o building cxx object cmakefiles rocksdb dir utilities blob db blob log format cc o building cxx object cmakefiles rocksdb dir utilities blob db ttl extractor cc o building cxx object cmakefiles rocksdb dir utilities checkpoint checkpoint impl cc o building cxx object cmakefiles rocksdb dir utilities col buf decoder cc o building cxx object cmakefiles rocksdb dir utilities col buf encoder cc o building cxx object cmakefiles rocksdb dir utilities column aware encoding util cc o building cxx object cmakefiles rocksdb dir utilities date tiered date tiered db impl cc o building cxx object cmakefiles rocksdb dir utilities debug cc o building cxx object cmakefiles rocksdb dir utilities document document db cc o building cxx object cmakefiles rocksdb dir utilities document json document cc o building cxx object cmakefiles rocksdb dir utilities env timed cc o building cxx object cmakefiles rocksdb dir utilities geodb geodb impl cc o building cxx object cmakefiles rocksdb dir utilities memory memory util cc o building cxx object cmakefiles rocksdb dir utilities merge operators cc o building cxx object cmakefiles rocksdb dir utilities options options util cc o building cxx object cmakefiles rocksdb dir utilities persistent cache block cache tier cc o building cxx object cmakefiles rocksdb dir utilities persistent cache block cache tier file cc o building cxx object cmakefiles rocksdb dir utilities persistent cache block cache tier metadata cc o building cxx object cmakefiles rocksdb dir utilities persistent cache volatile tier impl cc o building cxx object cmakefiles rocksdb dir utilities redis redis lists cc o building cxx object cmakefiles rocksdb dir utilities simulator cache sim cache cc o building cxx object cmakefiles rocksdb dir utilities spatialdb spatial db cc o building cxx object cmakefiles rocksdb dir utilities transactions optimistic transaction db impl cc o building cxx object cmakefiles rocksdb dir utilities transactions optimistic transaction cc o building cxx object cmakefiles rocksdb dir utilities transactions pessimistic transaction cc o building cxx object cmakefiles rocksdb dir utilities transactions pessimistic transaction db cc o building cxx object cmakefiles rocksdb dir utilities transactions snapshot checker cc o building cxx object cmakefiles rocksdb dir utilities transactions transaction base cc o building cxx object cmakefiles rocksdb dir utilities transactions transaction lock mgr cc o building cxx object cmakefiles rocksdb dir utilities transactions transaction util cc o building cxx object cmakefiles rocksdb dir utilities transactions write prepared txn cc o building cxx object cmakefiles rocksdb dir utilities ttl db ttl impl cc o building cxx object cmakefiles rocksdb dir utilities write batch with index write batch with index cc o building cxx object cmakefiles rocksdb dir utilities write batch with index write batch with index internal cc o building cxx object cmakefiles rocksdb dir port port posix cc o building cxx object cmakefiles rocksdb dir env env posix cc o building cxx object cmakefiles rocksdb dir env io posix cc o linking cxx static library librocksdb a built target rocksdb go test v tags make pc linux gnu ldflags x github com cockroachdb cockroach pkg build typ development x github com cockroachdb cockroach pkg build tag dirty x github com cockroachdb cockroach pkg build utctime x github com cockroachdb cockroach pkg build rev x github com cockroachdb cockroach pkg build cgotargettriple pc linux gnu i github com cockroachdb cockroach pkg workload jsonload can t load package package github com cockroachdb cockroach pkg workload jsonload cannot find package github com cockroachdb cockroach pkg workload jsonload in any of usr local go src github com cockroachdb cockroach pkg workload jsonload from goroot go src github com cockroachdb cockroach pkg workload jsonload from gopath makefile recipe for target gotestdashi failed make error
1
245,176
20,750,851,199
IssuesEvent
2022-03-15 07:19:31
pingcap/tiflow
https://api.github.com/repos/pingcap/tiflow
opened
Kafka many_pk_or_uk fails
component/test
### Which jobs are flaking? kafka-integration-test ### Which test(s) are flaking? many_pk_or_uk ### Jenkins logs or GitHub Actions link ```log https://ci2.pingcap.net/blue/organizations/jenkins/cdc_ghpr_kafka_integration_test/detail/cdc_ghpr_kafka_integration_test/9780/pipeline#step-1266-log-310 ``` [Uploading log-many_pk_or_uk-cyclic_ab.tar.gz…]() ### Anything else we need to know - Does this test exist for other branches as well? - Has there been a high frequency of failure lately?
1.0
Kafka many_pk_or_uk fails - ### Which jobs are flaking? kafka-integration-test ### Which test(s) are flaking? many_pk_or_uk ### Jenkins logs or GitHub Actions link ```log https://ci2.pingcap.net/blue/organizations/jenkins/cdc_ghpr_kafka_integration_test/detail/cdc_ghpr_kafka_integration_test/9780/pipeline#step-1266-log-310 ``` [Uploading log-many_pk_or_uk-cyclic_ab.tar.gz…]() ### Anything else we need to know - Does this test exist for other branches as well? - Has there been a high frequency of failure lately?
test
kafka many pk or uk fails which jobs are flaking kafka integration test which test s are flaking many pk or uk jenkins logs or github actions link log anything else we need to know does this test exist for other branches as well has there been a high frequency of failure lately
1
92,458
8,364,392,470
IssuesEvent
2018-10-03 22:50:13
pg-irc/pathways-backend
https://api.github.com/repos/pg-irc/pathways-backend
closed
As a site manager, I would like to collection usage metrics to see how users are interacting with my site.
Epic User Testing
SEO, usage, success criteria, shareability. This should be done before a release so we can determine what is working, and what is not.
1.0
As a site manager, I would like to collection usage metrics to see how users are interacting with my site. - SEO, usage, success criteria, shareability. This should be done before a release so we can determine what is working, and what is not.
test
as a site manager i would like to collection usage metrics to see how users are interacting with my site seo usage success criteria shareability this should be done before a release so we can determine what is working and what is not
1
23,549
4,955,888,407
IssuesEvent
2016-12-01 21:45:01
OfficeDev/office-ui-fabric-js
https://api.github.com/repos/OfficeDev/office-ui-fabric-js
opened
Panel documentation needed
documentation
The code in **Using this component** does not work as expected. I see the button but clicking on it does nothing. Can you provide updated sample code please? Also, can you provide additional documentation on how to use: •Variants •States Thanks!
1.0
Panel documentation needed - The code in **Using this component** does not work as expected. I see the button but clicking on it does nothing. Can you provide updated sample code please? Also, can you provide additional documentation on how to use: •Variants •States Thanks!
non_test
panel documentation needed the code in using this component does not work as expected i see the button but clicking on it does nothing can you provide updated sample code please also can you provide additional documentation on how to use •variants •states thanks
0
86,396
24,837,943,326
IssuesEvent
2022-10-26 10:20:19
sandboxie-plus/Sandboxie
https://api.github.com/repos/sandboxie-plus/Sandboxie
closed
Processes broken out of sandbox can have weird command line arguments
Confirmation pending fixed in next build
### Describe what you noticed and did 1. Create a sandbox. 2. Specify a breakout process, for example notepad.exe. 3. Run PowerShell in the sandbox. 4. Enter `start notepad` in the PowerShell window. 5. Notepad will complain "The system cannot find the path specified". 6. Check the spawned Notepad process using tools like Process Explorer. Its command line arguments will have a weird string in it. The string can be different each time you launch Notepad this way, and it seems like a random piece of data in memory. Examples of the weird command line args: - stem32\cdpusersvc.dll,-101š - stem32\drivers\fltmgr.sys,¾ - em32\themeui.dll,-2682 Trying to run Notepad in this sandbox in different ways can have different results. (Although in all cases Notepad can break out successfully) - Start Notepad by Sandboxie's Run. Works as usual. - Enter `notepad` or `start notepad` in Command Prompt. Works as usual. - Enter `notepad` (not `start notepad`) in PowerShell. Works as usual. - Enter `(New-Object -ComObject Wscript.Shell).Run("notepad")` in PowerShell. Weird command line args appear. - Tried the WMI way of process creation, but failed. ### How often did you encounter it so far? _No response_ ### Affected program Not relevant ### Download link Not relevant ### Where is the program located? Not relevant to my request. ### Expected behavior The command line arguments should be passed correctly. ### What is your Windows edition and version? Windows 10 Home 21H2 64-bit ### In which Windows account you have this problem? A local or Microsoft account without special changes. ### Please mention any installed security software Windows Defender, Huorong ### What version of Sandboxie are you running? v1.4.2 64-bit ### Is it a new installation of Sandboxie? I just updated Sandboxie from a previous version (to be specified). ### Is it a regression? _No response_ ### In which sandbox type you have this problem? In a standard isolation sandbox (yellow sandbox icon). ### Can you reproduce this problem on an empty sandbox? I can confirm it also on an empty sandbox. ### Did you previously enable some security policy settings outside Sandboxie? _No response_ ### Crash dump _No response_ ### Trace log _No response_ ### Sandboxie.ini configuration _No response_
1.0
Processes broken out of sandbox can have weird command line arguments - ### Describe what you noticed and did 1. Create a sandbox. 2. Specify a breakout process, for example notepad.exe. 3. Run PowerShell in the sandbox. 4. Enter `start notepad` in the PowerShell window. 5. Notepad will complain "The system cannot find the path specified". 6. Check the spawned Notepad process using tools like Process Explorer. Its command line arguments will have a weird string in it. The string can be different each time you launch Notepad this way, and it seems like a random piece of data in memory. Examples of the weird command line args: - stem32\cdpusersvc.dll,-101š - stem32\drivers\fltmgr.sys,¾ - em32\themeui.dll,-2682 Trying to run Notepad in this sandbox in different ways can have different results. (Although in all cases Notepad can break out successfully) - Start Notepad by Sandboxie's Run. Works as usual. - Enter `notepad` or `start notepad` in Command Prompt. Works as usual. - Enter `notepad` (not `start notepad`) in PowerShell. Works as usual. - Enter `(New-Object -ComObject Wscript.Shell).Run("notepad")` in PowerShell. Weird command line args appear. - Tried the WMI way of process creation, but failed. ### How often did you encounter it so far? _No response_ ### Affected program Not relevant ### Download link Not relevant ### Where is the program located? Not relevant to my request. ### Expected behavior The command line arguments should be passed correctly. ### What is your Windows edition and version? Windows 10 Home 21H2 64-bit ### In which Windows account you have this problem? A local or Microsoft account without special changes. ### Please mention any installed security software Windows Defender, Huorong ### What version of Sandboxie are you running? v1.4.2 64-bit ### Is it a new installation of Sandboxie? I just updated Sandboxie from a previous version (to be specified). ### Is it a regression? _No response_ ### In which sandbox type you have this problem? In a standard isolation sandbox (yellow sandbox icon). ### Can you reproduce this problem on an empty sandbox? I can confirm it also on an empty sandbox. ### Did you previously enable some security policy settings outside Sandboxie? _No response_ ### Crash dump _No response_ ### Trace log _No response_ ### Sandboxie.ini configuration _No response_
non_test
processes broken out of sandbox can have weird command line arguments describe what you noticed and did create a sandbox specify a breakout process for example notepad exe run powershell in the sandbox enter start notepad in the powershell window notepad will complain the system cannot find the path specified check the spawned notepad process using tools like process explorer its command line arguments will have a weird string in it the string can be different each time you launch notepad this way and it seems like a random piece of data in memory examples of the weird command line args cdpusersvc dll š drivers fltmgr sys ¾ themeui dll trying to run notepad in this sandbox in different ways can have different results although in all cases notepad can break out successfully start notepad by sandboxie s run works as usual enter notepad or start notepad in command prompt works as usual enter notepad not start notepad in powershell works as usual enter new object comobject wscript shell run notepad in powershell weird command line args appear tried the wmi way of process creation but failed how often did you encounter it so far no response affected program not relevant download link not relevant where is the program located not relevant to my request expected behavior the command line arguments should be passed correctly what is your windows edition and version windows home bit in which windows account you have this problem a local or microsoft account without special changes please mention any installed security software windows defender huorong what version of sandboxie are you running bit is it a new installation of sandboxie i just updated sandboxie from a previous version to be specified is it a regression no response in which sandbox type you have this problem in a standard isolation sandbox yellow sandbox icon can you reproduce this problem on an empty sandbox i can confirm it also on an empty sandbox did you previously enable some security policy settings outside sandboxie no response crash dump no response trace log no response sandboxie ini configuration no response
0
683,093
23,368,244,864
IssuesEvent
2022-08-10 17:16:38
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.bet365.gr - Site does not load
priority-normal type-fastclick severity-critical browser-fenix engine-gecko regression diagnosis-priority-p1
<!-- @browser: Firefox Mobile 105.0 --> <!-- @ua_header: Mozilla/5.0 (Android 12; Mobile; rv:105.0) Gecko/105.0 Firefox/105.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/108297 --> <!-- @extra_labels: browser-fenix, type-fastclick --> **URL**: https://www.bet365.gr/#/HO/ **Browser / Version**: Firefox Mobile 105.0 **Operating System**: Android 12 **Tested Another Browser**: Yes Edge **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: Page not loading at all. Cannot log in etc <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220730092553</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li><li>hasFastClick: true</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2022/7/bfb88c3f-8d37-4cae-854a-9433355c2901) _From [webcompat.com](https://webcompat.com/) with ❤️_
2.0
www.bet365.gr - Site does not load - <!-- @browser: Firefox Mobile 105.0 --> <!-- @ua_header: Mozilla/5.0 (Android 12; Mobile; rv:105.0) Gecko/105.0 Firefox/105.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/108297 --> <!-- @extra_labels: browser-fenix, type-fastclick --> **URL**: https://www.bet365.gr/#/HO/ **Browser / Version**: Firefox Mobile 105.0 **Operating System**: Android 12 **Tested Another Browser**: Yes Edge **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: Page not loading at all. Cannot log in etc <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220730092553</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li><li>hasFastClick: true</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2022/7/bfb88c3f-8d37-4cae-854a-9433355c2901) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
site does not load url browser version firefox mobile operating system android tested another browser yes edge problem type site is not usable description page not loading correctly steps to reproduce page not loading at all cannot log in etc browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false hasfastclick true from with ❤️
0
141,555
11,425,461,601
IssuesEvent
2020-02-03 19:53:03
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
[CI] IndexAuditUpgradeIT.testAuditLogs fails
:Security/Audit >test-failure v6.4.2
## https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+6.4+bwc-tests/8/console ## Reproduction line does not reproduce locally ``` ./gradlew :x-pack:qa:rolling-upgrade:with-system-key:v6.3.1#upgradedClusterTestRunner \ -Dtests.seed=57D936B12E792F96 \ -Dtests.class=org.elasticsearch.upgrades.IndexAuditUpgradeIT \ -Dtests.method="testAuditLogs" \ -Dtests.security.manager=true \ -Dtests.locale=bg \ -Dtests.timezone=Pacific/Marquesas ``` ## Example relevant log: ``` 00:00:46 > Throwable #1: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2343, key=upgraded-node-0}, {doc_count=2175, key=node-1}, {doc_count=2106, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at __randomizedtesting.SeedInfo.seed([57D936B12E792F96:300A75E395A08B0D]:0) 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:830) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:804) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.testAuditLogs(IndexAuditUpgradeIT.java:68) 00:00:46 > at java.lang.Thread.run(Thread.java:748) 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=1983, key=upgraded-node-0}, {doc_count=1794, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=1983, key=upgraded-node-0}, {doc_count=1794, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=1983, key=upgraded-node-0}, {doc_count=1794, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 2> NOTE: leaving temporary files on disk at: /var/lib/jenkins/workspace/elastic+elasticsearch+6.4+bwc-tests/x-pack/qa/rolling-upgrade/with-system-key/build/testrun/v6.3.1#upgradedClusterTestRunner/J0/temp/org.elasticsearch.upgrades.IndexAuditUpgradeIT_57D936B12E792F96-001 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=1983, key=upgraded-node-0}, {doc_count=1794, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 2> NOTE: test params are: codec=Asserting(Lucene70): {}, docValues:{}, maxPointsInLeafNode=174, maxMBSortInHeap=7.47073894872366, sim=RandomSimilarity(queryNorm=true): {}, locale=bg, timezone=Pacific/Marquesas 00:00:46 2> NOTE: Linux 4.4.0-1061-aws amd64/Oracle Corporation 1.8.0_181 (64-bit)/cpus=16,threads=1,free=433534592,total=514850816 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 2> NOTE: All tests run in this JVM: [IndexingIT, RollupIDUpgradeIT, TokenBackwardsCompatibilityIT, IndexAuditUpgradeIT] 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=1983, key=upgraded-node-0}, {doc_count=1794, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=1983, key=upgraded-node-0}, {doc_count=1794, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=1983, key=upgraded-node-0}, {doc_count=1794, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=1983, key=upgraded-node-0}, {doc_count=1794, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=1983, key=upgraded-node-0}, {doc_count=1794, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=1983, key=upgraded-node-0}, {doc_count=1794, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=2025, key=upgraded-node-0}, {doc_count=1841, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=2141, key=upgraded-node-0}, {doc_count=1933, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2212, key=upgraded-node-0}, {doc_count=2175, key=node-1}, {doc_count=2000, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 Completed [4/6] in 10.40s, 1 test, 1 failure <<< FAILURES! ``` ## Frequency It happened 8 times in August, not super frequent but seems consistent.
1.0
[CI] IndexAuditUpgradeIT.testAuditLogs fails - ## https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+6.4+bwc-tests/8/console ## Reproduction line does not reproduce locally ``` ./gradlew :x-pack:qa:rolling-upgrade:with-system-key:v6.3.1#upgradedClusterTestRunner \ -Dtests.seed=57D936B12E792F96 \ -Dtests.class=org.elasticsearch.upgrades.IndexAuditUpgradeIT \ -Dtests.method="testAuditLogs" \ -Dtests.security.manager=true \ -Dtests.locale=bg \ -Dtests.timezone=Pacific/Marquesas ``` ## Example relevant log: ``` 00:00:46 > Throwable #1: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2343, key=upgraded-node-0}, {doc_count=2175, key=node-1}, {doc_count=2106, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at __randomizedtesting.SeedInfo.seed([57D936B12E792F96:300A75E395A08B0D]:0) 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:830) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:804) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.testAuditLogs(IndexAuditUpgradeIT.java:68) 00:00:46 > at java.lang.Thread.run(Thread.java:748) 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=1983, key=upgraded-node-0}, {doc_count=1794, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=1983, key=upgraded-node-0}, {doc_count=1794, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=1983, key=upgraded-node-0}, {doc_count=1794, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 2> NOTE: leaving temporary files on disk at: /var/lib/jenkins/workspace/elastic+elasticsearch+6.4+bwc-tests/x-pack/qa/rolling-upgrade/with-system-key/build/testrun/v6.3.1#upgradedClusterTestRunner/J0/temp/org.elasticsearch.upgrades.IndexAuditUpgradeIT_57D936B12E792F96-001 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=1983, key=upgraded-node-0}, {doc_count=1794, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 2> NOTE: test params are: codec=Asserting(Lucene70): {}, docValues:{}, maxPointsInLeafNode=174, maxMBSortInHeap=7.47073894872366, sim=RandomSimilarity(queryNorm=true): {}, locale=bg, timezone=Pacific/Marquesas 00:00:46 2> NOTE: Linux 4.4.0-1061-aws amd64/Oracle Corporation 1.8.0_181 (64-bit)/cpus=16,threads=1,free=433534592,total=514850816 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 2> NOTE: All tests run in this JVM: [IndexingIT, RollupIDUpgradeIT, TokenBackwardsCompatibilityIT, IndexAuditUpgradeIT] 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=1983, key=upgraded-node-0}, {doc_count=1794, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=1983, key=upgraded-node-0}, {doc_count=1794, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=1983, key=upgraded-node-0}, {doc_count=1794, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=1983, key=upgraded-node-0}, {doc_count=1794, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=1983, key=upgraded-node-0}, {doc_count=1794, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=1983, key=upgraded-node-0}, {doc_count=1794, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=2025, key=upgraded-node-0}, {doc_count=1841, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2175, key=node-1}, {doc_count=2141, key=upgraded-node-0}, {doc_count=1933, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 > Suppressed: java.lang.AssertionError: Found node buckets [{doc_count=2431, key=node-2}, {doc_count=2212, key=upgraded-node-0}, {doc_count=2175, key=node-1}, {doc_count=2000, key=upgraded-node-1}, {doc_count=891, key=node-0}] 00:00:46 > Expected: a collection with size <6> 00:00:46 > but: collection size was <5> 00:00:46 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.assertNumUniqueNodeNameBuckets(IndexAuditUpgradeIT.java:126) 00:00:46 > at org.elasticsearch.upgrades.IndexAuditUpgradeIT.lambda$testAuditLogs$0(IndexAuditUpgradeIT.java:70) 00:00:46 > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:818) 00:00:46 > ... 39 more 00:00:46 Completed [4/6] in 10.40s, 1 test, 1 failure <<< FAILURES! ``` ## Frequency It happened 8 times in August, not super frequent but seems consistent.
test
indexauditupgradeit testauditlogs fails reproduction line does not reproduce locally gradlew x pack qa rolling upgrade with system key upgradedclustertestrunner dtests seed dtests class org elasticsearch upgrades indexauditupgradeit dtests method testauditlogs dtests security manager true dtests locale bg dtests timezone pacific marquesas example relevant log throwable java lang assertionerror found node buckets expected a collection with size but collection size was at randomizedtesting seedinfo seed at org hamcrest matcherassert assertthat matcherassert java at org elasticsearch upgrades indexauditupgradeit assertnumuniquenodenamebuckets indexauditupgradeit java at org elasticsearch upgrades indexauditupgradeit lambda testauditlogs indexauditupgradeit java at org elasticsearch test estestcase assertbusy estestcase java at org elasticsearch test estestcase assertbusy estestcase java at org elasticsearch upgrades indexauditupgradeit testauditlogs indexauditupgradeit java at java lang thread run thread java suppressed java lang assertionerror found node buckets expected a collection with size but collection size was at org hamcrest matcherassert assertthat matcherassert java at org elasticsearch upgrades indexauditupgradeit assertnumuniquenodenamebuckets indexauditupgradeit java at org elasticsearch upgrades indexauditupgradeit lambda testauditlogs indexauditupgradeit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror found node buckets expected a collection with size but collection size was at org hamcrest matcherassert assertthat matcherassert java at org elasticsearch upgrades indexauditupgradeit assertnumuniquenodenamebuckets indexauditupgradeit java at org elasticsearch upgrades indexauditupgradeit lambda testauditlogs indexauditupgradeit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror found node buckets expected a collection with size but collection size was at org hamcrest matcherassert assertthat matcherassert java at org elasticsearch upgrades indexauditupgradeit assertnumuniquenodenamebuckets indexauditupgradeit java at org elasticsearch upgrades indexauditupgradeit lambda testauditlogs indexauditupgradeit java note leaving temporary files on disk at var lib jenkins workspace elastic elasticsearch bwc tests x pack qa rolling upgrade with system key build testrun upgradedclustertestrunner temp org elasticsearch upgrades indexauditupgradeit at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror found node buckets note test params are codec asserting docvalues maxpointsinleafnode maxmbsortinheap sim randomsimilarity querynorm true locale bg timezone pacific marquesas note linux aws oracle corporation bit cpus threads free total expected a collection with size but collection size was at org hamcrest matcherassert assertthat matcherassert java at org elasticsearch upgrades indexauditupgradeit assertnumuniquenodenamebuckets indexauditupgradeit java note all tests run in this jvm at org elasticsearch upgrades indexauditupgradeit lambda testauditlogs indexauditupgradeit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror found node buckets expected a collection with size but collection size was at org hamcrest matcherassert assertthat matcherassert java at org elasticsearch upgrades indexauditupgradeit assertnumuniquenodenamebuckets indexauditupgradeit java at org elasticsearch upgrades indexauditupgradeit lambda testauditlogs indexauditupgradeit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror found node buckets expected a collection with size but collection size was at org hamcrest matcherassert assertthat matcherassert java at org elasticsearch upgrades indexauditupgradeit assertnumuniquenodenamebuckets indexauditupgradeit java at org elasticsearch upgrades indexauditupgradeit lambda testauditlogs indexauditupgradeit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror found node buckets expected a collection with size but collection size was at org hamcrest matcherassert assertthat matcherassert java at org elasticsearch upgrades indexauditupgradeit assertnumuniquenodenamebuckets indexauditupgradeit java at org elasticsearch upgrades indexauditupgradeit lambda testauditlogs indexauditupgradeit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror found node buckets expected a collection with size but collection size was at org hamcrest matcherassert assertthat matcherassert java at org elasticsearch upgrades indexauditupgradeit assertnumuniquenodenamebuckets indexauditupgradeit java at org elasticsearch upgrades indexauditupgradeit lambda testauditlogs indexauditupgradeit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror found node buckets expected a collection with size but collection size was at org hamcrest matcherassert assertthat matcherassert java at org elasticsearch upgrades indexauditupgradeit assertnumuniquenodenamebuckets indexauditupgradeit java at org elasticsearch upgrades indexauditupgradeit lambda testauditlogs indexauditupgradeit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror found node buckets expected a collection with size but collection size was at org hamcrest matcherassert assertthat matcherassert java at org elasticsearch upgrades indexauditupgradeit assertnumuniquenodenamebuckets indexauditupgradeit java at org elasticsearch upgrades indexauditupgradeit lambda testauditlogs indexauditupgradeit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror found node buckets expected a collection with size but collection size was at org hamcrest matcherassert assertthat matcherassert java at org elasticsearch upgrades indexauditupgradeit assertnumuniquenodenamebuckets indexauditupgradeit java at org elasticsearch upgrades indexauditupgradeit lambda testauditlogs indexauditupgradeit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror found node buckets expected a collection with size but collection size was at org hamcrest matcherassert assertthat matcherassert java at org elasticsearch upgrades indexauditupgradeit assertnumuniquenodenamebuckets indexauditupgradeit java at org elasticsearch upgrades indexauditupgradeit lambda testauditlogs indexauditupgradeit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror found node buckets expected a collection with size but collection size was at org hamcrest matcherassert assertthat matcherassert java at org elasticsearch upgrades indexauditupgradeit assertnumuniquenodenamebuckets indexauditupgradeit java at org elasticsearch upgrades indexauditupgradeit lambda testauditlogs indexauditupgradeit java at org elasticsearch test estestcase assertbusy estestcase java more completed in test failure failures frequency it happened times in august not super frequent but seems consistent
1
348,310
31,496,229,844
IssuesEvent
2023-08-31 02:44:05
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
Failing test: Serverless Observability Functional Tests.x-pack/test_serverless/functional/test_suites/observability/discover_log_explorer/dataset_selector·ts - serverless observability UI Discover Log-Explorer profile Dataset Selector without installed integrations or uncategorized data streams when navigating into Uncategorized data streams should display an empty prompt for no data streams
failed-test
A test failed on a tracked branch ``` Error: expected 3 to equal 0 at Assertion.assert (expect.js:100:11) at Assertion.apply (expect.js:227:8) at Assertion.be (expect.js:69:22) at Context.<anonymous> (dataset_selector.ts:145:54) at processTicksAndRejections (node:internal/process/task_queues:95:5) at Object.apply (wrap_function.js:73:16) ``` First failure: [CI Build - main](https://buildkite.com/elastic/kibana-serverless/builds/2436#018a4965-9765-464b-ad4a-d5d23daa4d9e) <!-- kibanaCiData = {"failed-test":{"test.class":"Serverless Observability Functional Tests.x-pack/test_serverless/functional/test_suites/observability/discover_log_explorer/dataset_selector·ts","test.name":"serverless observability UI Discover Log-Explorer profile Dataset Selector without installed integrations or uncategorized data streams when navigating into Uncategorized data streams should display an empty prompt for no data streams","test.failCount":1}} -->
1.0
Failing test: Serverless Observability Functional Tests.x-pack/test_serverless/functional/test_suites/observability/discover_log_explorer/dataset_selector·ts - serverless observability UI Discover Log-Explorer profile Dataset Selector without installed integrations or uncategorized data streams when navigating into Uncategorized data streams should display an empty prompt for no data streams - A test failed on a tracked branch ``` Error: expected 3 to equal 0 at Assertion.assert (expect.js:100:11) at Assertion.apply (expect.js:227:8) at Assertion.be (expect.js:69:22) at Context.<anonymous> (dataset_selector.ts:145:54) at processTicksAndRejections (node:internal/process/task_queues:95:5) at Object.apply (wrap_function.js:73:16) ``` First failure: [CI Build - main](https://buildkite.com/elastic/kibana-serverless/builds/2436#018a4965-9765-464b-ad4a-d5d23daa4d9e) <!-- kibanaCiData = {"failed-test":{"test.class":"Serverless Observability Functional Tests.x-pack/test_serverless/functional/test_suites/observability/discover_log_explorer/dataset_selector·ts","test.name":"serverless observability UI Discover Log-Explorer profile Dataset Selector without installed integrations or uncategorized data streams when navigating into Uncategorized data streams should display an empty prompt for no data streams","test.failCount":1}} -->
test
failing test serverless observability functional tests x pack test serverless functional test suites observability discover log explorer dataset selector·ts serverless observability ui discover log explorer profile dataset selector without installed integrations or uncategorized data streams when navigating into uncategorized data streams should display an empty prompt for no data streams a test failed on a tracked branch error expected to equal at assertion assert expect js at assertion apply expect js at assertion be expect js at context dataset selector ts at processticksandrejections node internal process task queues at object apply wrap function js first failure
1
340,999
30,560,517,147
IssuesEvent
2023-07-20 14:20:43
flutter/flutter
https://api.github.com/repos/flutter/flutter
opened
[camera] Integration tests require user interaction
a: tests team platform-android p: camera package team-android
I'm currently having to exclude the `camera` and `camera_android` tests from the new emulator CI tests I'm bringing up, because when they run a system dialog pops up asking for camera permissions. When running locally, if I tap it in the emulator everything runs fine. It looks like this is probably a regression from https://github.com/flutter/packages/pull/4348, because we have code in `test_driver/integration_test.dart` in the `camera` packages that is clearly supposed to handle this, by running adb commands once the process starts. Some options I can think of, although there may well be others: - Teach the repo tooling to use `flutter drive` for specific packages that opt out of `flutter test`. This will regress us in a way though, since `flutter drive` logs warnings that we should be using `flutter test`, and it's not clear to me if `drive` will even continue to be supported for non-web. - Find an alternate solution to grant permissions. E.g., can we grant them before the app is even installed, perhaps as part of emulator creation, or a new pre-test-run-script option in the repo tooling? I don't know what `adb` allows in that regard. - Add functionality to `flutter test` to allow for this in some way if possible (e.g., by allowing a custom driver even though one isn't normally used). I'll be disabling this test in the emulator CI bringup for now to avoid blocking the rest of that work, via a new script/configs/exclude_integration_android_emulator.yaml file. When this is fixed, the exclusion should be remove from that file. /cc @bparrishMines since he's been looking at the way integration tests run, the `drive` vs `test` differences, etc.
1.0
[camera] Integration tests require user interaction - I'm currently having to exclude the `camera` and `camera_android` tests from the new emulator CI tests I'm bringing up, because when they run a system dialog pops up asking for camera permissions. When running locally, if I tap it in the emulator everything runs fine. It looks like this is probably a regression from https://github.com/flutter/packages/pull/4348, because we have code in `test_driver/integration_test.dart` in the `camera` packages that is clearly supposed to handle this, by running adb commands once the process starts. Some options I can think of, although there may well be others: - Teach the repo tooling to use `flutter drive` for specific packages that opt out of `flutter test`. This will regress us in a way though, since `flutter drive` logs warnings that we should be using `flutter test`, and it's not clear to me if `drive` will even continue to be supported for non-web. - Find an alternate solution to grant permissions. E.g., can we grant them before the app is even installed, perhaps as part of emulator creation, or a new pre-test-run-script option in the repo tooling? I don't know what `adb` allows in that regard. - Add functionality to `flutter test` to allow for this in some way if possible (e.g., by allowing a custom driver even though one isn't normally used). I'll be disabling this test in the emulator CI bringup for now to avoid blocking the rest of that work, via a new script/configs/exclude_integration_android_emulator.yaml file. When this is fixed, the exclusion should be remove from that file. /cc @bparrishMines since he's been looking at the way integration tests run, the `drive` vs `test` differences, etc.
test
integration tests require user interaction i m currently having to exclude the camera and camera android tests from the new emulator ci tests i m bringing up because when they run a system dialog pops up asking for camera permissions when running locally if i tap it in the emulator everything runs fine it looks like this is probably a regression from because we have code in test driver integration test dart in the camera packages that is clearly supposed to handle this by running adb commands once the process starts some options i can think of although there may well be others teach the repo tooling to use flutter drive for specific packages that opt out of flutter test this will regress us in a way though since flutter drive logs warnings that we should be using flutter test and it s not clear to me if drive will even continue to be supported for non web find an alternate solution to grant permissions e g can we grant them before the app is even installed perhaps as part of emulator creation or a new pre test run script option in the repo tooling i don t know what adb allows in that regard add functionality to flutter test to allow for this in some way if possible e g by allowing a custom driver even though one isn t normally used i ll be disabling this test in the emulator ci bringup for now to avoid blocking the rest of that work via a new script configs exclude integration android emulator yaml file when this is fixed the exclusion should be remove from that file cc bparrishmines since he s been looking at the way integration tests run the drive vs test differences etc
1
152,660
12,124,010,900
IssuesEvent
2020-04-22 13:36:01
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
storage: TestNodeLivenessStatusMap failed under stress
C-test-failure O-robot
SHA: https://github.com/cockroachdb/cockroach/commits/4d8126d6a72d4092c334d9d483e9bc22e3fb98f7 Parameters: ``` TAGS= GOFLAGS= ``` To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stressrace instead of stress and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stress TESTS=TestNodeLivenessStatusMap PKG=github.com/cockroachdb/cockroach/pkg/storage TESTTIMEOUT=5m STRESSFLAGS='-stderr=false -maxtime 20m -timeout 10m' ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1863704&tab=buildLog ``` I200409 06:27:00.796062 108108 storage/node_liveness_test.go:825 [in test] done shutting down node 2 I200409 06:27:00.796096 108108 storage/node_liveness_test.go:828 [in test] decommissioning node 3 I200409 06:27:00.810725 109288 rpc/nodedialer/nodedialer.go:95 [ct-client] unable to connect to n2: failed to grpc dial n2 at 127.0.0.1:33027: initial connection heartbeat failed: rpc error: code = Unavailable desc = all SubConns are in TransientFailure, latest connection error: connection error: desc = "transport: Error while dialing dial tcp 127.0.0.1:33027: connect: connection refused" I200409 06:27:00.813443 108108 sql/event_log.go:126 [in test] Event: "node_decommissioned", target: 3, info: {} I200409 06:27:00.813536 108108 storage/node_liveness_test.go:832 [in test] done decommissioning node 3 I200409 06:27:00.813594 108108 storage/node_liveness_test.go:835 [in test] decommissioning and shutting down node 4 I200409 06:27:00.831184 109024 rpc/nodedialer/nodedialer.go:95 [ct-client] unable to connect to n2: failed to grpc dial n2 at 127.0.0.1:33027: initial connection heartbeat failed: rpc error: code = Unavailable desc = all SubConns are in TransientFailure, latest connection error: connection error: desc = "transport: Error while dialing dial tcp 127.0.0.1:33027: connect: connection refused" I200409 06:27:00.834830 108108 sql/event_log.go:126 [in test] Event: "node_decommissioned", target: 4, info: {} I200409 06:27:00.834885 108108 util/stop/stopper.go:548 quiescing; tasks left: 2 [async] closedts-subscription 1 ts.poller: poll 1 [async] closedts-rangefeed-subscriber W200409 06:27:00.835112 109262 ts/db.go:195 [n4,ts-poll] error writing time series data: failed to send RPC: sending to all 1 replicas failed; last error: <nil> rpc error: code = Unavailable desc = transport is closing I200409 06:27:00.835156 108108 util/stop/stopper.go:548 quiescing; tasks left: 2 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I200409 06:27:00.841346 109165 rpc/nodedialer/nodedialer.go:95 [ct-client] unable to connect to n2: failed to grpc dial n2 at 127.0.0.1:33027: context canceled I200409 06:27:00.841430 109163 rpc/nodedialer/nodedialer.go:95 [ct-client] unable to connect to n3: failed to grpc dial n3 at 127.0.0.1:35013: context canceled I200409 06:27:00.841481 109164 rpc/nodedialer/nodedialer.go:95 [ct-client] unable to connect to n1: failed to grpc dial n1 at 127.0.0.1:41201: context canceled I200409 06:27:00.841983 108108 util/stop/stopper.go:548 quiescing; tasks left: 2 [async] closedts-subscription I200409 06:27:00.842019 108108 util/stop/stopper.go:548 quiescing; tasks left: 1 [async] closedts-subscription I200409 06:27:00.867781 108108 storage/node_liveness_test.go:840 [in test] done removing node 4 I200409 06:27:00.867845 108108 storage/node_liveness_test.go:842 [in test] checking status map --- FAIL: TestNodeLivenessStatusMap (10.50s) === RUN TestNodeLivenessStatusMap/LIVE === PAUSE TestNodeLivenessStatusMap/LIVE === CONT TestNodeLivenessStatusMap/LIVE I200409 06:27:10.676813 109356 storage/node_liveness_test.go:874 [in test] checking expected status for node 1 --- FAIL: TestNodeLivenessStatusMap/LIVE (45.09s) node_liveness_test.go:863: condition failed to evaluate within 45s: unexpected status: got DEAD, expected LIVE goroutine 109356 [running]: runtime/debug.Stack(0xa7a358200, 0xc4219f5260, 0x2c58b40) /usr/local/go/src/runtime/debug/stack.go:24 +0xa7 github.com/cockroachdb/cockroach/pkg/testutils.SucceedsSoon(0x2cad5e0, 0xc420ffa0f0, 0xc4219f5170) /go/src/github.com/cockroachdb/cockroach/pkg/testutils/soon.go:50 +0x10d github.com/cockroachdb/cockroach/pkg/storage_test.TestNodeLivenessStatusMap.func1(0xc420ffa0f0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/node_liveness_test.go:863 +0xe2 testing.tRunner(0xc420ffa0f0, 0xc421350db0) /usr/local/go/src/testing/testing.go:777 +0xd0 created by testing.(*T).Run /usr/local/go/src/testing/testing.go:824 +0x2e0 ```
1.0
storage: TestNodeLivenessStatusMap failed under stress - SHA: https://github.com/cockroachdb/cockroach/commits/4d8126d6a72d4092c334d9d483e9bc22e3fb98f7 Parameters: ``` TAGS= GOFLAGS= ``` To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stressrace instead of stress and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stress TESTS=TestNodeLivenessStatusMap PKG=github.com/cockroachdb/cockroach/pkg/storage TESTTIMEOUT=5m STRESSFLAGS='-stderr=false -maxtime 20m -timeout 10m' ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1863704&tab=buildLog ``` I200409 06:27:00.796062 108108 storage/node_liveness_test.go:825 [in test] done shutting down node 2 I200409 06:27:00.796096 108108 storage/node_liveness_test.go:828 [in test] decommissioning node 3 I200409 06:27:00.810725 109288 rpc/nodedialer/nodedialer.go:95 [ct-client] unable to connect to n2: failed to grpc dial n2 at 127.0.0.1:33027: initial connection heartbeat failed: rpc error: code = Unavailable desc = all SubConns are in TransientFailure, latest connection error: connection error: desc = "transport: Error while dialing dial tcp 127.0.0.1:33027: connect: connection refused" I200409 06:27:00.813443 108108 sql/event_log.go:126 [in test] Event: "node_decommissioned", target: 3, info: {} I200409 06:27:00.813536 108108 storage/node_liveness_test.go:832 [in test] done decommissioning node 3 I200409 06:27:00.813594 108108 storage/node_liveness_test.go:835 [in test] decommissioning and shutting down node 4 I200409 06:27:00.831184 109024 rpc/nodedialer/nodedialer.go:95 [ct-client] unable to connect to n2: failed to grpc dial n2 at 127.0.0.1:33027: initial connection heartbeat failed: rpc error: code = Unavailable desc = all SubConns are in TransientFailure, latest connection error: connection error: desc = "transport: Error while dialing dial tcp 127.0.0.1:33027: connect: connection refused" I200409 06:27:00.834830 108108 sql/event_log.go:126 [in test] Event: "node_decommissioned", target: 4, info: {} I200409 06:27:00.834885 108108 util/stop/stopper.go:548 quiescing; tasks left: 2 [async] closedts-subscription 1 ts.poller: poll 1 [async] closedts-rangefeed-subscriber W200409 06:27:00.835112 109262 ts/db.go:195 [n4,ts-poll] error writing time series data: failed to send RPC: sending to all 1 replicas failed; last error: <nil> rpc error: code = Unavailable desc = transport is closing I200409 06:27:00.835156 108108 util/stop/stopper.go:548 quiescing; tasks left: 2 [async] closedts-subscription 1 [async] closedts-rangefeed-subscriber I200409 06:27:00.841346 109165 rpc/nodedialer/nodedialer.go:95 [ct-client] unable to connect to n2: failed to grpc dial n2 at 127.0.0.1:33027: context canceled I200409 06:27:00.841430 109163 rpc/nodedialer/nodedialer.go:95 [ct-client] unable to connect to n3: failed to grpc dial n3 at 127.0.0.1:35013: context canceled I200409 06:27:00.841481 109164 rpc/nodedialer/nodedialer.go:95 [ct-client] unable to connect to n1: failed to grpc dial n1 at 127.0.0.1:41201: context canceled I200409 06:27:00.841983 108108 util/stop/stopper.go:548 quiescing; tasks left: 2 [async] closedts-subscription I200409 06:27:00.842019 108108 util/stop/stopper.go:548 quiescing; tasks left: 1 [async] closedts-subscription I200409 06:27:00.867781 108108 storage/node_liveness_test.go:840 [in test] done removing node 4 I200409 06:27:00.867845 108108 storage/node_liveness_test.go:842 [in test] checking status map --- FAIL: TestNodeLivenessStatusMap (10.50s) === RUN TestNodeLivenessStatusMap/LIVE === PAUSE TestNodeLivenessStatusMap/LIVE === CONT TestNodeLivenessStatusMap/LIVE I200409 06:27:10.676813 109356 storage/node_liveness_test.go:874 [in test] checking expected status for node 1 --- FAIL: TestNodeLivenessStatusMap/LIVE (45.09s) node_liveness_test.go:863: condition failed to evaluate within 45s: unexpected status: got DEAD, expected LIVE goroutine 109356 [running]: runtime/debug.Stack(0xa7a358200, 0xc4219f5260, 0x2c58b40) /usr/local/go/src/runtime/debug/stack.go:24 +0xa7 github.com/cockroachdb/cockroach/pkg/testutils.SucceedsSoon(0x2cad5e0, 0xc420ffa0f0, 0xc4219f5170) /go/src/github.com/cockroachdb/cockroach/pkg/testutils/soon.go:50 +0x10d github.com/cockroachdb/cockroach/pkg/storage_test.TestNodeLivenessStatusMap.func1(0xc420ffa0f0) /go/src/github.com/cockroachdb/cockroach/pkg/storage/node_liveness_test.go:863 +0xe2 testing.tRunner(0xc420ffa0f0, 0xc421350db0) /usr/local/go/src/testing/testing.go:777 +0xd0 created by testing.(*T).Run /usr/local/go/src/testing/testing.go:824 +0x2e0 ```
test
storage testnodelivenessstatusmap failed under stress sha parameters tags goflags to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stressrace instead of stress and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach stdbuf ol el make stress tests testnodelivenessstatusmap pkg github com cockroachdb cockroach pkg storage testtimeout stressflags stderr false maxtime timeout failed test storage node liveness test go done shutting down node storage node liveness test go decommissioning node rpc nodedialer nodedialer go unable to connect to failed to grpc dial at initial connection heartbeat failed rpc error code unavailable desc all subconns are in transientfailure latest connection error connection error desc transport error while dialing dial tcp connect connection refused sql event log go event node decommissioned target info storage node liveness test go done decommissioning node storage node liveness test go decommissioning and shutting down node rpc nodedialer nodedialer go unable to connect to failed to grpc dial at initial connection heartbeat failed rpc error code unavailable desc all subconns are in transientfailure latest connection error connection error desc transport error while dialing dial tcp connect connection refused sql event log go event node decommissioned target info util stop stopper go quiescing tasks left closedts subscription ts poller poll closedts rangefeed subscriber ts db go error writing time series data failed to send rpc sending to all replicas failed last error rpc error code unavailable desc transport is closing util stop stopper go quiescing tasks left closedts subscription closedts rangefeed subscriber rpc nodedialer nodedialer go unable to connect to failed to grpc dial at context canceled rpc nodedialer nodedialer go unable to connect to failed to grpc dial at context canceled rpc nodedialer nodedialer go unable to connect to failed to grpc dial at context canceled util stop stopper go quiescing tasks left closedts subscription util stop stopper go quiescing tasks left closedts subscription storage node liveness test go done removing node storage node liveness test go checking status map fail testnodelivenessstatusmap run testnodelivenessstatusmap live pause testnodelivenessstatusmap live cont testnodelivenessstatusmap live storage node liveness test go checking expected status for node fail testnodelivenessstatusmap live node liveness test go condition failed to evaluate within unexpected status got dead expected live goroutine runtime debug stack usr local go src runtime debug stack go github com cockroachdb cockroach pkg testutils succeedssoon go src github com cockroachdb cockroach pkg testutils soon go github com cockroachdb cockroach pkg storage test testnodelivenessstatusmap go src github com cockroachdb cockroach pkg storage node liveness test go testing trunner usr local go src testing testing go created by testing t run usr local go src testing testing go
1
220,397
17,192,691,774
IssuesEvent
2021-07-16 13:16:59
kkomelin/laravel-translatable-string-exporter
https://api.github.com/repos/kkomelin/laravel-translatable-string-exporter
closed
Exclude short translation keys from export
enhancement ready for testing
Thank you for the excellent tool. Would it be possible to exclude patterns that use short keys? For example, I use `__('menus.tickets')` for my menus and there is a [corresponding short key file](https://laravel.com/docs/8.x/localization#using-short-keys) like this: ``` <?php // resources/lang/ja/menus.php return [ 'tickets' => 'チケット', ]; ``` However, when I run `php artisan translatable:export ja` it creates this line: `"menus.tickets": "menus.tickets",` which I then have to remove manually. I'm aware it's tricky to detect whether a short key exists but maybe an option to ignore `[a-z]+.[a-z]+` patterns would catch most cases. Or ignore the patterns if the file exists 1. `php artisan translatable:export ja` 2. Pattern is `([a-z]+).[a-z]+` such as `menu.tickets` 3. If `\1` exists such as `lang/ja/menus` then ignore. This could be set as an option.
1.0
Exclude short translation keys from export - Thank you for the excellent tool. Would it be possible to exclude patterns that use short keys? For example, I use `__('menus.tickets')` for my menus and there is a [corresponding short key file](https://laravel.com/docs/8.x/localization#using-short-keys) like this: ``` <?php // resources/lang/ja/menus.php return [ 'tickets' => 'チケット', ]; ``` However, when I run `php artisan translatable:export ja` it creates this line: `"menus.tickets": "menus.tickets",` which I then have to remove manually. I'm aware it's tricky to detect whether a short key exists but maybe an option to ignore `[a-z]+.[a-z]+` patterns would catch most cases. Or ignore the patterns if the file exists 1. `php artisan translatable:export ja` 2. Pattern is `([a-z]+).[a-z]+` such as `menu.tickets` 3. If `\1` exists such as `lang/ja/menus` then ignore. This could be set as an option.
test
exclude short translation keys from export thank you for the excellent tool would it be possible to exclude patterns that use short keys for example i use menus tickets for my menus and there is a like this php resources lang ja menus php return tickets チケット however when i run php artisan translatable export ja it creates this line menus tickets menus tickets which i then have to remove manually i m aware it s tricky to detect whether a short key exists but maybe an option to ignore patterns would catch most cases or ignore the patterns if the file exists php artisan translatable export ja pattern is such as menu tickets if exists such as lang ja menus then ignore this could be set as an option
1
232,168
18,848,356,024
IssuesEvent
2021-11-11 17:25:54
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Manual test run on Windows x64 & x86 for Tor 1.0.21
tests feature/tor OS/Windows QA/Yes release-notes/exclude OS/Desktop
### **`Windows 11 x64`** ### Tor Client Updater - [x] For development go-update-server. Run brave-browser with `--user-data-dir=component-dev --use-dev-goupdater-url` (These flags are only available in v1.7.x). Once the crx is pushed to production run without these flags. - [x] Navigate to `brave://components` and verify `Tor Client Updater (OS)` is updated successfully. - [x] Open `New Private Window with Tor` and confirm that it starts without any errors. - [x] Navigate to `check.torproject.org` and verify that tor is working successfully. - [x] Navigate to `brave.com` and `http://3g2upl4pq6kufc4m.onion/` to check if the sites work correctly. - [x] Load `brave.com` and `mail.protonmail.com` in a regular Window/Tab and ensure that clicking on `Open in Tor` correctly launches a Tor window and opens the appropriate `.onion` website. ### Windows - [x] Setup Windows SDK by downloading - https://developer.microsoft.com/en-us/windows/downloads/windows-10-sdk/ - [x] Run `"C:\Program Files (x86)\Windows Kits\10\bin\x86\signtool.exe" verify /pa "C:\Users\<user>\AppData\Local\BraveSoftware\Brave-Browser-<channel>\User Data\cpoalefficncklhjfpglfiplenlpccdb\<version>\tor-<version-tor>-win-brave-<version-brave>.exe"` to verify if the signature of the binary is correct. ### **`Windows 10 x64`** ### Tor Client Updater - [x] For development go-update-server. Run brave-browser with `--user-data-dir=component-dev --use-dev-goupdater-url` (These flags are only available in v1.7.x). Once the crx is pushed to production run without these flags. - [x] Navigate to `brave://components` and verify `Tor Client Updater (OS)` is updated successfully. - [x] Open `New Private Window with Tor` and confirm that it starts without any errors. - [x] Navigate to `check.torproject.org` and verify that tor is working successfully. - [x] Navigate to `brave.com` and `http://3g2upl4pq6kufc4m.onion/` to check if the sites work correctly. - [x] Load `brave.com` and `mail.protonmail.com` in a regular Window/Tab and ensure that clicking on `Open in Tor` correctly launches a Tor window and opens the appropriate `.onion` website. ### Windows - [x] Setup Windows SDK by downloading - https://developer.microsoft.com/en-us/windows/downloads/windows-10-sdk/ - [x] Run `"C:\Program Files (x86)\Windows Kits\10\bin\x86\signtool.exe" verify /pa "C:\Users\<user>\AppData\Local\BraveSoftware\Brave-Browser-<channel>\User Data\cpoalefficncklhjfpglfiplenlpccdb\<version>\tor-<version-tor>-win-brave-<version-brave>.exe"` to verify if the signature of the binary is correct. ### **`Windows 10 x86`** ### Tor Client Updater - [x] For development go-update-server. Run brave-browser with `--user-data-dir=component-dev --use-dev-goupdater-url` (These flags are only available in v1.7.x). Once the crx is pushed to production run without these flags. - [x] Navigate to `brave://components` and verify `Tor Client Updater (OS)` is updated successfully. - [x] Open `New Private Window with Tor` and confirm that it starts without any errors. - [x] Navigate to `check.torproject.org` and verify that tor is working successfully. - [x] Navigate to `brave.com` and `http://3g2upl4pq6kufc4m.onion/` to check if the sites work correctly. - [x] Load `brave.com` and `mail.protonmail.com` in a regular Window/Tab and ensure that clicking on `Open in Tor` correctly launches a Tor window and opens the appropriate `.onion` website. ### Windows - [x] Setup Windows SDK by downloading - https://developer.microsoft.com/en-us/windows/downloads/windows-10-sdk/ - [x] Run `"C:\Program Files (x86)\Windows Kits\10\bin\x86\signtool.exe" verify /pa "C:\Users\<user>\AppData\Local\BraveSoftware\Brave-Browser-<channel>\User Data\cpoalefficncklhjfpglfiplenlpccdb\<version>\tor-<version-tor>-win-brave-<version-brave>.exe"` to verify if the signature of the binary is correct.
1.0
Manual test run on Windows x64 & x86 for Tor 1.0.21 - ### **`Windows 11 x64`** ### Tor Client Updater - [x] For development go-update-server. Run brave-browser with `--user-data-dir=component-dev --use-dev-goupdater-url` (These flags are only available in v1.7.x). Once the crx is pushed to production run without these flags. - [x] Navigate to `brave://components` and verify `Tor Client Updater (OS)` is updated successfully. - [x] Open `New Private Window with Tor` and confirm that it starts without any errors. - [x] Navigate to `check.torproject.org` and verify that tor is working successfully. - [x] Navigate to `brave.com` and `http://3g2upl4pq6kufc4m.onion/` to check if the sites work correctly. - [x] Load `brave.com` and `mail.protonmail.com` in a regular Window/Tab and ensure that clicking on `Open in Tor` correctly launches a Tor window and opens the appropriate `.onion` website. ### Windows - [x] Setup Windows SDK by downloading - https://developer.microsoft.com/en-us/windows/downloads/windows-10-sdk/ - [x] Run `"C:\Program Files (x86)\Windows Kits\10\bin\x86\signtool.exe" verify /pa "C:\Users\<user>\AppData\Local\BraveSoftware\Brave-Browser-<channel>\User Data\cpoalefficncklhjfpglfiplenlpccdb\<version>\tor-<version-tor>-win-brave-<version-brave>.exe"` to verify if the signature of the binary is correct. ### **`Windows 10 x64`** ### Tor Client Updater - [x] For development go-update-server. Run brave-browser with `--user-data-dir=component-dev --use-dev-goupdater-url` (These flags are only available in v1.7.x). Once the crx is pushed to production run without these flags. - [x] Navigate to `brave://components` and verify `Tor Client Updater (OS)` is updated successfully. - [x] Open `New Private Window with Tor` and confirm that it starts without any errors. - [x] Navigate to `check.torproject.org` and verify that tor is working successfully. - [x] Navigate to `brave.com` and `http://3g2upl4pq6kufc4m.onion/` to check if the sites work correctly. - [x] Load `brave.com` and `mail.protonmail.com` in a regular Window/Tab and ensure that clicking on `Open in Tor` correctly launches a Tor window and opens the appropriate `.onion` website. ### Windows - [x] Setup Windows SDK by downloading - https://developer.microsoft.com/en-us/windows/downloads/windows-10-sdk/ - [x] Run `"C:\Program Files (x86)\Windows Kits\10\bin\x86\signtool.exe" verify /pa "C:\Users\<user>\AppData\Local\BraveSoftware\Brave-Browser-<channel>\User Data\cpoalefficncklhjfpglfiplenlpccdb\<version>\tor-<version-tor>-win-brave-<version-brave>.exe"` to verify if the signature of the binary is correct. ### **`Windows 10 x86`** ### Tor Client Updater - [x] For development go-update-server. Run brave-browser with `--user-data-dir=component-dev --use-dev-goupdater-url` (These flags are only available in v1.7.x). Once the crx is pushed to production run without these flags. - [x] Navigate to `brave://components` and verify `Tor Client Updater (OS)` is updated successfully. - [x] Open `New Private Window with Tor` and confirm that it starts without any errors. - [x] Navigate to `check.torproject.org` and verify that tor is working successfully. - [x] Navigate to `brave.com` and `http://3g2upl4pq6kufc4m.onion/` to check if the sites work correctly. - [x] Load `brave.com` and `mail.protonmail.com` in a regular Window/Tab and ensure that clicking on `Open in Tor` correctly launches a Tor window and opens the appropriate `.onion` website. ### Windows - [x] Setup Windows SDK by downloading - https://developer.microsoft.com/en-us/windows/downloads/windows-10-sdk/ - [x] Run `"C:\Program Files (x86)\Windows Kits\10\bin\x86\signtool.exe" verify /pa "C:\Users\<user>\AppData\Local\BraveSoftware\Brave-Browser-<channel>\User Data\cpoalefficncklhjfpglfiplenlpccdb\<version>\tor-<version-tor>-win-brave-<version-brave>.exe"` to verify if the signature of the binary is correct.
test
manual test run on windows for tor windows tor client updater for development go update server run brave browser with user data dir component dev use dev goupdater url these flags are only available in x once the crx is pushed to production run without these flags navigate to brave components and verify tor client updater os is updated successfully open new private window with tor and confirm that it starts without any errors navigate to check torproject org and verify that tor is working successfully navigate to brave com and to check if the sites work correctly load brave com and mail protonmail com in a regular window tab and ensure that clicking on open in tor correctly launches a tor window and opens the appropriate onion website windows setup windows sdk by downloading run c program files windows kits bin signtool exe verify pa c users appdata local bravesoftware brave browser user data cpoalefficncklhjfpglfiplenlpccdb tor win brave exe to verify if the signature of the binary is correct windows tor client updater for development go update server run brave browser with user data dir component dev use dev goupdater url these flags are only available in x once the crx is pushed to production run without these flags navigate to brave components and verify tor client updater os is updated successfully open new private window with tor and confirm that it starts without any errors navigate to check torproject org and verify that tor is working successfully navigate to brave com and to check if the sites work correctly load brave com and mail protonmail com in a regular window tab and ensure that clicking on open in tor correctly launches a tor window and opens the appropriate onion website windows setup windows sdk by downloading run c program files windows kits bin signtool exe verify pa c users appdata local bravesoftware brave browser user data cpoalefficncklhjfpglfiplenlpccdb tor win brave exe to verify if the signature of the binary is correct windows tor client updater for development go update server run brave browser with user data dir component dev use dev goupdater url these flags are only available in x once the crx is pushed to production run without these flags navigate to brave components and verify tor client updater os is updated successfully open new private window with tor and confirm that it starts without any errors navigate to check torproject org and verify that tor is working successfully navigate to brave com and to check if the sites work correctly load brave com and mail protonmail com in a regular window tab and ensure that clicking on open in tor correctly launches a tor window and opens the appropriate onion website windows setup windows sdk by downloading run c program files windows kits bin signtool exe verify pa c users appdata local bravesoftware brave browser user data cpoalefficncklhjfpglfiplenlpccdb tor win brave exe to verify if the signature of the binary is correct
1
208,692
15,914,324,823
IssuesEvent
2021-04-13 00:21:52
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
closed
There is an extra option in the 'Storage Access Policy Detected' dialog against one SAS attached file share
:gear: files :heavy_check_mark: merged 🧪 testing
**Storage Explorer Version**: 1.19.0 **Build Number**: 20210409.17 **Branch**: main **Platform/OS**: Windows 10/ Linux Ubuntu 18.04/ MacOS Catalina **Architecture**: ia32/x64 **Regression From**: Not a regression ## Steps to Reproduce ## 1. Expand one storage account -> File Shares. 2. Select one file share -> Right click it then select 'Manage Access Policies...' -> Add one access policy with 'List, Read' permissions. 3. Attach the file share via SAS which is created by the access policy. 4. Open the attached file share -> Try to upload one file. 5. Observe the prompt dialog -> Check there is no extra option. ## Expected Experience ## There is no extra option 'Optimize block size (requires write permission)'. ## Actual Experience ## There is an extra option 'Optimize block size (requires write permission)'. ![Screen Shot 2021-04-12 at 3 27 20 PM](https://user-images.githubusercontent.com/41351993/114359628-c102e180-9ba6-11eb-8103-2bef4abaa88a.png)
1.0
There is an extra option in the 'Storage Access Policy Detected' dialog against one SAS attached file share - **Storage Explorer Version**: 1.19.0 **Build Number**: 20210409.17 **Branch**: main **Platform/OS**: Windows 10/ Linux Ubuntu 18.04/ MacOS Catalina **Architecture**: ia32/x64 **Regression From**: Not a regression ## Steps to Reproduce ## 1. Expand one storage account -> File Shares. 2. Select one file share -> Right click it then select 'Manage Access Policies...' -> Add one access policy with 'List, Read' permissions. 3. Attach the file share via SAS which is created by the access policy. 4. Open the attached file share -> Try to upload one file. 5. Observe the prompt dialog -> Check there is no extra option. ## Expected Experience ## There is no extra option 'Optimize block size (requires write permission)'. ## Actual Experience ## There is an extra option 'Optimize block size (requires write permission)'. ![Screen Shot 2021-04-12 at 3 27 20 PM](https://user-images.githubusercontent.com/41351993/114359628-c102e180-9ba6-11eb-8103-2bef4abaa88a.png)
test
there is an extra option in the storage access policy detected dialog against one sas attached file share storage explorer version build number branch main platform os windows linux ubuntu macos catalina architecture regression from not a regression steps to reproduce expand one storage account file shares select one file share right click it then select manage access policies add one access policy with list read permissions attach the file share via sas which is created by the access policy open the attached file share try to upload one file observe the prompt dialog check there is no extra option expected experience there is no extra option optimize block size requires write permission actual experience there is an extra option optimize block size requires write permission
1
580,322
17,226,643,864
IssuesEvent
2021-07-20 03:20:31
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
Completions are failing with No Such Element Exception
Area/Completion Priority/Blocker SwanLakeDump Team/LanguageServer Type/Bug
**Description:** Completions fail with No Such Element Exception for the following source snippet. ``` function testFunction() { function hello; <cursor> } ``` **Affected Versions:** Swanlake beta2 at least
1.0
Completions are failing with No Such Element Exception - **Description:** Completions fail with No Such Element Exception for the following source snippet. ``` function testFunction() { function hello; <cursor> } ``` **Affected Versions:** Swanlake beta2 at least
non_test
completions are failing with no such element exception description completions fail with no such element exception for the following source snippet function testfunction function hello affected versions swanlake at least
0
28,656
4,425,517,179
IssuesEvent
2016-08-16 15:39:26
TransNexus/NexOSS
https://api.github.com/repos/TransNexus/NexOSS
closed
Improvements that need testing
needs testing
@JimDalton please test these new features to ensure correct functionality (check to mark complete): - [x] New login interface - [x] Redirect to the previous page on login - [x] Filtering of table is no longer silent Please add small improvements not related to specific pages to this issue.
1.0
Improvements that need testing - @JimDalton please test these new features to ensure correct functionality (check to mark complete): - [x] New login interface - [x] Redirect to the previous page on login - [x] Filtering of table is no longer silent Please add small improvements not related to specific pages to this issue.
test
improvements that need testing jimdalton please test these new features to ensure correct functionality check to mark complete new login interface redirect to the previous page on login filtering of table is no longer silent please add small improvements not related to specific pages to this issue
1
9,921
7,034,783,490
IssuesEvent
2017-12-27 18:58:14
JabRef/jabref
https://api.github.com/repos/JabRef/jabref
reopened
Performance problems with new groups search/filter and large databases
bug groups performance
JabRef 4.0.0-dev--snapshot--2017-05-18--master--018173ebd Windows 10 10.0 amd64 Java 1.8.0_131 The newly implemented groups filter/search (https://github.com/JabRef/jabref/pull/2588) exhibits massive performance problems when used on a large database (>10,000 entries, ~1,000 static groups). 1. Open large database 2. Start to enter any term you want to search for in the "Filter Groups" part of the groups panel. 3. You will immediately see a massive performance decrease. It takes minutes to just complete entering the word you want to search for (because the filtering starts immediately). The groups filtering itself also takes minutes to complete. During that time JabRef behaves as if it had crashed. This issue has been reported before (https://github.com/JabRef/jabref/pull/2588#issuecomment-283042778, https://github.com/JabRef/jabref/pull/2588#issuecomment-302629336) but I have been advised to open a new ticket so that the issue is not forgotten.
True
Performance problems with new groups search/filter and large databases - JabRef 4.0.0-dev--snapshot--2017-05-18--master--018173ebd Windows 10 10.0 amd64 Java 1.8.0_131 The newly implemented groups filter/search (https://github.com/JabRef/jabref/pull/2588) exhibits massive performance problems when used on a large database (>10,000 entries, ~1,000 static groups). 1. Open large database 2. Start to enter any term you want to search for in the "Filter Groups" part of the groups panel. 3. You will immediately see a massive performance decrease. It takes minutes to just complete entering the word you want to search for (because the filtering starts immediately). The groups filtering itself also takes minutes to complete. During that time JabRef behaves as if it had crashed. This issue has been reported before (https://github.com/JabRef/jabref/pull/2588#issuecomment-283042778, https://github.com/JabRef/jabref/pull/2588#issuecomment-302629336) but I have been advised to open a new ticket so that the issue is not forgotten.
non_test
performance problems with new groups search filter and large databases jabref dev snapshot master windows java the newly implemented groups filter search exhibits massive performance problems when used on a large database entries static groups open large database start to enter any term you want to search for in the filter groups part of the groups panel you will immediately see a massive performance decrease it takes minutes to just complete entering the word you want to search for because the filtering starts immediately the groups filtering itself also takes minutes to complete during that time jabref behaves as if it had crashed this issue has been reported before but i have been advised to open a new ticket so that the issue is not forgotten
0
270,695
20,605,697,017
IssuesEvent
2022-03-06 23:14:36
bounswe/bounswe2022group4
https://api.github.com/repos/bounswe/bounswe2022group4
closed
Create a personal page for yourself - Yusuf Bayındır
Category - To Do Category - Documentation Priority - Low whom: individual
Create a personal page for yourself and link it to the homepage.
1.0
Create a personal page for yourself - Yusuf Bayındır - Create a personal page for yourself and link it to the homepage.
non_test
create a personal page for yourself yusuf bayındır create a personal page for yourself and link it to the homepage
0
102,204
16,548,287,473
IssuesEvent
2021-05-28 04:36:17
samq-ghdemo/Java-Demo
https://api.github.com/repos/samq-ghdemo/Java-Demo
opened
CVE-2018-14042 (Medium) detected in bootstrap-3.3.7.min.js
security vulnerability
## CVE-2018-14042 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: Java-Demo/target/easybuggy-1-SNAPSHOT/dfi/style_bootstrap.html</p> <p>Path to vulnerable library: Java-Demo/target/easybuggy-1-SNAPSHOT/dfi/style_bootstrap.html,Java-Demo/.extract/webapps/ROOT/dfi/style_bootstrap.html,Java-Demo/src/main/webapp/dfi/style_bootstrap.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/Java-Demo/commit/ebe5ad9510b919e1dd226c70afc5e1115136f8cf">ebe5ad9510b919e1dd226c70afc5e1115136f8cf</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip. <p>Publish Date: 2018-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042>CVE-2018-14042</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/twbs/bootstrap/pull/26630">https://github.com/twbs/bootstrap/pull/26630</a></p> <p>Release Date: 2018-07-13</p> <p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"twitter-bootstrap","packageVersion":"3.3.7","packageFilePaths":["/target/easybuggy-1-SNAPSHOT/dfi/style_bootstrap.html","/.extract/webapps/ROOT/dfi/style_bootstrap.html","/src/main/webapp/dfi/style_bootstrap.html"],"isTransitiveDependency":false,"dependencyTree":"twitter-bootstrap:3.3.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-14042","vulnerabilityDetails":"In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2018-14042 (Medium) detected in bootstrap-3.3.7.min.js - ## CVE-2018-14042 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: Java-Demo/target/easybuggy-1-SNAPSHOT/dfi/style_bootstrap.html</p> <p>Path to vulnerable library: Java-Demo/target/easybuggy-1-SNAPSHOT/dfi/style_bootstrap.html,Java-Demo/.extract/webapps/ROOT/dfi/style_bootstrap.html,Java-Demo/src/main/webapp/dfi/style_bootstrap.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/Java-Demo/commit/ebe5ad9510b919e1dd226c70afc5e1115136f8cf">ebe5ad9510b919e1dd226c70afc5e1115136f8cf</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip. <p>Publish Date: 2018-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042>CVE-2018-14042</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/twbs/bootstrap/pull/26630">https://github.com/twbs/bootstrap/pull/26630</a></p> <p>Release Date: 2018-07-13</p> <p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"twitter-bootstrap","packageVersion":"3.3.7","packageFilePaths":["/target/easybuggy-1-SNAPSHOT/dfi/style_bootstrap.html","/.extract/webapps/ROOT/dfi/style_bootstrap.html","/src/main/webapp/dfi/style_bootstrap.html"],"isTransitiveDependency":false,"dependencyTree":"twitter-bootstrap:3.3.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-14042","vulnerabilityDetails":"In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_test
cve medium detected in bootstrap min js cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file java demo target easybuggy snapshot dfi style bootstrap html path to vulnerable library java demo target easybuggy snapshot dfi style bootstrap html java demo extract webapps root dfi style bootstrap html java demo src main webapp dfi style bootstrap html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch main vulnerability details in bootstrap before xss is possible in the data container property of tooltip publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org webjars npm bootstrap org webjars bootstrap isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree twitter bootstrap isminimumfixversionavailable true minimumfixversion org webjars npm bootstrap org webjars bootstrap basebranches vulnerabilityidentifier cve vulnerabilitydetails in bootstrap before xss is possible in the data container property of tooltip vulnerabilityurl
0
296,057
25,524,753,009
IssuesEvent
2022-11-29 00:37:24
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
DISABLED test_comprehensive_fft_ihfft2_cuda_float64 (__main__.TestDecompCUDA)
module: cuda triaged module: flaky-tests skipped module: primTorch module: decompositions
Platforms: rocm This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_comprehensive_fft_ihfft2_cuda_float64&suite=TestDecompCUDA) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/8860676341). Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 failures and 1 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT BE ALARMED IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_comprehensive_fft_ihfft2_cuda_float64` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. cc @ezyang @mruberry @ngimel @Lezcano @fdrocha @SherlockNoMad
1.0
DISABLED test_comprehensive_fft_ihfft2_cuda_float64 (__main__.TestDecompCUDA) - Platforms: rocm This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_comprehensive_fft_ihfft2_cuda_float64&suite=TestDecompCUDA) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/8860676341). Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 failures and 1 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT BE ALARMED IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_comprehensive_fft_ihfft2_cuda_float64` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. cc @ezyang @mruberry @ngimel @Lezcano @fdrocha @SherlockNoMad
test
disabled test comprehensive fft cuda main testdecompcuda platforms rocm this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not be alarmed if the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test comprehensive fft cuda there should be several instances run as flaky tests are rerun in ci from which you can study the logs cc ezyang mruberry ngimel lezcano fdrocha sherlocknomad
1
219,286
17,084,292,138
IssuesEvent
2021-07-08 09:46:10
Tencent/bk-ci
https://api.github.com/repos/Tencent/bk-ci
closed
feat: 流水线插件新增“失败时可跳过”的选项
area/ci/backend area/ci/frontend kind/feat/product priority/important-soon stage/test stage/uat test/passed uat/passed
<!-- Please only use this template for submitting enhancement requests --> **What would you like to be added**: 流水线插件新增“失败时可跳过”的选项 **Why is this needed**: 部分插件的失败结果已经通过第三方系统特殊处理过,希望能继续执行当前流水线流程,避免浪费
2.0
feat: 流水线插件新增“失败时可跳过”的选项 - <!-- Please only use this template for submitting enhancement requests --> **What would you like to be added**: 流水线插件新增“失败时可跳过”的选项 **Why is this needed**: 部分插件的失败结果已经通过第三方系统特殊处理过,希望能继续执行当前流水线流程,避免浪费
test
feat 流水线插件新增“失败时可跳过”的选项 what would you like to be added 流水线插件新增“失败时可跳过”的选项 why is this needed 部分插件的失败结果已经通过第三方系统特殊处理过,希望能继续执行当前流水线流程,避免浪费
1
64,236
15,829,377,628
IssuesEvent
2021-04-06 11:06:10
iterative/dvc
https://api.github.com/repos/iterative/dvc
closed
deb/rpm: build granular packages
build enhancement help wanted p3-nice-to-have
1. Big package size. 2. Librarys duplication on client system. 3. Need to upgrade package, when dependency is upgraded. 4. The patches of the distribution developers are not used. As example: #2768 5. Perhaps you build the package by hand? 6. You can find many other information about that issue. What is your current way to build packages? What about the system for automatically assembling the correct packages? I can help you with that. **Roadmap:** - [x] 1.1. ~~Determine the initial status.~~ - [x] 1.2. ~~Build all dependencies in auto mode.~~ - [x] 1.3. ~~Build the primary package: python3-dvc~~ - [x] 1.4. Build all extra dependencies in auto mode. - [x] 1.5. Issue #2815 is merged. - [x] 1.6. Now new package system is ready for use. - [ ] 2.1. Make `debian/` for all packages. - [ ] 2.2. Make pull requests to the upstreams. - [ ] 2.3. Build doc packages. - [ ] 2.4. Passing Debian lint checks. - [ ] 2.5. Now DVC is ready for Debian Experimental. **Current status:** Already in Debian repository: * python3-ply * python3-future * python3-colorama * python3-configobj * python3-setuptools * python3-schema * python3-requests * python3-distro * python3-appdirs * python3-inflect * python3-humanize * python3-pathspec * python3-shortuuid * python3-tqdm * python3-packaging * python3-yaml * python3-ruamel.yaml * python3-git * python3-networkx * python3-flufl.lock * python3-voluptuous Not in Debian: - [x] configparser: * https://github.com/jaraco/configparser/ * Do you use this package instead of the built-in? - [x] nanotime: * https://github.com/jbenet/nanotime/tree/master/python * Perhaps abandoned. - [x] jsonpath-ng * https://github.com/h2non/jsonpath-ng * Perhaps abandoned. * Debian dependencies: * python3-ply * python3-six * python3-decorator - [x] grandalf * https://github.com/bdcht/grandalf * Debian dependencies: * python3-pyparsing * python3-future - [x] treelib * https://github.com/caesar0301/treelib - [x] funcy * https://github.com/Suor/funcy - [x] ~asciimatics~ (Thrown out of code!) * https://github.com/peterbrittain/asciimatics * Debian dependencies: * python3-pyfiglet * python3-future * python3-wcwidth * python3-pillow
1.0
deb/rpm: build granular packages - 1. Big package size. 2. Librarys duplication on client system. 3. Need to upgrade package, when dependency is upgraded. 4. The patches of the distribution developers are not used. As example: #2768 5. Perhaps you build the package by hand? 6. You can find many other information about that issue. What is your current way to build packages? What about the system for automatically assembling the correct packages? I can help you with that. **Roadmap:** - [x] 1.1. ~~Determine the initial status.~~ - [x] 1.2. ~~Build all dependencies in auto mode.~~ - [x] 1.3. ~~Build the primary package: python3-dvc~~ - [x] 1.4. Build all extra dependencies in auto mode. - [x] 1.5. Issue #2815 is merged. - [x] 1.6. Now new package system is ready for use. - [ ] 2.1. Make `debian/` for all packages. - [ ] 2.2. Make pull requests to the upstreams. - [ ] 2.3. Build doc packages. - [ ] 2.4. Passing Debian lint checks. - [ ] 2.5. Now DVC is ready for Debian Experimental. **Current status:** Already in Debian repository: * python3-ply * python3-future * python3-colorama * python3-configobj * python3-setuptools * python3-schema * python3-requests * python3-distro * python3-appdirs * python3-inflect * python3-humanize * python3-pathspec * python3-shortuuid * python3-tqdm * python3-packaging * python3-yaml * python3-ruamel.yaml * python3-git * python3-networkx * python3-flufl.lock * python3-voluptuous Not in Debian: - [x] configparser: * https://github.com/jaraco/configparser/ * Do you use this package instead of the built-in? - [x] nanotime: * https://github.com/jbenet/nanotime/tree/master/python * Perhaps abandoned. - [x] jsonpath-ng * https://github.com/h2non/jsonpath-ng * Perhaps abandoned. * Debian dependencies: * python3-ply * python3-six * python3-decorator - [x] grandalf * https://github.com/bdcht/grandalf * Debian dependencies: * python3-pyparsing * python3-future - [x] treelib * https://github.com/caesar0301/treelib - [x] funcy * https://github.com/Suor/funcy - [x] ~asciimatics~ (Thrown out of code!) * https://github.com/peterbrittain/asciimatics * Debian dependencies: * python3-pyfiglet * python3-future * python3-wcwidth * python3-pillow
non_test
deb rpm build granular packages big package size librarys duplication on client system need to upgrade package when dependency is upgraded the patches of the distribution developers are not used as example perhaps you build the package by hand you can find many other information about that issue what is your current way to build packages what about the system for automatically assembling the correct packages i can help you with that roadmap determine the initial status build all dependencies in auto mode build the primary package dvc build all extra dependencies in auto mode issue is merged now new package system is ready for use make debian for all packages make pull requests to the upstreams build doc packages passing debian lint checks now dvc is ready for debian experimental current status already in debian repository ply future colorama configobj setuptools schema requests distro appdirs inflect humanize pathspec shortuuid tqdm packaging yaml ruamel yaml git networkx flufl lock voluptuous not in debian configparser do you use this package instead of the built in nanotime perhaps abandoned jsonpath ng perhaps abandoned debian dependencies ply six decorator grandalf debian dependencies pyparsing future treelib funcy asciimatics thrown out of code debian dependencies pyfiglet future wcwidth pillow
0
374,509
11,091,561,738
IssuesEvent
2019-12-15 13:16:21
servicemesher/istio-official-translation
https://api.github.com/repos/servicemesher/istio-official-translation
closed
/news/_index.md
finished lang/zh priority/P3 sync/update version/1.4
Source File: [/news/_index.md](https://github.com/istio/istio.io/tree/master/content/en/news/_index.md) Diff: ~~~diff diff --git a/content/en/news/_index.md b/content/en/news/_index.md index 586c99fe4..c38e915f2 100644 --- a/content/en/news/_index.md +++ b/content/en/news/_index.md @@ -5,7 +5,4 @@ linktitle: News sidebar_multicard: true icon: bullhorn decoration: pill -outputs: - - html - - rss --- ~~~
1.0
/news/_index.md - Source File: [/news/_index.md](https://github.com/istio/istio.io/tree/master/content/en/news/_index.md) Diff: ~~~diff diff --git a/content/en/news/_index.md b/content/en/news/_index.md index 586c99fe4..c38e915f2 100644 --- a/content/en/news/_index.md +++ b/content/en/news/_index.md @@ -5,7 +5,4 @@ linktitle: News sidebar_multicard: true icon: bullhorn decoration: pill -outputs: - - html - - rss --- ~~~
non_test
news index md source file diff diff diff git a content en news index md b content en news index md index a content en news index md b content en news index md linktitle news sidebar multicard true icon bullhorn decoration pill outputs html rss
0
40,451
5,292,359,402
IssuesEvent
2017-02-09 01:49:17
GoogleCloudPlatform/google-cloud-eclipse
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-eclipse
reopened
Builds failing with Oxygen
high priority testing
Oxygen builds are frequently failing on Travis: ``` No output has been received in the last 10m0s, this potentially indicates a stalled build or something wrong with the build itself. Check the details on how to adjust your build configuration on: https://docs.travis-ci.com/user/common-build-problems/#Build-times-out-because-no-output-was-received ``` From the [failed build logs](https://github.com/GoogleCloudPlatform/google-cloud-eclipse/files/753310/failedbuild.txt), it seems to be happening in the integration tests, and there are a lot of puzzling output. It may be related to #1314
1.0
Builds failing with Oxygen - Oxygen builds are frequently failing on Travis: ``` No output has been received in the last 10m0s, this potentially indicates a stalled build or something wrong with the build itself. Check the details on how to adjust your build configuration on: https://docs.travis-ci.com/user/common-build-problems/#Build-times-out-because-no-output-was-received ``` From the [failed build logs](https://github.com/GoogleCloudPlatform/google-cloud-eclipse/files/753310/failedbuild.txt), it seems to be happening in the integration tests, and there are a lot of puzzling output. It may be related to #1314
test
builds failing with oxygen oxygen builds are frequently failing on travis no output has been received in the last this potentially indicates a stalled build or something wrong with the build itself check the details on how to adjust your build configuration on from the it seems to be happening in the integration tests and there are a lot of puzzling output it may be related to
1
243,039
20,359,260,943
IssuesEvent
2022-02-20 12:46:46
mennaelkashef/eShop
https://api.github.com/repos/mennaelkashef/eShop
opened
Test bug
Hello! RULE-GOT-APPLIED DOES-NOT-CONTAIN-STRING Rule-works-on-convert-to-bug test instabug
# :clipboard: Bug Details >Test bug key | value --|-- Reported At | 2022-02-20 12:45:48 UTC Categories | Report a bug Tags | test, Hello!, RULE-GOT-APPLIED, DOES-NOT-CONTAIN-STRING, Rule-works-on-convert-to-bug, instabug App Version | 1.1 (1) Session Duration | 7 Device | Google sdk_gphone64_arm64, OS Level 32 Display | 1080x2340 (xhdpi) Location | Cairo, Egypt (en) ## :point_right: [View Full Bug Report on Instabug](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8491?utm_source=github&utm_medium=integrations) :point_left: ___ # :iphone: View Hierarchy This bug was reported from **com.example.app.bugReporting.BugReportingFragment** Find its interactive view hierarchy with all its subviews here: :point_right: **[Check View Hierarchy](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8491?show-hierarchy-view=true&utm_source=github&utm_medium=integrations)** :point_left: ___ # :chart_with_downwards_trend: Session Profiler Here is what the app was doing right before the bug was reported: Key | Value --|-- Used Memory | 56.2% - 1.08/1.92 GB Used Storage | 80.7% - 0.61/0.76 GB Connectivity | WiFi Battery | 100% - unplugged Orientation | portrait Find all the changes that happened in the parameters mentioned above during the last 60 seconds before the bug was reported here: :point_right: **[View Full Session Profiler](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8491?show-session-profiler=true&utm_source=github&utm_medium=integrations)** :point_left: ___ # :bust_in_silhouette: User Info ### User Attributes ``` key_name 1969707769: key value bla bla bla la key_name 432458846: key value bla bla bla la key_name 1912315229: key value bla bla bla la key_name 151057475: key value bla bla bla la key_name -1243046529: key value bla bla bla la key_name -1015826291: key value bla bla bla la ``` ___ # :mag_right: Logs ### User Steps Here are the last 10 steps done by the user right before the bug was reported: ``` 12:45:45 In activity com.example.app.main.MainActivity: fragment com.example.app.main.MainFragment was paused. 12:45:45 In activity com.example.app.main.MainActivity: fragment com.example.app.main.MainFragment was stopped. 12:45:45 In activity com.example.app.main.MainActivity: fragment com.example.app.bugReporting.BugReportingFragment was attached. 12:45:45 In activity com.example.app.main.MainActivity: fragment com.example.app.bugReporting.BugReportingFragment was created. 12:45:45 In activity com.example.app.main.MainActivity: fragment com.example.app.bugReporting.BugReportingFragment was started. 12:45:45 In activity com.example.app.main.MainActivity: fragment com.example.app.bugReporting.BugReportingFragment was resumed. 12:45:47 Tap in "All Options" of type "androidx.appcompat.widget.AppCompatTextView" in "com.example.app.main.MainActivity" 12:45:47 com.example.app.main.MainActivity was paused. 12:45:47 In activity com.example.app.main.MainActivity: fragment com.example.app.bugReporting.BugReportingFragment was paused. 12:45:48 Tap in "androidx.constraintlayout.widget.ConstraintLayout" in "com.example.app.main.MainActivity" ``` Find all the user steps done by the user throughout the session here: :point_right: **[View All User Steps](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8491?show-logs=user_steps&utm_source=github&utm_medium=integrations)** :point_left: ### Console Log Here are the last 10 console logs logged right before the bug was reported: ``` 12:45:55 D/EGL_emulation( 6089): app_time_stats: avg=63.65ms min=12.95ms max=501.82ms count=16 12:45:56 D/EGL_emulation( 6089): app_time_stats: avg=67.48ms min=15.03ms max=497.89ms count=15 12:45:57 D/EGL_emulation( 6089): app_time_stats: avg=346.84ms min=149.12ms max=505.78ms count=3 12:45:59 D/EGL_emulation( 6089): app_time_stats: avg=83.04ms min=14.84ms max=502.52ms count=15 12:46:00 D/EGL_emulation( 6089): app_time_stats: avg=500.35ms min=499.02ms max=501.68ms count=2 12:46:01 D/EGL_emulation( 6089): app_time_stats: avg=508.55ms min=497.90ms max=519.19ms count=2 12:46:02 D/EGL_emulation( 6089): app_time_stats: avg=432.75ms min=299.34ms max=500.72ms count=3 12:46:02 D/IB-BaseReportingPresenter( 6089): checkUserEmailValid :empty-email 12:46:02 D/IB-ActionsOrchestrator( 6089): runAction 12:46:02 D/IB-AttachmentsUtility( 6089): encryptAttachments ``` Find all the logged console logs throughout the session here: :point_right: **[View All Console Log](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8491?show-logs=console_log&utm_source=github&utm_medium=integrations)** :point_left: ___ # :camera: Images [![image attachment](https://d38gnqwzxziyyy.cloudfront.net/attachments/bugs/17891606/f25310b7347b4157b48918e811bf3138_one_bug_thumb/25099288/bug_1645361147267_.jpg?Expires=4801034804&Signature=VELvXqZtTCopQ1Q~-QyPRDIXgSCoAIU7mytEPXV2z404bI9UDwtd6aEM~tSRtIa3LnS84A-eVDJeuMmLo9NhbtV2W9VYNvUD~0ch3CMFJPvuir~0BLQYWxdbaFGg0v5TRQnkOlo9F3bSt2lBcNdwES2s1mJYKttfk2DqzpM1le7s-ms6dcnbKqmPNvUw~Xxy7LACztzcrXWuLPCBpwoB-eOlydy6-uKDHVWrvN9-lwvmVLS5tpKPWuitsnfygij0Upk26cVR4HFXjEziWZzYu5BH5zNVorkVp3VnoVLmkJlHZc9CeZpR5ev55sdLilJff5P1vcS69QM5tljjii~nKg__&Key-Pair-Id=APKAIXAG65U6UUX7JAQQ)](https://d38gnqwzxziyyy.cloudfront.net/attachments/bugs/17891606/f25310b7347b4157b48918e811bf3138_original/25099288/bug_1645361147267_.jpg?Expires=4801034804&Signature=auhcHLAj-uBMYGyqA4Tu0HkJlt-843XRPoEbecGQR9zXm4B-FQ2XvRLhJnDoSLIT6itOyMWpc0os-AbCUrRm1AASeYhD5OnQdnRxRV1zBfx0mSEnAX21XjiDrVU9-phU4MRVWMHazG6lak3xLpFZ0VkMxE7KebJdQAxqtAABN8XohawFVrDAo3oTtzDwBdKjq5LxwldlRmdn0Gh5KLx0rzbXQ8Wf5~LuBsoMzebPelj5TuhGpMNShMns9hoqkG0Ld6pyEMerkgCjQzCDB0wQOrfxlsHp4~-xjWvE9mjj2R34PHhTt49tlhAkY1eirBokTpG8Zx4VNygA5D8YvadD1A__&Key-Pair-Id=APKAIXAG65U6UUX7JAQQ) ___ # :warning: Looking for More Details? 1. **Network Log**: we are unable to capture your network requests automatically. If you are using HttpUrlConnection or Okhttp requests, [**check the details mentioned here**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations#section-network-logs). 2. **User Events**: start capturing custom User Events to send them along with each report. [**Find all the details in the docs**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations). 3. **Instabug Log**: start adding Instabug logs to see them right inside each report you receive. [**Find all the details in the docs**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations).
1.0
Test bug - # :clipboard: Bug Details >Test bug key | value --|-- Reported At | 2022-02-20 12:45:48 UTC Categories | Report a bug Tags | test, Hello!, RULE-GOT-APPLIED, DOES-NOT-CONTAIN-STRING, Rule-works-on-convert-to-bug, instabug App Version | 1.1 (1) Session Duration | 7 Device | Google sdk_gphone64_arm64, OS Level 32 Display | 1080x2340 (xhdpi) Location | Cairo, Egypt (en) ## :point_right: [View Full Bug Report on Instabug](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8491?utm_source=github&utm_medium=integrations) :point_left: ___ # :iphone: View Hierarchy This bug was reported from **com.example.app.bugReporting.BugReportingFragment** Find its interactive view hierarchy with all its subviews here: :point_right: **[Check View Hierarchy](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8491?show-hierarchy-view=true&utm_source=github&utm_medium=integrations)** :point_left: ___ # :chart_with_downwards_trend: Session Profiler Here is what the app was doing right before the bug was reported: Key | Value --|-- Used Memory | 56.2% - 1.08/1.92 GB Used Storage | 80.7% - 0.61/0.76 GB Connectivity | WiFi Battery | 100% - unplugged Orientation | portrait Find all the changes that happened in the parameters mentioned above during the last 60 seconds before the bug was reported here: :point_right: **[View Full Session Profiler](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8491?show-session-profiler=true&utm_source=github&utm_medium=integrations)** :point_left: ___ # :bust_in_silhouette: User Info ### User Attributes ``` key_name 1969707769: key value bla bla bla la key_name 432458846: key value bla bla bla la key_name 1912315229: key value bla bla bla la key_name 151057475: key value bla bla bla la key_name -1243046529: key value bla bla bla la key_name -1015826291: key value bla bla bla la ``` ___ # :mag_right: Logs ### User Steps Here are the last 10 steps done by the user right before the bug was reported: ``` 12:45:45 In activity com.example.app.main.MainActivity: fragment com.example.app.main.MainFragment was paused. 12:45:45 In activity com.example.app.main.MainActivity: fragment com.example.app.main.MainFragment was stopped. 12:45:45 In activity com.example.app.main.MainActivity: fragment com.example.app.bugReporting.BugReportingFragment was attached. 12:45:45 In activity com.example.app.main.MainActivity: fragment com.example.app.bugReporting.BugReportingFragment was created. 12:45:45 In activity com.example.app.main.MainActivity: fragment com.example.app.bugReporting.BugReportingFragment was started. 12:45:45 In activity com.example.app.main.MainActivity: fragment com.example.app.bugReporting.BugReportingFragment was resumed. 12:45:47 Tap in "All Options" of type "androidx.appcompat.widget.AppCompatTextView" in "com.example.app.main.MainActivity" 12:45:47 com.example.app.main.MainActivity was paused. 12:45:47 In activity com.example.app.main.MainActivity: fragment com.example.app.bugReporting.BugReportingFragment was paused. 12:45:48 Tap in "androidx.constraintlayout.widget.ConstraintLayout" in "com.example.app.main.MainActivity" ``` Find all the user steps done by the user throughout the session here: :point_right: **[View All User Steps](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8491?show-logs=user_steps&utm_source=github&utm_medium=integrations)** :point_left: ### Console Log Here are the last 10 console logs logged right before the bug was reported: ``` 12:45:55 D/EGL_emulation( 6089): app_time_stats: avg=63.65ms min=12.95ms max=501.82ms count=16 12:45:56 D/EGL_emulation( 6089): app_time_stats: avg=67.48ms min=15.03ms max=497.89ms count=15 12:45:57 D/EGL_emulation( 6089): app_time_stats: avg=346.84ms min=149.12ms max=505.78ms count=3 12:45:59 D/EGL_emulation( 6089): app_time_stats: avg=83.04ms min=14.84ms max=502.52ms count=15 12:46:00 D/EGL_emulation( 6089): app_time_stats: avg=500.35ms min=499.02ms max=501.68ms count=2 12:46:01 D/EGL_emulation( 6089): app_time_stats: avg=508.55ms min=497.90ms max=519.19ms count=2 12:46:02 D/EGL_emulation( 6089): app_time_stats: avg=432.75ms min=299.34ms max=500.72ms count=3 12:46:02 D/IB-BaseReportingPresenter( 6089): checkUserEmailValid :empty-email 12:46:02 D/IB-ActionsOrchestrator( 6089): runAction 12:46:02 D/IB-AttachmentsUtility( 6089): encryptAttachments ``` Find all the logged console logs throughout the session here: :point_right: **[View All Console Log](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8491?show-logs=console_log&utm_source=github&utm_medium=integrations)** :point_left: ___ # :camera: Images [![image attachment](https://d38gnqwzxziyyy.cloudfront.net/attachments/bugs/17891606/f25310b7347b4157b48918e811bf3138_one_bug_thumb/25099288/bug_1645361147267_.jpg?Expires=4801034804&Signature=VELvXqZtTCopQ1Q~-QyPRDIXgSCoAIU7mytEPXV2z404bI9UDwtd6aEM~tSRtIa3LnS84A-eVDJeuMmLo9NhbtV2W9VYNvUD~0ch3CMFJPvuir~0BLQYWxdbaFGg0v5TRQnkOlo9F3bSt2lBcNdwES2s1mJYKttfk2DqzpM1le7s-ms6dcnbKqmPNvUw~Xxy7LACztzcrXWuLPCBpwoB-eOlydy6-uKDHVWrvN9-lwvmVLS5tpKPWuitsnfygij0Upk26cVR4HFXjEziWZzYu5BH5zNVorkVp3VnoVLmkJlHZc9CeZpR5ev55sdLilJff5P1vcS69QM5tljjii~nKg__&Key-Pair-Id=APKAIXAG65U6UUX7JAQQ)](https://d38gnqwzxziyyy.cloudfront.net/attachments/bugs/17891606/f25310b7347b4157b48918e811bf3138_original/25099288/bug_1645361147267_.jpg?Expires=4801034804&Signature=auhcHLAj-uBMYGyqA4Tu0HkJlt-843XRPoEbecGQR9zXm4B-FQ2XvRLhJnDoSLIT6itOyMWpc0os-AbCUrRm1AASeYhD5OnQdnRxRV1zBfx0mSEnAX21XjiDrVU9-phU4MRVWMHazG6lak3xLpFZ0VkMxE7KebJdQAxqtAABN8XohawFVrDAo3oTtzDwBdKjq5LxwldlRmdn0Gh5KLx0rzbXQ8Wf5~LuBsoMzebPelj5TuhGpMNShMns9hoqkG0Ld6pyEMerkgCjQzCDB0wQOrfxlsHp4~-xjWvE9mjj2R34PHhTt49tlhAkY1eirBokTpG8Zx4VNygA5D8YvadD1A__&Key-Pair-Id=APKAIXAG65U6UUX7JAQQ) ___ # :warning: Looking for More Details? 1. **Network Log**: we are unable to capture your network requests automatically. If you are using HttpUrlConnection or Okhttp requests, [**check the details mentioned here**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations#section-network-logs). 2. **User Events**: start capturing custom User Events to send them along with each report. [**Find all the details in the docs**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations). 3. **Instabug Log**: start adding Instabug logs to see them right inside each report you receive. [**Find all the details in the docs**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations).
test
test bug clipboard bug details test bug key value reported at utc categories report a bug tags test hello rule got applied does not contain string rule works on convert to bug instabug app version session duration device google sdk os level display xhdpi location cairo egypt en point right point left iphone view hierarchy this bug was reported from com example app bugreporting bugreportingfragment find its interactive view hierarchy with all its subviews here point right point left chart with downwards trend session profiler here is what the app was doing right before the bug was reported key value used memory gb used storage gb connectivity wifi battery unplugged orientation portrait find all the changes that happened in the parameters mentioned above during the last seconds before the bug was reported here point right point left bust in silhouette user info user attributes key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la mag right logs user steps here are the last steps done by the user right before the bug was reported in activity com example app main mainactivity fragment com example app main mainfragment was paused in activity com example app main mainactivity fragment com example app main mainfragment was stopped in activity com example app main mainactivity fragment com example app bugreporting bugreportingfragment was attached in activity com example app main mainactivity fragment com example app bugreporting bugreportingfragment was created in activity com example app main mainactivity fragment com example app bugreporting bugreportingfragment was started in activity com example app main mainactivity fragment com example app bugreporting bugreportingfragment was resumed tap in all options of type androidx appcompat widget appcompattextview in com example app main mainactivity com example app main mainactivity was paused in activity com example app main mainactivity fragment com example app bugreporting bugreportingfragment was paused tap in androidx constraintlayout widget constraintlayout in com example app main mainactivity find all the user steps done by the user throughout the session here point right point left console log here are the last console logs logged right before the bug was reported d egl emulation app time stats avg min max count d egl emulation app time stats avg min max count d egl emulation app time stats avg min max count d egl emulation app time stats avg min max count d egl emulation app time stats avg min max count d egl emulation app time stats avg min max count d egl emulation app time stats avg min max count d ib basereportingpresenter checkuseremailvalid empty email d ib actionsorchestrator runaction d ib attachmentsutility encryptattachments find all the logged console logs throughout the session here point right point left camera images warning looking for more details network log we are unable to capture your network requests automatically if you are using httpurlconnection or okhttp requests user events start capturing custom user events to send them along with each report instabug log start adding instabug logs to see them right inside each report you receive
1