Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
19,230
3,155,728,719
IssuesEvent
2015-09-17 10:33:40
bigbluebutton/bigbluebutton
https://api.github.com/repos/bigbluebutton/bigbluebutton
closed
If try to change microphone when entering in BBB, it enters in a loop
Defect Normal Priority
Originally reported on Google Code with ID 1896 ``` When you enter in BBB, you select No when you receive the question: have you heard your voice? Then you select a different microphone and push Next. It never finishes. You have to reload the page. ``` Reported by `smoral@adhoclearning.com` on 2015-03-06 14:44:53
1.0
If try to change microphone when entering in BBB, it enters in a loop - Originally reported on Google Code with ID 1896 ``` When you enter in BBB, you select No when you receive the question: have you heard your voice? Then you select a different microphone and push Next. It never finishes. You have to reload the page. ``` Reported by `smoral@adhoclearning.com` on 2015-03-06 14:44:53
defect
if try to change microphone when entering in bbb it enters in a loop originally reported on google code with id when you enter in bbb you select no when you receive the question have you heard your voice then you select a different microphone and push next it never finishes you have to reload the page reported by smoral adhoclearning com on
1
89,638
25,862,949,281
IssuesEvent
2022-12-13 18:21:35
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
libtorch_python.so: undefined symbol: PyInstanceMethod_Type
module: build triaged topic: build
### ๐Ÿ› Describe the bug I run into this error vvv ``` C++ exception with description "/var/lib/jenkins/multipy/multipy/runtime/loader.cpp:780: libnvfuser_python.so: could not load library, dlopen says: /opt/conda/lib/python3.10/site-packages/torch/lib/libtorch_python.so: undefined symbol: PyInstanceMethod_Type" thrown in the test body. ``` link of failing CI: https://github.com/pytorch/pytorch/actions/runs/3586511045/jobs/6036274959 cmake lines related to the object: https://github.com/pytorch/pytorch/blob/d8335704a54f89e93f562cda7b49048e1eab1ebb/torch/csrc/jit/codegen/cuda/CMakeLists.txt#L165-L200 It's a little strange to me how the error was thrown from libtorch_python.so. This feels like a pytorch build issue. I'm also a little bit uncertain how this is invoked. In the setup, I have `libnvfuser_python.so` which depends on both `libnvfuser_codegen.so` and `libtorch_python.so`. While `libnvfuser_codegen.so` is dlopened by torch and does not have dependency on `libnvfuser_python.so`. ### Versions This is failing on CI job: [linux-bionic-cuda11.6-py3.10-gcc7 / test (deploy, 1, 1, linux.4xlarge.nvidia.gpu)](https://github.com/pytorch/pytorch/actions/runs/3590997095/jobs/6045485004) cc @malfet @seemethere
2.0
libtorch_python.so: undefined symbol: PyInstanceMethod_Type - ### ๐Ÿ› Describe the bug I run into this error vvv ``` C++ exception with description "/var/lib/jenkins/multipy/multipy/runtime/loader.cpp:780: libnvfuser_python.so: could not load library, dlopen says: /opt/conda/lib/python3.10/site-packages/torch/lib/libtorch_python.so: undefined symbol: PyInstanceMethod_Type" thrown in the test body. ``` link of failing CI: https://github.com/pytorch/pytorch/actions/runs/3586511045/jobs/6036274959 cmake lines related to the object: https://github.com/pytorch/pytorch/blob/d8335704a54f89e93f562cda7b49048e1eab1ebb/torch/csrc/jit/codegen/cuda/CMakeLists.txt#L165-L200 It's a little strange to me how the error was thrown from libtorch_python.so. This feels like a pytorch build issue. I'm also a little bit uncertain how this is invoked. In the setup, I have `libnvfuser_python.so` which depends on both `libnvfuser_codegen.so` and `libtorch_python.so`. While `libnvfuser_codegen.so` is dlopened by torch and does not have dependency on `libnvfuser_python.so`. ### Versions This is failing on CI job: [linux-bionic-cuda11.6-py3.10-gcc7 / test (deploy, 1, 1, linux.4xlarge.nvidia.gpu)](https://github.com/pytorch/pytorch/actions/runs/3590997095/jobs/6045485004) cc @malfet @seemethere
non_defect
libtorch python so undefined symbol pyinstancemethod type ๐Ÿ› describe the bug i run into this error vvv c exception with description var lib jenkins multipy multipy runtime loader cpp libnvfuser python so could not load library dlopen says opt conda lib site packages torch lib libtorch python so undefined symbol pyinstancemethod type thrown in the test body link of failing ci cmake lines related to the object it s a little strange to me how the error was thrown from libtorch python so this feels like a pytorch build issue i m also a little bit uncertain how this is invoked in the setup i have libnvfuser python so which depends on both libnvfuser codegen so and libtorch python so while libnvfuser codegen so is dlopened by torch and does not have dependency on libnvfuser python so versions this is failing on ci job cc malfet seemethere
0
2,538
2,607,926,695
IssuesEvent
2015-02-26 00:24:56
chrsmithdemos/minify
https://api.github.com/repos/chrsmithdemos/minify
closed
Add a GET parameter for debugging
auto-migrated Priority-Medium Type-Defect
``` A way to see the original js without minifying it, mainly for debugging/ developing purposes: In the .htaccess, we need the QSA flag: RewriteRule ^(.*\.(css|js))$ /minify.php?files=$1 [L,NC,QSA] And in the minify.php, simply: if (isset($_GET['debug'])) { echo $minify->combine(!($_GET['debug'])); } else { echo $minify->combine(); } So http://www.example.com/js/blabla.js?debug=1 should return the original js, without minifying it. Congrats for the software! It helps me a lot! Victor ``` ----- Original issue reported on code.google.com by `espiga...@gmail.com` on 21 Jul 2007 at 3:24
1.0
Add a GET parameter for debugging - ``` A way to see the original js without minifying it, mainly for debugging/ developing purposes: In the .htaccess, we need the QSA flag: RewriteRule ^(.*\.(css|js))$ /minify.php?files=$1 [L,NC,QSA] And in the minify.php, simply: if (isset($_GET['debug'])) { echo $minify->combine(!($_GET['debug'])); } else { echo $minify->combine(); } So http://www.example.com/js/blabla.js?debug=1 should return the original js, without minifying it. Congrats for the software! It helps me a lot! Victor ``` ----- Original issue reported on code.google.com by `espiga...@gmail.com` on 21 Jul 2007 at 3:24
defect
add a get parameter for debugging a way to see the original js without minifying it mainly for debugging developing purposes in the htaccess we need the qsa flag rewriterule css js minify php files and in the minify php simply if isset get echo minify combine get else echo minify combine so should return the original js without minifying it congrats for the software it helps me a lot victor original issue reported on code google com by espiga gmail com on jul at
1
14,040
24,277,096,750
IssuesEvent
2022-09-28 14:34:26
CS3219-AY2223S1/cs3219-project-ay2223s1-g5
https://api.github.com/repos/CS3219-AY2223S1/cs3219-project-ay2223s1-g5
closed
[FR-EDITOR-1] The system should have a code editor which allows users to code together in real time.
functional requirement P1
- [x] #185 - [x] #186
1.0
[FR-EDITOR-1] The system should have a code editor which allows users to code together in real time. - - [x] #185 - [x] #186
non_defect
the system should have a code editor which allows users to code together in real time
0
118,714
4,752,110,756
IssuesEvent
2016-10-23 08:37:54
CS2103AUG2016-T13-C2/main
https://api.github.com/repos/CS2103AUG2016-T13-C2/main
closed
bugs in undo delete
priority.high type.bug
should not add at the bottom, and also carry the information of completion
1.0
bugs in undo delete - should not add at the bottom, and also carry the information of completion
non_defect
bugs in undo delete should not add at the bottom and also carry the information of completion
0
55,954
14,860,098,508
IssuesEvent
2021-01-18 19:48:48
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
SelectOneMenu: JS error calling disable()
defect
**Describe the defect** Found during Integration Testing... When calling the widget `disable()` method throws the following JS error using the Showcase. ![image](https://user-images.githubusercontent.com/4399574/104950852-726eee00-598f-11eb-8f2a-13ab332ae729.png) This can happen whenver the items are rendered on the client. ```java .attr("renderPanelContentOnClient", menu.getVar() == null, false); ``` **Environment:** - PF Version: _10.0_ **Expected behavior** The SelectOneMenu should be disabled. **Example XHTML** ```html <p:selectOneMenu id="hideNoselectionOption" value="#{selectOneMenuView.hideNoSelectOption}" hideNoSelectionOption="#{not empty selectOneMenuView.hideNoSelectOption}"> <p:ajax update="@this" process="@this"/> <f:selectItem itemLabel="Select One" itemValue="#{null}" noSelectionOption="true"/> <f:selectItem itemLabel="Option1" itemValue="Option1"/> <f:selectItem itemLabel="Option2" itemValue="Option2"/> <f:selectItem itemLabel="Option3" itemValue="Option3"/> </p:selectOneMenu> ```
1.0
SelectOneMenu: JS error calling disable() - **Describe the defect** Found during Integration Testing... When calling the widget `disable()` method throws the following JS error using the Showcase. ![image](https://user-images.githubusercontent.com/4399574/104950852-726eee00-598f-11eb-8f2a-13ab332ae729.png) This can happen whenver the items are rendered on the client. ```java .attr("renderPanelContentOnClient", menu.getVar() == null, false); ``` **Environment:** - PF Version: _10.0_ **Expected behavior** The SelectOneMenu should be disabled. **Example XHTML** ```html <p:selectOneMenu id="hideNoselectionOption" value="#{selectOneMenuView.hideNoSelectOption}" hideNoSelectionOption="#{not empty selectOneMenuView.hideNoSelectOption}"> <p:ajax update="@this" process="@this"/> <f:selectItem itemLabel="Select One" itemValue="#{null}" noSelectionOption="true"/> <f:selectItem itemLabel="Option1" itemValue="Option1"/> <f:selectItem itemLabel="Option2" itemValue="Option2"/> <f:selectItem itemLabel="Option3" itemValue="Option3"/> </p:selectOneMenu> ```
defect
selectonemenu js error calling disable describe the defect found during integration testing when calling the widget disable method throws the following js error using the showcase this can happen whenver the items are rendered on the client java attr renderpanelcontentonclient menu getvar null false environment pf version expected behavior the selectonemenu should be disabled example xhtml html p selectonemenu id hidenoselectionoption value selectonemenuview hidenoselectoption hidenoselectionoption not empty selectonemenuview hidenoselectoption
1
54,377
13,632,224,963
IssuesEvent
2020-09-24 19:17:43
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
opened
Rebuild taxonomy entity index programmatically on a regular basis
Content governance Dashboards and UI Defect Drupal engineering Good first task
## Description Occasionally we're seeing cloned pages with edited Section taxonomy terms show up in the wrong /taxonomy/term/TID view We've found that rebuilding the index seems to fix this. ## Acceptance Criteria - [ ] Rebuild taxonomy entity index on a regular basis.
1.0
Rebuild taxonomy entity index programmatically on a regular basis - ## Description Occasionally we're seeing cloned pages with edited Section taxonomy terms show up in the wrong /taxonomy/term/TID view We've found that rebuilding the index seems to fix this. ## Acceptance Criteria - [ ] Rebuild taxonomy entity index on a regular basis.
defect
rebuild taxonomy entity index programmatically on a regular basis description occasionally we re seeing cloned pages with edited section taxonomy terms show up in the wrong taxonomy term tid view we ve found that rebuilding the index seems to fix this acceptance criteria rebuild taxonomy entity index on a regular basis
1
237,125
18,153,028,726
IssuesEvent
2021-09-26 15:47:47
kimym56/42Project
https://api.github.com/repos/kimym56/42Project
opened
0927 PanResponder
documentation
## **Done** - ์•„๋ž˜ ๋งํฌ ์ฝ”๋“œ(๋‚ ์งœ ๋ฒ”์œ„๋ฅผ dragging์œผ๋กœ ์ง€์ •) ํ…Œ์ŠคํŠธ & ์•ˆ๋“œ๋กœ์ด๋“œ location value ๋ฌธ์ œ ํ•ด๊ฒฐ https://blog.bam.tech/developer-news/how-to-handle-user-gestures-in-react-native-with-panresponder ## **To Do** - ์•„๋ž˜ ๋งํฌ ์ฝ”๋“œ(๋‚ ์งœ ๋ฒ”์œ„๋ฅผ touch(select)๋กœ ์ง€์ •)๋ž‘ ์œ„ ๋งํฌ ์ฝ”๋“œ ๊ฒฐ์ • https://react-day-picker.js.org/examples/selected-range/
1.0
0927 PanResponder - ## **Done** - ์•„๋ž˜ ๋งํฌ ์ฝ”๋“œ(๋‚ ์งœ ๋ฒ”์œ„๋ฅผ dragging์œผ๋กœ ์ง€์ •) ํ…Œ์ŠคํŠธ & ์•ˆ๋“œ๋กœ์ด๋“œ location value ๋ฌธ์ œ ํ•ด๊ฒฐ https://blog.bam.tech/developer-news/how-to-handle-user-gestures-in-react-native-with-panresponder ## **To Do** - ์•„๋ž˜ ๋งํฌ ์ฝ”๋“œ(๋‚ ์งœ ๋ฒ”์œ„๋ฅผ touch(select)๋กœ ์ง€์ •)๋ž‘ ์œ„ ๋งํฌ ์ฝ”๋“œ ๊ฒฐ์ • https://react-day-picker.js.org/examples/selected-range/
non_defect
panresponder done ์•„๋ž˜ ๋งํฌ ์ฝ”๋“œ ๋‚ ์งœ ๋ฒ”์œ„๋ฅผ dragging์œผ๋กœ ์ง€์ • ํ…Œ์ŠคํŠธ ์•ˆ๋“œ๋กœ์ด๋“œ location value ๋ฌธ์ œ ํ•ด๊ฒฐ to do ์•„๋ž˜ ๋งํฌ ์ฝ”๋“œ ๋‚ ์งœ ๋ฒ”์œ„๋ฅผ touch select ๋กœ ์ง€์ • ๋ž‘ ์œ„ ๋งํฌ ์ฝ”๋“œ ๊ฒฐ์ •
0
77,218
26,855,486,153
IssuesEvent
2023-02-03 14:19:50
obophenotype/cell-ontology
https://api.github.com/repos/obophenotype/cell-ontology
closed
names that are too broad for their definitions (ciliary epithelial cells)
Priority-Medium Type-Defect auto-migrated typo text definition
``` TICKET: These terms have names that are too broad for their definitions, or definitions that are two narrow for their names: id: CL:0002304 name: non-pigmented ciliary epithelial cell def: "A multi-ciliated cell of the retina that lacks visual pigment and contributes to aqueous humor by secreting chloride ions. This cell type maintains gap junctions with pigmented epithelial cells." [GOC:tfm, PMID:15106942] [Term] id: CL:0002303 name: pigmented ciliary epithelial cell def: "A ciliated epithelial cell of the retina, this cell type uptakes sodium chloride and passes it to non-pigmented ciliary epithelial cells." [PMID:15106942] ``` Original issue reported on code.google.com by `dosu...@gmail.com` on 27 Feb 2012 at 5:41
1.0
names that are too broad for their definitions (ciliary epithelial cells) - ``` TICKET: These terms have names that are too broad for their definitions, or definitions that are two narrow for their names: id: CL:0002304 name: non-pigmented ciliary epithelial cell def: "A multi-ciliated cell of the retina that lacks visual pigment and contributes to aqueous humor by secreting chloride ions. This cell type maintains gap junctions with pigmented epithelial cells." [GOC:tfm, PMID:15106942] [Term] id: CL:0002303 name: pigmented ciliary epithelial cell def: "A ciliated epithelial cell of the retina, this cell type uptakes sodium chloride and passes it to non-pigmented ciliary epithelial cells." [PMID:15106942] ``` Original issue reported on code.google.com by `dosu...@gmail.com` on 27 Feb 2012 at 5:41
defect
names that are too broad for their definitions ciliary epithelial cells ticket these terms have names that are too broad for their definitions or definitions that are two narrow for their names id cl name non pigmented ciliary epithelial cell def a multi ciliated cell of the retina that lacks visual pigment and contributes to aqueous humor by secreting chloride ions this cell type maintains gap junctions with pigmented epithelial cells goc tfm pmid id cl name pigmented ciliary epithelial cell def a ciliated epithelial cell of the retina this cell type uptakes sodium chloride and passes it to non pigmented ciliary epithelial cells original issue reported on code google com by dosu gmail com on feb at
1
13,832
2,787,212,491
IssuesEvent
2015-05-08 02:51:48
mblanchette/maven-java-formatter-plugin
https://api.github.com/repos/mblanchette/maven-java-formatter-plugin
closed
Update embedded Eclipse to 4.4 to allow for Java 8 sources.
auto-migrated Priority-Medium Type-Defect
``` Version 0.4 embeds an older Eclipse which does not know about Java 8. The recently released Eclipse 4.4 does, so it would be really nice if version 0.5 used this instead. ``` Original issue reported on code.google.com by `thorbjo...@gmail.com` on 27 Jun 2014 at 9:24
1.0
Update embedded Eclipse to 4.4 to allow for Java 8 sources. - ``` Version 0.4 embeds an older Eclipse which does not know about Java 8. The recently released Eclipse 4.4 does, so it would be really nice if version 0.5 used this instead. ``` Original issue reported on code.google.com by `thorbjo...@gmail.com` on 27 Jun 2014 at 9:24
defect
update embedded eclipse to to allow for java sources version embeds an older eclipse which does not know about java the recently released eclipse does so it would be really nice if version used this instead original issue reported on code google com by thorbjo gmail com on jun at
1
84,817
10,418,983,553
IssuesEvent
2019-09-15 13:17:01
square/okhttp
https://api.github.com/repos/square/okhttp
closed
Improve/Fix visibility of sample recipes e.g. Certificate Pinning
documentation
The docs on https://square.github.io/okhttp/https/#certificate-pinning say: ``` public SSLContext sslContextForTrustedCertificates(InputStream in) { ... // Full source omitted. See sample. } ``` but there is no such sample. https://github.com/square/okhttp/tree/master/samples/guide/src/main/java/okhttp3/recipes
1.0
Improve/Fix visibility of sample recipes e.g. Certificate Pinning - The docs on https://square.github.io/okhttp/https/#certificate-pinning say: ``` public SSLContext sslContextForTrustedCertificates(InputStream in) { ... // Full source omitted. See sample. } ``` but there is no such sample. https://github.com/square/okhttp/tree/master/samples/guide/src/main/java/okhttp3/recipes
non_defect
improve fix visibility of sample recipes e g certificate pinning the docs on say public sslcontext sslcontextfortrustedcertificates inputstream in full source omitted see sample but there is no such sample
0
68,760
7,109,041,991
IssuesEvent
2018-01-17 03:35:14
bitcoinjs/bitcoinjs-lib
https://api.github.com/repos/bitcoinjs/bitcoinjs-lib
closed
BIP143 rejects uncompressed public keys in P2WPKH or P2WSH
bug testing
Not sure if I am misunderstanding something here but does this not only check if the output is p2wpkh and not p2wsh as well? i.e. is correct? ``` if (kpPubKey.length !== 33 && (input.signType === scriptTypes.P2WPKH || input.signType === scriptTypes.P2WSH)) throw new Error('BIP143 rejects uncompressed public keys in P2WPKH or P2WSH') ``` https://github.com/bitcoinjs/bitcoinjs-lib/blob/86cd4a44a1524686d993f35b02a1ed331aaa551f/src/transaction_builder.js#L709
1.0
BIP143 rejects uncompressed public keys in P2WPKH or P2WSH - Not sure if I am misunderstanding something here but does this not only check if the output is p2wpkh and not p2wsh as well? i.e. is correct? ``` if (kpPubKey.length !== 33 && (input.signType === scriptTypes.P2WPKH || input.signType === scriptTypes.P2WSH)) throw new Error('BIP143 rejects uncompressed public keys in P2WPKH or P2WSH') ``` https://github.com/bitcoinjs/bitcoinjs-lib/blob/86cd4a44a1524686d993f35b02a1ed331aaa551f/src/transaction_builder.js#L709
non_defect
rejects uncompressed public keys in or not sure if i am misunderstanding something here but does this not only check if the output is and not as well i e is correct if kppubkey length input signtype scripttypes input signtype scripttypes throw new error rejects uncompressed public keys in or
0
5,609
3,970,223,416
IssuesEvent
2016-05-04 05:44:51
kolliSuman/issues
https://api.github.com/repos/kolliSuman/issues
closed
QA_Introduction to Programmable Logic Controller_Self evaluation_smk
Category: Usability Developed By: VLEAD Release Number: Production Severity: S3 Status: Open
Defect Description : In Self Evaluation page of "Introduction to Programmable Logic Controller and Introduction to Digital I/O Interface to PLCโ€ when we click on get start button the Content on the top of the page is not clear where Content should be present with no distortions Actual Result : In Self Evaluation page of "Introduction to Programmable Logic Controller and Introduction to Digital I/O Interface to PLCโ€ when we click on get start button the Content on the top of the page is not clear Environment : OS: Windows 7, Ubuntu-16.04,Centos-6 Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0 Bandwidth : 100Mbps Hardware Configuration:8GBRAM , Processor:i5 Test Step Link: https://github.com/Virtual-Labs/industrial-electrical-drives-nitk/blob/master/test-cases/integration_test-cases/Introduction%20to%20Programmable%20Logic%20Controller/Introduction%20to%20Programmable%20Logic%20Controller_04_Self%20evaluation_smk.org
True
QA_Introduction to Programmable Logic Controller_Self evaluation_smk - Defect Description : In Self Evaluation page of "Introduction to Programmable Logic Controller and Introduction to Digital I/O Interface to PLCโ€ when we click on get start button the Content on the top of the page is not clear where Content should be present with no distortions Actual Result : In Self Evaluation page of "Introduction to Programmable Logic Controller and Introduction to Digital I/O Interface to PLCโ€ when we click on get start button the Content on the top of the page is not clear Environment : OS: Windows 7, Ubuntu-16.04,Centos-6 Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0 Bandwidth : 100Mbps Hardware Configuration:8GBRAM , Processor:i5 Test Step Link: https://github.com/Virtual-Labs/industrial-electrical-drives-nitk/blob/master/test-cases/integration_test-cases/Introduction%20to%20Programmable%20Logic%20Controller/Introduction%20to%20Programmable%20Logic%20Controller_04_Self%20evaluation_smk.org
non_defect
qa introduction to programmable logic controller self evaluation smk defect description in self evaluation page of introduction to programmable logic controller and introduction to digital i o interface to plcโ€ when we click on get start button the content on the top of the page is not clear where content should be present with no distortions actual result in self evaluation page of introduction to programmable logic controller and introduction to digital i o interface to plcโ€ when we click on get start button the content on the top of the page is not clear environment os windows ubuntu centos browsers firefox chrome chromium bandwidth hardware configuration processor test step link
0
63,182
17,409,860,851
IssuesEvent
2021-08-03 10:52:56
AtlasOfLivingAustralia/collectory
https://api.github.com/repos/AtlasOfLivingAustralia/collectory
closed
download button a dataset page doesn't work
priority-medium status-new type-defect
_From @mbohun on August 19, 2014 12:33_ _migrated from:_ https://code.google.com/p/ala/issues/detail?id=524 _date:_ Tue Jan 14 20:41:20 2014 _author:_ milo_nic...@hotmail.com --- e.g. clicking the download records button on [http://collections.ala.org.au/public/show/dr365](http://collections.ala.org.au/public/show/dr365) produces an error - A reasonTypeId must be provided. _Copied from original issue: AtlasOfLivingAustralia/biocache-hubs#55_
1.0
download button a dataset page doesn't work - _From @mbohun on August 19, 2014 12:33_ _migrated from:_ https://code.google.com/p/ala/issues/detail?id=524 _date:_ Tue Jan 14 20:41:20 2014 _author:_ milo_nic...@hotmail.com --- e.g. clicking the download records button on [http://collections.ala.org.au/public/show/dr365](http://collections.ala.org.au/public/show/dr365) produces an error - A reasonTypeId must be provided. _Copied from original issue: AtlasOfLivingAustralia/biocache-hubs#55_
defect
download button a dataset page doesn t work from mbohun on august migrated from date tue jan author milo nic hotmail com e g clicking the download records button on produces an error a reasontypeid must be provided copied from original issue atlasoflivingaustralia biocache hubs
1
203,692
15,380,550,885
IssuesEvent
2021-03-02 21:16:28
radicle-dev/radicle-link
https://api.github.com/repos/radicle-dev/radicle-link
closed
Split integration tests
brooming help wanted testing
`propagation_basic` was meant to test some, err, basic things. Before it gets out of hand, we should split into a couple more modules. As those tests are becoming more heavy on the IOs, this also prepares for running them concurrently and/or behind a `slow` feature.
1.0
Split integration tests - `propagation_basic` was meant to test some, err, basic things. Before it gets out of hand, we should split into a couple more modules. As those tests are becoming more heavy on the IOs, this also prepares for running them concurrently and/or behind a `slow` feature.
non_defect
split integration tests propagation basic was meant to test some err basic things before it gets out of hand we should split into a couple more modules as those tests are becoming more heavy on the ios this also prepares for running them concurrently and or behind a slow feature
0
2,772
2,607,944,887
IssuesEvent
2015-02-26 00:32:52
chrsmithdemos/switchlist
https://api.github.com/repos/chrsmithdemos/switchlist
opened
Add explicit checks for duplicate town names.
auto-migrated Priority-Medium Type-Defect
``` There's currently no validation that two towns don't have the same names. This would mess up the "train stops" code that assumes only one town exists with a given name. ``` ----- Original issue reported on code.google.com by `rwbowdi...@gmail.com` on 24 Apr 2011 at 5:25
1.0
Add explicit checks for duplicate town names. - ``` There's currently no validation that two towns don't have the same names. This would mess up the "train stops" code that assumes only one town exists with a given name. ``` ----- Original issue reported on code.google.com by `rwbowdi...@gmail.com` on 24 Apr 2011 at 5:25
defect
add explicit checks for duplicate town names there s currently no validation that two towns don t have the same names this would mess up the train stops code that assumes only one town exists with a given name original issue reported on code google com by rwbowdi gmail com on apr at
1
27,303
4,958,098,713
IssuesEvent
2016-12-02 08:27:19
TNGSB/eWallet
https://api.github.com/repos/TNGSB/eWallet
closed
eWallet_MobileApp(Reload) #105
Defect - High (Sev-2)
[Defect_Mobile App #105.xlsx](https://github.com/TNGSB/eWallet/files/593552/Defect_Mobile.App.105.xlsx) Defect Description : System displayed a successful message and automatically redirect user to "Credit/Debit Card" page before the countdown counter time out Test Description : To verify error message when user exceeded the countdown timer for credit card detail input Refer attachment for POT
1.0
eWallet_MobileApp(Reload) #105 - [Defect_Mobile App #105.xlsx](https://github.com/TNGSB/eWallet/files/593552/Defect_Mobile.App.105.xlsx) Defect Description : System displayed a successful message and automatically redirect user to "Credit/Debit Card" page before the countdown counter time out Test Description : To verify error message when user exceeded the countdown timer for credit card detail input Refer attachment for POT
defect
ewallet mobileapp reload defect description system displayed a successful message and automatically redirect user to credit debit card page before the countdown counter time out test description to verify error message when user exceeded the countdown timer for credit card detail input refer attachment for pot
1
11,012
2,622,955,112
IssuesEvent
2015-03-04 09:03:13
folded/carve
https://api.github.com/repos/folded/carve
opened
Carve fails on (apparantly) valid input
auto-migrated Priority-Medium Type-Defect
``` Just run: >intersect "(_0a.ply | _1a.ply | _2a.ply) A_MINUS_B _3s.ply" with the attached files. Files attached were created with carve by intersecting several simple models. Viewing tool supplied with carve opens them without any problems. ``` Original issue reported on code.google.com by `ru.el...@gmail.com` on 17 May 2010 at 2:31 Attachments: * [_0a.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-17/comment-0/_0a.ply) * [_1a.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-17/comment-0/_1a.ply) * [_2a.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-17/comment-0/_2a.ply) * [_3s.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-17/comment-0/_3s.ply)
1.0
Carve fails on (apparantly) valid input - ``` Just run: >intersect "(_0a.ply | _1a.ply | _2a.ply) A_MINUS_B _3s.ply" with the attached files. Files attached were created with carve by intersecting several simple models. Viewing tool supplied with carve opens them without any problems. ``` Original issue reported on code.google.com by `ru.el...@gmail.com` on 17 May 2010 at 2:31 Attachments: * [_0a.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-17/comment-0/_0a.ply) * [_1a.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-17/comment-0/_1a.ply) * [_2a.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-17/comment-0/_2a.ply) * [_3s.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-17/comment-0/_3s.ply)
defect
carve fails on apparantly valid input just run intersect ply ply ply a minus b ply with the attached files files attached were created with carve by intersecting several simple models viewing tool supplied with carve opens them without any problems original issue reported on code google com by ru el gmail com on may at attachments
1
108,794
23,664,428,850
IssuesEvent
2022-08-26 19:07:06
unoplatform/uno
https://api.github.com/repos/unoplatform/uno
opened
Enable XAML Trimming for `net6.0-[ios|android|macos|maccatalyst]` and GTK targets
kind/enhancement area/code-generation
## What would you like to be added: Support for [XAML Trimming](https://platform.uno/docs/articles/features/resources-trimming.html?) for `net6.0-[ios|android|macos|maccatalyst]` and GTK targets. ## Why is this needed: Reduce the size of applications. ## Anything else we need to know? WPF is explicitly excluded from this list as trimming is forcibly disabled by the SDK. Work branch: https://github.com/unoplatform/uno/tree/dev/eb/xaml-trimming-net6 ## Related issues - https://github.com/dotnet/sdk/issues/27492 - https://github.com/xamarin/xamarin-android/issues/7301
1.0
Enable XAML Trimming for `net6.0-[ios|android|macos|maccatalyst]` and GTK targets - ## What would you like to be added: Support for [XAML Trimming](https://platform.uno/docs/articles/features/resources-trimming.html?) for `net6.0-[ios|android|macos|maccatalyst]` and GTK targets. ## Why is this needed: Reduce the size of applications. ## Anything else we need to know? WPF is explicitly excluded from this list as trimming is forcibly disabled by the SDK. Work branch: https://github.com/unoplatform/uno/tree/dev/eb/xaml-trimming-net6 ## Related issues - https://github.com/dotnet/sdk/issues/27492 - https://github.com/xamarin/xamarin-android/issues/7301
non_defect
enable xaml trimming for and gtk targets what would you like to be added support for for and gtk targets why is this needed reduce the size of applications anything else we need to know wpf is explicitly excluded from this list as trimming is forcibly disabled by the sdk work branch related issues
0
18,922
3,734,406,963
IssuesEvent
2016-03-08 06:41:42
kumulsoft/Fixed-Assets
https://api.github.com/repos/kumulsoft/Fixed-Assets
closed
Asset Transactions - Grid Page - Enhancement
enhancement Fixed HIGH Ready for testing
Remove the 'Trans Detail' column to give space to add two new columns 'Centre and Custodian) ![image](https://cloud.githubusercontent.com/assets/10192106/13310137/232bdbf2-dbcf-11e5-95e7-169430324517.png)
1.0
Asset Transactions - Grid Page - Enhancement - Remove the 'Trans Detail' column to give space to add two new columns 'Centre and Custodian) ![image](https://cloud.githubusercontent.com/assets/10192106/13310137/232bdbf2-dbcf-11e5-95e7-169430324517.png)
non_defect
asset transactions grid page enhancement remove the trans detail column to give space to add two new columns centre and custodian
0
62,085
17,023,847,476
IssuesEvent
2021-07-03 04:09:20
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Only 1 of 2 streets with same name found
Component: nominatim Priority: major Resolution: duplicate Type: defect
**[Submitted to the original trac issue database at 11.18pm, Thursday, 3rd January 2013]** http://nominatim.openstreetmap.org/search.php?q=tannenweg+lenzkirch&viewbox=8.19%2C47.88%2C8.21%2C47.87&polygon=1 Only one of the 2 residential streets named Tannenweg are found. The streets are merging in a Y shape in their western parts. The error leads to a wrong bounding box for zooming. Thus depending on screen resolution and street geometry it might not be obvious to a human user that a second street with the same name exists (is only shown partially but the label off-screen).
1.0
Only 1 of 2 streets with same name found - **[Submitted to the original trac issue database at 11.18pm, Thursday, 3rd January 2013]** http://nominatim.openstreetmap.org/search.php?q=tannenweg+lenzkirch&viewbox=8.19%2C47.88%2C8.21%2C47.87&polygon=1 Only one of the 2 residential streets named Tannenweg are found. The streets are merging in a Y shape in their western parts. The error leads to a wrong bounding box for zooming. Thus depending on screen resolution and street geometry it might not be obvious to a human user that a second street with the same name exists (is only shown partially but the label off-screen).
defect
only of streets with same name found only one of the residential streets named tannenweg are found the streets are merging in a y shape in their western parts the error leads to a wrong bounding box for zooming thus depending on screen resolution and street geometry it might not be obvious to a human user that a second street with the same name exists is only shown partially but the label off screen
1
435,622
30,510,415,808
IssuesEvent
2023-07-18 20:21:19
Nnoemis/Test1
https://api.github.com/repos/Nnoemis/Test1
opened
How would you test a 5 kg capacity grocery shopping paper bag? Describe the tests that you could perform.
documentation
1. Insert 3 kg of oranges. 2. Insert 4.5 kg of oranges. 3.Insert 5 kg of oranges. 4. Insert + 100 g of oranges. 5. Insert + 200 g of oranges. 6. Every next time insert + 100 g of oranges and stop when paper bag is damaged. (Bounding testing) 7. Try the these steps with a wet paper bag. 8. Try the bag`s volume. Sourse [https://simeon.svet-bg.com/index.php/forums/topic/think-testing-5-kg-bag/](url)
1.0
How would you test a 5 kg capacity grocery shopping paper bag? Describe the tests that you could perform. - 1. Insert 3 kg of oranges. 2. Insert 4.5 kg of oranges. 3.Insert 5 kg of oranges. 4. Insert + 100 g of oranges. 5. Insert + 200 g of oranges. 6. Every next time insert + 100 g of oranges and stop when paper bag is damaged. (Bounding testing) 7. Try the these steps with a wet paper bag. 8. Try the bag`s volume. Sourse [https://simeon.svet-bg.com/index.php/forums/topic/think-testing-5-kg-bag/](url)
non_defect
how would you test a kg capacity grocery shopping paper bag describe the tests that you could perform insert kg of oranges insert kg of oranges insert kg of oranges insert g of oranges insert g of oranges every next time insert g of oranges and stop when paper bag is damaged bounding testing try the these steps with a wet paper bag try the bag s volume sourse url
0
41,834
10,679,235,358
IssuesEvent
2019-10-21 18:51:50
techo/voluntariado-eventual
https://api.github.com/repos/techo/voluntariado-eventual
closed
Error al mostrar puntos de encuentro en la inscripciรณn
Defecto
**Describรญ el error** Como usuario al querer inscribirme en una actividad los puntos de encuentro se muestran incorrectamente. **Para reproducirlo** Pasos para reproducir el comportamiento: 1. Cargar una nueva actividad. 2. Cargar 2 puntos de encuentro en diferentes provincias 3. Como usuario ir a la actividad a inscribirme 4. Ver que las pronvincias de los puntos de encuentro se muestran incorrectamente. **Comportamiento esperando** Que las provincias se vean igual que en la carga de actividad. **Capturas de pantalla** ![image](https://user-images.githubusercontent.com/94343/66665466-f8e13f80-ec24-11e9-9747-c4a4f57117d2.png)
1.0
Error al mostrar puntos de encuentro en la inscripciรณn - **Describรญ el error** Como usuario al querer inscribirme en una actividad los puntos de encuentro se muestran incorrectamente. **Para reproducirlo** Pasos para reproducir el comportamiento: 1. Cargar una nueva actividad. 2. Cargar 2 puntos de encuentro en diferentes provincias 3. Como usuario ir a la actividad a inscribirme 4. Ver que las pronvincias de los puntos de encuentro se muestran incorrectamente. **Comportamiento esperando** Que las provincias se vean igual que en la carga de actividad. **Capturas de pantalla** ![image](https://user-images.githubusercontent.com/94343/66665466-f8e13f80-ec24-11e9-9747-c4a4f57117d2.png)
defect
error al mostrar puntos de encuentro en la inscripciรณn describรญ el error como usuario al querer inscribirme en una actividad los puntos de encuentro se muestran incorrectamente para reproducirlo pasos para reproducir el comportamiento cargar una nueva actividad cargar puntos de encuentro en diferentes provincias como usuario ir a la actividad a inscribirme ver que las pronvincias de los puntos de encuentro se muestran incorrectamente comportamiento esperando que las provincias se vean igual que en la carga de actividad capturas de pantalla
1
74,877
25,379,761,270
IssuesEvent
2022-11-21 16:37:21
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
opened
Liquibase 4.17 compatibility
T: Defect
### Expected behavior Jooq Liquibase extension to be compatible with recent version of Liquibase ### Actual behavior The Jooq liquibase extension uses `FileSystemResourceAccessor` which was [removed](https://github.com/liquibase/liquibase/issues/3478) in Liquibase 4.17. ### Steps to reproduce the problem Log output from an attempt to run the Gradle plugin `generateJooq` task: ``` matt @ ghost-s1 in ~/Workspace/github.com/mattupstate/acme (โŽˆk3d-acme-dev/default) on git:main+ $ ./gradlew acme-data:acme-data-sql:generateJooq > Task :acme-data:acme-data-sql:generateJooq FAILED Exception in thread "main" java.lang.NoClassDefFoundError: liquibase/resource/FileSystemResourceAccessor at org.jooq.meta.extensions.liquibase.LiquibaseDatabase.export(LiquibaseDatabase.java:155) at org.jooq.meta.extensions.AbstractInterpretingDatabase.connection(AbstractInterpretingDatabase.java:100) at org.jooq.meta.extensions.AbstractInterpretingDatabase.create0(AbstractInterpretingDatabase.java:77) at org.jooq.meta.AbstractDatabase.create(AbstractDatabase.java:369) at org.jooq.meta.AbstractDatabase.create(AbstractDatabase.java:359) at org.jooq.meta.AbstractDatabase.setConnection(AbstractDatabase.java:337) at org.jooq.codegen.GenerationTool.run0(GenerationTool.java:553) at org.jooq.codegen.GenerationTool.run(GenerationTool.java:240) at org.jooq.codegen.GenerationTool.generate(GenerationTool.java:235) at org.jooq.codegen.GenerationTool.main(GenerationTool.java:207) Caused by: java.lang.ClassNotFoundException: liquibase.resource.FileSystemResourceAccessor at java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:641) at java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:188) at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:520) ... 10 more ``` ### jOOQ Version 3.17.4 ### Database product and version N/A ### Java Version N/A ### OS Version N/A ### JDBC driver name and version (include name if unofficial driver) N/A
1.0
Liquibase 4.17 compatibility - ### Expected behavior Jooq Liquibase extension to be compatible with recent version of Liquibase ### Actual behavior The Jooq liquibase extension uses `FileSystemResourceAccessor` which was [removed](https://github.com/liquibase/liquibase/issues/3478) in Liquibase 4.17. ### Steps to reproduce the problem Log output from an attempt to run the Gradle plugin `generateJooq` task: ``` matt @ ghost-s1 in ~/Workspace/github.com/mattupstate/acme (โŽˆk3d-acme-dev/default) on git:main+ $ ./gradlew acme-data:acme-data-sql:generateJooq > Task :acme-data:acme-data-sql:generateJooq FAILED Exception in thread "main" java.lang.NoClassDefFoundError: liquibase/resource/FileSystemResourceAccessor at org.jooq.meta.extensions.liquibase.LiquibaseDatabase.export(LiquibaseDatabase.java:155) at org.jooq.meta.extensions.AbstractInterpretingDatabase.connection(AbstractInterpretingDatabase.java:100) at org.jooq.meta.extensions.AbstractInterpretingDatabase.create0(AbstractInterpretingDatabase.java:77) at org.jooq.meta.AbstractDatabase.create(AbstractDatabase.java:369) at org.jooq.meta.AbstractDatabase.create(AbstractDatabase.java:359) at org.jooq.meta.AbstractDatabase.setConnection(AbstractDatabase.java:337) at org.jooq.codegen.GenerationTool.run0(GenerationTool.java:553) at org.jooq.codegen.GenerationTool.run(GenerationTool.java:240) at org.jooq.codegen.GenerationTool.generate(GenerationTool.java:235) at org.jooq.codegen.GenerationTool.main(GenerationTool.java:207) Caused by: java.lang.ClassNotFoundException: liquibase.resource.FileSystemResourceAccessor at java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:641) at java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:188) at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:520) ... 10 more ``` ### jOOQ Version 3.17.4 ### Database product and version N/A ### Java Version N/A ### OS Version N/A ### JDBC driver name and version (include name if unofficial driver) N/A
defect
liquibase compatibility expected behavior jooq liquibase extension to be compatible with recent version of liquibase actual behavior the jooq liquibase extension uses filesystemresourceaccessor which was in liquibase steps to reproduce the problem log output from an attempt to run the gradle plugin generatejooq task matt ghost in workspace github com mattupstate acme โŽˆ acme dev default on git main gradlew acme data acme data sql generatejooq task acme data acme data sql generatejooq failed exception in thread main java lang noclassdeffounderror liquibase resource filesystemresourceaccessor at org jooq meta extensions liquibase liquibasedatabase export liquibasedatabase java at org jooq meta extensions abstractinterpretingdatabase connection abstractinterpretingdatabase java at org jooq meta extensions abstractinterpretingdatabase abstractinterpretingdatabase java at org jooq meta abstractdatabase create abstractdatabase java at org jooq meta abstractdatabase create abstractdatabase java at org jooq meta abstractdatabase setconnection abstractdatabase java at org jooq codegen generationtool generationtool java at org jooq codegen generationtool run generationtool java at org jooq codegen generationtool generate generationtool java at org jooq codegen generationtool main generationtool java caused by java lang classnotfoundexception liquibase resource filesystemresourceaccessor at java base jdk internal loader builtinclassloader loadclass builtinclassloader java at java base jdk internal loader classloaders appclassloader loadclass classloaders java at java base java lang classloader loadclass classloader java more jooq version database product and version n a java version n a os version n a jdbc driver name and version include name if unofficial driver n a
1
15,881
2,869,088,191
IssuesEvent
2015-06-05 23:14:17
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
Collapse together errors of the same kind
Area-Pkg Pkg-PolymerDevExp PolymerMilestone-Next Priority-Medium Triaged Type-Defect
It's common to see one warning/error repeated in many places. For example the href vs \_href might happen a lot now that we introduce the warning for it. It would be great to combine all occurrences of the same error together. Some ideas: - for the log_injector UI, we can combine it graphically. - for the pub-build/pub-serve output in stdout, we can delay reporting warnings until the end of the phase, collect all warnings together and report them as such. At least we can do that on a file-per-file basis.
1.0
Collapse together errors of the same kind - It's common to see one warning/error repeated in many places. For example the href vs \_href might happen a lot now that we introduce the warning for it. It would be great to combine all occurrences of the same error together. Some ideas: - for the log_injector UI, we can combine it graphically. - for the pub-build/pub-serve output in stdout, we can delay reporting warnings until the end of the phase, collect all warnings together and report them as such. At least we can do that on a file-per-file basis.
defect
collapse together errors of the same kind it s common to see one warning error repeated in many places for example the href vs href might happen a lot now that we introduce the warning for it it would be great to combine all occurrences of the same error together some ideas for the log injector ui we can combine it graphically for the pub build pub serve output in stdout we can delay reporting warnings until the end of the phase collect all warnings together and report them as such at least we can do that on a file per file basis
1
4,357
10,965,734,316
IssuesEvent
2019-11-28 04:13:34
fga-eps-mds/2019.2-Over26
https://api.github.com/repos/fga-eps-mds/2019.2-Over26
closed
Atualizar diagramas do Documento de Arquitetura
Architecture Documentation EPS
## Descriรงรฃo da Mudanรงa * <!--- Forneรงa um resumo geral da _issue_ --> ร‰ necessรกrio adequar os diagramas presentes no Documento de Arquitetura ร  atual estrutura do projeto. ## Checklist * <!-- Essa checklist propรตe a criaรงรฃo de uma boa issue --> <!-- Se a issue รฉ sobre uma histรณria de usuรกrio, seu nome deve ser "USXX - Nome da histรณria--> <!-- Se a issue รฉ sobre um bug, seu nome deve ser "BF - Nome curto do bug"--> <!-- Se a issue รฉ sobre outra tarefa o nome deve ser uma simples descriรงรฃo da tarefa--> - [x] Esta issue tem um nome significativo. - [x] O nome da issue estรก no padrรฃo. - [x] Esta issue tem uma descriรงรฃo de fรกcil entendimento. - [x] Esta issue tem uma boa definiรงรฃo de critรฉrios de aceitaรงรฃo. - [x] Esta issue tem labels associadas. - [ ] Esta issue estรก associada ร  uma milestone. - [ ] Esta issue tem uma pontuaรงรฃo estimada. ## Tarefas * <!-- Adicione aqui as tarefas necessรกrias para concluir a issue --> - [x] Atualizar diagrama de classes - [x] Atualizar diagrama lรณgico - [x] Atualizar diagrama de pacotes ## Critรฉrios de Aceitaรงรฃo * <!-- Liste aqui o conjunto de aspectos mecessรกrios para considerar a atividade como completa--> <!-- Os itens serรฃo adicionados pelo Product Owner --> - [x] Diagramas atualizados.
1.0
Atualizar diagramas do Documento de Arquitetura - ## Descriรงรฃo da Mudanรงa * <!--- Forneรงa um resumo geral da _issue_ --> ร‰ necessรกrio adequar os diagramas presentes no Documento de Arquitetura ร  atual estrutura do projeto. ## Checklist * <!-- Essa checklist propรตe a criaรงรฃo de uma boa issue --> <!-- Se a issue รฉ sobre uma histรณria de usuรกrio, seu nome deve ser "USXX - Nome da histรณria--> <!-- Se a issue รฉ sobre um bug, seu nome deve ser "BF - Nome curto do bug"--> <!-- Se a issue รฉ sobre outra tarefa o nome deve ser uma simples descriรงรฃo da tarefa--> - [x] Esta issue tem um nome significativo. - [x] O nome da issue estรก no padrรฃo. - [x] Esta issue tem uma descriรงรฃo de fรกcil entendimento. - [x] Esta issue tem uma boa definiรงรฃo de critรฉrios de aceitaรงรฃo. - [x] Esta issue tem labels associadas. - [ ] Esta issue estรก associada ร  uma milestone. - [ ] Esta issue tem uma pontuaรงรฃo estimada. ## Tarefas * <!-- Adicione aqui as tarefas necessรกrias para concluir a issue --> - [x] Atualizar diagrama de classes - [x] Atualizar diagrama lรณgico - [x] Atualizar diagrama de pacotes ## Critรฉrios de Aceitaรงรฃo * <!-- Liste aqui o conjunto de aspectos mecessรกrios para considerar a atividade como completa--> <!-- Os itens serรฃo adicionados pelo Product Owner --> - [x] Diagramas atualizados.
non_defect
atualizar diagramas do documento de arquitetura descriรงรฃo da mudanรงa รฉ necessรกrio adequar os diagramas presentes no documento de arquitetura ร  atual estrutura do projeto checklist esta issue tem um nome significativo o nome da issue estรก no padrรฃo esta issue tem uma descriรงรฃo de fรกcil entendimento esta issue tem uma boa definiรงรฃo de critรฉrios de aceitaรงรฃo esta issue tem labels associadas esta issue estรก associada ร  uma milestone esta issue tem uma pontuaรงรฃo estimada tarefas atualizar diagrama de classes atualizar diagrama lรณgico atualizar diagrama de pacotes critรฉrios de aceitaรงรฃo diagramas atualizados
0
44,751
12,372,216,940
IssuesEvent
2020-05-18 19:59:26
FoldingAtHome/fah-issues
https://api.github.com/repos/FoldingAtHome/fah-issues
closed
Dialog Box opened from FahCore22.exe, entry point clReleaseDevice could not be located
FAHClient defect
## Environment * OS: Windows 7 Enterprise 64 build 7601 SP1 * FAH version: April 28, 2020 ## What were you trying to do Folding on FAH for Covid-19 ## What happened NVIDIA GEOFORCE GT 610 as GPU:0 hangs, throws dialog box, FahCore_22.exe - Entry Point Not Found. White X in Red circle: The procedure entry point clReleaseDevice could not be located in the dynamic link library OpenCL.dll. "OK" the dialog box, and a second box is thrown of same type OK that one and GPU process indicator turns to solid yellow. And this cycle continues over and over. ## To Reproduce just keeps happening each time.
1.0
Dialog Box opened from FahCore22.exe, entry point clReleaseDevice could not be located - ## Environment * OS: Windows 7 Enterprise 64 build 7601 SP1 * FAH version: April 28, 2020 ## What were you trying to do Folding on FAH for Covid-19 ## What happened NVIDIA GEOFORCE GT 610 as GPU:0 hangs, throws dialog box, FahCore_22.exe - Entry Point Not Found. White X in Red circle: The procedure entry point clReleaseDevice could not be located in the dynamic link library OpenCL.dll. "OK" the dialog box, and a second box is thrown of same type OK that one and GPU process indicator turns to solid yellow. And this cycle continues over and over. ## To Reproduce just keeps happening each time.
defect
dialog box opened from exe entry point clreleasedevice could not be located environment os windows enterprise build fah version april what were you trying to do folding on fah for covid what happened nvidia geoforce gt as gpu hangs throws dialog box fahcore exe entry point not found white x in red circle the procedure entry point clreleasedevice could not be located in the dynamic link library opencl dll ok the dialog box and a second box is thrown of same type ok that one and gpu process indicator turns to solid yellow and this cycle continues over and over to reproduce just keeps happening each time
1
46,293
24,466,746,399
IssuesEvent
2022-10-07 15:39:27
Kitware/vtk-js
https://api.github.com/repos/Kitware/vtk-js
closed
Setting a model variable as object always trigger modified()
type: bug ๐Ÿž type: performance โšก๏ธ
### High-level description If `macro.set(publicAPI, model, ['foo']);` and `foo` is an `Object` (e.g. `{a: 1, b: 2}`, then calling `setFoo({a: 1, b:2})` will always trigger `modified()`. ### Steps to reproduce ``` extend(...){ macro.set(publicAPI, model, ['foo']); } ... myobj.setFoo({a: 1, b: 2}); myobj.getMTime(); // =>1234 myobj.setFoo({a: 1, b: 2}); myobj.getMTime(); // =>1235 ``` See [here](https://github.com/Kitware/vtk-js/blob/master/Sources/Widgets/Core/WidgetManager/index.js#L239) for a performance hit that modifies a WidgetRepresentation each time the mouse is moved which will always `requestData()` on the widget representation ### Detailed behavior This is because `macro.setter()` does a shallow equal comparison `if (model[field] !== value) {` instead of a deep equal. ### Expected behavior I can see 2 options: - Add deep equal check if `value` is an object in `macro.setter()` - Create `macro.setObject(publicAPI, model, ['foo']);` convenient setter similar to `macro.setArray()` ### Environment - **vtk.js**: master
True
Setting a model variable as object always trigger modified() - ### High-level description If `macro.set(publicAPI, model, ['foo']);` and `foo` is an `Object` (e.g. `{a: 1, b: 2}`, then calling `setFoo({a: 1, b:2})` will always trigger `modified()`. ### Steps to reproduce ``` extend(...){ macro.set(publicAPI, model, ['foo']); } ... myobj.setFoo({a: 1, b: 2}); myobj.getMTime(); // =>1234 myobj.setFoo({a: 1, b: 2}); myobj.getMTime(); // =>1235 ``` See [here](https://github.com/Kitware/vtk-js/blob/master/Sources/Widgets/Core/WidgetManager/index.js#L239) for a performance hit that modifies a WidgetRepresentation each time the mouse is moved which will always `requestData()` on the widget representation ### Detailed behavior This is because `macro.setter()` does a shallow equal comparison `if (model[field] !== value) {` instead of a deep equal. ### Expected behavior I can see 2 options: - Add deep equal check if `value` is an object in `macro.setter()` - Create `macro.setObject(publicAPI, model, ['foo']);` convenient setter similar to `macro.setArray()` ### Environment - **vtk.js**: master
non_defect
setting a model variable as object always trigger modified high level description if macro set publicapi model and foo is an object e g a b then calling setfoo a b will always trigger modified steps to reproduce extend macro set publicapi model myobj setfoo a b myobj getmtime myobj setfoo a b myobj getmtime see for a performance hit that modifies a widgetrepresentation each time the mouse is moved which will always requestdata on the widget representation detailed behavior this is because macro setter does a shallow equal comparison if model value instead of a deep equal expected behavior i can see options add deep equal check if value is an object in macro setter create macro setobject publicapi model convenient setter similar to macro setarray environment vtk js master
0
47,552
13,056,241,652
IssuesEvent
2020-07-30 04:05:51
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
steamshovel file open slowness (Trac #735)
Migrated from Trac combo core defect
"After investigating this problem a bit more, we found out how to get around the slowness we've been experiencing when using steamshovel in the offline-software meta-project. When we just run steamshovel, without passing the files in the command line, but opening them from the GUI, things work fine. When loading the files from command line, it goes very slowly." Possibly something different between cmd line loading and file menu loading. Migrated from https://code.icecube.wisc.edu/ticket/735 ```json { "status": "closed", "changetime": "2015-07-10T14:08:08", "description": "\"After investigating this problem a bit more, we found out how to get around the slowness we've been experiencing when using steamshovel in the offline-software meta-project. When we just run steamshovel, without passing the files in the command line, but opening them from the GUI, things work fine. When loading the files from command line, it goes very slowly.\"\n\nPossibly something different between cmd line loading and file menu loading.", "reporter": "david.schultz", "cc": "", "resolution": "fixed", "_ts": "1436537288260585", "component": "combo core", "summary": "steamshovel file open slowness", "priority": "normal", "keywords": "", "time": "2014-08-13T15:04:32", "milestone": "", "owner": "hdembinski", "type": "defect" } ```
1.0
steamshovel file open slowness (Trac #735) - "After investigating this problem a bit more, we found out how to get around the slowness we've been experiencing when using steamshovel in the offline-software meta-project. When we just run steamshovel, without passing the files in the command line, but opening them from the GUI, things work fine. When loading the files from command line, it goes very slowly." Possibly something different between cmd line loading and file menu loading. Migrated from https://code.icecube.wisc.edu/ticket/735 ```json { "status": "closed", "changetime": "2015-07-10T14:08:08", "description": "\"After investigating this problem a bit more, we found out how to get around the slowness we've been experiencing when using steamshovel in the offline-software meta-project. When we just run steamshovel, without passing the files in the command line, but opening them from the GUI, things work fine. When loading the files from command line, it goes very slowly.\"\n\nPossibly something different between cmd line loading and file menu loading.", "reporter": "david.schultz", "cc": "", "resolution": "fixed", "_ts": "1436537288260585", "component": "combo core", "summary": "steamshovel file open slowness", "priority": "normal", "keywords": "", "time": "2014-08-13T15:04:32", "milestone": "", "owner": "hdembinski", "type": "defect" } ```
defect
steamshovel file open slowness trac after investigating this problem a bit more we found out how to get around the slowness we ve been experiencing when using steamshovel in the offline software meta project when we just run steamshovel without passing the files in the command line but opening them from the gui things work fine when loading the files from command line it goes very slowly possibly something different between cmd line loading and file menu loading migrated from json status closed changetime description after investigating this problem a bit more we found out how to get around the slowness we ve been experiencing when using steamshovel in the offline software meta project when we just run steamshovel without passing the files in the command line but opening them from the gui things work fine when loading the files from command line it goes very slowly n npossibly something different between cmd line loading and file menu loading reporter david schultz cc resolution fixed ts component combo core summary steamshovel file open slowness priority normal keywords time milestone owner hdembinski type defect
1
21,020
3,442,784,529
IssuesEvent
2015-12-15 00:20:22
prettydiff/prettydiff
https://api.github.com/repos/prettydiff/prettydiff
closed
Conditions in tags get rearranged in twig templates
Defect QA
I've been using atom-beautify on twig templates, only to realize that it messed up some logic parts of the templates. When beautifying twig templates, conditions inside of tags get interpreted as attributes, changing **if [...] endif** to **endif [...] if**. ```twig <html> <body> <a href="/linktarget.html" {% if active %}class="active"{% endif %}>linktext</a> </body> </html> ``` turns into ```twig <html> <body> <a class="active" href="/linktarget.html" {% endif %} {% if active %}>linktext</a> </body> </html> ```
1.0
Conditions in tags get rearranged in twig templates - I've been using atom-beautify on twig templates, only to realize that it messed up some logic parts of the templates. When beautifying twig templates, conditions inside of tags get interpreted as attributes, changing **if [...] endif** to **endif [...] if**. ```twig <html> <body> <a href="/linktarget.html" {% if active %}class="active"{% endif %}>linktext</a> </body> </html> ``` turns into ```twig <html> <body> <a class="active" href="/linktarget.html" {% endif %} {% if active %}>linktext</a> </body> </html> ```
defect
conditions in tags get rearranged in twig templates i ve been using atom beautify on twig templates only to realize that it messed up some logic parts of the templates when beautifying twig templates conditions inside of tags get interpreted as attributes changing if endif to endif if twig linktext turns into twig linktext
1
290,241
32,045,665,693
IssuesEvent
2023-09-23 01:34:13
Chiencc/asuswrt-gt-ac5300
https://api.github.com/repos/Chiencc/asuswrt-gt-ac5300
reopened
jquery.mobile-1.4.5.min.js: 1 vulnerabilities (highest severity is: 6.5)
Mend: dependency security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery.mobile-1.4.5.min.js</b></p></summary> <p>Touch-Optimized Web Framework for Smartphones & Tablets</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery-mobile/1.4.5/jquery.mobile.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery-mobile/1.4.5/jquery.mobile.min.js</a></p> <p>Path to dependency file: /release/src/router/www/sysdep/VZW-AC1300/www/QIS_wizard.htm</p> <p>Path to vulnerable library: /release/src/router/www/sysdep/BLUECAVE/www/mobile/js/jquery.mobile.js,/release/src/router/www/sysdep/VZW-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC2200/www/js/jquery.mobile.js,/release/src/router/www/sysdep/BLUECAVE/www/./mobile/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1750/www/js/jquery.mobile.js</p> <p> <p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (jquery.mobile version) | Remediation Possible** | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [WS-2019-0136](https://github.com/jquery/jquery-mobile/commit/b0d9cc758a48f13321750d7409fb7655dcdf2b50) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.5 | jquery.mobile-1.4.5.min.js | Direct | BMC.NET - 1.0.3;org.webjars:jquery-mobile - 1.3.0-1,1.4.3;jquery.mobile - 1.3.0;jQWidgets_Framework - 8.0.0,6.0.6 | &#10060; | <p>**In some cases, Remediation PR cannot be created automatically for a vulnerability despite the availability of remediation</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> WS-2019-0136</summary> ### Vulnerable Library - <b>jquery.mobile-1.4.5.min.js</b></p> <p>Touch-Optimized Web Framework for Smartphones & Tablets</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery-mobile/1.4.5/jquery.mobile.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery-mobile/1.4.5/jquery.mobile.min.js</a></p> <p>Path to dependency file: /release/src/router/www/sysdep/VZW-AC1300/www/QIS_wizard.htm</p> <p>Path to vulnerable library: /release/src/router/www/sysdep/BLUECAVE/www/mobile/js/jquery.mobile.js,/release/src/router/www/sysdep/VZW-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC2200/www/js/jquery.mobile.js,/release/src/router/www/sysdep/BLUECAVE/www/./mobile/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1750/www/js/jquery.mobile.js</p> <p> Dependency Hierarchy: - :x: **jquery.mobile-1.4.5.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> All versions of Jquery mobile have an open redirect that leads to cross-site scripting when the endpoint reflects user input. <p>Publish Date: 2019-06-13 <p>URL: <a href=https://github.com/jquery/jquery-mobile/commit/b0d9cc758a48f13321750d7409fb7655dcdf2b50>WS-2019-0136</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/WS-2019-0136">https://nvd.nist.gov/vuln/detail/WS-2019-0136</a></p> <p>Release Date: 2019-06-13</p> <p>Fix Resolution: BMC.NET - 1.0.3;org.webjars:jquery-mobile - 1.3.0-1,1.4.3;jquery.mobile - 1.3.0;jQWidgets_Framework - 8.0.0,6.0.6</p> </p> <p></p> </details>
True
jquery.mobile-1.4.5.min.js: 1 vulnerabilities (highest severity is: 6.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery.mobile-1.4.5.min.js</b></p></summary> <p>Touch-Optimized Web Framework for Smartphones & Tablets</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery-mobile/1.4.5/jquery.mobile.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery-mobile/1.4.5/jquery.mobile.min.js</a></p> <p>Path to dependency file: /release/src/router/www/sysdep/VZW-AC1300/www/QIS_wizard.htm</p> <p>Path to vulnerable library: /release/src/router/www/sysdep/BLUECAVE/www/mobile/js/jquery.mobile.js,/release/src/router/www/sysdep/VZW-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC2200/www/js/jquery.mobile.js,/release/src/router/www/sysdep/BLUECAVE/www/./mobile/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1750/www/js/jquery.mobile.js</p> <p> <p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (jquery.mobile version) | Remediation Possible** | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [WS-2019-0136](https://github.com/jquery/jquery-mobile/commit/b0d9cc758a48f13321750d7409fb7655dcdf2b50) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.5 | jquery.mobile-1.4.5.min.js | Direct | BMC.NET - 1.0.3;org.webjars:jquery-mobile - 1.3.0-1,1.4.3;jquery.mobile - 1.3.0;jQWidgets_Framework - 8.0.0,6.0.6 | &#10060; | <p>**In some cases, Remediation PR cannot be created automatically for a vulnerability despite the availability of remediation</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> WS-2019-0136</summary> ### Vulnerable Library - <b>jquery.mobile-1.4.5.min.js</b></p> <p>Touch-Optimized Web Framework for Smartphones & Tablets</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery-mobile/1.4.5/jquery.mobile.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery-mobile/1.4.5/jquery.mobile.min.js</a></p> <p>Path to dependency file: /release/src/router/www/sysdep/VZW-AC1300/www/QIS_wizard.htm</p> <p>Path to vulnerable library: /release/src/router/www/sysdep/BLUECAVE/www/mobile/js/jquery.mobile.js,/release/src/router/www/sysdep/VZW-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC2200/www/js/jquery.mobile.js,/release/src/router/www/sysdep/BLUECAVE/www/./mobile/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1750/www/js/jquery.mobile.js</p> <p> Dependency Hierarchy: - :x: **jquery.mobile-1.4.5.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> All versions of Jquery mobile have an open redirect that leads to cross-site scripting when the endpoint reflects user input. <p>Publish Date: 2019-06-13 <p>URL: <a href=https://github.com/jquery/jquery-mobile/commit/b0d9cc758a48f13321750d7409fb7655dcdf2b50>WS-2019-0136</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/WS-2019-0136">https://nvd.nist.gov/vuln/detail/WS-2019-0136</a></p> <p>Release Date: 2019-06-13</p> <p>Fix Resolution: BMC.NET - 1.0.3;org.webjars:jquery-mobile - 1.3.0-1,1.4.3;jquery.mobile - 1.3.0;jQWidgets_Framework - 8.0.0,6.0.6</p> </p> <p></p> </details>
non_defect
jquery mobile min js vulnerabilities highest severity is vulnerable library jquery mobile min js touch optimized web framework for smartphones tablets library home page a href path to dependency file release src router www sysdep vzw www qis wizard htm path to vulnerable library release src router www sysdep bluecave www mobile js jquery mobile js release src router www sysdep vzw www js jquery mobile js release src router www sysdep map www js jquery mobile js release src router www sysdep map www js jquery mobile js release src router www sysdep bluecave www mobile js jquery mobile js release src router www sysdep map www js jquery mobile js release src router www sysdep map www js jquery mobile js found in head commit a href vulnerabilities cve severity cvss dependency type fixed in jquery mobile version remediation possible medium jquery mobile min js direct bmc net org webjars jquery mobile jquery mobile jqwidgets framework in some cases remediation pr cannot be created automatically for a vulnerability despite the availability of remediation details ws vulnerable library jquery mobile min js touch optimized web framework for smartphones tablets library home page a href path to dependency file release src router www sysdep vzw www qis wizard htm path to vulnerable library release src router www sysdep bluecave www mobile js jquery mobile js release src router www sysdep vzw www js jquery mobile js release src router www sysdep map www js jquery mobile js release src router www sysdep map www js jquery mobile js release src router www sysdep bluecave www mobile js jquery mobile js release src router www sysdep map www js jquery mobile js release src router www sysdep map www js jquery mobile js dependency hierarchy x jquery mobile min js vulnerable library found in head commit a href found in base branch master vulnerability details all versions of jquery mobile have an open redirect that leads to cross site scripting when the endpoint reflects user input publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bmc net org webjars jquery mobile jquery mobile jqwidgets framework
0
24,675
17,600,369,614
IssuesEvent
2021-08-17 11:04:21
google/web-stories-wp
https://api.github.com/repos/google/web-stories-wp
closed
Packages: Import React dependency from own React package
P2 Type: Infrastructure Pod: WP & Infra Package: React
<!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ --> ## Feature Description <!-- A clear and concise description of what the problem is and what you want to happen. --> We created our own dedicated `react` package in #8159. Now, we want to re-export `react` and `react-dom/server` functions from this package and make any package depend on this one. Similar to `@wordpress/element`. This makes it easier to maintain the React dependency and update it in the future. ## Alternatives Considered <!-- A clear and concise description of any alternative solutions or features you've considered. --> ## Additional Context <!-- Add any other context or screenshots about the feature request. --> --- _Do not alter or remove anything below. The following sections will be managed by moderators only._ ## Acceptance Criteria <!-- One or more bullet points for acceptance criteria. --> ## Implementation Brief <!-- One or more bullet points for how to technically implement the feature. -->
1.0
Packages: Import React dependency from own React package - <!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ --> ## Feature Description <!-- A clear and concise description of what the problem is and what you want to happen. --> We created our own dedicated `react` package in #8159. Now, we want to re-export `react` and `react-dom/server` functions from this package and make any package depend on this one. Similar to `@wordpress/element`. This makes it easier to maintain the React dependency and update it in the future. ## Alternatives Considered <!-- A clear and concise description of any alternative solutions or features you've considered. --> ## Additional Context <!-- Add any other context or screenshots about the feature request. --> --- _Do not alter or remove anything below. The following sections will be managed by moderators only._ ## Acceptance Criteria <!-- One or more bullet points for acceptance criteria. --> ## Implementation Brief <!-- One or more bullet points for how to technically implement the feature. -->
non_defect
packages import react dependency from own react package feature description we created our own dedicated react package in now we want to re export react and react dom server functions from this package and make any package depend on this one similar to wordpress element this makes it easier to maintain the react dependency and update it in the future alternatives considered additional context do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria implementation brief
0
72,881
31,769,581,331
IssuesEvent
2023-09-12 10:53:48
gauravrs18/issue_onboarding
https://api.github.com/repos/gauravrs18/issue_onboarding
closed
dev-angular-style-account-services-new-connection-component-approve-component -consumer-details-component -application-component -map-component
CX-account-services
dev-angular-style-account-services-new-connection-component-approve-component -consumer-details-component -application-component -map-component
1.0
dev-angular-style-account-services-new-connection-component-approve-component -consumer-details-component -application-component -map-component - dev-angular-style-account-services-new-connection-component-approve-component -consumer-details-component -application-component -map-component
non_defect
dev angular style account services new connection component approve component consumer details component application component map component dev angular style account services new connection component approve component consumer details component application component map component
0
15,552
5,143,603,622
IssuesEvent
2017-01-12 16:24:05
threadly/threadly
https://api.github.com/repos/threadly/threadly
closed
Perform actions to release version 4.9.0
non-code related
4.9.0 will be the last feature expansion release before 5.0.0 is released. This issue is to track the release process for 4.9.0. Once this issue is closed, it will indicate that 4.9.0 has been released, and is available. Items to update: - Homepage stable download - Homepage javadocs - Homepage news - Wiki homepage for java docs - Wiki page for complete list of javadocs links - Add release information to github Code operations: - Deploy artifact to sonatype - Tag git - Keep build forever in jenkins - rebase `unstable_staging` on top of `master`
1.0
Perform actions to release version 4.9.0 - 4.9.0 will be the last feature expansion release before 5.0.0 is released. This issue is to track the release process for 4.9.0. Once this issue is closed, it will indicate that 4.9.0 has been released, and is available. Items to update: - Homepage stable download - Homepage javadocs - Homepage news - Wiki homepage for java docs - Wiki page for complete list of javadocs links - Add release information to github Code operations: - Deploy artifact to sonatype - Tag git - Keep build forever in jenkins - rebase `unstable_staging` on top of `master`
non_defect
perform actions to release version will be the last feature expansion release before is released this issue is to track the release process for once this issue is closed it will indicate that has been released and is available items to update homepage stable download homepage javadocs homepage news wiki homepage for java docs wiki page for complete list of javadocs links add release information to github code operations deploy artifact to sonatype tag git keep build forever in jenkins rebase unstable staging on top of master
0
556,627
16,487,403,020
IssuesEvent
2021-05-24 20:13:02
mlflow/mlflow
https://api.github.com/repos/mlflow/mlflow
closed
[BUG] INTERNAL_SERVER_ERROR when sorting the runs table by metric or by date
area/tracking area/uiux bug priority/important-soon
### System information - **Have I written custom code (as opposed to using a stock example script provided in MLflow)**: no - **OS Platform and Distribution (e.g., Linux Ubuntu 16.04)**: Linux Debian - **MLflow installed from (source or binary)**: binary - **MLflow version (run ``mlflow --version``)**: 1.7.0 - **Python version**: 3.7 - **npm version, if running the dev UI**: - - **Exact command to reproduce**: run the mlflow server and log parameters and metrics ### Describe the problem When I try to sort the runs by any metric or date (parameters, versions and users work) I get an `INTERNAL_SERVER_ERROR` and the whole table disappears. Then, if I sort by a parameter, for example, the table comes back. **UPDATE:** ![V4-Model-MLflow-Experiment-Google-Chrome-2020-04-07-10-18-17-_online-video-cutter com_](https://user-images.githubusercontent.com/6637618/78646910-10d70a80-78ba-11ea-82ea-a4678a3337fe.gif) ### Code to reproduce issue run the mlflow server and log parameters and metrics ### Other info / logs This is the trace-back I got from the server: ``` 2020/03/04 09:40:50 ERROR mlflow.server: Exception on /ajax-api/2.0/preview/mlflow/runs/search [POST] Traceback (most recent call last): File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 1246, in _execute_context cursor, statement, parameters, context File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/default.py", line 588, in do_execute cursor.execute(statement, parameters) pyodbc.ProgrammingError: ('42000', "[42000] [Microsoft][ODBC Driver 17 for SQL Server][SQL Server]Incorrect syntax near '1'. (102) (SQLExecDirectW)") The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/usr/local/lib/python3.7/site-packages/mlflow/store/db/utils.py", line 73, in make_managed_session yield session File "/usr/local/lib/python3.7/site-packages/mlflow/store/tracking/sqlalchemy_store.py", line 651, in _search_runs .offset(offset).limit(max_results).all() File "/usr/local/lib/python3.7/site-packages/sqlalchemy/orm/query.py", line 3233, in all return list(self) File "/usr/local/lib/python3.7/site-packages/sqlalchemy/orm/query.py", line 3389, in __iter__ return self._execute_and_instances(context) File "/usr/local/lib/python3.7/site-packages/sqlalchemy/orm/query.py", line 3414, in _execute_and_instances result = conn.execute(querycontext.statement, self._params) File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 982, in execute return meth(self, multiparams, params) File "/usr/local/lib/python3.7/site-packages/sqlalchemy/sql/elements.py", line 293, in _execute_on_connection return connection._execute_clauseelement(self, multiparams, params) File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 1101, in _execute_clauseelement distilled_params, File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 1250, in _execute_context e, statement, parameters, cursor, context File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 1476, in _handle_dbapi_exception util.raise_from_cause(sqlalchemy_exception, exc_info) File "/usr/local/lib/python3.7/site-packages/sqlalchemy/util/compat.py", line 398, in raise_from_cause reraise(type(exception), exception, tb=exc_tb, cause=cause) File "/usr/local/lib/python3.7/site-packages/sqlalchemy/util/compat.py", line 152, in reraise raise value.with_traceback(tb) File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 1246, in _execute_context cursor, statement, parameters, context File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/default.py", line 588, in do_execute cursor.execute(statement, parameters) sqlalchemy.exc.ProgrammingError: (pyodbc.ProgrammingError) ('42000', "[42000] [Microsoft][ODBC Driver 17 for SQL Server][SQL Server]Incorrect syntax near '1'. (102) (SQLExecDirectW)") [SQL: SELECT DISTINCT TOP 100 runs.run_uuid AS runs_run_uuid, runs.name AS runs_name, runs.source_type AS runs_source_type, runs.source_name AS runs_source_name, runs.entry_point_name AS runs_entry_point_name, runs.user_id AS runs_user_id, runs.status AS runs_status, runs.start_time AS runs_start_time, runs.end_time AS runs_end_time, runs.source_version AS runs_source_version, runs.lifecycle_stage AS runs_lifecycle_stage, runs.artifact_uri AS runs_artifact_uri, runs.experiment_id AS runs_experiment_id, CASE WHEN (anon_1.is_nan IS 1) THEN ? WHEN (anon_1.value IS NULL) THEN ? ELSE ? END AS clause_1, anon_1.value AS anon_1_value FROM runs LEFT OUTER JOIN (SELECT latest_metrics.[key] AS [key], latest_metrics.value AS value, latest_metrics.timestamp AS timestamp, latest_metrics.step AS step, latest_metrics.is_nan AS is_nan, latest_metrics.run_uuid AS run_uuid FROM latest_metrics WHERE latest_metrics.[key] = ?) AS anon_1 ON runs.run_uuid = anon_1.run_uuid WHERE runs.experiment_id IN (?) AND runs.lifecycle_stage IN (?) ORDER BY clause_1, anon_1.value, runs.start_time DESC, runs.run_uuid] [parameters: (1, 1, 0, 'auc', '4', 'active')] (Background on this error at: http://sqlalche.me/e/f405) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.7/site-packages/mlflow/server/handlers.py", line 155, in wrapper return func(*args, **kwargs) File "/usr/local/lib/python3.7/site-packages/mlflow/server/handlers.py", line 394, in _search_runs max_results, order_by, page_token) File "/usr/local/lib/python3.7/site-packages/mlflow/store/tracking/abstract_store.py", line 230, in search_runs order_by, page_token) File "/usr/local/lib/python3.7/site-packages/mlflow/store/tracking/sqlalchemy_store.py", line 654, in _search_runs next_page_token = compute_next_token(len(runs)) File "/usr/local/lib/python3.7/contextlib.py", line 130, in __exit__ self.gen.throw(type, value, traceback) File "/usr/local/lib/python3.7/site-packages/mlflow/store/db/utils.py", line 80, in make_managed_session raise MlflowException(message=e, error_code=INTERNAL_ERROR) mlflow.exceptions.MlflowException: (pyodbc.ProgrammingError) ('42000', "[42000] [Microsoft][ODBC Driver 17 for SQL Server][SQL Server]Incorrect syntax near '1'. (102) (SQLExecDirectW)") [SQL: SELECT DISTINCT TOP 100 runs.run_uuid AS runs_run_uuid, runs.name AS runs_name, runs.source_type AS runs_source_type, runs.source_name AS runs_source_name, runs.entry_point_name AS runs_entry_point_name, runs.user_id AS runs_user_id, runs.status AS runs_status, runs.start_time AS runs_start_time, runs.end_time AS runs_end_time, runs.source_version AS runs_source_version, runs.lifecycle_stage AS runs_lifecycle_stage, runs.artifact_uri AS runs_artifact_uri, runs.experiment_id AS runs_experiment_id, CASE WHEN (anon_1.is_nan IS 1) THEN ? WHEN (anon_1.value IS NULL) THEN ? ELSE ? END AS clause_1, anon_1.value AS anon_1_value FROM runs LEFT OUTER JOIN (SELECT latest_metrics.[key] AS [key], latest_metrics.value AS value, latest_metrics.timestamp AS timestamp, latest_metrics.step AS step, latest_metrics.is_nan AS is_nan, latest_metrics.run_uuid AS run_uuid FROM latest_metrics WHERE latest_metrics.[key] = ?) AS anon_1 ON runs.run_uuid = anon_1.run_uuid WHERE runs.experiment_id IN (?) AND runs.lifecycle_stage IN (?) ORDER BY clause_1, anon_1.value, runs.start_time DESC, runs.run_uuid] [parameters: (1, 1, 0, 'auc', '4', 'active')] (Background on this error at: http://sqlalche.me/e/f405) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.7/site-packages/flask/app.py", line 2446, in wsgi_app response = self.full_dispatch_request() File "/usr/local/lib/python3.7/site-packages/flask/app.py", line 1951, in full_dispatch_request rv = self.handle_user_exception(e) File "/usr/local/lib/python3.7/site-packages/flask/app.py", line 1820, in handle_user_exception reraise(exc_type, exc_value, tb) File "/usr/local/lib/python3.7/site-packages/flask/_compat.py", line 39, in reraise raise value File "/usr/local/lib/python3.7/site-packages/flask/app.py", line 1949, in full_dispatch_request rv = self.dispatch_request() File "/usr/local/lib/python3.7/site-packages/flask/app.py", line 1935, in dispatch_request return self.view_functions[rule.endpoint](**req.view_args) File "/usr/local/lib/python3.7/site-packages/mlflow/server/handlers.py", line 158, in wrapper response.set_data(e.serialize_as_json()) File "/usr/local/lib/python3.7/site-packages/mlflow/exceptions.py", line 50, in serialize_as_json return json.dumps(exception_dict) File "/usr/local/lib/python3.7/json/__init__.py", line 231, in dumps return _default_encoder.encode(obj) File "/usr/local/lib/python3.7/json/encoder.py", line 199, in encode chunks = self.iterencode(o, _one_shot=True) File "/usr/local/lib/python3.7/json/encoder.py", line 257, in iterencode return _iterencode(o, 0) File "/usr/local/lib/python3.7/json/encoder.py", line 179, in default raise TypeError(f'Object of type {o.__class__.__name__} ' TypeError: Object of type ProgrammingError is not JSON serializable ``` ### **UPDATE** I've updated the version to v1.7.2 and the bug is still there.
1.0
[BUG] INTERNAL_SERVER_ERROR when sorting the runs table by metric or by date - ### System information - **Have I written custom code (as opposed to using a stock example script provided in MLflow)**: no - **OS Platform and Distribution (e.g., Linux Ubuntu 16.04)**: Linux Debian - **MLflow installed from (source or binary)**: binary - **MLflow version (run ``mlflow --version``)**: 1.7.0 - **Python version**: 3.7 - **npm version, if running the dev UI**: - - **Exact command to reproduce**: run the mlflow server and log parameters and metrics ### Describe the problem When I try to sort the runs by any metric or date (parameters, versions and users work) I get an `INTERNAL_SERVER_ERROR` and the whole table disappears. Then, if I sort by a parameter, for example, the table comes back. **UPDATE:** ![V4-Model-MLflow-Experiment-Google-Chrome-2020-04-07-10-18-17-_online-video-cutter com_](https://user-images.githubusercontent.com/6637618/78646910-10d70a80-78ba-11ea-82ea-a4678a3337fe.gif) ### Code to reproduce issue run the mlflow server and log parameters and metrics ### Other info / logs This is the trace-back I got from the server: ``` 2020/03/04 09:40:50 ERROR mlflow.server: Exception on /ajax-api/2.0/preview/mlflow/runs/search [POST] Traceback (most recent call last): File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 1246, in _execute_context cursor, statement, parameters, context File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/default.py", line 588, in do_execute cursor.execute(statement, parameters) pyodbc.ProgrammingError: ('42000', "[42000] [Microsoft][ODBC Driver 17 for SQL Server][SQL Server]Incorrect syntax near '1'. (102) (SQLExecDirectW)") The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/usr/local/lib/python3.7/site-packages/mlflow/store/db/utils.py", line 73, in make_managed_session yield session File "/usr/local/lib/python3.7/site-packages/mlflow/store/tracking/sqlalchemy_store.py", line 651, in _search_runs .offset(offset).limit(max_results).all() File "/usr/local/lib/python3.7/site-packages/sqlalchemy/orm/query.py", line 3233, in all return list(self) File "/usr/local/lib/python3.7/site-packages/sqlalchemy/orm/query.py", line 3389, in __iter__ return self._execute_and_instances(context) File "/usr/local/lib/python3.7/site-packages/sqlalchemy/orm/query.py", line 3414, in _execute_and_instances result = conn.execute(querycontext.statement, self._params) File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 982, in execute return meth(self, multiparams, params) File "/usr/local/lib/python3.7/site-packages/sqlalchemy/sql/elements.py", line 293, in _execute_on_connection return connection._execute_clauseelement(self, multiparams, params) File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 1101, in _execute_clauseelement distilled_params, File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 1250, in _execute_context e, statement, parameters, cursor, context File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 1476, in _handle_dbapi_exception util.raise_from_cause(sqlalchemy_exception, exc_info) File "/usr/local/lib/python3.7/site-packages/sqlalchemy/util/compat.py", line 398, in raise_from_cause reraise(type(exception), exception, tb=exc_tb, cause=cause) File "/usr/local/lib/python3.7/site-packages/sqlalchemy/util/compat.py", line 152, in reraise raise value.with_traceback(tb) File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 1246, in _execute_context cursor, statement, parameters, context File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/default.py", line 588, in do_execute cursor.execute(statement, parameters) sqlalchemy.exc.ProgrammingError: (pyodbc.ProgrammingError) ('42000', "[42000] [Microsoft][ODBC Driver 17 for SQL Server][SQL Server]Incorrect syntax near '1'. (102) (SQLExecDirectW)") [SQL: SELECT DISTINCT TOP 100 runs.run_uuid AS runs_run_uuid, runs.name AS runs_name, runs.source_type AS runs_source_type, runs.source_name AS runs_source_name, runs.entry_point_name AS runs_entry_point_name, runs.user_id AS runs_user_id, runs.status AS runs_status, runs.start_time AS runs_start_time, runs.end_time AS runs_end_time, runs.source_version AS runs_source_version, runs.lifecycle_stage AS runs_lifecycle_stage, runs.artifact_uri AS runs_artifact_uri, runs.experiment_id AS runs_experiment_id, CASE WHEN (anon_1.is_nan IS 1) THEN ? WHEN (anon_1.value IS NULL) THEN ? ELSE ? END AS clause_1, anon_1.value AS anon_1_value FROM runs LEFT OUTER JOIN (SELECT latest_metrics.[key] AS [key], latest_metrics.value AS value, latest_metrics.timestamp AS timestamp, latest_metrics.step AS step, latest_metrics.is_nan AS is_nan, latest_metrics.run_uuid AS run_uuid FROM latest_metrics WHERE latest_metrics.[key] = ?) AS anon_1 ON runs.run_uuid = anon_1.run_uuid WHERE runs.experiment_id IN (?) AND runs.lifecycle_stage IN (?) ORDER BY clause_1, anon_1.value, runs.start_time DESC, runs.run_uuid] [parameters: (1, 1, 0, 'auc', '4', 'active')] (Background on this error at: http://sqlalche.me/e/f405) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.7/site-packages/mlflow/server/handlers.py", line 155, in wrapper return func(*args, **kwargs) File "/usr/local/lib/python3.7/site-packages/mlflow/server/handlers.py", line 394, in _search_runs max_results, order_by, page_token) File "/usr/local/lib/python3.7/site-packages/mlflow/store/tracking/abstract_store.py", line 230, in search_runs order_by, page_token) File "/usr/local/lib/python3.7/site-packages/mlflow/store/tracking/sqlalchemy_store.py", line 654, in _search_runs next_page_token = compute_next_token(len(runs)) File "/usr/local/lib/python3.7/contextlib.py", line 130, in __exit__ self.gen.throw(type, value, traceback) File "/usr/local/lib/python3.7/site-packages/mlflow/store/db/utils.py", line 80, in make_managed_session raise MlflowException(message=e, error_code=INTERNAL_ERROR) mlflow.exceptions.MlflowException: (pyodbc.ProgrammingError) ('42000', "[42000] [Microsoft][ODBC Driver 17 for SQL Server][SQL Server]Incorrect syntax near '1'. (102) (SQLExecDirectW)") [SQL: SELECT DISTINCT TOP 100 runs.run_uuid AS runs_run_uuid, runs.name AS runs_name, runs.source_type AS runs_source_type, runs.source_name AS runs_source_name, runs.entry_point_name AS runs_entry_point_name, runs.user_id AS runs_user_id, runs.status AS runs_status, runs.start_time AS runs_start_time, runs.end_time AS runs_end_time, runs.source_version AS runs_source_version, runs.lifecycle_stage AS runs_lifecycle_stage, runs.artifact_uri AS runs_artifact_uri, runs.experiment_id AS runs_experiment_id, CASE WHEN (anon_1.is_nan IS 1) THEN ? WHEN (anon_1.value IS NULL) THEN ? ELSE ? END AS clause_1, anon_1.value AS anon_1_value FROM runs LEFT OUTER JOIN (SELECT latest_metrics.[key] AS [key], latest_metrics.value AS value, latest_metrics.timestamp AS timestamp, latest_metrics.step AS step, latest_metrics.is_nan AS is_nan, latest_metrics.run_uuid AS run_uuid FROM latest_metrics WHERE latest_metrics.[key] = ?) AS anon_1 ON runs.run_uuid = anon_1.run_uuid WHERE runs.experiment_id IN (?) AND runs.lifecycle_stage IN (?) ORDER BY clause_1, anon_1.value, runs.start_time DESC, runs.run_uuid] [parameters: (1, 1, 0, 'auc', '4', 'active')] (Background on this error at: http://sqlalche.me/e/f405) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.7/site-packages/flask/app.py", line 2446, in wsgi_app response = self.full_dispatch_request() File "/usr/local/lib/python3.7/site-packages/flask/app.py", line 1951, in full_dispatch_request rv = self.handle_user_exception(e) File "/usr/local/lib/python3.7/site-packages/flask/app.py", line 1820, in handle_user_exception reraise(exc_type, exc_value, tb) File "/usr/local/lib/python3.7/site-packages/flask/_compat.py", line 39, in reraise raise value File "/usr/local/lib/python3.7/site-packages/flask/app.py", line 1949, in full_dispatch_request rv = self.dispatch_request() File "/usr/local/lib/python3.7/site-packages/flask/app.py", line 1935, in dispatch_request return self.view_functions[rule.endpoint](**req.view_args) File "/usr/local/lib/python3.7/site-packages/mlflow/server/handlers.py", line 158, in wrapper response.set_data(e.serialize_as_json()) File "/usr/local/lib/python3.7/site-packages/mlflow/exceptions.py", line 50, in serialize_as_json return json.dumps(exception_dict) File "/usr/local/lib/python3.7/json/__init__.py", line 231, in dumps return _default_encoder.encode(obj) File "/usr/local/lib/python3.7/json/encoder.py", line 199, in encode chunks = self.iterencode(o, _one_shot=True) File "/usr/local/lib/python3.7/json/encoder.py", line 257, in iterencode return _iterencode(o, 0) File "/usr/local/lib/python3.7/json/encoder.py", line 179, in default raise TypeError(f'Object of type {o.__class__.__name__} ' TypeError: Object of type ProgrammingError is not JSON serializable ``` ### **UPDATE** I've updated the version to v1.7.2 and the bug is still there.
non_defect
internal server error when sorting the runs table by metric or by date system information have i written custom code as opposed to using a stock example script provided in mlflow no os platform and distribution e g linux ubuntu linux debian mlflow installed from source or binary binary mlflow version run mlflow version python version npm version if running the dev ui exact command to reproduce run the mlflow server and log parameters and metrics describe the problem when i try to sort the runs by any metric or date parameters versions and users work i get an internal server error and the whole table disappears then if i sort by a parameter for example the table comes back update code to reproduce issue run the mlflow server and log parameters and metrics other info logs this is the trace back i got from the server error mlflow server exception on ajax api preview mlflow runs search traceback most recent call last file usr local lib site packages sqlalchemy engine base py line in execute context cursor statement parameters context file usr local lib site packages sqlalchemy engine default py line in do execute cursor execute statement parameters pyodbc programmingerror incorrect syntax near sqlexecdirectw the above exception was the direct cause of the following exception traceback most recent call last file usr local lib site packages mlflow store db utils py line in make managed session yield session file usr local lib site packages mlflow store tracking sqlalchemy store py line in search runs offset offset limit max results all file usr local lib site packages sqlalchemy orm query py line in all return list self file usr local lib site packages sqlalchemy orm query py line in iter return self execute and instances context file usr local lib site packages sqlalchemy orm query py line in execute and instances result conn execute querycontext statement self params file usr local lib site packages sqlalchemy engine base py line in execute return meth self multiparams params file usr local lib site packages sqlalchemy sql elements py line in execute on connection return connection execute clauseelement self multiparams params file usr local lib site packages sqlalchemy engine base py line in execute clauseelement distilled params file usr local lib site packages sqlalchemy engine base py line in execute context e statement parameters cursor context file usr local lib site packages sqlalchemy engine base py line in handle dbapi exception util raise from cause sqlalchemy exception exc info file usr local lib site packages sqlalchemy util compat py line in raise from cause reraise type exception exception tb exc tb cause cause file usr local lib site packages sqlalchemy util compat py line in reraise raise value with traceback tb file usr local lib site packages sqlalchemy engine base py line in execute context cursor statement parameters context file usr local lib site packages sqlalchemy engine default py line in do execute cursor execute statement parameters sqlalchemy exc programmingerror pyodbc programmingerror incorrect syntax near sqlexecdirectw sql select distinct top runs run uuid as runs run uuid runs name as runs name runs source type as runs source type runs source name as runs source name runs entry point name as runs entry point name runs user id as runs user id runs status as runs status runs start time as runs start time runs end time as runs end time runs source version as runs source version runs lifecycle stage as runs lifecycle stage runs artifact uri as runs artifact uri runs experiment id as runs experiment id case when anon is nan is then when anon value is null then else end as clause anon value as anon value from runs left outer join select latest metrics as latest metrics value as value latest metrics timestamp as timestamp latest metrics step as step latest metrics is nan as is nan latest metrics run uuid as run uuid from latest metrics where latest metrics as anon on runs run uuid anon run uuid where runs experiment id in and runs lifecycle stage in order by clause anon value runs start time desc runs run uuid background on this error at during handling of the above exception another exception occurred traceback most recent call last file usr local lib site packages mlflow server handlers py line in wrapper return func args kwargs file usr local lib site packages mlflow server handlers py line in search runs max results order by page token file usr local lib site packages mlflow store tracking abstract store py line in search runs order by page token file usr local lib site packages mlflow store tracking sqlalchemy store py line in search runs next page token compute next token len runs file usr local lib contextlib py line in exit self gen throw type value traceback file usr local lib site packages mlflow store db utils py line in make managed session raise mlflowexception message e error code internal error mlflow exceptions mlflowexception pyodbc programmingerror incorrect syntax near sqlexecdirectw sql select distinct top runs run uuid as runs run uuid runs name as runs name runs source type as runs source type runs source name as runs source name runs entry point name as runs entry point name runs user id as runs user id runs status as runs status runs start time as runs start time runs end time as runs end time runs source version as runs source version runs lifecycle stage as runs lifecycle stage runs artifact uri as runs artifact uri runs experiment id as runs experiment id case when anon is nan is then when anon value is null then else end as clause anon value as anon value from runs left outer join select latest metrics as latest metrics value as value latest metrics timestamp as timestamp latest metrics step as step latest metrics is nan as is nan latest metrics run uuid as run uuid from latest metrics where latest metrics as anon on runs run uuid anon run uuid where runs experiment id in and runs lifecycle stage in order by clause anon value runs start time desc runs run uuid background on this error at during handling of the above exception another exception occurred traceback most recent call last file usr local lib site packages flask app py line in wsgi app response self full dispatch request file usr local lib site packages flask app py line in full dispatch request rv self handle user exception e file usr local lib site packages flask app py line in handle user exception reraise exc type exc value tb file usr local lib site packages flask compat py line in reraise raise value file usr local lib site packages flask app py line in full dispatch request rv self dispatch request file usr local lib site packages flask app py line in dispatch request return self view functions req view args file usr local lib site packages mlflow server handlers py line in wrapper response set data e serialize as json file usr local lib site packages mlflow exceptions py line in serialize as json return json dumps exception dict file usr local lib json init py line in dumps return default encoder encode obj file usr local lib json encoder py line in encode chunks self iterencode o one shot true file usr local lib json encoder py line in iterencode return iterencode o file usr local lib json encoder py line in default raise typeerror f object of type o class name typeerror object of type programmingerror is not json serializable update i ve updated the version to and the bug is still there
0
95,111
3,934,256,902
IssuesEvent
2016-04-25 21:56:59
ceylon/ceylon-ide-eclipse
https://api.github.com/repos/ceylon/ceylon-ide-eclipse
closed
initializer annotation and highlight range are completely broken
bug on last release bug on master high priority
`AdditionalAnnotationCreator` is busted :-(
1.0
initializer annotation and highlight range are completely broken - `AdditionalAnnotationCreator` is busted :-(
non_defect
initializer annotation and highlight range are completely broken additionalannotationcreator is busted
0
9,812
2,615,175,313
IssuesEvent
2015-03-01 06:58:36
chrsmith/reaver-wps
https://api.github.com/repos/chrsmith/reaver-wps
opened
1.4 reaver issue with AWUS036H.
auto-migrated Priority-Triage Type-Defect
``` I have gone through just about ever post out there about wireless not working after 2 minutes etc. Here are my findings from the past week of fiddling around with this. I have installed two AWUS036H usb cards on the same laptop running backtrack 5 r3 (i have also tested this with Kali linux 1.0.7 with same results) Hard drive and live cd. all the findings are exactly the same. all of the operating systems have 1.4 reaver on them. My problem is, when i start using reaver 1.4 within couple minutes my awus036h card rtl8781L along with all the wireless cards stop working. I know its difficult to imagine. The only real work around i have found is when reaver comes up with error Warning! Received timeout occured (then reaver hangs and stops working) is to unplug the card from usb, plug it back in and put it in monitor mode then continue the attack until it locks up again. Things that i have tried: 1) rule out all software conflicts, did airmon-ng start (wlan0 and wlan1 cards which are both my awus036h cards).. kill all possible conflicts.. run airodump-ng on one of the cards while keeping the other card on another terminal window so i have two terminals running airdump while on one of those runs reaver as well. when reaver locks up, both of the monitored interfaces go down and by going down i mean all input like packet, data, and all that goes flat dead. there is no software conflict. i have ran iwconfig before and after reaver takes a shit and no configuration has been changed with interfaces. nothing got switched between managed or monitored modes, there is plenty of power to both cards, i have used the Y adapter cables to input additional power to the cards so thats not the issue as well. I should also mention i have an internal wireless card that is intel something.. which still continued to run airdump after reaver has taken out my awus036h card driver. (while both of the cards are seased, i switched reaver to my intel wireless card mon3 interface (actual adapter is wlan3) it locked that up too! within 2 mins use. I have pretty much done all i can but it seems reaver 1.4 triggers some kind of glitch when used with awus036h card which brings down all the wireless interfaces no matter that they are if you point reaver to the divice.. it will bring it down until you physically unplug it and put it back in OR reboot. then you have another 1-2 minutes before it locks up again. I have not had any issues using awus036hA the aerthos chipset that keeps running just fine. but rtl8717l is another story. oh and also i have completely uninstalled dhcp client, disabled network monitor, suplements, configured /network/interface so none of any software causes any kind of conflit or even looks at the wireless interfaces. is there anything else i've missed? and no i have not been able to find an actual solution to this problem although i have read a lot of posts and problems may sound similar but not in their entirety. yes i have tried every single one of the suggested fixes and ran out of options but to stock up on new wireless cards. any suggestions? ``` Original issue reported on code.google.com by `berezini...@gmail.com` on 9 Jul 2014 at 9:50
1.0
1.4 reaver issue with AWUS036H. - ``` I have gone through just about ever post out there about wireless not working after 2 minutes etc. Here are my findings from the past week of fiddling around with this. I have installed two AWUS036H usb cards on the same laptop running backtrack 5 r3 (i have also tested this with Kali linux 1.0.7 with same results) Hard drive and live cd. all the findings are exactly the same. all of the operating systems have 1.4 reaver on them. My problem is, when i start using reaver 1.4 within couple minutes my awus036h card rtl8781L along with all the wireless cards stop working. I know its difficult to imagine. The only real work around i have found is when reaver comes up with error Warning! Received timeout occured (then reaver hangs and stops working) is to unplug the card from usb, plug it back in and put it in monitor mode then continue the attack until it locks up again. Things that i have tried: 1) rule out all software conflicts, did airmon-ng start (wlan0 and wlan1 cards which are both my awus036h cards).. kill all possible conflicts.. run airodump-ng on one of the cards while keeping the other card on another terminal window so i have two terminals running airdump while on one of those runs reaver as well. when reaver locks up, both of the monitored interfaces go down and by going down i mean all input like packet, data, and all that goes flat dead. there is no software conflict. i have ran iwconfig before and after reaver takes a shit and no configuration has been changed with interfaces. nothing got switched between managed or monitored modes, there is plenty of power to both cards, i have used the Y adapter cables to input additional power to the cards so thats not the issue as well. I should also mention i have an internal wireless card that is intel something.. which still continued to run airdump after reaver has taken out my awus036h card driver. (while both of the cards are seased, i switched reaver to my intel wireless card mon3 interface (actual adapter is wlan3) it locked that up too! within 2 mins use. I have pretty much done all i can but it seems reaver 1.4 triggers some kind of glitch when used with awus036h card which brings down all the wireless interfaces no matter that they are if you point reaver to the divice.. it will bring it down until you physically unplug it and put it back in OR reboot. then you have another 1-2 minutes before it locks up again. I have not had any issues using awus036hA the aerthos chipset that keeps running just fine. but rtl8717l is another story. oh and also i have completely uninstalled dhcp client, disabled network monitor, suplements, configured /network/interface so none of any software causes any kind of conflit or even looks at the wireless interfaces. is there anything else i've missed? and no i have not been able to find an actual solution to this problem although i have read a lot of posts and problems may sound similar but not in their entirety. yes i have tried every single one of the suggested fixes and ran out of options but to stock up on new wireless cards. any suggestions? ``` Original issue reported on code.google.com by `berezini...@gmail.com` on 9 Jul 2014 at 9:50
defect
reaver issue with i have gone through just about ever post out there about wireless not working after minutes etc here are my findings from the past week of fiddling around with this i have installed two usb cards on the same laptop running backtrack i have also tested this with kali linux with same results hard drive and live cd all the findings are exactly the same all of the operating systems have reaver on them my problem is when i start using reaver within couple minutes my card along with all the wireless cards stop working i know its difficult to imagine the only real work around i have found is when reaver comes up with error warning received timeout occured then reaver hangs and stops working is to unplug the card from usb plug it back in and put it in monitor mode then continue the attack until it locks up again things that i have tried rule out all software conflicts did airmon ng start and cards which are both my cards kill all possible conflicts run airodump ng on one of the cards while keeping the other card on another terminal window so i have two terminals running airdump while on one of those runs reaver as well when reaver locks up both of the monitored interfaces go down and by going down i mean all input like packet data and all that goes flat dead there is no software conflict i have ran iwconfig before and after reaver takes a shit and no configuration has been changed with interfaces nothing got switched between managed or monitored modes there is plenty of power to both cards i have used the y adapter cables to input additional power to the cards so thats not the issue as well i should also mention i have an internal wireless card that is intel something which still continued to run airdump after reaver has taken out my card driver while both of the cards are seased i switched reaver to my intel wireless card interface actual adapter is it locked that up too within mins use i have pretty much done all i can but it seems reaver triggers some kind of glitch when used with card which brings down all the wireless interfaces no matter that they are if you point reaver to the divice it will bring it down until you physically unplug it and put it back in or reboot then you have another minutes before it locks up again i have not had any issues using the aerthos chipset that keeps running just fine but is another story oh and also i have completely uninstalled dhcp client disabled network monitor suplements configured network interface so none of any software causes any kind of conflit or even looks at the wireless interfaces is there anything else i ve missed and no i have not been able to find an actual solution to this problem although i have read a lot of posts and problems may sound similar but not in their entirety yes i have tried every single one of the suggested fixes and ran out of options but to stock up on new wireless cards any suggestions original issue reported on code google com by berezini gmail com on jul at
1
48,476
13,094,936,504
IssuesEvent
2020-08-03 13:18:59
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
CacheSimpleConfig should accept cache name in its constructors
Module: ICache Source: Community Team: Core Type: Defect
Other configs - such as `MapConfig` or `QueueConfig` have a constructor with a name. This allow to use constructs such as: `config.addMapConfig(new MapConfig("myMap").setFoo(foo));` CacheSimpleConfig does not have this constructor. It should have it for the same of consistency.
1.0
CacheSimpleConfig should accept cache name in its constructors - Other configs - such as `MapConfig` or `QueueConfig` have a constructor with a name. This allow to use constructs such as: `config.addMapConfig(new MapConfig("myMap").setFoo(foo));` CacheSimpleConfig does not have this constructor. It should have it for the same of consistency.
defect
cachesimpleconfig should accept cache name in its constructors other configs such as mapconfig or queueconfig have a constructor with a name this allow to use constructs such as config addmapconfig new mapconfig mymap setfoo foo cachesimpleconfig does not have this constructor it should have it for the same of consistency
1
18,118
3,024,309,808
IssuesEvent
2015-08-02 13:46:35
aayush93/xbt
https://api.github.com/repos/aayush93/xbt
closed
IPv6 time
auto-migrated Priority-Medium Type-Defect
``` How about IPv6 support ? ``` Original issue reported on code.google.com by `allfe...@yandex.ru` on 21 Nov 2012 at 6:47
1.0
IPv6 time - ``` How about IPv6 support ? ``` Original issue reported on code.google.com by `allfe...@yandex.ru` on 21 Nov 2012 at 6:47
defect
time how about support original issue reported on code google com by allfe yandex ru on nov at
1
64,294
18,397,282,207
IssuesEvent
2021-10-12 12:50:02
vector-im/element-android
https://api.github.com/repos/vector-im/element-android
closed
Emails are not listed in the screen notification by email
T-Defect feature:settings
At least an email is associated to the account - Clear cache from Settings/General clear cache - Go to settings/notifications - Observe that the section "Email notification" displays the "no email" placeholder - Click on this placeholder - The emails are loaded - Go back - The emails are display and user can enable Email notification I think we have to force a refresh of the 3Pid as per https://github.com/vector-im/element-android/blob/develop/vector/src/main/java/im/vector/app/features/settings/threepids/ThreePidsSettingsViewModel.kt#L106 and this is not done in https://github.com/vector-im/element-android/blob/develop/vector/src/main/java/im/vector/app/features/settings/notifications/VectorSettingsNotificationPreferenceFragment.kt#L424 - [ ] So I think this is just about changing the parameter in https://github.com/vector-im/element-android/blob/develop/vector/src/main/java/im/vector/app/features/settings/notifications/VectorSettingsNotificationPreferenceFragment.kt#L424 - [ ] Also the pushers has to be refreshed, if configured on another session of the same user, the displayed information are not correct when entering the screen (user has to go to the debug screen about pushers to get the latest pushers, and then the information will be correct.
1.0
Emails are not listed in the screen notification by email - At least an email is associated to the account - Clear cache from Settings/General clear cache - Go to settings/notifications - Observe that the section "Email notification" displays the "no email" placeholder - Click on this placeholder - The emails are loaded - Go back - The emails are display and user can enable Email notification I think we have to force a refresh of the 3Pid as per https://github.com/vector-im/element-android/blob/develop/vector/src/main/java/im/vector/app/features/settings/threepids/ThreePidsSettingsViewModel.kt#L106 and this is not done in https://github.com/vector-im/element-android/blob/develop/vector/src/main/java/im/vector/app/features/settings/notifications/VectorSettingsNotificationPreferenceFragment.kt#L424 - [ ] So I think this is just about changing the parameter in https://github.com/vector-im/element-android/blob/develop/vector/src/main/java/im/vector/app/features/settings/notifications/VectorSettingsNotificationPreferenceFragment.kt#L424 - [ ] Also the pushers has to be refreshed, if configured on another session of the same user, the displayed information are not correct when entering the screen (user has to go to the debug screen about pushers to get the latest pushers, and then the information will be correct.
defect
emails are not listed in the screen notification by email at least an email is associated to the account clear cache from settings general clear cache go to settings notifications observe that the section email notification displays the no email placeholder click on this placeholder the emails are loaded go back the emails are display and user can enable email notification i think we have to force a refresh of the as per and this is not done in so i think this is just about changing the parameter in also the pushers has to be refreshed if configured on another session of the same user the displayed information are not correct when entering the screen user has to go to the debug screen about pushers to get the latest pushers and then the information will be correct
1
223,087
17,103,014,843
IssuesEvent
2021-07-09 13:55:02
OpenSlides/OpenSlides
https://api.github.com/repos/OpenSlides/OpenSlides
opened
Provide info about used environment variables
OpenSlides4 documentation
Each service should provide information about all used/recognized environment variables and its defaults so we do not have to guess during deployment. Proposal: We use a uniform format for this definitions and every service should additionally mention it in its README. E. g. every service should have a `default.env` file with content like this: ``` # Environment variables used in this service # This variable is used for something more or less important. FOO_BAR=42 # This variable can be used in a nice way ... BAZ_BAR=foo ... ```
1.0
Provide info about used environment variables - Each service should provide information about all used/recognized environment variables and its defaults so we do not have to guess during deployment. Proposal: We use a uniform format for this definitions and every service should additionally mention it in its README. E. g. every service should have a `default.env` file with content like this: ``` # Environment variables used in this service # This variable is used for something more or less important. FOO_BAR=42 # This variable can be used in a nice way ... BAZ_BAR=foo ... ```
non_defect
provide info about used environment variables each service should provide information about all used recognized environment variables and its defaults so we do not have to guess during deployment proposal we use a uniform format for this definitions and every service should additionally mention it in its readme e g every service should have a default env file with content like this environment variables used in this service this variable is used for something more or less important foo bar this variable can be used in a nice way baz bar foo
0
38,652
8,951,492,190
IssuesEvent
2019-01-25 14:07:13
hazelcast/hazelcast-cpp-client
https://api.github.com/repos/hazelcast/hazelcast-cpp-client
closed
MixedMapAPITestInstance/MixedMapAPITest.testGetAllPutAll/0 sometimes fail
Estimation: M Priority: High Type: Defect
During nightly test(https://hazelcast-l337.ci.cloudbees.com/job/cpp-linux-nightly-64-SHARED-Debug/737/#), we observe test failure: ``` [0;31m[ FAILED ] [mMixedMapAPITestInstance/MixedMapAPITest.testGetAllPutAll/0, where GetParam() = 0x28ac7f0 [0;31m[ FAILED ] [mMixedMapAPITestInstance/MixedMapAPITest.testGetAllPutAll/1, where GetParam() = 0x28ac480 ``` The error is: ```C++ exception with description "ExceptionMessage {Connection closed.} at InvocationService::cleanResources" thrown in the test body.```
1.0
MixedMapAPITestInstance/MixedMapAPITest.testGetAllPutAll/0 sometimes fail - During nightly test(https://hazelcast-l337.ci.cloudbees.com/job/cpp-linux-nightly-64-SHARED-Debug/737/#), we observe test failure: ``` [0;31m[ FAILED ] [mMixedMapAPITestInstance/MixedMapAPITest.testGetAllPutAll/0, where GetParam() = 0x28ac7f0 [0;31m[ FAILED ] [mMixedMapAPITestInstance/MixedMapAPITest.testGetAllPutAll/1, where GetParam() = 0x28ac480 ``` The error is: ```C++ exception with description "ExceptionMessage {Connection closed.} at InvocationService::cleanResources" thrown in the test body.```
defect
mixedmapapitestinstance mixedmapapitest testgetallputall sometimes fail during nightly test we observe test failure mmixedmapapitestinstance mixedmapapitest testgetallputall where getparam mmixedmapapitestinstance mixedmapapitest testgetallputall where getparam the error is c exception with description exceptionmessage connection closed at invocationservice cleanresources thrown in the test body
1
22,352
3,640,405,441
IssuesEvent
2016-02-13 00:01:48
netty/netty
https://api.github.com/repos/netty/netty
closed
HTTP/2 codec may not always call Http2Connection$Listener.onStreamRemoved
defect
If `Http2Connection$Listener.onStreamAdded` is called it is not always the case that `Http2Connection$Listener.onStreamRemoved` will be called. Some use cases (including `InboundHttp2ToHttpAdapter`) depend upon the `onStreamRemoved` method to be called in order to clean up allocated buffers. [Http2ConnectionHandler.channelInactive](https://github.com/netty/netty/blob/4.1/codec-http2/src/main/java/io/netty/handler/codec/http2/Http2ConnectionHandler.java#L180) currently only iterates over active streams, and this does not account for streams which may still exist but are not active.
1.0
HTTP/2 codec may not always call Http2Connection$Listener.onStreamRemoved - If `Http2Connection$Listener.onStreamAdded` is called it is not always the case that `Http2Connection$Listener.onStreamRemoved` will be called. Some use cases (including `InboundHttp2ToHttpAdapter`) depend upon the `onStreamRemoved` method to be called in order to clean up allocated buffers. [Http2ConnectionHandler.channelInactive](https://github.com/netty/netty/blob/4.1/codec-http2/src/main/java/io/netty/handler/codec/http2/Http2ConnectionHandler.java#L180) currently only iterates over active streams, and this does not account for streams which may still exist but are not active.
defect
http codec may not always call listener onstreamremoved if listener onstreamadded is called it is not always the case that listener onstreamremoved will be called some use cases including depend upon the onstreamremoved method to be called in order to clean up allocated buffers currently only iterates over active streams and this does not account for streams which may still exist but are not active
1
102,664
4,157,996,019
IssuesEvent
2016-06-16 23:43:26
expressjs/serve-index
https://api.github.com/repos/expressjs/serve-index
closed
nsp security error on 1.7.3
deps low priority
Looking at [master](https://github.com/expressjs/serve-index/blob/master/package.json#L9), it looks like all that needs to be done is cut a new version, but we're seeing this error in https://github.com/redfin/react-server/issues/291 ``` dougwade packages/react-server-cli โ€นreact-server-cli-nsp*โ€บ ยป nsp check (+) 1 vulnerabilities found โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ โ”‚ Regular Expression Denial of Service โ”‚ โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค โ”‚ Name โ”‚ negotiator โ”‚ โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค โ”‚ Installed โ”‚ 0.5.3 โ”‚ โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค โ”‚ Vulnerable โ”‚ <= 0.6.0 โ”‚ โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค โ”‚ Patched โ”‚ >= 0.6.1 โ”‚ โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค โ”‚ Path โ”‚ react-server-cli@0.3.2 > webpack-dev-server@1.14.1 > serve-index@1.7.3 > accepts@1.2.13 > negotiator@0.5.3 โ”‚ โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค โ”‚ More Info โ”‚ https://nodesecurity.io/advisories/106 โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ ```
1.0
nsp security error on 1.7.3 - Looking at [master](https://github.com/expressjs/serve-index/blob/master/package.json#L9), it looks like all that needs to be done is cut a new version, but we're seeing this error in https://github.com/redfin/react-server/issues/291 ``` dougwade packages/react-server-cli โ€นreact-server-cli-nsp*โ€บ ยป nsp check (+) 1 vulnerabilities found โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ โ”‚ Regular Expression Denial of Service โ”‚ โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค โ”‚ Name โ”‚ negotiator โ”‚ โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค โ”‚ Installed โ”‚ 0.5.3 โ”‚ โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค โ”‚ Vulnerable โ”‚ <= 0.6.0 โ”‚ โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค โ”‚ Patched โ”‚ >= 0.6.1 โ”‚ โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค โ”‚ Path โ”‚ react-server-cli@0.3.2 > webpack-dev-server@1.14.1 > serve-index@1.7.3 > accepts@1.2.13 > negotiator@0.5.3 โ”‚ โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค โ”‚ More Info โ”‚ https://nodesecurity.io/advisories/106 โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ ```
non_defect
nsp security error on looking at it looks like all that needs to be done is cut a new version but we re seeing this error in dougwade packages react server cli โ€นreact server cli nsp โ€บ ยป nsp check vulnerabilities found โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ฌโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ โ”‚ regular expression denial of service โ”‚ โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค โ”‚ name โ”‚ negotiator โ”‚ โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค โ”‚ installed โ”‚ โ”‚ โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค โ”‚ vulnerable โ”‚ โ”‚ โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค โ”‚ patched โ”‚ โ”‚ โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค โ”‚ path โ”‚ react server cli webpack dev server serve index accepts negotiator โ”‚ โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ผโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ค โ”‚ more info โ”‚ โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜
0
51,818
13,211,319,155
IssuesEvent
2020-08-15 22:16:52
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
example scripts for credo fail (Trac #1118)
Incomplete Migration Migrated from Trac combo reconstruction defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1118">https://code.icecube.wisc.edu/projects/icecube/ticket/1118</a>, reported by kjmeagherand owned by jtatar</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:11:57", "_ts": "1550067117911749", "description": "credol3.py and nugen2107l2b.py require particleforge\ntest1.py and test2.py use old I3Units import\n\nno meta-project level documentation either", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "time": "2015-08-17T13:34:45", "component": "combo reconstruction", "summary": "example scripts for credo fail", "priority": "normal", "keywords": "", "milestone": "", "owner": "jtatar", "type": "defect" } ``` </p> </details>
1.0
example scripts for credo fail (Trac #1118) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1118">https://code.icecube.wisc.edu/projects/icecube/ticket/1118</a>, reported by kjmeagherand owned by jtatar</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:11:57", "_ts": "1550067117911749", "description": "credol3.py and nugen2107l2b.py require particleforge\ntest1.py and test2.py use old I3Units import\n\nno meta-project level documentation either", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "time": "2015-08-17T13:34:45", "component": "combo reconstruction", "summary": "example scripts for credo fail", "priority": "normal", "keywords": "", "milestone": "", "owner": "jtatar", "type": "defect" } ``` </p> </details>
defect
example scripts for credo fail trac migrated from json status closed changetime ts description py and py require particleforge py and py use old import n nno meta project level documentation either reporter kjmeagher cc resolution fixed time component combo reconstruction summary example scripts for credo fail priority normal keywords milestone owner jtatar type defect
1
255,089
27,484,737,617
IssuesEvent
2023-03-04 01:13:46
panasalap/linux-4.1.15
https://api.github.com/repos/panasalap/linux-4.1.15
opened
CVE-2016-4581 (Medium) detected in linux-yocto-devv4.2.8
security vulnerability
## CVE-2016-4581 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-devv4.2.8</b></p></summary> <p> <p>Linux Embedded Kernel - tracks the next mainline release</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-dev>https://git.yoctoproject.org/git/linux-yocto-dev</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/pnode.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/pnode.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> fs/pnode.c in the Linux kernel before 4.5.4 does not properly traverse a mount propagation tree in a certain case involving a slave mount, which allows local users to cause a denial of service (NULL pointer dereference and OOPS) via a crafted series of mount system calls. <p>Publish Date: 2016-05-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-4581>CVE-2016-4581</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-4581">https://nvd.nist.gov/vuln/detail/CVE-2016-4581</a></p> <p>Release Date: 2016-05-23</p> <p>Fix Resolution: 4.5.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2016-4581 (Medium) detected in linux-yocto-devv4.2.8 - ## CVE-2016-4581 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-devv4.2.8</b></p></summary> <p> <p>Linux Embedded Kernel - tracks the next mainline release</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-dev>https://git.yoctoproject.org/git/linux-yocto-dev</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/pnode.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/pnode.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> fs/pnode.c in the Linux kernel before 4.5.4 does not properly traverse a mount propagation tree in a certain case involving a slave mount, which allows local users to cause a denial of service (NULL pointer dereference and OOPS) via a crafted series of mount system calls. <p>Publish Date: 2016-05-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-4581>CVE-2016-4581</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-4581">https://nvd.nist.gov/vuln/detail/CVE-2016-4581</a></p> <p>Release Date: 2016-05-23</p> <p>Fix Resolution: 4.5.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in linux yocto cve medium severity vulnerability vulnerable library linux yocto linux embedded kernel tracks the next mainline release library home page a href found in base branch master vulnerable source files fs pnode c fs pnode c vulnerability details fs pnode c in the linux kernel before does not properly traverse a mount propagation tree in a certain case involving a slave mount which allows local users to cause a denial of service null pointer dereference and oops via a crafted series of mount system calls publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
126,064
26,775,795,044
IssuesEvent
2023-01-31 17:03:09
Clueless-Community/seamless-ui
https://api.github.com/repos/Clueless-Community/seamless-ui
opened
Make Accordion
codepeak 22
Need to improve this component on `\Accordion\src\accordion-3.html` ## Improvements required : Make an accordion which opens on click and only one can remain open at a time, rest should be closed ![image](https://user-images.githubusercontent.com/98400348/215831472-7077ff32-c363-4814-8141-b8cb477020b6.png)
1.0
Make Accordion - Need to improve this component on `\Accordion\src\accordion-3.html` ## Improvements required : Make an accordion which opens on click and only one can remain open at a time, rest should be closed ![image](https://user-images.githubusercontent.com/98400348/215831472-7077ff32-c363-4814-8141-b8cb477020b6.png)
non_defect
make accordion need to improve this component on accordion src accordion html improvements required make an accordion which opens on click and only one can remain open at a time rest should be closed
0
65,760
3,240,202,109
IssuesEvent
2015-10-15 01:29:08
BYU-ARCLITE/Ayamel-Examples
https://api.github.com/repos/BYU-ARCLITE/Ayamel-Examples
closed
Annotation Editor Fails to Register Change
Annotations Bug duplicate Priority 2
When making changes to annotations, they are saved as shown below, but the editor fails to register that the "save" took place: ![image](https://cloud.githubusercontent.com/assets/7268293/10493680/abb31a5e-726f-11e5-9cce-d857675e718b.png) After seeing that screen and clicking on the arrow to return to the content, the system pops up this message: ![image](https://cloud.githubusercontent.com/assets/7268293/10493735/f250bade-726f-11e5-9c1a-f75413d4297e.png)
1.0
Annotation Editor Fails to Register Change - When making changes to annotations, they are saved as shown below, but the editor fails to register that the "save" took place: ![image](https://cloud.githubusercontent.com/assets/7268293/10493680/abb31a5e-726f-11e5-9cce-d857675e718b.png) After seeing that screen and clicking on the arrow to return to the content, the system pops up this message: ![image](https://cloud.githubusercontent.com/assets/7268293/10493735/f250bade-726f-11e5-9c1a-f75413d4297e.png)
non_defect
annotation editor fails to register change when making changes to annotations they are saved as shown below but the editor fails to register that the save took place after seeing that screen and clicking on the arrow to return to the content the system pops up this message
0
762
2,587,972,319
IssuesEvent
2015-02-17 21:49:28
chrsmith/codesearch
https://api.github.com/repos/chrsmith/codesearch
opened
csearch: print statistics of -f fregexp hits if -verbose [PATCH]
auto-migrated Priority-Medium Type-Defect
``` csearch already prints the trigrams and number of files matched from the index if -verbose mode is on. This patch preprocesses the list of fileids so we can print how many were matched by the filename regex. It does double the number of calls to ix.Name(fileid), but that call looks cheap and any slowdown would be dwarfed by the disk access during the grep stage anyway. ``` ----- Original issue reported on code.google.com by `dgryski` on 17 Feb 2012 at 1:50 Attachments: * [filename-regex-stats.patch](https://storage.googleapis.com/google-code-attachments/codesearch/issue-14/comment-0/filename-regex-stats.patch)
1.0
csearch: print statistics of -f fregexp hits if -verbose [PATCH] - ``` csearch already prints the trigrams and number of files matched from the index if -verbose mode is on. This patch preprocesses the list of fileids so we can print how many were matched by the filename regex. It does double the number of calls to ix.Name(fileid), but that call looks cheap and any slowdown would be dwarfed by the disk access during the grep stage anyway. ``` ----- Original issue reported on code.google.com by `dgryski` on 17 Feb 2012 at 1:50 Attachments: * [filename-regex-stats.patch](https://storage.googleapis.com/google-code-attachments/codesearch/issue-14/comment-0/filename-regex-stats.patch)
defect
csearch print statistics of f fregexp hits if verbose csearch already prints the trigrams and number of files matched from the index if verbose mode is on this patch preprocesses the list of fileids so we can print how many were matched by the filename regex it does double the number of calls to ix name fileid but that call looks cheap and any slowdown would be dwarfed by the disk access during the grep stage anyway original issue reported on code google com by dgryski on feb at attachments
1
55,519
14,530,821,787
IssuesEvent
2020-12-14 19:50:26
mozilla-extensions/normandy-devtools
https://api.github.com/repos/mozilla-extensions/normandy-devtools
closed
The dropdowns from the Normandy Devtools are not fixed in their position when scrolling through the page
[QA]:Normal issue defect
[Affected versions]: - Firefox Release 83.0 (Build ID: 20201112153044) [Affected Platforms]: - Windows 10 x64 [Prerequisites]: - Have the Firefox browser installed. - Have the [Normandy Devtools V2.2.0](https://github.com/mozilla-extensions/normandy-devtools/releases) addon installed. - Have VPN access to the staging environment. [Steps to reproduce]: 1. Open the Normandy Devtools and choose the Stage environment. 2. Click on the โ€œCreate Recipeโ€ button. 3. Click on any dropdown field from the page. 4. Scroll the page and observe the behavior. [Expected result]: - The dropdown remains fixed on its area. [Actual result]: - The dropdown is floating on the page while scrolling. [Notes]: - This issue is reproducible with the โ€œActionโ€, โ€œSampling Typeโ€, โ€œCountriesโ€, and โ€œLocalesโ€ dropdowns. - Attached a screen recording of the issue. ![dropdowns_Normandy](https://user-images.githubusercontent.com/50919288/99554506-ef014280-29c7-11eb-8a1c-b40304bcd75e.gif)
1.0
The dropdowns from the Normandy Devtools are not fixed in their position when scrolling through the page - [Affected versions]: - Firefox Release 83.0 (Build ID: 20201112153044) [Affected Platforms]: - Windows 10 x64 [Prerequisites]: - Have the Firefox browser installed. - Have the [Normandy Devtools V2.2.0](https://github.com/mozilla-extensions/normandy-devtools/releases) addon installed. - Have VPN access to the staging environment. [Steps to reproduce]: 1. Open the Normandy Devtools and choose the Stage environment. 2. Click on the โ€œCreate Recipeโ€ button. 3. Click on any dropdown field from the page. 4. Scroll the page and observe the behavior. [Expected result]: - The dropdown remains fixed on its area. [Actual result]: - The dropdown is floating on the page while scrolling. [Notes]: - This issue is reproducible with the โ€œActionโ€, โ€œSampling Typeโ€, โ€œCountriesโ€, and โ€œLocalesโ€ dropdowns. - Attached a screen recording of the issue. ![dropdowns_Normandy](https://user-images.githubusercontent.com/50919288/99554506-ef014280-29c7-11eb-8a1c-b40304bcd75e.gif)
defect
the dropdowns from the normandy devtools are not fixed in their position when scrolling through the page firefox release build id windows have the firefox browser installed have the addon installed have vpn access to the staging environment open the normandy devtools and choose the stage environment click on the โ€œcreate recipeโ€ button click on any dropdown field from the page scroll the page and observe the behavior the dropdown remains fixed on its area the dropdown is floating on the page while scrolling this issue is reproducible with the โ€œactionโ€ โ€œsampling typeโ€ โ€œcountriesโ€ and โ€œlocalesโ€ dropdowns attached a screen recording of the issue
1
58,864
16,836,181,954
IssuesEvent
2021-06-18 12:24:10
BOINC/boinc
https://api.github.com/repos/BOINC/boinc
closed
Right clicking the BOINC Manager logo on the titlebar on Windows causes it to instantly crash
C: Manager R: duplicate T: Defect
**Describe the bug** A clear and concise description of what the bug is. The BOINC Manager will instantly crash if the titlebar logo is right clicked. **Steps To Reproduce** 1. Open the BOINC Manager 2. Right click the BOINC Manager logo in the titlebar **Expected behavior** A clear and concise description of what you expected to happen. Standard titlebar context menu items **Screenshots** If applicable, add screenshots to help explain your problem. **System Information** - OS: Windows 10 - BOINC Version: 7.16.11 **Additional context** Add any other context about the problem here.
1.0
Right clicking the BOINC Manager logo on the titlebar on Windows causes it to instantly crash - **Describe the bug** A clear and concise description of what the bug is. The BOINC Manager will instantly crash if the titlebar logo is right clicked. **Steps To Reproduce** 1. Open the BOINC Manager 2. Right click the BOINC Manager logo in the titlebar **Expected behavior** A clear and concise description of what you expected to happen. Standard titlebar context menu items **Screenshots** If applicable, add screenshots to help explain your problem. **System Information** - OS: Windows 10 - BOINC Version: 7.16.11 **Additional context** Add any other context about the problem here.
defect
right clicking the boinc manager logo on the titlebar on windows causes it to instantly crash describe the bug a clear and concise description of what the bug is the boinc manager will instantly crash if the titlebar logo is right clicked steps to reproduce open the boinc manager right click the boinc manager logo in the titlebar expected behavior a clear and concise description of what you expected to happen standard titlebar context menu items screenshots if applicable add screenshots to help explain your problem system information os windows boinc version additional context add any other context about the problem here
1
25,190
2,677,853,450
IssuesEvent
2015-03-26 04:43:01
JukkaL/mypy
https://api.github.com/repos/JukkaL/mypy
closed
Runtime access to type annotations
feature pep484 priority
The current implementation erases type information from __annotations__ and friends at runtime: In [2]: def f(x: List[int]) -> int: ...: pass In [3]: f.__annotations__ Out[3]: {'return': int, 'x': list} ## lost x: List[int] Is this by design, and if so, why? Is there some reason to not keep an accurate run-time representation of type annotations? Could that information be made available uniformly on every typeable thing? In [4]: class A: ...: i = Undefined(int) In [5]: A.__annotations__ ## lost i: int Perhaps even include a run-time representation of some inferred types e.g. all types of all instance attributes might be made available via class.__annotations__ (or something like it), regardless of how those types were declared (on __init__ or via class attributes), or even inferred (by mypy from body of __init__ or otherwise). Thoughts?
1.0
Runtime access to type annotations - The current implementation erases type information from __annotations__ and friends at runtime: In [2]: def f(x: List[int]) -> int: ...: pass In [3]: f.__annotations__ Out[3]: {'return': int, 'x': list} ## lost x: List[int] Is this by design, and if so, why? Is there some reason to not keep an accurate run-time representation of type annotations? Could that information be made available uniformly on every typeable thing? In [4]: class A: ...: i = Undefined(int) In [5]: A.__annotations__ ## lost i: int Perhaps even include a run-time representation of some inferred types e.g. all types of all instance attributes might be made available via class.__annotations__ (or something like it), regardless of how those types were declared (on __init__ or via class attributes), or even inferred (by mypy from body of __init__ or otherwise). Thoughts?
non_defect
runtime access to type annotations the current implementation erases type information from annotations and friends at runtime in def f x list int pass in f annotations out return int x list lost x list is this by design and if so why is there some reason to not keep an accurate run time representation of type annotations could that information be made available uniformly on every typeable thing in class a i undefined int in a annotations lost i int perhaps even include a run time representation of some inferred types e g all types of all instance attributes might be made available via class annotations or something like it regardless of how those types were declared on init or via class attributes or even inferred by mypy from body of init or otherwise thoughts
0
142,719
21,875,074,689
IssuesEvent
2022-05-19 09:24:20
flutter/flutter
https://api.github.com/repos/flutter/flutter
closed
Time picker with incorrect time separator format for locale `id`
framework f: material design f: date/time picker a: internationalization
Should be `.`, not `:`. b/229057815 b/231987861
1.0
Time picker with incorrect time separator format for locale `id` - Should be `.`, not `:`. b/229057815 b/231987861
non_defect
time picker with incorrect time separator format for locale id should be not b b
0
67,459
20,961,613,639
IssuesEvent
2022-03-27 21:49:42
abedmaatalla/sipdroid
https://api.github.com/repos/abedmaatalla/sipdroid
closed
C2DM
Priority-Medium Type-Defect auto-migrated
``` Feature request: Have you considered adding C2DM (cloud-to-device-messaging) to siproid and pbxes.org to replace the TCP standby connection that is currently optimal? ``` Original issue reported on code.google.com by `ezelspin...@gmail.com` on 17 Dec 2011 at 9:00
1.0
C2DM - ``` Feature request: Have you considered adding C2DM (cloud-to-device-messaging) to siproid and pbxes.org to replace the TCP standby connection that is currently optimal? ``` Original issue reported on code.google.com by `ezelspin...@gmail.com` on 17 Dec 2011 at 9:00
defect
feature request have you considered adding cloud to device messaging to siproid and pbxes org to replace the tcp standby connection that is currently optimal original issue reported on code google com by ezelspin gmail com on dec at
1
64,480
18,684,729,310
IssuesEvent
2021-11-01 10:56:18
obophenotype/cell-ontology
https://api.github.com/repos/obophenotype/cell-ontology
closed
Project: Review all cases where cell types have multiple develops_from relatiionships
Priority-Medium Type-Defect auto-migrated autoclosed-unfixed
``` Assuming develops_from indicates lineage then the following is problematic cell1 subclassOf develops_from some cell2 cell1 subClassOf develops_from some cell3 e.g. id: CL:0000140 name: odontocyte relationship: develops_from CL:0000060 ! odontoblast relationship: develops_from CL:0000134 ! mesenchymal stem cell unless cell2 subClassOf cell3 OR cell3 SubClassOf cell2 (in which case the pattern reflects redundancy and one axioms should be deleted) or cell2 and cell3 overlap and all cell1 develop from some cell in that is in a class defined by the intersection of cell2 and cell3. I suspect that this is rarely the case where this pattern appears in CL. More likely these are attempts at recording some/some relationships. All such cases should therefore be reviewed. ``` Original issue reported on code.google.com by `dosu...@gmail.com` on 16 Jul 2014 at 12:13
1.0
Project: Review all cases where cell types have multiple develops_from relatiionships - ``` Assuming develops_from indicates lineage then the following is problematic cell1 subclassOf develops_from some cell2 cell1 subClassOf develops_from some cell3 e.g. id: CL:0000140 name: odontocyte relationship: develops_from CL:0000060 ! odontoblast relationship: develops_from CL:0000134 ! mesenchymal stem cell unless cell2 subClassOf cell3 OR cell3 SubClassOf cell2 (in which case the pattern reflects redundancy and one axioms should be deleted) or cell2 and cell3 overlap and all cell1 develop from some cell in that is in a class defined by the intersection of cell2 and cell3. I suspect that this is rarely the case where this pattern appears in CL. More likely these are attempts at recording some/some relationships. All such cases should therefore be reviewed. ``` Original issue reported on code.google.com by `dosu...@gmail.com` on 16 Jul 2014 at 12:13
defect
project review all cases where cell types have multiple develops from relatiionships assuming develops from indicates lineage then the following is problematic subclassof develops from some subclassof develops from some e g id cl name odontocyte relationship develops from cl odontoblast relationship develops from cl mesenchymal stem cell unless subclassof or subclassof in which case the pattern reflects redundancy and one axioms should be deleted or and overlap and all develop from some cell in that is in a class defined by the intersection of and i suspect that this is rarely the case where this pattern appears in cl more likely these are attempts at recording some some relationships all such cases should therefore be reviewed original issue reported on code google com by dosu gmail com on jul at
1
11,392
2,649,862,610
IssuesEvent
2015-03-15 11:09:50
dakkanner/ra-duty-scheduler
https://api.github.com/repos/dakkanner/ra-duty-scheduler
closed
Weekend EOT additional assigned days
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? Create schedule where the end of term lands on a Friday or saturday. What is the expected output? What do you see instead? It schedules whoever should be on for the whole weekend, making them have an inflated count of days. ``` Original issue reported on code.google.com by `dakkan...@gmail.com` on 21 Nov 2014 at 9:20
1.0
Weekend EOT additional assigned days - ``` What steps will reproduce the problem? Create schedule where the end of term lands on a Friday or saturday. What is the expected output? What do you see instead? It schedules whoever should be on for the whole weekend, making them have an inflated count of days. ``` Original issue reported on code.google.com by `dakkan...@gmail.com` on 21 Nov 2014 at 9:20
defect
weekend eot additional assigned days what steps will reproduce the problem create schedule where the end of term lands on a friday or saturday what is the expected output what do you see instead it schedules whoever should be on for the whole weekend making them have an inflated count of days original issue reported on code google com by dakkan gmail com on nov at
1
603,507
18,668,490,905
IssuesEvent
2021-10-30 08:35:48
renovatebot/renovate
https://api.github.com/repos/renovatebot/renovate
closed
Renovate GitHub-native automerge debug message when `merge` method is not allowed
type:bug priority-4-low platform:github status:ready
### How are you running Renovate? WhiteSource Renovate hosted app on github.com ### Please select which platform you are using if self-hosting. _No response_ ### If you're self-hosting Renovate, tell us what version of Renovate you run. 28.8.3 ### Describe the bug ## Intro I'm trying out the new automerge on my `honkinggoose/gitgosling` repo, with Renovate version `28.8.3`. ## Steps to reproduce: 1. The repository had no pending PRs. 1. I requested a lockfile update via the Dashboard. 1. I wondered what the behavior is now, so I checked the logs from this run (pasted below in log section of bug report). 1. I was expecting Renovate to automerge the lockfile right away, as it's the only PR and it's fully up to date with `main` right when its created. 1. Instead I got a debug message: `"message": "Merge method merge commits are not allowed on this repository"` ## Repo settings On my repository I only allow `squash` merges: ![repo-merge-button-settings](https://user-images.githubusercontent.com/34918129/138567016-95d8f950-b0db-4aeb-b81c-edea8e0f99ff.png) I also have a branch protection rule stating that any PR branch must be up-to-date with `main` before it can be merged. ## Further context Link to PR that had the debug logs: https://github.com/HonkingGoose/git-gosling/pull/313 Note that Renovate did manage to automerge the PR in the end, but I think it used the Renovate fallback to do that. If GitHub-native automerge _requires_ that the normal merge is available we should document that, because we don't mention this right now. @rarkins thought: > Seems we need a way to detect merge type ### Relevant debug logs <details><summary>Relevant logs for initial run</summary> ``` DEBUG: getBranchStatus(renovate/lock-file-maintenance)(branch="renovate/lock-file-maintenance") DEBUG: branch status check result(branch="renovate/lock-file-maintenance") { "state": "pending", "statuses": [] } DEBUG: No check runs found(branch="renovate/lock-file-maintenance") { "result": { "total_count": 0, "check_runs": [] } } DEBUG: Branch status yellow(branch="renovate/lock-file-maintenance") DEBUG: Creating PR(branch="renovate/lock-file-maintenance") { "prTitle": "chore(deps): lock file maintenance" } DEBUG: Creating PR(branch="renovate/lock-file-maintenance") { "title": "chore(deps): lock file maintenance", "head": "HonkingGoose:renovate/lock-file-maintenance", "base": "main", "draft": false } DEBUG: PR created(branch="renovate/lock-file-maintenance") { "pr": 313, "draft": false } DEBUG: Adding labels '' to #313(branch="renovate/lock-file-maintenance") DEBUG: GitHub automerge unknown error(branch="renovate/lock-file-maintenance") { "prNumber": 313, "errors": [ { "type": "UNPROCESSABLE", "path": [ "enablePullRequestAutoMerge" ], "locations": [ { "line": 5, "column": 3 } ], "message": "Merge method merge commits are not allowed on this repository" } ] } INFO: PR created(branch="renovate/lock-file-maintenance") { "pr": 313, "prTitle": "chore(deps): lock file maintenance" } DEBUG: Skipping assignees and reviewers as automerge=true(branch="renovate/lock-file-maintenance") DEBUG: Created Pull Request #313(branch="renovate/lock-file-maintenance") DEBUG: PR is configured for automerge(branch="renovate/lock-file-maintenance") DEBUG: PR is not ready for merge(branch="renovate/lock-file-maintenance") DEBUG: getBranchPr(renovate/lock-file-maintenance) DEBUG: findPr(renovate/lock-file-maintenance, undefined, open) DEBUG: Found PR #313 DEBUG: PR not found in open or closed PRs list - trying to fetch it directly { "prNo": 313 } DEBUG: Ensuring Dependency Dashboard DEBUG: ensureIssue(Dependency Dashboard) DEBUG: Patching issue DEBUG: Issue updated DEBUG: Removing any stale branches DEBUG: config.repoIsOnboarded=true DEBUG: Branch lists { "branchList": [ "renovate/lock-file-maintenance", "renovate/npm-8.x" ], "renovateBranches": [ "renovate/lock-file-maintenance" ] } DEBUG: remainingBranches= DEBUG: No branches to clean up DEBUG: Repository timing splits (milliseconds) { "splits": { "init": 2411, "extract": 2967, "lookup": 2479, "update": 130426 }, "total": 139426 } ``` </details> <details><summary>Relevant logs for run that automerged the lockfile PR</summary> ``` DEBUG: Branch status green(branch="renovate/lock-file-maintenance") DEBUG: Processing existing PR(branch="renovate/lock-file-maintenance") DEBUG: Pull Request #313 does not need updating(branch="renovate/lock-file-maintenance") DEBUG: PR is configured for automerge(branch="renovate/lock-file-maintenance") DEBUG: resolveBranchStatus(branchName=renovate/lock-file-maintenance, ignoreTests=false)(branch="renovate/lock-file-maintenance") DEBUG: getBranchStatus(renovate/lock-file-maintenance)(branch="renovate/lock-file-maintenance") DEBUG: branch status check result(branch="renovate/lock-file-maintenance") { "state": "success", "statuses": [ { "url": "https://api.github.com/repos/HonkingGoose/git-gosling/statuses/2b9cedd52417d823105802e402932d23f354d9fb", "avatar_url": "https://avatars.githubusercontent.com/in/8329?v=4", "id": 14944469753, "node_id": "SC_kwDOEbfyB88AAAADesKC-Q", "state": "success", "description": "Deployment has completed", "target_url": "https://vercel.com/honkinggoose/git-gosling/EBFgt4sMCXh7utZs9nu4zRzTQLYM", "context": "Vercel", "created_at": "2021-10-23T07:14:57Z", "updated_at": "2021-10-23T07:14:57Z" } ] } DEBUG: check runs result(branch="renovate/lock-file-maintenance") { "checkRuns": [ { "name": "build (14.x)", "status": "completed", "conclusion": "success" }, { "name": "lint (14.x)", "status": "completed", "conclusion": "success" } ] } DEBUG: Branch status green(branch="renovate/lock-file-maintenance") DEBUG: Automerging #313 with strategy auto(branch="renovate/lock-file-maintenance") DEBUG: mergePr(313, renovate/lock-file-maintenance)(branch="renovate/lock-file-maintenance") DEBUG: mergePr(branch="renovate/lock-file-maintenance") { "options": { "body": { "merge_method": "squash" } }, "url": "repos/HonkingGoose/git-gosling/pulls/313/merge" } DEBUG: PR merged(branch="renovate/lock-file-maintenance") { "automergeResult": { "sha": "ec34406e5fa1722a1a1efa8a9add78593a5f9bc6", "merged": true, "message": "Pull Request successfully merged" }, "pr": 313 } INFO: PR automerged(branch="renovate/lock-file-maintenance") { "pr": 313, "prTitle": "chore(deps): lock file maintenance" } DEBUG: No remote branch to delete(branch="renovate/lock-file-maintenance") { "branchName": "renovate/lock-file-maintenance" } DEBUG: No local branch to delete(branch="renovate/lock-file-maintenance") { "branchName": "renovate/lock-file-maintenance" } INFO: Renovating repository again after automerge result ``` </details> ### Have you created a minimal reproduction repository? No reproduction, but I have linked to a public repo where it occurs
1.0
Renovate GitHub-native automerge debug message when `merge` method is not allowed - ### How are you running Renovate? WhiteSource Renovate hosted app on github.com ### Please select which platform you are using if self-hosting. _No response_ ### If you're self-hosting Renovate, tell us what version of Renovate you run. 28.8.3 ### Describe the bug ## Intro I'm trying out the new automerge on my `honkinggoose/gitgosling` repo, with Renovate version `28.8.3`. ## Steps to reproduce: 1. The repository had no pending PRs. 1. I requested a lockfile update via the Dashboard. 1. I wondered what the behavior is now, so I checked the logs from this run (pasted below in log section of bug report). 1. I was expecting Renovate to automerge the lockfile right away, as it's the only PR and it's fully up to date with `main` right when its created. 1. Instead I got a debug message: `"message": "Merge method merge commits are not allowed on this repository"` ## Repo settings On my repository I only allow `squash` merges: ![repo-merge-button-settings](https://user-images.githubusercontent.com/34918129/138567016-95d8f950-b0db-4aeb-b81c-edea8e0f99ff.png) I also have a branch protection rule stating that any PR branch must be up-to-date with `main` before it can be merged. ## Further context Link to PR that had the debug logs: https://github.com/HonkingGoose/git-gosling/pull/313 Note that Renovate did manage to automerge the PR in the end, but I think it used the Renovate fallback to do that. If GitHub-native automerge _requires_ that the normal merge is available we should document that, because we don't mention this right now. @rarkins thought: > Seems we need a way to detect merge type ### Relevant debug logs <details><summary>Relevant logs for initial run</summary> ``` DEBUG: getBranchStatus(renovate/lock-file-maintenance)(branch="renovate/lock-file-maintenance") DEBUG: branch status check result(branch="renovate/lock-file-maintenance") { "state": "pending", "statuses": [] } DEBUG: No check runs found(branch="renovate/lock-file-maintenance") { "result": { "total_count": 0, "check_runs": [] } } DEBUG: Branch status yellow(branch="renovate/lock-file-maintenance") DEBUG: Creating PR(branch="renovate/lock-file-maintenance") { "prTitle": "chore(deps): lock file maintenance" } DEBUG: Creating PR(branch="renovate/lock-file-maintenance") { "title": "chore(deps): lock file maintenance", "head": "HonkingGoose:renovate/lock-file-maintenance", "base": "main", "draft": false } DEBUG: PR created(branch="renovate/lock-file-maintenance") { "pr": 313, "draft": false } DEBUG: Adding labels '' to #313(branch="renovate/lock-file-maintenance") DEBUG: GitHub automerge unknown error(branch="renovate/lock-file-maintenance") { "prNumber": 313, "errors": [ { "type": "UNPROCESSABLE", "path": [ "enablePullRequestAutoMerge" ], "locations": [ { "line": 5, "column": 3 } ], "message": "Merge method merge commits are not allowed on this repository" } ] } INFO: PR created(branch="renovate/lock-file-maintenance") { "pr": 313, "prTitle": "chore(deps): lock file maintenance" } DEBUG: Skipping assignees and reviewers as automerge=true(branch="renovate/lock-file-maintenance") DEBUG: Created Pull Request #313(branch="renovate/lock-file-maintenance") DEBUG: PR is configured for automerge(branch="renovate/lock-file-maintenance") DEBUG: PR is not ready for merge(branch="renovate/lock-file-maintenance") DEBUG: getBranchPr(renovate/lock-file-maintenance) DEBUG: findPr(renovate/lock-file-maintenance, undefined, open) DEBUG: Found PR #313 DEBUG: PR not found in open or closed PRs list - trying to fetch it directly { "prNo": 313 } DEBUG: Ensuring Dependency Dashboard DEBUG: ensureIssue(Dependency Dashboard) DEBUG: Patching issue DEBUG: Issue updated DEBUG: Removing any stale branches DEBUG: config.repoIsOnboarded=true DEBUG: Branch lists { "branchList": [ "renovate/lock-file-maintenance", "renovate/npm-8.x" ], "renovateBranches": [ "renovate/lock-file-maintenance" ] } DEBUG: remainingBranches= DEBUG: No branches to clean up DEBUG: Repository timing splits (milliseconds) { "splits": { "init": 2411, "extract": 2967, "lookup": 2479, "update": 130426 }, "total": 139426 } ``` </details> <details><summary>Relevant logs for run that automerged the lockfile PR</summary> ``` DEBUG: Branch status green(branch="renovate/lock-file-maintenance") DEBUG: Processing existing PR(branch="renovate/lock-file-maintenance") DEBUG: Pull Request #313 does not need updating(branch="renovate/lock-file-maintenance") DEBUG: PR is configured for automerge(branch="renovate/lock-file-maintenance") DEBUG: resolveBranchStatus(branchName=renovate/lock-file-maintenance, ignoreTests=false)(branch="renovate/lock-file-maintenance") DEBUG: getBranchStatus(renovate/lock-file-maintenance)(branch="renovate/lock-file-maintenance") DEBUG: branch status check result(branch="renovate/lock-file-maintenance") { "state": "success", "statuses": [ { "url": "https://api.github.com/repos/HonkingGoose/git-gosling/statuses/2b9cedd52417d823105802e402932d23f354d9fb", "avatar_url": "https://avatars.githubusercontent.com/in/8329?v=4", "id": 14944469753, "node_id": "SC_kwDOEbfyB88AAAADesKC-Q", "state": "success", "description": "Deployment has completed", "target_url": "https://vercel.com/honkinggoose/git-gosling/EBFgt4sMCXh7utZs9nu4zRzTQLYM", "context": "Vercel", "created_at": "2021-10-23T07:14:57Z", "updated_at": "2021-10-23T07:14:57Z" } ] } DEBUG: check runs result(branch="renovate/lock-file-maintenance") { "checkRuns": [ { "name": "build (14.x)", "status": "completed", "conclusion": "success" }, { "name": "lint (14.x)", "status": "completed", "conclusion": "success" } ] } DEBUG: Branch status green(branch="renovate/lock-file-maintenance") DEBUG: Automerging #313 with strategy auto(branch="renovate/lock-file-maintenance") DEBUG: mergePr(313, renovate/lock-file-maintenance)(branch="renovate/lock-file-maintenance") DEBUG: mergePr(branch="renovate/lock-file-maintenance") { "options": { "body": { "merge_method": "squash" } }, "url": "repos/HonkingGoose/git-gosling/pulls/313/merge" } DEBUG: PR merged(branch="renovate/lock-file-maintenance") { "automergeResult": { "sha": "ec34406e5fa1722a1a1efa8a9add78593a5f9bc6", "merged": true, "message": "Pull Request successfully merged" }, "pr": 313 } INFO: PR automerged(branch="renovate/lock-file-maintenance") { "pr": 313, "prTitle": "chore(deps): lock file maintenance" } DEBUG: No remote branch to delete(branch="renovate/lock-file-maintenance") { "branchName": "renovate/lock-file-maintenance" } DEBUG: No local branch to delete(branch="renovate/lock-file-maintenance") { "branchName": "renovate/lock-file-maintenance" } INFO: Renovating repository again after automerge result ``` </details> ### Have you created a minimal reproduction repository? No reproduction, but I have linked to a public repo where it occurs
non_defect
renovate github native automerge debug message when merge method is not allowed how are you running renovate whitesource renovate hosted app on github com please select which platform you are using if self hosting no response if you re self hosting renovate tell us what version of renovate you run describe the bug intro i m trying out the new automerge on my honkinggoose gitgosling repo with renovate version steps to reproduce the repository had no pending prs i requested a lockfile update via the dashboard i wondered what the behavior is now so i checked the logs from this run pasted below in log section of bug report i was expecting renovate to automerge the lockfile right away as it s the only pr and it s fully up to date with main right when its created instead i got a debug message message merge method merge commits are not allowed on this repository repo settings on my repository i only allow squash merges i also have a branch protection rule stating that any pr branch must be up to date with main before it can be merged further context link to pr that had the debug logs note that renovate did manage to automerge the pr in the end but i think it used the renovate fallback to do that if github native automerge requires that the normal merge is available we should document that because we don t mention this right now rarkins thought seems we need a way to detect merge type relevant debug logs relevant logs for initial run debug getbranchstatus renovate lock file maintenance branch renovate lock file maintenance debug branch status check result branch renovate lock file maintenance state pending statuses debug no check runs found branch renovate lock file maintenance result total count check runs debug branch status yellow branch renovate lock file maintenance debug creating pr branch renovate lock file maintenance prtitle chore deps lock file maintenance debug creating pr branch renovate lock file maintenance title chore deps lock file maintenance head honkinggoose renovate lock file maintenance base main draft false debug pr created branch renovate lock file maintenance pr draft false debug adding labels to branch renovate lock file maintenance debug github automerge unknown error branch renovate lock file maintenance prnumber errors type unprocessable path enablepullrequestautomerge locations line column message merge method merge commits are not allowed on this repository info pr created branch renovate lock file maintenance pr prtitle chore deps lock file maintenance debug skipping assignees and reviewers as automerge true branch renovate lock file maintenance debug created pull request branch renovate lock file maintenance debug pr is configured for automerge branch renovate lock file maintenance debug pr is not ready for merge branch renovate lock file maintenance debug getbranchpr renovate lock file maintenance debug findpr renovate lock file maintenance undefined open debug found pr debug pr not found in open or closed prs list trying to fetch it directly prno debug ensuring dependency dashboard debug ensureissue dependency dashboard debug patching issue debug issue updated debug removing any stale branches debug config repoisonboarded true debug branch lists branchlist renovate lock file maintenance renovate npm x renovatebranches renovate lock file maintenance debug remainingbranches debug no branches to clean up debug repository timing splits milliseconds splits init extract lookup update total relevant logs for run that automerged the lockfile pr debug branch status green branch renovate lock file maintenance debug processing existing pr branch renovate lock file maintenance debug pull request does not need updating branch renovate lock file maintenance debug pr is configured for automerge branch renovate lock file maintenance debug resolvebranchstatus branchname renovate lock file maintenance ignoretests false branch renovate lock file maintenance debug getbranchstatus renovate lock file maintenance branch renovate lock file maintenance debug branch status check result branch renovate lock file maintenance state success statuses url avatar url id node id sc q state success description deployment has completed target url context vercel created at updated at debug check runs result branch renovate lock file maintenance checkruns name build x status completed conclusion success name lint x status completed conclusion success debug branch status green branch renovate lock file maintenance debug automerging with strategy auto branch renovate lock file maintenance debug mergepr renovate lock file maintenance branch renovate lock file maintenance debug mergepr branch renovate lock file maintenance options body merge method squash url repos honkinggoose git gosling pulls merge debug pr merged branch renovate lock file maintenance automergeresult sha merged true message pull request successfully merged pr info pr automerged branch renovate lock file maintenance pr prtitle chore deps lock file maintenance debug no remote branch to delete branch renovate lock file maintenance branchname renovate lock file maintenance debug no local branch to delete branch renovate lock file maintenance branchname renovate lock file maintenance info renovating repository again after automerge result have you created a minimal reproduction repository no reproduction but i have linked to a public repo where it occurs
0
344,437
30,746,789,385
IssuesEvent
2023-07-28 15:37:49
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Failing test: Jest Integration Tests.src/core/server/integration_tests/saved_objects/migrations/group3 - SO type registrations does not remove types from registrations without updating excludeOnUpgradeQuery
Team:Core failed-test
A test failed on a tracked branch ``` Error: Missing version for public endpoint GET /api/observability_onboarding/custom_logs/step/{name} at parseEndpoint (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/kbn-server-route-repository/src/parse_endpoint.ts:23:11) at /var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/x-pack/plugins/observability_onboarding/server/routes/register_routes.ts:39:47 at Array.forEach (<anonymous>) at forEach (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/x-pack/plugins/observability_onboarding/server/routes/register_routes.ts:37:10) at ObservabilityOnboardingPlugin.setup (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/x-pack/plugins/observability_onboarding/server/plugin.ts:66:19) at PluginWrapper.setup (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/core/plugins/core-plugins-server-internal/src/plugin.ts:105:26) at PluginsSystem.setup [as setupPlugins] (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/core/plugins/core-plugins-server-internal/src/plugins_system.ts:131:40) at PluginsService.setupPlugins [as setup] (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/core/plugins/core-plugins-server-internal/src/plugins_service.ts:166:52) at Server.setup (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/core/root/core-root-server-internal/src/server.ts:348:26) at Root.setup (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/core/root/core-root-server-internal/src/root/index.ts:66:14) at Object.<anonymous> (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/src/core/server/integration_tests/saved_objects/migrations/group3/type_registrations.test.ts:151:19) ``` First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/30745#01884a02-35b0-4943-a777-bd69fc053f9a) <!-- kibanaCiData = {"failed-test":{"test.class":"Jest Integration Tests.src/core/server/integration_tests/saved_objects/migrations/group3","test.name":"SO type registrations does not remove types from registrations without updating excludeOnUpgradeQuery","test.failCount":3}} -->
1.0
Failing test: Jest Integration Tests.src/core/server/integration_tests/saved_objects/migrations/group3 - SO type registrations does not remove types from registrations without updating excludeOnUpgradeQuery - A test failed on a tracked branch ``` Error: Missing version for public endpoint GET /api/observability_onboarding/custom_logs/step/{name} at parseEndpoint (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/kbn-server-route-repository/src/parse_endpoint.ts:23:11) at /var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/x-pack/plugins/observability_onboarding/server/routes/register_routes.ts:39:47 at Array.forEach (<anonymous>) at forEach (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/x-pack/plugins/observability_onboarding/server/routes/register_routes.ts:37:10) at ObservabilityOnboardingPlugin.setup (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/x-pack/plugins/observability_onboarding/server/plugin.ts:66:19) at PluginWrapper.setup (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/core/plugins/core-plugins-server-internal/src/plugin.ts:105:26) at PluginsSystem.setup [as setupPlugins] (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/core/plugins/core-plugins-server-internal/src/plugins_system.ts:131:40) at PluginsService.setupPlugins [as setup] (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/core/plugins/core-plugins-server-internal/src/plugins_service.ts:166:52) at Server.setup (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/core/root/core-root-server-internal/src/server.ts:348:26) at Root.setup (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/core/root/core-root-server-internal/src/root/index.ts:66:14) at Object.<anonymous> (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/src/core/server/integration_tests/saved_objects/migrations/group3/type_registrations.test.ts:151:19) ``` First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/30745#01884a02-35b0-4943-a777-bd69fc053f9a) <!-- kibanaCiData = {"failed-test":{"test.class":"Jest Integration Tests.src/core/server/integration_tests/saved_objects/migrations/group3","test.name":"SO type registrations does not remove types from registrations without updating excludeOnUpgradeQuery","test.failCount":3}} -->
non_defect
failing test jest integration tests src core server integration tests saved objects migrations so type registrations does not remove types from registrations without updating excludeonupgradequery a test failed on a tracked branch error missing version for public endpoint get api observability onboarding custom logs step name at parseendpoint var lib buildkite agent builds kb spot elastic kibana on merge kibana packages kbn server route repository src parse endpoint ts at var lib buildkite agent builds kb spot elastic kibana on merge kibana x pack plugins observability onboarding server routes register routes ts at array foreach at foreach var lib buildkite agent builds kb spot elastic kibana on merge kibana x pack plugins observability onboarding server routes register routes ts at observabilityonboardingplugin setup var lib buildkite agent builds kb spot elastic kibana on merge kibana x pack plugins observability onboarding server plugin ts at pluginwrapper setup var lib buildkite agent builds kb spot elastic kibana on merge kibana packages core plugins core plugins server internal src plugin ts at pluginssystem setup var lib buildkite agent builds kb spot elastic kibana on merge kibana packages core plugins core plugins server internal src plugins system ts at pluginsservice setupplugins var lib buildkite agent builds kb spot elastic kibana on merge kibana packages core plugins core plugins server internal src plugins service ts at server setup var lib buildkite agent builds kb spot elastic kibana on merge kibana packages core root core root server internal src server ts at root setup var lib buildkite agent builds kb spot elastic kibana on merge kibana packages core root core root server internal src root index ts at object var lib buildkite agent builds kb spot elastic kibana on merge kibana src core server integration tests saved objects migrations type registrations test ts first failure
0
721
2,587,731,740
IssuesEvent
2015-02-17 20:18:55
chrsmith/codesearch
https://api.github.com/repos/chrsmith/codesearch
opened
cindex fails on the first run: mmap() returns EINVAL
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. build all utils (or download binaries from this site) 2. be sure that there is no .csearchindex (or it has zero size) 3. run `cindex /usr/include` or any other path `cindex` fails, because mmap returns error with EINVAL code. This happens inside mmap_linux.go, on line 24. That function (mmapFile) calls mmap with zero size, which fails. I had manually copy ~/.csearchindex\~ to ~/.csearchindex. That fixed the issue. I expected cindex to work right out of the box, even with no .csearchindex. I'm using: Gentoo Base System release 2.0.3, Linux 3.1.6, amd64. Go repository is the newest (built a few minutes ago from head). ``` ----- Original issue reported on code.google.com by zaitsev....@gmail.com on 19 Jan 2012 at 8:10
1.0
cindex fails on the first run: mmap() returns EINVAL - ``` What steps will reproduce the problem? 1. build all utils (or download binaries from this site) 2. be sure that there is no .csearchindex (or it has zero size) 3. run `cindex /usr/include` or any other path `cindex` fails, because mmap returns error with EINVAL code. This happens inside mmap_linux.go, on line 24. That function (mmapFile) calls mmap with zero size, which fails. I had manually copy ~/.csearchindex\~ to ~/.csearchindex. That fixed the issue. I expected cindex to work right out of the box, even with no .csearchindex. I'm using: Gentoo Base System release 2.0.3, Linux 3.1.6, amd64. Go repository is the newest (built a few minutes ago from head). ``` ----- Original issue reported on code.google.com by zaitsev....@gmail.com on 19 Jan 2012 at 8:10
defect
cindex fails on the first run mmap returns einval what steps will reproduce the problem build all utils or download binaries from this site be sure that there is no csearchindex or it has zero size run cindex usr include or any other path cindex fails because mmap returns error with einval code this happens inside mmap linux go on line that function mmapfile calls mmap with zero size which fails i had manually copy csearchindex to csearchindex that fixed the issue i expected cindex to work right out of the box even with no csearchindex i m using gentoo base system release linux go repository is the newest built a few minutes ago from head original issue reported on code google com by zaitsev gmail com on jan at
1
75,119
9,819,146,358
IssuesEvent
2019-06-13 21:08:15
bradyrx/climpred
https://api.github.com/repos/bradyrx/climpred
closed
Consolidate bootstrap API in sphinx
documentation
As a start, I just dropped all public functions from `bootstrap` into the API. I imagine only a few of these are supposed to be called by the user. @aaronspring, if you could slim it down to only what is necessary, that would be great. Unless all of them should be on there.
1.0
Consolidate bootstrap API in sphinx - As a start, I just dropped all public functions from `bootstrap` into the API. I imagine only a few of these are supposed to be called by the user. @aaronspring, if you could slim it down to only what is necessary, that would be great. Unless all of them should be on there.
non_defect
consolidate bootstrap api in sphinx as a start i just dropped all public functions from bootstrap into the api i imagine only a few of these are supposed to be called by the user aaronspring if you could slim it down to only what is necessary that would be great unless all of them should be on there
0
36,654
8,049,079,421
IssuesEvent
2018-08-01 08:59:58
zotonic/zotonic
https://api.github.com/repos/zotonic/zotonic
opened
Concurrent category hierarchy save problem
core defect
It is possible that multiple concurrent category hierarchy saves lead to sql constraint problems. Normally this does not happen when you enable new modules manually, but when you install new sites via scripts it can happen that multiple modules with new categories in their datamodel save a new hierarchy concurrently. The problem can be solved by making sure `hierarchy:save/3` can't run concurrently by either running it from a gen_server or as a job. ```erlang 5:59:39.716 [error] channelwww z_db:300 z_db error {error,error,<<"23505">>,<<"duplicate key value violates unique constraint \"hierarchy_pkey\"">>,[{detail,<<"Key (name, id)=($category, 345) already exists.">>}]} in query "\n insert into hierarchy\n (name, id, parent_id, lvl, nr, lft, rght)\n values\n ($1, $2, $3, $4, $5, $6, $7)" with ['$category',345,104,2,4998046,4998046,4998046] 15:59:39.717 [error] gen_server <0.1364.0> terminated with reason: no match of right hand value {rollback,{{error,{error,error,<<"23505">>,<<"duplicate key value violates unique constraint \"hierarchy_pkey\"">>,[{detail,<<"Key (name, id)=($category, 345) already exists.">>}]}},[{z_db,'-q/4-fun-2-',5,[{file,"src/db/z_db.erl"},{line,301}]},{timer,tc,2,[{file,"timer.erl"},{line,181}]},{z_db,with_connection,3,[{file,"src/db/z_db.erl"},{line,211}]},{lists,foreach,2,[{file,"lists.erl"},{line,1338}]},{m_hierarchy,save_nocheck_trans,3,[{file,"src/models/m_hierarchy.erl"},{line,295}]},{z_db,'-transaction1/2-fun-0-',...},...]}} in m_hierarchy:save_nocheck/3 line 261 ```
1.0
Concurrent category hierarchy save problem - It is possible that multiple concurrent category hierarchy saves lead to sql constraint problems. Normally this does not happen when you enable new modules manually, but when you install new sites via scripts it can happen that multiple modules with new categories in their datamodel save a new hierarchy concurrently. The problem can be solved by making sure `hierarchy:save/3` can't run concurrently by either running it from a gen_server or as a job. ```erlang 5:59:39.716 [error] channelwww z_db:300 z_db error {error,error,<<"23505">>,<<"duplicate key value violates unique constraint \"hierarchy_pkey\"">>,[{detail,<<"Key (name, id)=($category, 345) already exists.">>}]} in query "\n insert into hierarchy\n (name, id, parent_id, lvl, nr, lft, rght)\n values\n ($1, $2, $3, $4, $5, $6, $7)" with ['$category',345,104,2,4998046,4998046,4998046] 15:59:39.717 [error] gen_server <0.1364.0> terminated with reason: no match of right hand value {rollback,{{error,{error,error,<<"23505">>,<<"duplicate key value violates unique constraint \"hierarchy_pkey\"">>,[{detail,<<"Key (name, id)=($category, 345) already exists.">>}]}},[{z_db,'-q/4-fun-2-',5,[{file,"src/db/z_db.erl"},{line,301}]},{timer,tc,2,[{file,"timer.erl"},{line,181}]},{z_db,with_connection,3,[{file,"src/db/z_db.erl"},{line,211}]},{lists,foreach,2,[{file,"lists.erl"},{line,1338}]},{m_hierarchy,save_nocheck_trans,3,[{file,"src/models/m_hierarchy.erl"},{line,295}]},{z_db,'-transaction1/2-fun-0-',...},...]}} in m_hierarchy:save_nocheck/3 line 261 ```
defect
concurrent category hierarchy save problem it is possible that multiple concurrent category hierarchy saves lead to sql constraint problems normally this does not happen when you enable new modules manually but when you install new sites via scripts it can happen that multiple modules with new categories in their datamodel save a new hierarchy concurrently the problem can be solved by making sure hierarchy save can t run concurrently by either running it from a gen server or as a job erlang channelwww z db z db error error error in query n insert into hierarchy n name id parent id lvl nr lft rght n values n with gen server terminated with reason no match of right hand value rollback error error error timer tc z db with connection lists foreach m hierarchy save nocheck trans z db fun in m hierarchy save nocheck line
1
67,230
27,754,555,272
IssuesEvent
2023-03-16 00:38:44
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[OPORTUNIDADE] [REMOTO] Pessoa Desenvolvedora Back-end na Enjoei
BACK-END JAVASCRIPT JQUERY GIT POSTGRESQL RUBY AWS REMOTO BACKEND GITHUB RUBY ON RAILS MICROSERVICES ELASTICSEARCH AWS CLOUD HELP WANTED GOLANG Stale
### Sobre a empresa e desafio Criada em 2009 para ser um blog para conectar pessoas que queriam se desfazer de peรงas de roupas que nรฃo usavam mais. Em meses, a Enjoei se transformou em um social commerce que oferece um espaรงo exclusivo para cada cliente criar a sua โ€œlojinhaโ€. Hoje รฉ um marketplace, um app, um site, uma plataforma de tecnologia com foco em moda. Investida por Monashees+, Bessemer Venture Partners, Dynamo e Grupo Globo e que em 2020 levantou R$ 1,13 bilhรฃo em IPO. Histรณria incrรญvel, nรฃo รฉ? Pois bem, queremos conectar vocรช pessoa desenvolvedora ao prรณximo passo na sua carreira. Estamos com uma oportunidade, para fazer parte do time Enjoei e construir uma histรณria de sucesso, como Back-end Developer. Buscamos um perfil que tenha interesse genuรญno em produzir software de qualidade, que se preocupe com o produto, goste de desafios diรกrios e de solucionar questรตes de performance e de arquitetura para orientar e compartilhar conhecimento com pessoas desenvolvedoras menos experientes. ### Requerimentos Para este desafio, รฉ necessรกrio ter experiรชncia em desenvolvimento de preferรชncia em Ruby โ€“ Ruby on Rails, Javascript (Jquery) ou Golang, Git e Github, PostgresSQL, desenvolvimento em microservices, Elasticsearch e AWS ### Modelo de contraรงรฃo CLT ### Link Vamos nesse desafio juntos? Saiba mais -> https://byintera.in/21y
1.0
[OPORTUNIDADE] [REMOTO] Pessoa Desenvolvedora Back-end na Enjoei - ### Sobre a empresa e desafio Criada em 2009 para ser um blog para conectar pessoas que queriam se desfazer de peรงas de roupas que nรฃo usavam mais. Em meses, a Enjoei se transformou em um social commerce que oferece um espaรงo exclusivo para cada cliente criar a sua โ€œlojinhaโ€. Hoje รฉ um marketplace, um app, um site, uma plataforma de tecnologia com foco em moda. Investida por Monashees+, Bessemer Venture Partners, Dynamo e Grupo Globo e que em 2020 levantou R$ 1,13 bilhรฃo em IPO. Histรณria incrรญvel, nรฃo รฉ? Pois bem, queremos conectar vocรช pessoa desenvolvedora ao prรณximo passo na sua carreira. Estamos com uma oportunidade, para fazer parte do time Enjoei e construir uma histรณria de sucesso, como Back-end Developer. Buscamos um perfil que tenha interesse genuรญno em produzir software de qualidade, que se preocupe com o produto, goste de desafios diรกrios e de solucionar questรตes de performance e de arquitetura para orientar e compartilhar conhecimento com pessoas desenvolvedoras menos experientes. ### Requerimentos Para este desafio, รฉ necessรกrio ter experiรชncia em desenvolvimento de preferรชncia em Ruby โ€“ Ruby on Rails, Javascript (Jquery) ou Golang, Git e Github, PostgresSQL, desenvolvimento em microservices, Elasticsearch e AWS ### Modelo de contraรงรฃo CLT ### Link Vamos nesse desafio juntos? Saiba mais -> https://byintera.in/21y
non_defect
pessoa desenvolvedora back end na enjoei sobre a empresa e desafio criada em para ser um blog para conectar pessoas que queriam se desfazer de peรงas de roupas que nรฃo usavam mais em meses a enjoei se transformou em um social commerce que oferece um espaรงo exclusivo para cada cliente criar a sua โ€œlojinhaโ€ hoje รฉ um marketplace um app um site uma plataforma de tecnologia com foco em moda investida por monashees bessemer venture partners dynamo e grupo globo e que em levantou r bilhรฃo em ipo histรณria incrรญvel nรฃo รฉ pois bem queremos conectar vocรช pessoa desenvolvedora ao prรณximo passo na sua carreira estamos com uma oportunidade para fazer parte do time enjoei e construir uma histรณria de sucesso como back end developer buscamos um perfil que tenha interesse genuรญno em produzir software de qualidade que se preocupe com o produto goste de desafios diรกrios e de solucionar questรตes de performance e de arquitetura para orientar e compartilhar conhecimento com pessoas desenvolvedoras menos experientes requerimentos para este desafio รฉ necessรกrio ter experiรชncia em desenvolvimento de preferรชncia em ruby โ€“ ruby on rails javascript jquery ou golang git e github postgressql desenvolvimento em microservices elasticsearch e aws modelo de contraรงรฃo clt link vamos nesse desafio juntos saiba mais
0
28,384
4,389,865,150
IssuesEvent
2016-08-09 00:00:35
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
stress: failed test in cockroach/storage/storage.test: TestReplicateAddAndRemove
Robot test-failure
Binary: cockroach/static-tests.tar.gz sha: https://github.com/cockroachdb/cockroach/commits/c1114f920f13baf1df10cb555fd7516088249ae8 Stress build found a failed test: ``` === RUN TestReplicateAddAndRemove I160726 05:24:33.209159 storage/engine/rocksdb.go:353 opening in memory rocksdb instance W160726 05:24:33.210050 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node W160726 05:24:33.211342 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node I160726 05:24:33.211509 storage/store.go:2671 system config not yet available I160726 05:24:33.212090 storage/replica_command.go:1637 store=1:1 ([]=) range=1 [/Min-/Max): new range lease replica {1 1 1} 1970-01-01 00:00:00 +0000 UTC 900ms following replica {0 0 0} 1970-01-01 00:00:00 +0000 UTC 0 [physicalTime=1970-01-01 00:00:00 +0000 UTC] I160726 05:24:33.212416 storage/engine/rocksdb.go:353 opening in memory rocksdb instance W160726 05:24:33.213239 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node I160726 05:24:33.213720 storage/store.go:2671 system config not yet available I160726 05:24:33.213840 gossip/client.go:71 starting client to 127.0.0.1:38919 I160726 05:24:33.213844 storage/engine/rocksdb.go:353 opening in memory rocksdb instance W160726 05:24:33.214625 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node I160726 05:24:33.215180 storage/store.go:2671 system config not yet available I160726 05:24:33.215304 gossip/client.go:71 starting client to 127.0.0.1:38919 I160726 05:24:33.215322 storage/engine/rocksdb.go:353 opening in memory rocksdb instance W160726 05:24:33.216338 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node I160726 05:24:33.217256 storage/store.go:2671 system config not yet available I160726 05:24:33.217397 gossip/client.go:71 starting client to 127.0.0.1:38919 I160726 05:24:33.218819 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 12 in 243.521ยตs. encoded size=2580, 26 KV pairs, 2 log entries I160726 05:24:33.219148 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > next_replica_id:2 I160726 05:24:33.219734 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.220086 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:4 StoreID:4 ReplicaID:2} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2}] I160726 05:24:33.220386 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.220413 storage/replica_raftstorage.go:587 store=4:4 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 12 (encoded size=2580, 26 KV pairs, 2 log entries) I160726 05:24:33.220963 storage/replica_raftstorage.go:590 store=4:4 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 492.712ยตs I160726 05:24:33.221150 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 14 in 161.195ยตs. encoded size=3431, 29 KV pairs, 4 log entries I160726 05:24:33.221452 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > next_replica_id:3 I160726 05:24:33.222290 storage/replica_raftstorage.go:587 store=2:2 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 14 (encoded size=3431, 29 KV pairs, 4 log entries) I160726 05:24:33.222730 storage/replica_raftstorage.go:590 store=2:2 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 401.576ยตs I160726 05:24:33.223154 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.223391 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.223721 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> I160726 05:24:33.224071 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.224404 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:2 StoreID:2 ReplicaID:3} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2} {NodeID:2 StoreID:2 ReplicaID:3}] I160726 05:24:33.224628 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.225384 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.225797 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.226119 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> I160726 05:24:33.227423 storage/replica.go:1945 store=2:2 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.227676 storage/replica.go:1945 store=2:2 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.228066 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> I160726 05:24:33.229166 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 17 in 200.114ยตs. encoded size=4442, 33 KV pairs, 7 log entries I160726 05:24:33.229475 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > replicas:<node_id:2 store_id:2 replica_id:3 > next_replica_id:4 I160726 05:24:33.230504 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.230881 storage/replica_raftstorage.go:587 store=3:3 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 17 (encoded size=4442, 33 KV pairs, 7 log entries) I160726 05:24:33.230906 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:3 StoreID:3 ReplicaID:4} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2} {NodeID:2 StoreID:2 ReplicaID:3} {NodeID:3 StoreID:3 ReplicaID:4}] W160726 05:24:33.231050 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1} I160726 05:24:33.231186 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.231622 storage/replica_raftstorage.go:590 store=3:3 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 694.304ยตs I160726 05:24:33.231941 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> W160726 05:24:33.232215 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1} I160726 05:24:33.232549 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.232733 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > replicas:<node_id:2 store_id:2 replica_id:3 > replicas:<node_id:3 store_id:3 replica_id:4 > next_replica_id:5 I160726 05:24:33.232931 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> W160726 05:24:33.233887 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1} I160726 05:24:33.235119 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.235346 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.235687 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> I160726 05:24:33.235916 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.236198 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing REMOVE_REPLICA {NodeID:2 StoreID:2 ReplicaID:3} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2} {NodeID:3 StoreID:3 ReplicaID:4}] I160726 05:24:33.237944 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> W160726 05:24:33.237957 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1} I160726 05:24:33.238396 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.239552 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.240543 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> W160726 05:24:33.240997 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1} I160726 05:24:33.241153 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> I160726 05:24:33.241324 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.241659 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> W160726 05:24:33.266682 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: <nil> W160726 05:24:33.266805 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 0} closed by the remote: <nil> W160726 05:24:33.266829 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 0} closed by the remote: <nil> W160726 05:24:33.266852 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: <nil> W160726 05:24:33.266874 storage/replica.go:337 range 1: outgoing raft transport stream to {3 3 0} closed by the remote: <nil> W160726 05:24:33.266907 storage/replica.go:337 range 1: outgoing raft transport stream to {3 3 4} closed by the remote: <nil> W160726 05:24:33.266921 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: <nil> W160726 05:24:33.266941 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:1 StoreID:1 ReplicaID:1}: no store registered for {NodeID:3 StoreID:3 ReplicaID:4} I160726 05:24:33.267618 storage/engine/rocksdb.go:353 opening in memory rocksdb instance W160726 05:24:33.268403 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node I160726 05:24:33.269801 storage/store.go:2671 system config not yet available W160726 05:24:33.269940 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node I160726 05:24:33.270410 storage/replica_command.go:1637 store=1:1 ([]=) range=1 [/Min-/Max): new range lease replica {1 1 1} 1970-01-01 00:00:00 +0000 UTC 900ms following replica {0 0 0} 1970-01-01 00:00:00 +0000 UTC 0 [physicalTime=1970-01-01 00:00:00 +0000 UTC] I160726 05:24:33.270697 storage/engine/rocksdb.go:353 opening in memory rocksdb instance W160726 05:24:33.271184 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node I160726 05:24:33.271629 storage/store.go:2671 system config not yet available I160726 05:24:33.271813 gossip/client.go:71 starting client to 127.0.0.1:37700 I160726 05:24:33.271900 storage/engine/rocksdb.go:353 opening in memory rocksdb instance W160726 05:24:33.272505 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node I160726 05:24:33.273221 storage/store.go:2671 system config not yet available I160726 05:24:33.273356 gossip/client.go:71 starting client to 127.0.0.1:37700 I160726 05:24:33.273382 storage/engine/rocksdb.go:353 opening in memory rocksdb instance W160726 05:24:33.274134 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node I160726 05:24:33.274729 storage/store.go:2671 system config not yet available I160726 05:24:33.274752 gossip/client.go:71 starting client to 127.0.0.1:37700 I160726 05:24:33.276138 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 12 in 142.306ยตs. encoded size=2580, 26 KV pairs, 2 log entries I160726 05:24:33.276432 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > next_replica_id:2 I160726 05:24:33.277000 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.277368 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:4 StoreID:4 ReplicaID:2} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2}] I160726 05:24:33.277783 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.277899 storage/replica_raftstorage.go:587 store=4:4 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 12 (encoded size=2580, 26 KV pairs, 2 log entries) I160726 05:24:33.278272 storage/replica_raftstorage.go:590 store=4:4 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 332.578ยตs I160726 05:24:33.278606 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 14 in 131.608ยตs. encoded size=3431, 29 KV pairs, 4 log entries I160726 05:24:33.278815 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > next_replica_id:3 I160726 05:24:33.279760 storage/replica_raftstorage.go:587 store=2:2 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 14 (encoded size=3431, 29 KV pairs, 4 log entries) I160726 05:24:33.280204 storage/replica_raftstorage.go:590 store=2:2 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 399.401ยตs I160726 05:24:33.280443 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.280728 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.280936 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.281222 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> I160726 05:24:33.281321 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:2 StoreID:2 ReplicaID:3} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2} {NodeID:2 StoreID:2 ReplicaID:3}] I160726 05:24:33.281495 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.282152 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.282741 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.283186 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> I160726 05:24:33.284773 storage/replica.go:1945 store=2:2 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.285021 storage/replica.go:1945 store=2:2 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.285465 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> I160726 05:24:33.289996 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > replicas:<node_id:2 store_id:2 replica_id:3 > next_replica_id:4 W160726 05:24:33.291631 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1} I160726 05:24:33.291894 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.292323 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.292666 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing REMOVE_REPLICA {NodeID:2 StoreID:2 ReplicaID:3} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2}] W160726 05:24:33.293532 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1} I160726 05:24:33.294206 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> W160726 05:24:33.294991 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1} I160726 05:24:33.295056 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 19 in 184.365ยตs. encoded size=5306, 36 KV pairs, 9 log entries I160726 05:24:33.295408 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > next_replica_id:4 I160726 05:24:33.296003 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.296205 storage/replica_raftstorage.go:587 store=3:3 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 19 (encoded size=5306, 36 KV pairs, 9 log entries) I160726 05:24:33.296535 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> I160726 05:24:33.296757 storage/replica_raftstorage.go:590 store=3:3 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 506.141ยตs I160726 05:24:33.297024 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.297498 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:3 StoreID:3 ReplicaID:4} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2} {NodeID:3 StoreID:3 ReplicaID:4}] I160726 05:24:33.297547 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.298421 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.299146 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.299590 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> I160726 05:24:33.300598 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.300760 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.301169 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> W160726 05:24:33.704343 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:33.704394 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:34.204542 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:34.204557 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 I160726 05:24:34.270157 gossip/client.go:71 starting client to 127.0.0.1:37298 I160726 05:24:34.270721 gossip/client.go:95 closing client to node 3 (127.0.0.1:37298): stopping outgoing client to node 3 (127.0.0.1:37298); already have incoming W160726 05:24:34.504339 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:34.504431 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:34.804438 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:34.804451 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:35.304476 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:35.304622 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:35.604507 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:35.604522 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:35.904329 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:35.904397 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:36.204411 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:36.204476 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:36.504400 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:36.504468 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:36.804397 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:36.804411 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:37.304381 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:37.304395 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:37.604380 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:37.604402 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:38.104389 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:38.104405 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:38.504301 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:38.504315 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:38.804308 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:38.804363 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:39.304551 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:39.304564 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:39.604313 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:39.604324 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:40.104258 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:40.104336 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:40.604440 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:40.604453 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:41.004364 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:41.005420 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:41.304399 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:41.304471 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:41.704448 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:41.704472 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:42.204429 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:42.204482 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:42.604358 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:42.604371 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:43.004396 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:43.004418 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:43.504343 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:43.504373 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:43.804438 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:43.804511 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:44.304409 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:44.304424 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:44.604345 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:44.604361 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:45.004281 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:45.004314 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:45.504305 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:45.504352 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:45.904459 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:45.904486 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:46.204439 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:46.204465 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:46.704299 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:46.704315 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:47.204330 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:47.204471 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:47.604430 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:47.604445 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:48.004317 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:48.004330 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 I160726 05:24:48.388402 stopper.go:408 quiesceing; tasks left: 1 storage/replica_range_lease.go:167 1 storage/replica.go:2395 I160726 05:24:48.388490 stopper.go:408 quiesceing; tasks left: 1 storage/replica.go:2395 W160726 05:24:48.388557 storage/store.go:1282 error gossiping system config: range 1: replica node_id:2 store_id:2 replica_id:3 not lease holder; <nil> is W160726 05:24:48.388805 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: <nil> W160726 05:24:48.388860 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 0} closed by the remote: <nil> W160726 05:24:48.388891 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: <nil> W160726 05:24:48.388898 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 0} closed by the remote: <nil> W160726 05:24:48.388911 storage/replica.go:337 range 1: outgoing raft transport stream to {3 3 0} closed by the remote: rpc error: code = 13 desc = transport is closing W160726 05:24:48.388915 storage/replica.go:337 range 1: outgoing raft transport stream to {3 3 4} closed by the remote: <nil> W160726 05:24:48.388932 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: <nil> --- FAIL: TestReplicateAddAndRemove (15.18s) testing.go:115: condition failed to evaluate within 15s: addFirst: false, expected [39 0 39 39], got [39 5 39 39] ``` Run Details: ``` 0 runs so far, 0 failures, over 5s 0 runs so far, 0 failures, over 10s 0 runs so far, 0 failures, over 15s 0 runs so far, 0 failures, over 20s 0 runs so far, 0 failures, over 25s 0 runs so far, 0 failures, over 30s 0 runs so far, 0 failures, over 35s 0 runs so far, 0 failures, over 40s 0 runs so far, 0 failures, over 45s 0 runs so far, 0 failures, over 50s 0 runs so far, 0 failures, over 55s 0 runs so far, 0 failures, over 1m0s 0 runs so far, 0 failures, over 1m5s 0 runs so far, 0 failures, over 1m10s 0 runs so far, 0 failures, over 1m15s 6 runs so far, 0 failures, over 1m20s 16 runs so far, 0 failures, over 1m25s 16 runs so far, 0 failures, over 1m30s 16 runs so far, 0 failures, over 1m35s 16 runs so far, 0 failures, over 1m40s 16 runs so far, 0 failures, over 1m45s 16 runs so far, 0 failures, over 1m50s 16 runs so far, 0 failures, over 1m55s 16 runs so far, 0 failures, over 2m0s 16 runs so far, 0 failures, over 2m5s 16 runs so far, 0 failures, over 2m10s 16 runs so far, 0 failures, over 2m15s 16 runs so far, 0 failures, over 2m20s 16 runs so far, 0 failures, over 2m25s 16 runs so far, 0 failures, over 2m30s 16 runs so far, 0 failures, over 2m35s 20 runs so far, 0 failures, over 2m40s 30 runs so far, 0 failures, over 2m45s 32 runs so far, 0 failures, over 2m50s 32 runs so far, 0 failures, over 2m55s 32 runs so far, 0 failures, over 3m0s 32 runs so far, 0 failures, over 3m5s 32 runs so far, 0 failures, over 3m10s 32 runs so far, 0 failures, over 3m15s 32 runs so far, 0 failures, over 3m20s 32 runs so far, 0 failures, over 3m25s 32 runs so far, 0 failures, over 3m30s 32 runs so far, 0 failures, over 3m35s 32 runs so far, 0 failures, over 3m40s 32 runs so far, 0 failures, over 3m45s 32 runs so far, 0 failures, over 3m50s 32 runs so far, 0 failures, over 3m55s 35 runs so far, 0 failures, over 4m0s 44 runs so far, 0 failures, over 4m5s 48 runs so far, 0 failures, over 4m10s 48 runs so far, 0 failures, over 4m15s 48 runs so far, 0 failures, over 4m20s 48 runs so far, 0 failures, over 4m25s 48 runs so far, 0 failures, over 4m30s 48 runs so far, 0 failures, over 4m35s 48 runs so far, 0 failures, over 4m40s 48 runs so far, 0 failures, over 4m45s 48 runs so far, 0 failures, over 4m50s 48 runs so far, 0 failures, over 4m55s 48 runs so far, 0 failures, over 5m0s 48 runs so far, 0 failures, over 5m5s 48 runs so far, 0 failures, over 5m10s 49 runs so far, 0 failures, over 5m15s 52 runs so far, 0 failures, over 5m20s 60 runs so far, 0 failures, over 5m25s 64 runs so far, 0 failures, over 5m30s 64 runs so far, 0 failures, over 5m35s 64 runs so far, 0 failures, over 5m40s 64 runs so far, 0 failures, over 5m45s 64 runs so far, 0 failures, over 5m50s 64 runs so far, 0 failures, over 5m55s 64 runs so far, 0 failures, over 6m0s 64 runs so far, 0 failures, over 6m5s 64 runs so far, 0 failures, over 6m10s 64 runs so far, 0 failures, over 6m15s 64 runs so far, 0 failures, over 6m20s 64 runs so far, 0 failures, over 6m25s 64 runs so far, 0 failures, over 6m30s 65 runs so far, 0 failures, over 6m35s 68 runs so far, 0 failures, over 6m40s 77 runs so far, 0 failures, over 6m45s 80 runs completed, 1 failures, over 6m50s FAIL ``` Please assign, take a look and update the issue accordingly.
1.0
stress: failed test in cockroach/storage/storage.test: TestReplicateAddAndRemove - Binary: cockroach/static-tests.tar.gz sha: https://github.com/cockroachdb/cockroach/commits/c1114f920f13baf1df10cb555fd7516088249ae8 Stress build found a failed test: ``` === RUN TestReplicateAddAndRemove I160726 05:24:33.209159 storage/engine/rocksdb.go:353 opening in memory rocksdb instance W160726 05:24:33.210050 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node W160726 05:24:33.211342 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node I160726 05:24:33.211509 storage/store.go:2671 system config not yet available I160726 05:24:33.212090 storage/replica_command.go:1637 store=1:1 ([]=) range=1 [/Min-/Max): new range lease replica {1 1 1} 1970-01-01 00:00:00 +0000 UTC 900ms following replica {0 0 0} 1970-01-01 00:00:00 +0000 UTC 0 [physicalTime=1970-01-01 00:00:00 +0000 UTC] I160726 05:24:33.212416 storage/engine/rocksdb.go:353 opening in memory rocksdb instance W160726 05:24:33.213239 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node I160726 05:24:33.213720 storage/store.go:2671 system config not yet available I160726 05:24:33.213840 gossip/client.go:71 starting client to 127.0.0.1:38919 I160726 05:24:33.213844 storage/engine/rocksdb.go:353 opening in memory rocksdb instance W160726 05:24:33.214625 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node I160726 05:24:33.215180 storage/store.go:2671 system config not yet available I160726 05:24:33.215304 gossip/client.go:71 starting client to 127.0.0.1:38919 I160726 05:24:33.215322 storage/engine/rocksdb.go:353 opening in memory rocksdb instance W160726 05:24:33.216338 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node I160726 05:24:33.217256 storage/store.go:2671 system config not yet available I160726 05:24:33.217397 gossip/client.go:71 starting client to 127.0.0.1:38919 I160726 05:24:33.218819 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 12 in 243.521ยตs. encoded size=2580, 26 KV pairs, 2 log entries I160726 05:24:33.219148 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > next_replica_id:2 I160726 05:24:33.219734 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.220086 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:4 StoreID:4 ReplicaID:2} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2}] I160726 05:24:33.220386 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.220413 storage/replica_raftstorage.go:587 store=4:4 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 12 (encoded size=2580, 26 KV pairs, 2 log entries) I160726 05:24:33.220963 storage/replica_raftstorage.go:590 store=4:4 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 492.712ยตs I160726 05:24:33.221150 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 14 in 161.195ยตs. encoded size=3431, 29 KV pairs, 4 log entries I160726 05:24:33.221452 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > next_replica_id:3 I160726 05:24:33.222290 storage/replica_raftstorage.go:587 store=2:2 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 14 (encoded size=3431, 29 KV pairs, 4 log entries) I160726 05:24:33.222730 storage/replica_raftstorage.go:590 store=2:2 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 401.576ยตs I160726 05:24:33.223154 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.223391 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.223721 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> I160726 05:24:33.224071 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.224404 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:2 StoreID:2 ReplicaID:3} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2} {NodeID:2 StoreID:2 ReplicaID:3}] I160726 05:24:33.224628 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.225384 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.225797 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.226119 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> I160726 05:24:33.227423 storage/replica.go:1945 store=2:2 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.227676 storage/replica.go:1945 store=2:2 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.228066 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> I160726 05:24:33.229166 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 17 in 200.114ยตs. encoded size=4442, 33 KV pairs, 7 log entries I160726 05:24:33.229475 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > replicas:<node_id:2 store_id:2 replica_id:3 > next_replica_id:4 I160726 05:24:33.230504 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.230881 storage/replica_raftstorage.go:587 store=3:3 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 17 (encoded size=4442, 33 KV pairs, 7 log entries) I160726 05:24:33.230906 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:3 StoreID:3 ReplicaID:4} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2} {NodeID:2 StoreID:2 ReplicaID:3} {NodeID:3 StoreID:3 ReplicaID:4}] W160726 05:24:33.231050 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1} I160726 05:24:33.231186 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.231622 storage/replica_raftstorage.go:590 store=3:3 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 694.304ยตs I160726 05:24:33.231941 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> W160726 05:24:33.232215 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1} I160726 05:24:33.232549 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.232733 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > replicas:<node_id:2 store_id:2 replica_id:3 > replicas:<node_id:3 store_id:3 replica_id:4 > next_replica_id:5 I160726 05:24:33.232931 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> W160726 05:24:33.233887 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1} I160726 05:24:33.235119 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.235346 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.235687 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> I160726 05:24:33.235916 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.236198 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing REMOVE_REPLICA {NodeID:2 StoreID:2 ReplicaID:3} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2} {NodeID:3 StoreID:3 ReplicaID:4}] I160726 05:24:33.237944 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> W160726 05:24:33.237957 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1} I160726 05:24:33.238396 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.239552 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.240543 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> W160726 05:24:33.240997 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1} I160726 05:24:33.241153 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> I160726 05:24:33.241324 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.241659 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> W160726 05:24:33.266682 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: <nil> W160726 05:24:33.266805 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 0} closed by the remote: <nil> W160726 05:24:33.266829 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 0} closed by the remote: <nil> W160726 05:24:33.266852 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: <nil> W160726 05:24:33.266874 storage/replica.go:337 range 1: outgoing raft transport stream to {3 3 0} closed by the remote: <nil> W160726 05:24:33.266907 storage/replica.go:337 range 1: outgoing raft transport stream to {3 3 4} closed by the remote: <nil> W160726 05:24:33.266921 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: <nil> W160726 05:24:33.266941 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:1 StoreID:1 ReplicaID:1}: no store registered for {NodeID:3 StoreID:3 ReplicaID:4} I160726 05:24:33.267618 storage/engine/rocksdb.go:353 opening in memory rocksdb instance W160726 05:24:33.268403 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node I160726 05:24:33.269801 storage/store.go:2671 system config not yet available W160726 05:24:33.269940 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node I160726 05:24:33.270410 storage/replica_command.go:1637 store=1:1 ([]=) range=1 [/Min-/Max): new range lease replica {1 1 1} 1970-01-01 00:00:00 +0000 UTC 900ms following replica {0 0 0} 1970-01-01 00:00:00 +0000 UTC 0 [physicalTime=1970-01-01 00:00:00 +0000 UTC] I160726 05:24:33.270697 storage/engine/rocksdb.go:353 opening in memory rocksdb instance W160726 05:24:33.271184 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node I160726 05:24:33.271629 storage/store.go:2671 system config not yet available I160726 05:24:33.271813 gossip/client.go:71 starting client to 127.0.0.1:37700 I160726 05:24:33.271900 storage/engine/rocksdb.go:353 opening in memory rocksdb instance W160726 05:24:33.272505 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node I160726 05:24:33.273221 storage/store.go:2671 system config not yet available I160726 05:24:33.273356 gossip/client.go:71 starting client to 127.0.0.1:37700 I160726 05:24:33.273382 storage/engine/rocksdb.go:353 opening in memory rocksdb instance W160726 05:24:33.274134 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node I160726 05:24:33.274729 storage/store.go:2671 system config not yet available I160726 05:24:33.274752 gossip/client.go:71 starting client to 127.0.0.1:37700 I160726 05:24:33.276138 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 12 in 142.306ยตs. encoded size=2580, 26 KV pairs, 2 log entries I160726 05:24:33.276432 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > next_replica_id:2 I160726 05:24:33.277000 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.277368 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:4 StoreID:4 ReplicaID:2} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2}] I160726 05:24:33.277783 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.277899 storage/replica_raftstorage.go:587 store=4:4 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 12 (encoded size=2580, 26 KV pairs, 2 log entries) I160726 05:24:33.278272 storage/replica_raftstorage.go:590 store=4:4 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 332.578ยตs I160726 05:24:33.278606 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 14 in 131.608ยตs. encoded size=3431, 29 KV pairs, 4 log entries I160726 05:24:33.278815 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > next_replica_id:3 I160726 05:24:33.279760 storage/replica_raftstorage.go:587 store=2:2 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 14 (encoded size=3431, 29 KV pairs, 4 log entries) I160726 05:24:33.280204 storage/replica_raftstorage.go:590 store=2:2 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 399.401ยตs I160726 05:24:33.280443 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.280728 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.280936 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.281222 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> I160726 05:24:33.281321 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:2 StoreID:2 ReplicaID:3} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2} {NodeID:2 StoreID:2 ReplicaID:3}] I160726 05:24:33.281495 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.282152 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.282741 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.283186 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> I160726 05:24:33.284773 storage/replica.go:1945 store=2:2 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.285021 storage/replica.go:1945 store=2:2 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.285465 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> I160726 05:24:33.289996 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > replicas:<node_id:2 store_id:2 replica_id:3 > next_replica_id:4 W160726 05:24:33.291631 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1} I160726 05:24:33.291894 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.292323 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.292666 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing REMOVE_REPLICA {NodeID:2 StoreID:2 ReplicaID:3} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2}] W160726 05:24:33.293532 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1} I160726 05:24:33.294206 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> W160726 05:24:33.294991 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1} I160726 05:24:33.295056 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 19 in 184.365ยตs. encoded size=5306, 36 KV pairs, 9 log entries I160726 05:24:33.295408 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > next_replica_id:4 I160726 05:24:33.296003 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.296205 storage/replica_raftstorage.go:587 store=3:3 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 19 (encoded size=5306, 36 KV pairs, 9 log entries) I160726 05:24:33.296535 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> I160726 05:24:33.296757 storage/replica_raftstorage.go:590 store=3:3 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 506.141ยตs I160726 05:24:33.297024 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.297498 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:3 StoreID:3 ReplicaID:4} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2} {NodeID:3 StoreID:3 ReplicaID:4}] I160726 05:24:33.297547 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.298421 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.299146 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.299590 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> I160726 05:24:33.300598 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil> I160726 05:24:33.300760 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil> I160726 05:24:33.301169 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil> W160726 05:24:33.704343 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:33.704394 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:34.204542 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:34.204557 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 I160726 05:24:34.270157 gossip/client.go:71 starting client to 127.0.0.1:37298 I160726 05:24:34.270721 gossip/client.go:95 closing client to node 3 (127.0.0.1:37298): stopping outgoing client to node 3 (127.0.0.1:37298); already have incoming W160726 05:24:34.504339 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:34.504431 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:34.804438 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:34.804451 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:35.304476 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:35.304622 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:35.604507 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:35.604522 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:35.904329 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:35.904397 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:36.204411 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:36.204476 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:36.504400 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:36.504468 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:36.804397 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:36.804411 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:37.304381 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:37.304395 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:37.604380 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:37.604402 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:38.104389 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:38.104405 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:38.504301 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:38.504315 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:38.804308 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:38.804363 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:39.304551 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:39.304564 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:39.604313 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:39.604324 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:40.104258 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:40.104336 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:40.604440 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:40.604453 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:41.004364 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:41.005420 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:41.304399 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:41.304471 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:41.704448 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:41.704472 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:42.204429 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:42.204482 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:42.604358 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:42.604371 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:43.004396 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:43.004418 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:43.504343 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:43.504373 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:43.804438 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:43.804511 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:44.304409 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:44.304424 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:44.604345 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:44.604361 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:45.004281 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:45.004314 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:45.504305 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:45.504352 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:45.904459 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:45.904486 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:46.204439 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:46.204465 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:46.704299 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:46.704315 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:47.204330 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:47.204471 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:47.604430 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:47.604445 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:48.004317 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 W160726 05:24:48.004330 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5 I160726 05:24:48.388402 stopper.go:408 quiesceing; tasks left: 1 storage/replica_range_lease.go:167 1 storage/replica.go:2395 I160726 05:24:48.388490 stopper.go:408 quiesceing; tasks left: 1 storage/replica.go:2395 W160726 05:24:48.388557 storage/store.go:1282 error gossiping system config: range 1: replica node_id:2 store_id:2 replica_id:3 not lease holder; <nil> is W160726 05:24:48.388805 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: <nil> W160726 05:24:48.388860 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 0} closed by the remote: <nil> W160726 05:24:48.388891 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: <nil> W160726 05:24:48.388898 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 0} closed by the remote: <nil> W160726 05:24:48.388911 storage/replica.go:337 range 1: outgoing raft transport stream to {3 3 0} closed by the remote: rpc error: code = 13 desc = transport is closing W160726 05:24:48.388915 storage/replica.go:337 range 1: outgoing raft transport stream to {3 3 4} closed by the remote: <nil> W160726 05:24:48.388932 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: <nil> --- FAIL: TestReplicateAddAndRemove (15.18s) testing.go:115: condition failed to evaluate within 15s: addFirst: false, expected [39 0 39 39], got [39 5 39 39] ``` Run Details: ``` 0 runs so far, 0 failures, over 5s 0 runs so far, 0 failures, over 10s 0 runs so far, 0 failures, over 15s 0 runs so far, 0 failures, over 20s 0 runs so far, 0 failures, over 25s 0 runs so far, 0 failures, over 30s 0 runs so far, 0 failures, over 35s 0 runs so far, 0 failures, over 40s 0 runs so far, 0 failures, over 45s 0 runs so far, 0 failures, over 50s 0 runs so far, 0 failures, over 55s 0 runs so far, 0 failures, over 1m0s 0 runs so far, 0 failures, over 1m5s 0 runs so far, 0 failures, over 1m10s 0 runs so far, 0 failures, over 1m15s 6 runs so far, 0 failures, over 1m20s 16 runs so far, 0 failures, over 1m25s 16 runs so far, 0 failures, over 1m30s 16 runs so far, 0 failures, over 1m35s 16 runs so far, 0 failures, over 1m40s 16 runs so far, 0 failures, over 1m45s 16 runs so far, 0 failures, over 1m50s 16 runs so far, 0 failures, over 1m55s 16 runs so far, 0 failures, over 2m0s 16 runs so far, 0 failures, over 2m5s 16 runs so far, 0 failures, over 2m10s 16 runs so far, 0 failures, over 2m15s 16 runs so far, 0 failures, over 2m20s 16 runs so far, 0 failures, over 2m25s 16 runs so far, 0 failures, over 2m30s 16 runs so far, 0 failures, over 2m35s 20 runs so far, 0 failures, over 2m40s 30 runs so far, 0 failures, over 2m45s 32 runs so far, 0 failures, over 2m50s 32 runs so far, 0 failures, over 2m55s 32 runs so far, 0 failures, over 3m0s 32 runs so far, 0 failures, over 3m5s 32 runs so far, 0 failures, over 3m10s 32 runs so far, 0 failures, over 3m15s 32 runs so far, 0 failures, over 3m20s 32 runs so far, 0 failures, over 3m25s 32 runs so far, 0 failures, over 3m30s 32 runs so far, 0 failures, over 3m35s 32 runs so far, 0 failures, over 3m40s 32 runs so far, 0 failures, over 3m45s 32 runs so far, 0 failures, over 3m50s 32 runs so far, 0 failures, over 3m55s 35 runs so far, 0 failures, over 4m0s 44 runs so far, 0 failures, over 4m5s 48 runs so far, 0 failures, over 4m10s 48 runs so far, 0 failures, over 4m15s 48 runs so far, 0 failures, over 4m20s 48 runs so far, 0 failures, over 4m25s 48 runs so far, 0 failures, over 4m30s 48 runs so far, 0 failures, over 4m35s 48 runs so far, 0 failures, over 4m40s 48 runs so far, 0 failures, over 4m45s 48 runs so far, 0 failures, over 4m50s 48 runs so far, 0 failures, over 4m55s 48 runs so far, 0 failures, over 5m0s 48 runs so far, 0 failures, over 5m5s 48 runs so far, 0 failures, over 5m10s 49 runs so far, 0 failures, over 5m15s 52 runs so far, 0 failures, over 5m20s 60 runs so far, 0 failures, over 5m25s 64 runs so far, 0 failures, over 5m30s 64 runs so far, 0 failures, over 5m35s 64 runs so far, 0 failures, over 5m40s 64 runs so far, 0 failures, over 5m45s 64 runs so far, 0 failures, over 5m50s 64 runs so far, 0 failures, over 5m55s 64 runs so far, 0 failures, over 6m0s 64 runs so far, 0 failures, over 6m5s 64 runs so far, 0 failures, over 6m10s 64 runs so far, 0 failures, over 6m15s 64 runs so far, 0 failures, over 6m20s 64 runs so far, 0 failures, over 6m25s 64 runs so far, 0 failures, over 6m30s 65 runs so far, 0 failures, over 6m35s 68 runs so far, 0 failures, over 6m40s 77 runs so far, 0 failures, over 6m45s 80 runs completed, 1 failures, over 6m50s FAIL ``` Please assign, take a look and update the issue accordingly.
non_defect
stress failed test in cockroach storage storage test testreplicateaddandremove binary cockroach static tests tar gz sha stress build found a failed test run testreplicateaddandremove storage engine rocksdb go opening in memory rocksdb instance gossip gossip go not connected to cluster use join to specify a connected node gossip gossip go not connected to cluster use join to specify a connected node storage store go system config not yet available storage replica command go store range storage engine rocksdb go opening in memory rocksdb instance gossip gossip go not connected to cluster use join to specify a connected node storage store go system config not yet available gossip client go starting client to storage engine rocksdb go opening in memory rocksdb instance gossip gossip go not connected to cluster use join to specify a connected node storage store go system config not yet available gossip client go starting client to storage engine rocksdb go opening in memory rocksdb instance gossip gossip go not connected to cluster use join to specify a connected node storage store go system config not yet available gossip client go starting client to storage replica raftstorage go generated snapshot for range at index in encoded size kv pairs log entries storage replica command go store range min max change replicas of read existing descriptor range id start key end key replicas next replica id storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica raftstorage go store range applying preemptive snapshot for range at index encoded size kv pairs log entries storage replica raftstorage go store range applied preemptive snapshot for range in storage replica raftstorage go generated snapshot for range at index in encoded size kv pairs log entries storage replica command go store range min max change replicas of read existing descriptor range id start key end key replicas replicas next replica id storage replica raftstorage go store range applying preemptive snapshot for range at index encoded size kv pairs log entries storage replica raftstorage go store range applied preemptive snapshot for range in storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica command go unable to gossip first range haslease false err storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica command go unable to gossip first range haslease false err storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica command go unable to gossip first range haslease false err storage replica raftstorage go generated snapshot for range at index in encoded size kv pairs log entries storage replica command go store range min max change replicas of read existing descriptor range id start key end key replicas replicas replicas next replica id storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica raftstorage go store range applying preemptive snapshot for range at index encoded size kv pairs log entries storage replica go store range storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc unable to accept raft message from nodeid storeid replicaid no store registered for nodeid storeid replicaid storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica raftstorage go store range applied preemptive snapshot for range in storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc unable to accept raft message from nodeid storeid replicaid no store registered for nodeid storeid replicaid storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica command go store range min max change replicas of read existing descriptor range id start key end key replicas replicas replicas replicas next replica id storage replica command go unable to gossip first range haslease false err storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc unable to accept raft message from nodeid storeid replicaid no store registered for nodeid storeid replicaid storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica command go unable to gossip first range haslease false err storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc unable to accept raft message from nodeid storeid replicaid no store registered for nodeid storeid replicaid storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc unable to accept raft message from nodeid storeid replicaid no store registered for nodeid storeid replicaid storage replica command go unable to gossip first range haslease false err storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica command go unable to gossip first range haslease false err storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc unable to accept raft message from nodeid storeid replicaid no store registered for nodeid storeid replicaid storage engine rocksdb go opening in memory rocksdb instance gossip gossip go not connected to cluster use join to specify a connected node storage store go system config not yet available gossip gossip go not connected to cluster use join to specify a connected node storage replica command go store range storage engine rocksdb go opening in memory rocksdb instance gossip gossip go not connected to cluster use join to specify a connected node storage store go system config not yet available gossip client go starting client to storage engine rocksdb go opening in memory rocksdb instance gossip gossip go not connected to cluster use join to specify a connected node storage store go system config not yet available gossip client go starting client to storage engine rocksdb go opening in memory rocksdb instance gossip gossip go not connected to cluster use join to specify a connected node storage store go system config not yet available gossip client go starting client to storage replica raftstorage go generated snapshot for range at index in encoded size kv pairs log entries storage replica command go store range min max change replicas of read existing descriptor range id start key end key replicas next replica id storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica raftstorage go store range applying preemptive snapshot for range at index encoded size kv pairs log entries storage replica raftstorage go store range applied preemptive snapshot for range in storage replica raftstorage go generated snapshot for range at index in encoded size kv pairs log entries storage replica command go store range min max change replicas of read existing descriptor range id start key end key replicas replicas next replica id storage replica raftstorage go store range applying preemptive snapshot for range at index encoded size kv pairs log entries storage replica raftstorage go store range applied preemptive snapshot for range in storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica command go unable to gossip first range haslease false err storage replica go store range storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica command go unable to gossip first range haslease false err storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica command go unable to gossip first range haslease false err storage replica command go store range min max change replicas of read existing descriptor range id start key end key replicas replicas replicas next replica id storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc unable to accept raft message from nodeid storeid replicaid no store registered for nodeid storeid replicaid storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc unable to accept raft message from nodeid storeid replicaid no store registered for nodeid storeid replicaid storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc unable to accept raft message from nodeid storeid replicaid no store registered for nodeid storeid replicaid storage replica raftstorage go generated snapshot for range at index in encoded size kv pairs log entries storage replica command go store range min max change replicas of read existing descriptor range id start key end key replicas replicas next replica id storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica raftstorage go store range applying preemptive snapshot for range at index encoded size kv pairs log entries storage replica command go unable to gossip first range haslease false err storage replica raftstorage go store range applied preemptive snapshot for range in storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica command go unable to gossip first range haslease false err storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica command go unable to gossip first range haslease false err storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid gossip client go starting client to gossip client go closing client to node stopping outgoing client to node already have incoming storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid stopper go quiesceing tasks left storage replica range lease go storage replica go stopper go quiesceing tasks left storage replica go storage store go error gossiping system config range replica node id store id replica id not lease holder is storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc transport is closing storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote fail testreplicateaddandremove testing go condition failed to evaluate within addfirst false expected got run details runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs completed failures over fail please assign take a look and update the issue accordingly
0
161,932
12,600,462,146
IssuesEvent
2020-06-11 08:11:50
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
[Failing Test] [sig-scheduling] SchedulerPreemption [Serial] PreemptionExecutionPath runs ReplicaSets to verify preemption running path [Conformance]
kind/failing-test priority/critical-urgent sig/scheduling
<!-- Please only use this template for submitting reports about continuously failing tests or jobs in Kubernetes CI --> **Which jobs are failing**: `Conformance - GCE - master (ci-kubernetes-gce-conformance-latest)` **Which test(s) are failing**: `[sig-scheduling] SchedulerPreemption [Serial] PreemptionExecutionPath runs ReplicaSets to verify preemption running path [Conformance] ` **Since when has it been failing**: `06-09-20 22:27 PDT` **Testgrid link**: https://testgrid.k8s.io/sig-release-master-blocking#Conformance%20-%20GCE%20-%20master **Reason for failure**: ``` /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Jun 9 15:35:40.256: failed pod observation expectations: rs1 had more than 2 pods created: 3 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/preemption.go:655 ``` **Anything else we need to know**: /sig scheduling /cc @kubernetes/ci-signal /priority critical-urgent /milestone v1.19
1.0
[Failing Test] [sig-scheduling] SchedulerPreemption [Serial] PreemptionExecutionPath runs ReplicaSets to verify preemption running path [Conformance] - <!-- Please only use this template for submitting reports about continuously failing tests or jobs in Kubernetes CI --> **Which jobs are failing**: `Conformance - GCE - master (ci-kubernetes-gce-conformance-latest)` **Which test(s) are failing**: `[sig-scheduling] SchedulerPreemption [Serial] PreemptionExecutionPath runs ReplicaSets to verify preemption running path [Conformance] ` **Since when has it been failing**: `06-09-20 22:27 PDT` **Testgrid link**: https://testgrid.k8s.io/sig-release-master-blocking#Conformance%20-%20GCE%20-%20master **Reason for failure**: ``` /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Jun 9 15:35:40.256: failed pod observation expectations: rs1 had more than 2 pods created: 3 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/preemption.go:655 ``` **Anything else we need to know**: /sig scheduling /cc @kubernetes/ci-signal /priority critical-urgent /milestone v1.19
non_defect
schedulerpreemption preemptionexecutionpath runs replicasets to verify preemption running path which jobs are failing conformance gce master ci kubernetes gce conformance latest which test s are failing schedulerpreemption preemptionexecutionpath runs replicasets to verify preemption running path since when has it been failing pdt testgrid link reason for failure go src io kubernetes output dockerized go src io kubernetes test framework framework go jun failed pod observation expectations had more than pods created go src io kubernetes output dockerized go src io kubernetes test scheduling preemption go anything else we need to know sig scheduling cc kubernetes ci signal priority critical urgent milestone
0
376,400
26,198,767,739
IssuesEvent
2023-01-03 15:39:12
taikoxyz/taiko-mono
https://api.github.com/repos/taikoxyz/taiko-mono
closed
feat(docs): add documentation on how to generate a private key
documentation good first issue dx
### Describe the feature request There are a lot of questions about how to generate a private key for this documentation: https://taiko.xyz/docs/alpha-1-testnet/deploy-a-contract#steps Let's add some steps on how to do it. There are two decent ways to do this, I would highly suggest to use a dummy account and not the private key of an account that has any real funds. You can create a dummy account keypair manually with openssl (something like this https://gist.github.com/miguelmota/3793b160992b4ea0b616497b8e5aee2f) , or you can just create an account with metamask and export the private key. We should keep in mind that not everyone has metamask. Perhaps we can even suggest usage of a useless private key if they don't care about the contract deployer (however doing this means that address will not show up when we scan the network for the POAP). ### Describe alternatives you've considered None ### Additional context None
1.0
feat(docs): add documentation on how to generate a private key - ### Describe the feature request There are a lot of questions about how to generate a private key for this documentation: https://taiko.xyz/docs/alpha-1-testnet/deploy-a-contract#steps Let's add some steps on how to do it. There are two decent ways to do this, I would highly suggest to use a dummy account and not the private key of an account that has any real funds. You can create a dummy account keypair manually with openssl (something like this https://gist.github.com/miguelmota/3793b160992b4ea0b616497b8e5aee2f) , or you can just create an account with metamask and export the private key. We should keep in mind that not everyone has metamask. Perhaps we can even suggest usage of a useless private key if they don't care about the contract deployer (however doing this means that address will not show up when we scan the network for the POAP). ### Describe alternatives you've considered None ### Additional context None
non_defect
feat docs add documentation on how to generate a private key describe the feature request there are a lot of questions about how to generate a private key for this documentation let s add some steps on how to do it there are two decent ways to do this i would highly suggest to use a dummy account and not the private key of an account that has any real funds you can create a dummy account keypair manually with openssl something like this or you can just create an account with metamask and export the private key we should keep in mind that not everyone has metamask perhaps we can even suggest usage of a useless private key if they don t care about the contract deployer however doing this means that address will not show up when we scan the network for the poap describe alternatives you ve considered none additional context none
0
41,328
10,417,616,108
IssuesEvent
2019-09-14 23:34:01
avereon/xenon
https://api.github.com/repos/avereon/xenon
closed
Guide tool icons not transparent
bug / error / defect
The guide tool icons do not appear to be using the alpha channel when rendering. This may be due to CSS configuration, guide renderer configuration or some other rendering configuration that does not allow those pixels to be translucent.
1.0
Guide tool icons not transparent - The guide tool icons do not appear to be using the alpha channel when rendering. This may be due to CSS configuration, guide renderer configuration or some other rendering configuration that does not allow those pixels to be translucent.
defect
guide tool icons not transparent the guide tool icons do not appear to be using the alpha channel when rendering this may be due to css configuration guide renderer configuration or some other rendering configuration that does not allow those pixels to be translucent
1
23,673
3,851,865,309
IssuesEvent
2016-04-06 05:27:47
GPF/imame4all
https://api.github.com/repos/GPF/imame4all
closed
Steelseries Free Controler support.
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Buy a SteelSeries Free Controller 2. Connect it in either Gamepad or Arcade (icade) mode 3. Basic buttons are mapped to coin and start and other buttons dont map right due to the way mame4ios seems to map buttons to functions and then those functions to game buttons and actions. When I remap the buttons, I have to assign "button 1" to "coin" (and other weird combinations) which messes up other things. What is the expected output? What do you see instead? I was hoping to remap a key to "coin" instead of the key being mapped to "coin" then me having to map "coin" to something else. Or if there was a way to support the weird order of keys the Steelseries Free spits out. What version of the product are you using? On what operating system? Latest as of Feb 27 (the Feb9release of mame4ios reloaded) Please provide any additional information below. I will donate a SteelSeries Free Controller if there is interest in support. Supporting the Free's Arcade mode should be supper simple as it sends key down/key up just like any other iCade device but the Free sends different combinations for buttons than the others making the orientations funny. The Free also can be a full analog gamepad controller with dual analog sticks. This is what would be awesome to have supported as well. Again, let me know if interest in support and I will donate a controller. ``` Original issue reported on code.google.com by `RyanStep...@gmail.com` on 28 Feb 2013 at 12:58
1.0
Steelseries Free Controler support. - ``` What steps will reproduce the problem? 1. Buy a SteelSeries Free Controller 2. Connect it in either Gamepad or Arcade (icade) mode 3. Basic buttons are mapped to coin and start and other buttons dont map right due to the way mame4ios seems to map buttons to functions and then those functions to game buttons and actions. When I remap the buttons, I have to assign "button 1" to "coin" (and other weird combinations) which messes up other things. What is the expected output? What do you see instead? I was hoping to remap a key to "coin" instead of the key being mapped to "coin" then me having to map "coin" to something else. Or if there was a way to support the weird order of keys the Steelseries Free spits out. What version of the product are you using? On what operating system? Latest as of Feb 27 (the Feb9release of mame4ios reloaded) Please provide any additional information below. I will donate a SteelSeries Free Controller if there is interest in support. Supporting the Free's Arcade mode should be supper simple as it sends key down/key up just like any other iCade device but the Free sends different combinations for buttons than the others making the orientations funny. The Free also can be a full analog gamepad controller with dual analog sticks. This is what would be awesome to have supported as well. Again, let me know if interest in support and I will donate a controller. ``` Original issue reported on code.google.com by `RyanStep...@gmail.com` on 28 Feb 2013 at 12:58
defect
steelseries free controler support what steps will reproduce the problem buy a steelseries free controller connect it in either gamepad or arcade icade mode basic buttons are mapped to coin and start and other buttons dont map right due to the way seems to map buttons to functions and then those functions to game buttons and actions when i remap the buttons i have to assign button to coin and other weird combinations which messes up other things what is the expected output what do you see instead i was hoping to remap a key to coin instead of the key being mapped to coin then me having to map coin to something else or if there was a way to support the weird order of keys the steelseries free spits out what version of the product are you using on what operating system latest as of feb the of reloaded please provide any additional information below i will donate a steelseries free controller if there is interest in support supporting the free s arcade mode should be supper simple as it sends key down key up just like any other icade device but the free sends different combinations for buttons than the others making the orientations funny the free also can be a full analog gamepad controller with dual analog sticks this is what would be awesome to have supported as well again let me know if interest in support and i will donate a controller original issue reported on code google com by ryanstep gmail com on feb at
1
51,719
13,211,296,750
IssuesEvent
2020-08-15 22:07:29
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
[photospline] - divzero - loop preconditions aren't checked (Trac #922)
Incomplete Migration Migrated from Trac combo reconstruction defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/922">https://code.icecube.wisc.edu/projects/icecube/ticket/922</a>, reported by negaand owned by jvansanten</em></summary> <p> ```json { "status": "closed", "changetime": "2015-04-12T17:41:42", "_ts": "1428860502423619", "description": "http://goo.gl/FFVVbF\n\nloop preconditions aren't checked allowing a potential divide-by-zero error to occur in 7 steps.\n\nfix: pre-check and hard fail if loop pre-conditions suck", "reporter": "nega", "cc": "", "resolution": "fixed", "time": "2015-04-10T04:17:37", "component": "combo reconstruction", "summary": "[photospline] - divzero - loop preconditions aren't checked", "priority": "normal", "keywords": "photospline divzero", "milestone": "", "owner": "jvansanten", "type": "defect" } ``` </p> </details>
1.0
[photospline] - divzero - loop preconditions aren't checked (Trac #922) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/922">https://code.icecube.wisc.edu/projects/icecube/ticket/922</a>, reported by negaand owned by jvansanten</em></summary> <p> ```json { "status": "closed", "changetime": "2015-04-12T17:41:42", "_ts": "1428860502423619", "description": "http://goo.gl/FFVVbF\n\nloop preconditions aren't checked allowing a potential divide-by-zero error to occur in 7 steps.\n\nfix: pre-check and hard fail if loop pre-conditions suck", "reporter": "nega", "cc": "", "resolution": "fixed", "time": "2015-04-10T04:17:37", "component": "combo reconstruction", "summary": "[photospline] - divzero - loop preconditions aren't checked", "priority": "normal", "keywords": "photospline divzero", "milestone": "", "owner": "jvansanten", "type": "defect" } ``` </p> </details>
defect
divzero loop preconditions aren t checked trac migrated from json status closed changetime ts description preconditions aren t checked allowing a potential divide by zero error to occur in steps n nfix pre check and hard fail if loop pre conditions suck reporter nega cc resolution fixed time component combo reconstruction summary divzero loop preconditions aren t checked priority normal keywords photospline divzero milestone owner jvansanten type defect
1
65,453
6,964,020,386
IssuesEvent
2017-12-08 19:49:09
legchikov/coding_interview
https://api.github.com/repos/legchikov/coding_interview
closed
[Arrays and Strings][Tests] Is Unique
test
Implement an algorithm to determine if a string has all unique characters. What if you cannot use additional data structures?
1.0
[Arrays and Strings][Tests] Is Unique - Implement an algorithm to determine if a string has all unique characters. What if you cannot use additional data structures?
non_defect
is unique implement an algorithm to determine if a string has all unique characters what if you cannot use additional data structures
0
32,834
6,953,397,182
IssuesEvent
2017-12-06 20:52:48
Dzhuneyt/jquery-tubular
https://api.github.com/repos/Dzhuneyt/jquery-tubular
closed
Video is not playing in Android device
auto-migrated Priority-Medium Type-Defect
``` I am using your tubular js for playing the video in background. it's working well few days ago. but now when i opened the website in my android phone the video doesn't show. and also the folllowing url doesn't show video in background in android device. http://www.seanmccambridge.com/tubular/ ``` Original issue reported on code.google.com by `jitendra...@technosoftwares.in` on 26 May 2014 at 2:21
1.0
Video is not playing in Android device - ``` I am using your tubular js for playing the video in background. it's working well few days ago. but now when i opened the website in my android phone the video doesn't show. and also the folllowing url doesn't show video in background in android device. http://www.seanmccambridge.com/tubular/ ``` Original issue reported on code.google.com by `jitendra...@technosoftwares.in` on 26 May 2014 at 2:21
defect
video is not playing in android device i am using your tubular js for playing the video in background it s working well few days ago but now when i opened the website in my android phone the video doesn t show and also the folllowing url doesn t show video in background in android device original issue reported on code google com by jitendra technosoftwares in on may at
1
160,836
6,103,449,580
IssuesEvent
2017-06-20 18:44:08
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
opened
[studio] Bulk publish tries to publish items that area already published
enhancement Priority: Medium
We need the bulk publish algorithm to the following if possible: - Query the database for all items that are not published - For that list, add item to publish queue Ping me if you have any questions.
1.0
[studio] Bulk publish tries to publish items that area already published - We need the bulk publish algorithm to the following if possible: - Query the database for all items that are not published - For that list, add item to publish queue Ping me if you have any questions.
non_defect
bulk publish tries to publish items that area already published we need the bulk publish algorithm to the following if possible query the database for all items that are not published for that list add item to publish queue ping me if you have any questions
0
71,792
23,804,601,000
IssuesEvent
2022-09-03 21:02:48
ascott18/TellMeWhen
https://api.github.com/repos/ascott18/TellMeWhen
closed
[Bug]: Swing Timer bug with flurry
T: defect S: resolved
### WoW Version TBC Classic ### TellMeWhen Version 9.2.4 r10 ### Describe the bug Swing timer is sometimes inaccurate, possibly due to when it checks for flurry being up or not. [Video of the behavior](https://www.youtube.com/watch?v=GHjEt8qJBGM) [screenshot](https://imgur.com/oZGoDgs) ### Export Strings ```text N/A ```
1.0
[Bug]: Swing Timer bug with flurry - ### WoW Version TBC Classic ### TellMeWhen Version 9.2.4 r10 ### Describe the bug Swing timer is sometimes inaccurate, possibly due to when it checks for flurry being up or not. [Video of the behavior](https://www.youtube.com/watch?v=GHjEt8qJBGM) [screenshot](https://imgur.com/oZGoDgs) ### Export Strings ```text N/A ```
defect
swing timer bug with flurry wow version tbc classic tellmewhen version describe the bug swing timer is sometimes inaccurate possibly due to when it checks for flurry being up or not export strings text n a
1
46,645
5,825,420,871
IssuesEvent
2017-05-07 21:18:14
mnc123/uapp
https://api.github.com/repos/mnc123/uapp
closed
Variaciรณn de pantalla de espera
Test
Cuando hay 5 usuarios en cola cambia la pรกgina de espera, esto me parece correcto pero hay que cambiarlo de vista cuando ya este adentro de los primeros 5 (o 4 no recuerdo), tambiรฉn ver en detalle cรณmo varรญa la pantalla de espera en estos casos.
1.0
Variaciรณn de pantalla de espera - Cuando hay 5 usuarios en cola cambia la pรกgina de espera, esto me parece correcto pero hay que cambiarlo de vista cuando ya este adentro de los primeros 5 (o 4 no recuerdo), tambiรฉn ver en detalle cรณmo varรญa la pantalla de espera en estos casos.
non_defect
variaciรณn de pantalla de espera cuando hay usuarios en cola cambia la pรกgina de espera esto me parece correcto pero hay que cambiarlo de vista cuando ya este adentro de los primeros o no recuerdo tambiรฉn ver en detalle cรณmo varรญa la pantalla de espera en estos casos
0
42,832
11,299,257,540
IssuesEvent
2020-01-17 10:47:23
vector-im/riot-web
https://api.github.com/repos/vector-im/riot-web
opened
Complete security incorrectly sends you through key backup restore
bug defect p1 story:1
Tom tested the complete security flow on his account with key backup, and when clicking start, he was sent into a restore key backup flow, but that's not intended. The expected thing is you are prompted for the SSSS passphrase and the device is verified.
1.0
Complete security incorrectly sends you through key backup restore - Tom tested the complete security flow on his account with key backup, and when clicking start, he was sent into a restore key backup flow, but that's not intended. The expected thing is you are prompted for the SSSS passphrase and the device is verified.
defect
complete security incorrectly sends you through key backup restore tom tested the complete security flow on his account with key backup and when clicking start he was sent into a restore key backup flow but that s not intended the expected thing is you are prompted for the ssss passphrase and the device is verified
1
357,524
25,176,399,412
IssuesEvent
2022-11-11 09:38:46
rama-pang/pe
https://api.github.com/repos/rama-pang/pe
opened
Numbering error in Use Case
severity.VeryLow type.DocumentationBug
In Use Case UC6 - Sorting the tutor list, Extensions section, number 1b; the next command after this has number "1a1" instead of "1b1". <!--session: 1668152096750-d2fd9f28-c252-473b-8fc4-fdb88cf3f718--> <!--Version: Web v3.4.4-->
1.0
Numbering error in Use Case - In Use Case UC6 - Sorting the tutor list, Extensions section, number 1b; the next command after this has number "1a1" instead of "1b1". <!--session: 1668152096750-d2fd9f28-c252-473b-8fc4-fdb88cf3f718--> <!--Version: Web v3.4.4-->
non_defect
numbering error in use case in use case sorting the tutor list extensions section number the next command after this has number instead of
0
126
2,511,982,053
IssuesEvent
2015-01-14 13:11:56
enthought/chaco
https://api.github.com/repos/enthought/chaco
opened
cntr.c could use an update
defect
The code in `cntr.c` is outdated and a little buggy (as I discovered when I tried to pass the optional mask to the `Cntr` type). It appears to be a clone of [this code](https://github.com/matplotlib/matplotlib/blob/master/src/cntr.c) in matplotlib, which looks a little more up-to-date.
1.0
cntr.c could use an update - The code in `cntr.c` is outdated and a little buggy (as I discovered when I tried to pass the optional mask to the `Cntr` type). It appears to be a clone of [this code](https://github.com/matplotlib/matplotlib/blob/master/src/cntr.c) in matplotlib, which looks a little more up-to-date.
defect
cntr c could use an update the code in cntr c is outdated and a little buggy as i discovered when i tried to pass the optional mask to the cntr type it appears to be a clone of in matplotlib which looks a little more up to date
1
55,814
23,609,640,486
IssuesEvent
2022-08-24 11:16:52
microsoft/vscode-cpptools
https://api.github.com/repos/microsoft/vscode-cpptools
closed
Language service is idle, "Invalid opened file instance. Ignoring IntelliSense message for file" for multiroot workspace
bug Language Service more info needed Feature: Multiroot Not Repro - Internal
Bug type: Language Service <!-- Prior to creating a bug report, please review: ๐Ÿ“ Existing issues at https://github.com/Microsoft/vscode-cpptools/issues ๐Ÿ“œ Our documentation at https://code.visualstudio.com/docs/languages/cpp ๐Ÿ“™ FAQs at https://code.visualstudio.com/docs/cpp/faq-cpp --> **Describe the bug** - OS and Version: Windows 10 21H2 - VS Code Version: 1.66 - C/C++ Extension Version: 1.9.7 - Other extensions you installed (and if the issue persists after disabling them): C/C++ Extension pack v1.1.0 (without CMake-related extensions). Issue still persists. - If using SSH remote, specify OS of remote machine: Windows Server 2019 - A clear and concise description of what the bug is, including information about the workspace (i.e. is the workspace a single project or multiple projects, size of the project, etc). **Steps to reproduce** 1. Connect via SSH to remote Windows OS; 2. Add the directory that is adjacent to the `${workspaceFolder}` in step 4 (it should not have `.vscode` folder unlike the next two); 3. Add several sub-directories of a `${workspaceFolder}`; 4. Add the `${workspaceFolder}` itself; 5. Switch "C_Cpp: Logging Level" to "Debug" 6. Hovering over C++ code will reveal `Unsupported Markup content received. Kind is:` 7. In "C/C++" Output window there should be `Invalid opened file instance. Ignoring IntelliSense message for file X` **Expected behavior** IntelliSense is expected to work on hover. **Code sample and logs** <!-- Please provide code sample, your c_cpp_properties.json and logs. --> - Configurations in `c_cpp_properties.json` ```json { "configurations": [ { "name": "Win32", "includePath": [ "${workspaceFolder}/subdir1/**", "${workspaceFolder}/subdir2/**" ], "defines": [ "_DEBUG", "UNICODE", "_UNICODE" ], "cStandard": "c17", "cppStandard": "c++17", } ], "version": 4 } ``` - Logs from running `C/C++: Log Diagnostics` from the VS Code command palette ``` -------- Diagnostics - 3/31/2022, 2:50:23 PM Version: 1.9.7 Current Configuration: { "name": "Win32", "includePath": [ "${workspaceFolder}/**", "C:/Program Files (x86)/Intel/oneAPI/compiler/latest/windows/include/sycl" ], "defines": [ "_DEBUG", "UNICODE", "_UNICODE" ], "windowsSdkVersion": "10.0.17763.0", "compilerPath": "cl.exe", "compilerPathIsExplicit": true, "cStandardIsExplicit": true, "cppStandardIsExplicit": true, "intelliSenseModeIsExplicit": false, "compilerArgs": [], "cStandard": "c17", "cppStandard": "c++17", "mergeConfigurations": false, "browse": { "path": [ "${workspaceFolder}/**", "C:/Program Files (x86)/Intel/oneAPI/compiler/latest/windows/include/sycl" ], "limitSymbolsToIncludedHeaders": true } } No active translation units. ``` - Logs from [the language server logging](https://code.visualstudio.com/docs/cpp/enable-logging-cpp#_enable-logging-for-the-language-server) ``` loggingLevel: Debug cpptools/didChangeCppProperties Attempting to get defaults from C++ compiler in "compilerPath" property: 'cl.exe' Attempting to get defaults from C compiler in "compilerPath" property: 'cl.exe' Folder: C:/PROGRAM FILES (X86)/MICROSOFT VISUAL STUDIO/2017/PROFESSIONAL/VC/TOOLS/MSVC/14.16.27023/ATLMFC/INCLUDE/* will be indexed Folder: C:/PROGRAM FILES (X86)/MICROSOFT VISUAL STUDIO/2017/PROFESSIONAL/VC/TOOLS/MSVC/14.16.27023/INCLUDE/* will be indexed Folder: C:/PROGRAM FILES (X86)/WINDOWS KITS/10/INCLUDE/10.0.17763.0/CPPWINRT/ will be indexed Folder: C:/PROGRAM FILES (X86)/WINDOWS KITS/10/INCLUDE/10.0.17763.0/SHARED/ will be indexed Folder: C:/PROGRAM FILES (X86)/WINDOWS KITS/10/INCLUDE/10.0.17763.0/UCRT/ will be indexed Folder: C:/PROGRAM FILES (X86)/WINDOWS KITS/10/INCLUDE/10.0.17763.0/UM/ will be indexed Folder: C:/PROGRAM FILES (X86)/WINDOWS KITS/10/INCLUDE/10.0.17763.0/WINRT/ will be indexed Folder: C:/<WORKSPACE FOLDER>/ will be indexed Folder: C:/PROGRAM FILES (X86)/INTEL/ONEAPI/COMPILER/LATEST/WINDOWS/INCLUDE/SYCL/ will be indexed textDocument/didOpen: <FILE1> cpptools/textEditorSelectionChange cpptools/activeDocumentChange: <FILE1> cpptools/getDocumentSymbols: <FILE1> (id: 2) cpptools/getCodeActions: <FILE1> (id: 3) cpptools/getFoldingRanges: <FILE1> (id: 4) cpptools/getSemanticTokens: <FILE1> (id: 5) textDocument/hover: <FILE1> (id: 6) Invalid opened file instance. Ignoring IntelliSense message for file <FILE1>. cpptools/getDiagnostics (id: 10) ``` **Additional context** This bug might be related to #8762, so steps to reproduce should be similar. **Possible workarounds** - exclude step 2;
1.0
Language service is idle, "Invalid opened file instance. Ignoring IntelliSense message for file" for multiroot workspace - Bug type: Language Service <!-- Prior to creating a bug report, please review: ๐Ÿ“ Existing issues at https://github.com/Microsoft/vscode-cpptools/issues ๐Ÿ“œ Our documentation at https://code.visualstudio.com/docs/languages/cpp ๐Ÿ“™ FAQs at https://code.visualstudio.com/docs/cpp/faq-cpp --> **Describe the bug** - OS and Version: Windows 10 21H2 - VS Code Version: 1.66 - C/C++ Extension Version: 1.9.7 - Other extensions you installed (and if the issue persists after disabling them): C/C++ Extension pack v1.1.0 (without CMake-related extensions). Issue still persists. - If using SSH remote, specify OS of remote machine: Windows Server 2019 - A clear and concise description of what the bug is, including information about the workspace (i.e. is the workspace a single project or multiple projects, size of the project, etc). **Steps to reproduce** 1. Connect via SSH to remote Windows OS; 2. Add the directory that is adjacent to the `${workspaceFolder}` in step 4 (it should not have `.vscode` folder unlike the next two); 3. Add several sub-directories of a `${workspaceFolder}`; 4. Add the `${workspaceFolder}` itself; 5. Switch "C_Cpp: Logging Level" to "Debug" 6. Hovering over C++ code will reveal `Unsupported Markup content received. Kind is:` 7. In "C/C++" Output window there should be `Invalid opened file instance. Ignoring IntelliSense message for file X` **Expected behavior** IntelliSense is expected to work on hover. **Code sample and logs** <!-- Please provide code sample, your c_cpp_properties.json and logs. --> - Configurations in `c_cpp_properties.json` ```json { "configurations": [ { "name": "Win32", "includePath": [ "${workspaceFolder}/subdir1/**", "${workspaceFolder}/subdir2/**" ], "defines": [ "_DEBUG", "UNICODE", "_UNICODE" ], "cStandard": "c17", "cppStandard": "c++17", } ], "version": 4 } ``` - Logs from running `C/C++: Log Diagnostics` from the VS Code command palette ``` -------- Diagnostics - 3/31/2022, 2:50:23 PM Version: 1.9.7 Current Configuration: { "name": "Win32", "includePath": [ "${workspaceFolder}/**", "C:/Program Files (x86)/Intel/oneAPI/compiler/latest/windows/include/sycl" ], "defines": [ "_DEBUG", "UNICODE", "_UNICODE" ], "windowsSdkVersion": "10.0.17763.0", "compilerPath": "cl.exe", "compilerPathIsExplicit": true, "cStandardIsExplicit": true, "cppStandardIsExplicit": true, "intelliSenseModeIsExplicit": false, "compilerArgs": [], "cStandard": "c17", "cppStandard": "c++17", "mergeConfigurations": false, "browse": { "path": [ "${workspaceFolder}/**", "C:/Program Files (x86)/Intel/oneAPI/compiler/latest/windows/include/sycl" ], "limitSymbolsToIncludedHeaders": true } } No active translation units. ``` - Logs from [the language server logging](https://code.visualstudio.com/docs/cpp/enable-logging-cpp#_enable-logging-for-the-language-server) ``` loggingLevel: Debug cpptools/didChangeCppProperties Attempting to get defaults from C++ compiler in "compilerPath" property: 'cl.exe' Attempting to get defaults from C compiler in "compilerPath" property: 'cl.exe' Folder: C:/PROGRAM FILES (X86)/MICROSOFT VISUAL STUDIO/2017/PROFESSIONAL/VC/TOOLS/MSVC/14.16.27023/ATLMFC/INCLUDE/* will be indexed Folder: C:/PROGRAM FILES (X86)/MICROSOFT VISUAL STUDIO/2017/PROFESSIONAL/VC/TOOLS/MSVC/14.16.27023/INCLUDE/* will be indexed Folder: C:/PROGRAM FILES (X86)/WINDOWS KITS/10/INCLUDE/10.0.17763.0/CPPWINRT/ will be indexed Folder: C:/PROGRAM FILES (X86)/WINDOWS KITS/10/INCLUDE/10.0.17763.0/SHARED/ will be indexed Folder: C:/PROGRAM FILES (X86)/WINDOWS KITS/10/INCLUDE/10.0.17763.0/UCRT/ will be indexed Folder: C:/PROGRAM FILES (X86)/WINDOWS KITS/10/INCLUDE/10.0.17763.0/UM/ will be indexed Folder: C:/PROGRAM FILES (X86)/WINDOWS KITS/10/INCLUDE/10.0.17763.0/WINRT/ will be indexed Folder: C:/<WORKSPACE FOLDER>/ will be indexed Folder: C:/PROGRAM FILES (X86)/INTEL/ONEAPI/COMPILER/LATEST/WINDOWS/INCLUDE/SYCL/ will be indexed textDocument/didOpen: <FILE1> cpptools/textEditorSelectionChange cpptools/activeDocumentChange: <FILE1> cpptools/getDocumentSymbols: <FILE1> (id: 2) cpptools/getCodeActions: <FILE1> (id: 3) cpptools/getFoldingRanges: <FILE1> (id: 4) cpptools/getSemanticTokens: <FILE1> (id: 5) textDocument/hover: <FILE1> (id: 6) Invalid opened file instance. Ignoring IntelliSense message for file <FILE1>. cpptools/getDiagnostics (id: 10) ``` **Additional context** This bug might be related to #8762, so steps to reproduce should be similar. **Possible workarounds** - exclude step 2;
non_defect
language service is idle invalid opened file instance ignoring intellisense message for file for multiroot workspace bug type language service prior to creating a bug report please review ๐Ÿ“ existing issues at ๐Ÿ“œ our documentation at ๐Ÿ“™ faqs at describe the bug os and version windows vs code version c c extension version other extensions you installed and if the issue persists after disabling them c c extension pack without cmake related extensions issue still persists if using ssh remote specify os of remote machine windows server a clear and concise description of what the bug is including information about the workspace i e is the workspace a single project or multiple projects size of the project etc steps to reproduce connect via ssh to remote windows os add the directory that is adjacent to the workspacefolder in step it should not have vscode folder unlike the next two add several sub directories of a workspacefolder add the workspacefolder itself switch c cpp logging level to debug hovering over c code will reveal unsupported markup content received kind is in c c output window there should be invalid opened file instance ignoring intellisense message for file x expected behavior intellisense is expected to work on hover code sample and logs configurations in c cpp properties json json configurations name includepath workspacefolder workspacefolder defines debug unicode unicode cstandard cppstandard c version logs from running c c log diagnostics from the vs code command palette diagnostics pm version current configuration name includepath workspacefolder c program files intel oneapi compiler latest windows include sycl defines debug unicode unicode windowssdkversion compilerpath cl exe compilerpathisexplicit true cstandardisexplicit true cppstandardisexplicit true intellisensemodeisexplicit false compilerargs cstandard cppstandard c mergeconfigurations false browse path workspacefolder c program files intel oneapi compiler latest windows include sycl limitsymbolstoincludedheaders true no active translation units logs from logginglevel debug cpptools didchangecppproperties attempting to get defaults from c compiler in compilerpath property cl exe attempting to get defaults from c compiler in compilerpath property cl exe folder c program files microsoft visual studio professional vc tools msvc atlmfc include will be indexed folder c program files microsoft visual studio professional vc tools msvc include will be indexed folder c program files windows kits include cppwinrt will be indexed folder c program files windows kits include shared will be indexed folder c program files windows kits include ucrt will be indexed folder c program files windows kits include um will be indexed folder c program files windows kits include winrt will be indexed folder c will be indexed folder c program files intel oneapi compiler latest windows include sycl will be indexed textdocument didopen cpptools texteditorselectionchange cpptools activedocumentchange cpptools getdocumentsymbols id cpptools getcodeactions id cpptools getfoldingranges id cpptools getsemantictokens id textdocument hover id invalid opened file instance ignoring intellisense message for file cpptools getdiagnostics id additional context this bug might be related to so steps to reproduce should be similar possible workarounds exclude step
0
309,736
23,304,228,421
IssuesEvent
2022-08-07 19:34:44
bitwes/Gut
https://api.github.com/repos/bitwes/Gut
closed
In the wiki, anchors for assert_typeof and assert_nottypeof are wrong
documentation
They link to https://github.com/bitwes/Gut/wiki/%24assert_typeof instead of https://github.com/bitwes/Gut/wiki/Asserts-and-Methods#assert_typeof
1.0
In the wiki, anchors for assert_typeof and assert_nottypeof are wrong - They link to https://github.com/bitwes/Gut/wiki/%24assert_typeof instead of https://github.com/bitwes/Gut/wiki/Asserts-and-Methods#assert_typeof
non_defect
in the wiki anchors for assert typeof and assert nottypeof are wrong they link to instead of
0
16,836
2,948,324,068
IssuesEvent
2015-07-06 01:32:22
Winetricks/winetricks
https://api.github.com/repos/Winetricks/winetricks
closed
New Steam Verb: "Magic: The Gathering - Duels of the Planeswalkers 2012"
auto-migrated Priority-Medium Type-Defect
``` Attached verb will install this game through Steam as per the instructions in AddingSteamGames, with no workarounds used. As of now however, sound does not work in-game, but this is an upstream bug in Wine 1.3.29 itself. There is also occasional crashes, but again I couldn't find a workaround in AppDB. (I'll report these when I get an opportunity to do so) Tested on Mac OS X 10.7.1 (Lion) under Wine 1.3.29 and Winetricks revision 20110629. ``` Original issue reported on code.google.com by `aersunst...@googlemail.com` on 9 Oct 2011 at 11:35 Attachments: * [mtg_dotp2012_steam.verb](https://storage.googleapis.com/google-code-attachments/winetricks/issue-135/comment-0/mtg_dotp2012_steam.verb)
1.0
New Steam Verb: "Magic: The Gathering - Duels of the Planeswalkers 2012" - ``` Attached verb will install this game through Steam as per the instructions in AddingSteamGames, with no workarounds used. As of now however, sound does not work in-game, but this is an upstream bug in Wine 1.3.29 itself. There is also occasional crashes, but again I couldn't find a workaround in AppDB. (I'll report these when I get an opportunity to do so) Tested on Mac OS X 10.7.1 (Lion) under Wine 1.3.29 and Winetricks revision 20110629. ``` Original issue reported on code.google.com by `aersunst...@googlemail.com` on 9 Oct 2011 at 11:35 Attachments: * [mtg_dotp2012_steam.verb](https://storage.googleapis.com/google-code-attachments/winetricks/issue-135/comment-0/mtg_dotp2012_steam.verb)
defect
new steam verb magic the gathering duels of the planeswalkers attached verb will install this game through steam as per the instructions in addingsteamgames with no workarounds used as of now however sound does not work in game but this is an upstream bug in wine itself there is also occasional crashes but again i couldn t find a workaround in appdb i ll report these when i get an opportunity to do so tested on mac os x lion under wine and winetricks revision original issue reported on code google com by aersunst googlemail com on oct at attachments
1
214,790
7,276,786,175
IssuesEvent
2018-02-21 17:21:12
TylerConlee/slab
https://api.github.com/repos/TylerConlee/slab
closed
Display CSAT as part of ticket notification
enhancement priority:normal
On the More Info screen, previous ticket CSATs can be displayed to give some context as to the attitude of the customer towards the support given. With a history of how the customer responds, an agent can then tailor the messaging even further to ensure the customer gets the best service possible.
1.0
Display CSAT as part of ticket notification - On the More Info screen, previous ticket CSATs can be displayed to give some context as to the attitude of the customer towards the support given. With a history of how the customer responds, an agent can then tailor the messaging even further to ensure the customer gets the best service possible.
non_defect
display csat as part of ticket notification on the more info screen previous ticket csats can be displayed to give some context as to the attitude of the customer towards the support given with a history of how the customer responds an agent can then tailor the messaging even further to ensure the customer gets the best service possible
0
31,148
4,691,651,642
IssuesEvent
2016-10-11 11:22:56
skaut/SRS
https://api.github.com/repos/skaut/SRS
closed
Hromadnรฉ pล™iล™azovรกnรญ rolรญ
Testovรกnรญ
Tak jako se nynรญ dรก hromadnฤ› uลพivatelลฏm zadat platba a tisk dokladลฏ v evidenci รบฤastnรญkลฏ, prosรญm o รบpravu, aby se hromadnฤ› daly zadรกvat i role.
1.0
Hromadnรฉ pล™iล™azovรกnรญ rolรญ - Tak jako se nynรญ dรก hromadnฤ› uลพivatelลฏm zadat platba a tisk dokladลฏ v evidenci รบฤastnรญkลฏ, prosรญm o รบpravu, aby se hromadnฤ› daly zadรกvat i role.
non_defect
hromadnรฉ pล™iล™azovรกnรญ rolรญ tak jako se nynรญ dรก hromadnฤ› uลพivatelลฏm zadat platba a tisk dokladลฏ v evidenci รบฤastnรญkลฏ prosรญm o รบpravu aby se hromadnฤ› daly zadรกvat i role
0
121,867
4,822,453,788
IssuesEvent
2016-11-05 21:22:28
ReactiveX/rxjs
https://api.github.com/repos/ReactiveX/rxjs
closed
Materialize: Notification property incorrectly is `exception` rather than `error` for `kind: "E"`
priority: critical type: bug
<!-- Thank you for raising your concerns, we appreciate your feedback and contributions to this repository. Before you continue, consider the following: If you have a "How do I do ...?" question, it is better for you and for us that this question is placed in [StackOverflow](http://stackoverflow.com/questions/tagged/rxjs5) or some chat channel. This way, you are making it easier for others to learn from your experiences too. These "Issues" are meant only for technical problems, bugs, and proposals related to the library. If your issue is a bug, please follow the format below: --> **RxJS version:** 5.0.0-rc.1 **Code to reproduce:** ```js Rx.Observable.throw(new Error('sad')) .materialize() .map(n => n.error.message) .subscribe(x => console.log(x)) ``` **Expected behavior:** logs `"sad"` **Actual behavior:** Throws cannot read "message" of undefined **Additional information:** Currently `error` is in the `exception` field. Required to upgrade falcor-router to Rx5
1.0
Materialize: Notification property incorrectly is `exception` rather than `error` for `kind: "E"` - <!-- Thank you for raising your concerns, we appreciate your feedback and contributions to this repository. Before you continue, consider the following: If you have a "How do I do ...?" question, it is better for you and for us that this question is placed in [StackOverflow](http://stackoverflow.com/questions/tagged/rxjs5) or some chat channel. This way, you are making it easier for others to learn from your experiences too. These "Issues" are meant only for technical problems, bugs, and proposals related to the library. If your issue is a bug, please follow the format below: --> **RxJS version:** 5.0.0-rc.1 **Code to reproduce:** ```js Rx.Observable.throw(new Error('sad')) .materialize() .map(n => n.error.message) .subscribe(x => console.log(x)) ``` **Expected behavior:** logs `"sad"` **Actual behavior:** Throws cannot read "message" of undefined **Additional information:** Currently `error` is in the `exception` field. Required to upgrade falcor-router to Rx5
non_defect
materialize notification property incorrectly is exception rather than error for kind e thank you for raising your concerns we appreciate your feedback and contributions to this repository before you continue consider the following if you have a how do i do question it is better for you and for us that this question is placed in or some chat channel this way you are making it easier for others to learn from your experiences too these issues are meant only for technical problems bugs and proposals related to the library if your issue is a bug please follow the format below rxjs version rc code to reproduce js rx observable throw new error sad materialize map n n error message subscribe x console log x expected behavior logs sad actual behavior throws cannot read message of undefined additional information currently error is in the exception field required to upgrade falcor router to
0
33,304
27,370,658,707
IssuesEvent
2023-02-27 23:10:10
pypa/packaging
https://api.github.com/repos/pypa/packaging
closed
Set up permissions to github workflows
infrastructure
I have suggested a PR to setuptools repo to harden permissions to the github workflows https://github.com/pypa/setuptools/pull/3833, if that's ok, I also would like to do the same to the packaging repo in order to increase supply-chain security by limiting the impact of an eventual compromised workflow. Let me know if a PR is also welcome in this repo and I'll submit it ASAP.
1.0
Set up permissions to github workflows - I have suggested a PR to setuptools repo to harden permissions to the github workflows https://github.com/pypa/setuptools/pull/3833, if that's ok, I also would like to do the same to the packaging repo in order to increase supply-chain security by limiting the impact of an eventual compromised workflow. Let me know if a PR is also welcome in this repo and I'll submit it ASAP.
non_defect
set up permissions to github workflows i have suggested a pr to setuptools repo to harden permissions to the github workflows if that s ok i also would like to do the same to the packaging repo in order to increase supply chain security by limiting the impact of an eventual compromised workflow let me know if a pr is also welcome in this repo and i ll submit it asap
0
64,420
18,670,712,558
IssuesEvent
2021-10-30 16:55:27
theseion/Fuel
https://api.github.com/repos/theseion/Fuel
closed
Fully implement DiskProxy
Priority-Medium auto-migrated Type-Defect stale
``` Fully implement DiskProxy (with any object as argument). it could be necessary to change our format to "headerStep [instStep refStep]+ trailerStep" ``` Original issue reported on code.google.com by `marianopeck` on 1 Dec 2011 at 12:17
1.0
Fully implement DiskProxy - ``` Fully implement DiskProxy (with any object as argument). it could be necessary to change our format to "headerStep [instStep refStep]+ trailerStep" ``` Original issue reported on code.google.com by `marianopeck` on 1 Dec 2011 at 12:17
defect
fully implement diskproxy fully implement diskproxy with any object as argument it could be necessary to change our format to headerstep trailerstep original issue reported on code google com by marianopeck on dec at
1
346,011
10,383,015,832
IssuesEvent
2019-09-10 08:46:50
acidanthera/bugtracker
https://api.github.com/repos/acidanthera/bugtracker
closed
ART Frequency, CPU Frequency and Clock Drift
priority:high project:oc
OCCPU is misreporting my CPU's frequency. I believe it's also causing the macOS system clock to run fast: ``` 00:226 00:113 OCCPU: Found Intel(R) Xeon(R) Gold 6136 CPU @ 3.00GHz 00:345 00:118 OCCPU: Signature 50654 Stepping 4 Model 55 Family 6 Type 0 ExtModel 5 ExtFamily 0 00:461 00:116 OCCPU: Detected Apple Processor Type: 0F -> 0F01 00:575 00:114 OCCPU: Ratio Min 12 Max 30 Current 36 Turbo 37 36 36 36 00:695 00:119 OCCPU: ART Frequency 2880000000 2880MHz 24000000 * 240 / 2 = 2880000000 00:821 00:126 OCCPU: TSC Frequency 2992972839 2992MHz 00:944 00:122 OCCPU: CPU Frequency 2880000000 2880MHz 01:068 00:123 OCCPU: FSB Frequency 96000000 96MHz 01:206 00:138 OCCPU: Pkg 1 Cores 12 Threads 24 ``` Looking at `OcCpu.c`, the following lines look suspect to me: ``` AsmCpuid (CPUID_TIME_STAMP_COUNTER, &CpuidEax, &CpuidEbx, NULL, NULL); if (CpuidEax > 0 && CpuidEbx > 0) { Cpu->CPUFrequency = MultU64x32 (BASE_ART_CLOCK_SOURCE, (UINT32) DivU64x32 (CpuidEbx, CpuidEax)); ``` https://github.com/acidanthera/OcSupportPkg/blob/3e98de1f89a9e946f026ecc24e4856e2096dc338/Library/OcCpuLib/OcCpuLib.c#L667 Should that be storing to `Cpu->ARTFrequency` instead? I tested by making this change and setting `Cpu->CPUFrequency` from `Cpu->TSCFrequency`. `About this Mac` now reports the correct frequency (3Ghz instead of 2.88Ghz). Also, the clock no longer drifts whereas it used to run a few seconds per minute faster than real-time.
1.0
ART Frequency, CPU Frequency and Clock Drift - OCCPU is misreporting my CPU's frequency. I believe it's also causing the macOS system clock to run fast: ``` 00:226 00:113 OCCPU: Found Intel(R) Xeon(R) Gold 6136 CPU @ 3.00GHz 00:345 00:118 OCCPU: Signature 50654 Stepping 4 Model 55 Family 6 Type 0 ExtModel 5 ExtFamily 0 00:461 00:116 OCCPU: Detected Apple Processor Type: 0F -> 0F01 00:575 00:114 OCCPU: Ratio Min 12 Max 30 Current 36 Turbo 37 36 36 36 00:695 00:119 OCCPU: ART Frequency 2880000000 2880MHz 24000000 * 240 / 2 = 2880000000 00:821 00:126 OCCPU: TSC Frequency 2992972839 2992MHz 00:944 00:122 OCCPU: CPU Frequency 2880000000 2880MHz 01:068 00:123 OCCPU: FSB Frequency 96000000 96MHz 01:206 00:138 OCCPU: Pkg 1 Cores 12 Threads 24 ``` Looking at `OcCpu.c`, the following lines look suspect to me: ``` AsmCpuid (CPUID_TIME_STAMP_COUNTER, &CpuidEax, &CpuidEbx, NULL, NULL); if (CpuidEax > 0 && CpuidEbx > 0) { Cpu->CPUFrequency = MultU64x32 (BASE_ART_CLOCK_SOURCE, (UINT32) DivU64x32 (CpuidEbx, CpuidEax)); ``` https://github.com/acidanthera/OcSupportPkg/blob/3e98de1f89a9e946f026ecc24e4856e2096dc338/Library/OcCpuLib/OcCpuLib.c#L667 Should that be storing to `Cpu->ARTFrequency` instead? I tested by making this change and setting `Cpu->CPUFrequency` from `Cpu->TSCFrequency`. `About this Mac` now reports the correct frequency (3Ghz instead of 2.88Ghz). Also, the clock no longer drifts whereas it used to run a few seconds per minute faster than real-time.
non_defect
art frequency cpu frequency and clock drift occpu is misreporting my cpu s frequency i believe it s also causing the macos system clock to run fast occpu found intel r xeon r gold cpu occpu signature stepping model family type extmodel extfamily occpu detected apple processor type occpu ratio min max current turbo occpu art frequency occpu tsc frequency occpu cpu frequency occpu fsb frequency occpu pkg cores threads looking at occpu c the following lines look suspect to me asmcpuid cpuid time stamp counter cpuideax cpuidebx null null if cpuideax cpuidebx cpu cpufrequency base art clock source cpuidebx cpuideax should that be storing to cpu artfrequency instead i tested by making this change and setting cpu cpufrequency from cpu tscfrequency about this mac now reports the correct frequency instead of also the clock no longer drifts whereas it used to run a few seconds per minute faster than real time
0
31,669
6,583,368,914
IssuesEvent
2017-09-13 05:15:18
kuenzign/dyknow-panel-extractor
https://api.github.com/repos/kuenzign/dyknow-panel-extractor
closed
Links in the about box do not load a web page
auto-migrated Priority-Medium Project-DPXCommon Type-Defect
_From @GoogleCodeExporter on March 15, 2015 13:14_ ``` When you click on a link in the about box, it does not load the intended web page in the default browser. ``` Original issue reported on code.google.com by `jjhatf02` on 28 Apr 2010 at 3:32 _Copied from original issue: JaredHatfield/dyknow-panel-extractor#11_
1.0
Links in the about box do not load a web page - _From @GoogleCodeExporter on March 15, 2015 13:14_ ``` When you click on a link in the about box, it does not load the intended web page in the default browser. ``` Original issue reported on code.google.com by `jjhatf02` on 28 Apr 2010 at 3:32 _Copied from original issue: JaredHatfield/dyknow-panel-extractor#11_
defect
links in the about box do not load a web page from googlecodeexporter on march when you click on a link in the about box it does not load the intended web page in the default browser original issue reported on code google com by on apr at copied from original issue jaredhatfield dyknow panel extractor
1
52,447
10,863,340,405
IssuesEvent
2019-11-14 14:57:50
USGCRP/gcis
https://api.github.com/repos/USGCRP/gcis
closed
Error building GCIS under Perl 5.22+
a bug context Core Code context IT priority low type technical
``` Can't load 'blib/arch/auto/Algorithm/Permute/Permute.so' for module Algorithm::Permute: blib/arch/auto/Algorithm/Permute/Permute.so: undefined symbol: PUSHBLOCK at /usr/local/lib/perl5/5.24.0/x86_64-linux/DynaLoader.pm line 193. at test.pl line 13. Compilation failed in require at test.pl line 13. BEGIN failed--compilation aborted at test.pl line 13. ``` Possibly related to using Perl 5.24. See [this issue](https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=825012).
1.0
Error building GCIS under Perl 5.22+ - ``` Can't load 'blib/arch/auto/Algorithm/Permute/Permute.so' for module Algorithm::Permute: blib/arch/auto/Algorithm/Permute/Permute.so: undefined symbol: PUSHBLOCK at /usr/local/lib/perl5/5.24.0/x86_64-linux/DynaLoader.pm line 193. at test.pl line 13. Compilation failed in require at test.pl line 13. BEGIN failed--compilation aborted at test.pl line 13. ``` Possibly related to using Perl 5.24. See [this issue](https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=825012).
non_defect
error building gcis under perl can t load blib arch auto algorithm permute permute so for module algorithm permute blib arch auto algorithm permute permute so undefined symbol pushblock at usr local lib linux dynaloader pm line at test pl line compilation failed in require at test pl line begin failed compilation aborted at test pl line possibly related to using perl see
0
20,690
3,407,672,115
IssuesEvent
2015-12-04 05:03:12
jarrodek/ChromeRestClient
https://api.github.com/repos/jarrodek/ChromeRestClient
closed
Unable to install the rest client
OpSys-Windows Priority-High Status-New Type-Defect
Originally reported on Google Code with ID 296 ``` What steps will reproduce the problem? 1. Access below url https://chrome.google.com/webstore/detail/advanced-rest-client/hgmloofddffdnphfgcellkdfbfbjeloo 2.click on FREE and proceed with installation 3.Installation fails with error "An error has occurred Advanced REST client (extension ID "hgmloofddffdnphfgcellkdfbfbjeloo") is blocked by the administrator." What is the expected output? What do you see instead? Installation of Advanced Rest client On what operating system, browser and browser version? Windows 2007, 64 bit, Chrome Version 37.0.2062.102 m Please provide any additional information below. The error is "An error has occurred Advanced REST client (extension ID "hgmloofddffdnphfgcellkdfbfbjeloo") is blocked by the administrator." ``` Reported by `VishalVShroff` on 2014-12-29 09:34:58
1.0
Unable to install the rest client - Originally reported on Google Code with ID 296 ``` What steps will reproduce the problem? 1. Access below url https://chrome.google.com/webstore/detail/advanced-rest-client/hgmloofddffdnphfgcellkdfbfbjeloo 2.click on FREE and proceed with installation 3.Installation fails with error "An error has occurred Advanced REST client (extension ID "hgmloofddffdnphfgcellkdfbfbjeloo") is blocked by the administrator." What is the expected output? What do you see instead? Installation of Advanced Rest client On what operating system, browser and browser version? Windows 2007, 64 bit, Chrome Version 37.0.2062.102 m Please provide any additional information below. The error is "An error has occurred Advanced REST client (extension ID "hgmloofddffdnphfgcellkdfbfbjeloo") is blocked by the administrator." ``` Reported by `VishalVShroff` on 2014-12-29 09:34:58
defect
unable to install the rest client originally reported on google code with id what steps will reproduce the problem access below url click on free and proceed with installation installation fails with error an error has occurred advanced rest client extension id hgmloofddffdnphfgcellkdfbfbjeloo is blocked by the administrator what is the expected output what do you see instead installation of advanced rest client on what operating system browser and browser version windows bit chrome version m please provide any additional information below the error is an error has occurred advanced rest client extension id hgmloofddffdnphfgcellkdfbfbjeloo is blocked by the administrator reported by vishalvshroff on
1
40,446
9,998,937,114
IssuesEvent
2019-07-12 09:24:31
contao/contao
https://api.github.com/repos/contao/contao
closed
new version of leafo/scssphp
defect
**Affected version(s)** all 4.4 and above **Description** there is a new version for leafo/scssphp v0.8.2 which fixes some major issues... could you please update the requirements
1.0
new version of leafo/scssphp - **Affected version(s)** all 4.4 and above **Description** there is a new version for leafo/scssphp v0.8.2 which fixes some major issues... could you please update the requirements
defect
new version of leafo scssphp affected version s all and above description there is a new version for leafo scssphp which fixes some major issues could you please update the requirements
1
374,756
26,130,450,486
IssuesEvent
2022-12-29 03:36:25
CameronRedmore/memory-deck
https://api.github.com/repos/CameronRedmore/memory-deck
opened
Create wiki pages
documentation enhancement
- [ ] How the plugin works - [ ] List dependencies - [ ] Instructions on how to run main.py manually for debugging purposes, potentially enable that functionality by default
1.0
Create wiki pages - - [ ] How the plugin works - [ ] List dependencies - [ ] Instructions on how to run main.py manually for debugging purposes, potentially enable that functionality by default
non_defect
create wiki pages how the plugin works list dependencies instructions on how to run main py manually for debugging purposes potentially enable that functionality by default
0
76,788
26,598,842,628
IssuesEvent
2023-01-23 14:26:12
vector-im/element-ios
https://api.github.com/repos/vector-im/element-ios
closed
1 translation error was found.
T-Defect A-Timeline S-Major O-Frequent
### Steps to reproduce n/a ### Outcome ![{1EE58C0F-D292-6620-FA47-03567BA0A3B9}](https://user-images.githubusercontent.com/48104960/210496447-1fea8556-ee16-4b10-bb63-710a3ebccf7f.png) **The target language is Simplified Chinese, for Traditional Chinese, the translation here is correct.** ### Your phone model iPhone 7 Plus ### Operating system version iOS 15.6.1 ### Application version Element iOS 1.9.14 ### Homeserver Synapse 1.74.0 ### Will you send logs? No
1.0
1 translation error was found. - ### Steps to reproduce n/a ### Outcome ![{1EE58C0F-D292-6620-FA47-03567BA0A3B9}](https://user-images.githubusercontent.com/48104960/210496447-1fea8556-ee16-4b10-bb63-710a3ebccf7f.png) **The target language is Simplified Chinese, for Traditional Chinese, the translation here is correct.** ### Your phone model iPhone 7 Plus ### Operating system version iOS 15.6.1 ### Application version Element iOS 1.9.14 ### Homeserver Synapse 1.74.0 ### Will you send logs? No
defect
translation error was found steps to reproduce n a outcome the target language is simplified chinese for traditional chinese the translation here is correct your phone model iphone plus operating system version ios application version element ios homeserver synapse will you send logs no
1
811,963
30,308,066,395
IssuesEvent
2023-07-10 10:52:20
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
securitycenter.snippets.snippets_mute_config_test: test_bulk_mute_findings failed
priority: p2 type: bug api: securitycenter samples flakybot: issue flakybot: flaky
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: d07634990df61b990c3a6908ce1502c2a6d69752 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/0f52d7e2-8f21-43d7-8582-2c6e0c4df6be), [Sponge](http://sponge2/0f52d7e2-8f21-43d7-8582-2c6e0c4df6be) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 72, in error_remapped_callable return callable_(*args, **kwargs) File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 1030, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 910, in _end_unary_response_blocking raise _InactiveRpcError(state) # pytype: disable=not-instantiable grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.INTERNAL details = "Internal error encountered." debug_error_string = "UNKNOWN:Error received from peer ipv4:74.125.195.95:443 {created_time:"2023-06-27T11:09:30.860723791+00:00", grpc_status:13, grpc_message:"Internal error encountered."}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/securitycenter/snippets/snippets_mute_config_test.py", line 121, in test_bulk_mute_findings f"projects/{PROJECT_ID}", f'resource.project_display_name="{PROJECT_ID}"' File "/workspace/securitycenter/snippets/snippets_mute_config.py", line 260, in bulk_mute_findings response = client.bulk_mute_findings(request) File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/cloud/securitycenter_v1/services/security_center/client.py", line 875, in bulk_mute_findings metadata=metadata, File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/gapic_v1/method.py", line 113, in __call__ return wrapped_func(*args, **kwargs) File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 74, in error_remapped_callable raise exceptions.from_grpc_error(exc) from exc google.api_core.exceptions.InternalServerError: 500 Internal error encountered.</pre></details>
1.0
securitycenter.snippets.snippets_mute_config_test: test_bulk_mute_findings failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: d07634990df61b990c3a6908ce1502c2a6d69752 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/0f52d7e2-8f21-43d7-8582-2c6e0c4df6be), [Sponge](http://sponge2/0f52d7e2-8f21-43d7-8582-2c6e0c4df6be) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 72, in error_remapped_callable return callable_(*args, **kwargs) File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 1030, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 910, in _end_unary_response_blocking raise _InactiveRpcError(state) # pytype: disable=not-instantiable grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.INTERNAL details = "Internal error encountered." debug_error_string = "UNKNOWN:Error received from peer ipv4:74.125.195.95:443 {created_time:"2023-06-27T11:09:30.860723791+00:00", grpc_status:13, grpc_message:"Internal error encountered."}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/securitycenter/snippets/snippets_mute_config_test.py", line 121, in test_bulk_mute_findings f"projects/{PROJECT_ID}", f'resource.project_display_name="{PROJECT_ID}"' File "/workspace/securitycenter/snippets/snippets_mute_config.py", line 260, in bulk_mute_findings response = client.bulk_mute_findings(request) File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/cloud/securitycenter_v1/services/security_center/client.py", line 875, in bulk_mute_findings metadata=metadata, File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/gapic_v1/method.py", line 113, in __call__ return wrapped_func(*args, **kwargs) File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 74, in error_remapped_callable raise exceptions.from_grpc_error(exc) from exc google.api_core.exceptions.InternalServerError: 500 Internal error encountered.</pre></details>
non_defect
securitycenter snippets snippets mute config test test bulk mute findings failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output traceback most recent call last file workspace securitycenter snippets nox py lib site packages google api core grpc helpers py line in error remapped callable return callable args kwargs file workspace securitycenter snippets nox py lib site packages grpc channel py line in call return end unary response blocking state call false none file workspace securitycenter snippets nox py lib site packages grpc channel py line in end unary response blocking raise inactiverpcerror state pytype disable not instantiable grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with status statuscode internal details internal error encountered debug error string unknown error received from peer created time grpc status grpc message internal error encountered the above exception was the direct cause of the following exception traceback most recent call last file workspace securitycenter snippets snippets mute config test py line in test bulk mute findings f projects project id f resource project display name project id file workspace securitycenter snippets snippets mute config py line in bulk mute findings response client bulk mute findings request file workspace securitycenter snippets nox py lib site packages google cloud securitycenter services security center client py line in bulk mute findings metadata metadata file workspace securitycenter snippets nox py lib site packages google api core gapic method py line in call return wrapped func args kwargs file workspace securitycenter snippets nox py lib site packages google api core grpc helpers py line in error remapped callable raise exceptions from grpc error exc from exc google api core exceptions internalservererror internal error encountered
0
55,852
8,032,543,746
IssuesEvent
2018-07-28 16:38:53
nf-core/cookiecutter
https://api.github.com/repos/nf-core/cookiecutter
opened
Documentation redundancy
documentation
The files `local.md` and `adding_your_own.md` have similar (if not identical) sections for Docker and Singularity - we should remove these in one location at least :-)
1.0
Documentation redundancy - The files `local.md` and `adding_your_own.md` have similar (if not identical) sections for Docker and Singularity - we should remove these in one location at least :-)
non_defect
documentation redundancy the files local md and adding your own md have similar if not identical sections for docker and singularity we should remove these in one location at least
0
45,197
2,920,859,454
IssuesEvent
2015-06-24 21:08:28
vickychijwani/quill
https://api.github.com/repos/vickychijwani/quill
reopened
Memory leak in WebViewFragment
bug help wanted high-priority
Steps to reproduce: 1. Open a post preview (the new WebView-based one) 2. Select some text 3. Tap back twice to come back to the post list 4. Open up the post again and watch LeakCanary dump a heap Note: - [x] Test fix on older Android versions (webkit vs chromium) - [x] Merge BrowserActivity and WebViewFragment (or fix and test *both*) ---- Unfortunately the heap trace (below) is unhelpful: ``` 06-23 03:53:53.152 22607-25907/me.vickychijwani.spectre D/LeakCanary๏น• In me.vickychijwani.spectre:1.0:1. * me.vickychijwani.spectre.view.PostViewActivity has leaked: * GC ROOT org.chromium.content.browser.input.PopupTouchHandleDrawable.mContext * leaks me.vickychijwani.spectre.view.PostViewActivity instance * Reference Key: c5347412-4c57-49a2-9df7-9760955cedeb * Device: LGE google Nexus 4 occam * Android Version: 5.1.1 API: 22 * Durations: watch=5124ms, gc=162ms, heap dump=11630ms, analysis=43875ms ``` Maybe analyzing the [actual heap dump](https://www.dropbox.com/s/zodudiuacbctjcr/quill_issue_75.hprof?dl=0) will help.
1.0
Memory leak in WebViewFragment - Steps to reproduce: 1. Open a post preview (the new WebView-based one) 2. Select some text 3. Tap back twice to come back to the post list 4. Open up the post again and watch LeakCanary dump a heap Note: - [x] Test fix on older Android versions (webkit vs chromium) - [x] Merge BrowserActivity and WebViewFragment (or fix and test *both*) ---- Unfortunately the heap trace (below) is unhelpful: ``` 06-23 03:53:53.152 22607-25907/me.vickychijwani.spectre D/LeakCanary๏น• In me.vickychijwani.spectre:1.0:1. * me.vickychijwani.spectre.view.PostViewActivity has leaked: * GC ROOT org.chromium.content.browser.input.PopupTouchHandleDrawable.mContext * leaks me.vickychijwani.spectre.view.PostViewActivity instance * Reference Key: c5347412-4c57-49a2-9df7-9760955cedeb * Device: LGE google Nexus 4 occam * Android Version: 5.1.1 API: 22 * Durations: watch=5124ms, gc=162ms, heap dump=11630ms, analysis=43875ms ``` Maybe analyzing the [actual heap dump](https://www.dropbox.com/s/zodudiuacbctjcr/quill_issue_75.hprof?dl=0) will help.
non_defect
memory leak in webviewfragment steps to reproduce open a post preview the new webview based one select some text tap back twice to come back to the post list open up the post again and watch leakcanary dump a heap note test fix on older android versions webkit vs chromium merge browseractivity and webviewfragment or fix and test both unfortunately the heap trace below is unhelpful me vickychijwani spectre d leakcanary๏น• in me vickychijwani spectre me vickychijwani spectre view postviewactivity has leaked gc root org chromium content browser input popuptouchhandledrawable mcontext leaks me vickychijwani spectre view postviewactivity instance reference key device lge google nexus occam android version api durations watch gc heap dump analysis maybe analyzing the will help
0
144,864
5,547,156,747
IssuesEvent
2017-03-23 04:11:23
ClaytonPassmore/ProjectOrange
https://api.github.com/repos/ClaytonPassmore/ProjectOrange
closed
Develop test case to validate NFP
Priority: Medium Server
We claimed that we should be able to handle 100 decks with 500 cards each. Generate test data so that we can verify this claim.
1.0
Develop test case to validate NFP - We claimed that we should be able to handle 100 decks with 500 cards each. Generate test data so that we can verify this claim.
non_defect
develop test case to validate nfp we claimed that we should be able to handle decks with cards each generate test data so that we can verify this claim
0
49,852
13,187,281,185
IssuesEvent
2020-08-13 02:55:03
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
Neutrino simulation event has a pulse width of zero (Trac #2180)
Incomplete Migration Migrated from Trac combo simulation defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2180">https://code.icecube.wisc.edu/ticket/2180</a>, reported by lwille and owned by juancarlos</em></summary> <p> ```json { "status": "closed", "changetime": "2019-03-27T16:25:07", "description": "I'm running into a Monopod error with about 10% of the files in Nancy's NuGen_new simulation set. The error is \"FATAL (millipede): Assertion failed: p->GetWidth() > 0 (MillipedeDOMCacheMap.cxx:379 in void MillipedeDOMCacheMap::UpdateData(const I3TimeWindow&, const I3RecoPulseSeriesMap&, const I3TimeWindowSeriesMap&, double, double, double, bool))\".\n\nLooking into this error, it appears that about 10% of the time, an event will have a single pulse with a width reported as 0. I'm uncertain if this is an issue with the simulation file processing or something else. An example of an event with a zero pulse width is here `/data/ana/Cscd/StartingEvents/NuGen_new/NuTau/medium_energy/IC86_flasher_p1=0.3_p2=0.0_domeff_081/l2/1/l2_00000248.i3.zst` Event 450.", "reporter": "lwille", "cc": "jvansanten, nwhitehorn", "resolution": "duplicate", "_ts": "1553703907985721", "component": "combo simulation", "summary": "Neutrino simulation event has a pulse width of zero", "priority": "normal", "keywords": "", "time": "2018-08-08T18:33:10", "milestone": "Vernal Equinox 2019", "owner": "juancarlos", "type": "defect" } ``` </p> </details>
1.0
Neutrino simulation event has a pulse width of zero (Trac #2180) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2180">https://code.icecube.wisc.edu/ticket/2180</a>, reported by lwille and owned by juancarlos</em></summary> <p> ```json { "status": "closed", "changetime": "2019-03-27T16:25:07", "description": "I'm running into a Monopod error with about 10% of the files in Nancy's NuGen_new simulation set. The error is \"FATAL (millipede): Assertion failed: p->GetWidth() > 0 (MillipedeDOMCacheMap.cxx:379 in void MillipedeDOMCacheMap::UpdateData(const I3TimeWindow&, const I3RecoPulseSeriesMap&, const I3TimeWindowSeriesMap&, double, double, double, bool))\".\n\nLooking into this error, it appears that about 10% of the time, an event will have a single pulse with a width reported as 0. I'm uncertain if this is an issue with the simulation file processing or something else. An example of an event with a zero pulse width is here `/data/ana/Cscd/StartingEvents/NuGen_new/NuTau/medium_energy/IC86_flasher_p1=0.3_p2=0.0_domeff_081/l2/1/l2_00000248.i3.zst` Event 450.", "reporter": "lwille", "cc": "jvansanten, nwhitehorn", "resolution": "duplicate", "_ts": "1553703907985721", "component": "combo simulation", "summary": "Neutrino simulation event has a pulse width of zero", "priority": "normal", "keywords": "", "time": "2018-08-08T18:33:10", "milestone": "Vernal Equinox 2019", "owner": "juancarlos", "type": "defect" } ``` </p> </details>
defect
neutrino simulation event has a pulse width of zero trac migrated from json status closed changetime description i m running into a monopod error with about of the files in nancy s nugen new simulation set the error is fatal millipede assertion failed p getwidth millipededomcachemap cxx in void millipededomcachemap updatedata const const const double double double bool n nlooking into this error it appears that about of the time an event will have a single pulse with a width reported as i m uncertain if this is an issue with the simulation file processing or something else an example of an event with a zero pulse width is here data ana cscd startingevents nugen new nutau medium energy flasher domeff zst event reporter lwille cc jvansanten nwhitehorn resolution duplicate ts component combo simulation summary neutrino simulation event has a pulse width of zero priority normal keywords time milestone vernal equinox owner juancarlos type defect
1
430,940
30,208,283,102
IssuesEvent
2023-07-05 10:58:20
scylladb/scylladb
https://api.github.com/repos/scylladb/scylladb
opened
doc: improve the docs for integration with Elasticsearch
Documentation
https://opensource.docs.scylladb.com/stable/using-scylla/integrations/integration-elasticsearch.html Feedback from @zseta : > My guess would be that the user might prefer a short and to-the-point page and not a list of links that lead to long-form content. It might be worthwhile to pull out the code bits from [this blog](https://www.scylladb.com/2019/03/07/scylla-and-elasticsearch-part-two-practical-examples-to-support-full-text-search-workloads/), and add it on [this page](https://opensource.docs.scylladb.com/stable/using-scylla/integrations/integration-elasticsearch.html) so the user can quickly grab the code that they need straight from docs. > > The [Databricks example](https://opensource.docs.scylladb.com/stable/using-scylla/integrations/integration-databricks.html) is a good one from this perspective, because it provides the instructions on-page and don't need to read long blogs to find the solution.
1.0
doc: improve the docs for integration with Elasticsearch - https://opensource.docs.scylladb.com/stable/using-scylla/integrations/integration-elasticsearch.html Feedback from @zseta : > My guess would be that the user might prefer a short and to-the-point page and not a list of links that lead to long-form content. It might be worthwhile to pull out the code bits from [this blog](https://www.scylladb.com/2019/03/07/scylla-and-elasticsearch-part-two-practical-examples-to-support-full-text-search-workloads/), and add it on [this page](https://opensource.docs.scylladb.com/stable/using-scylla/integrations/integration-elasticsearch.html) so the user can quickly grab the code that they need straight from docs. > > The [Databricks example](https://opensource.docs.scylladb.com/stable/using-scylla/integrations/integration-databricks.html) is a good one from this perspective, because it provides the instructions on-page and don't need to read long blogs to find the solution.
non_defect
doc improve the docs for integration with elasticsearch feedback from zseta my guess would be that the user might prefer a short and to the point page and not a list of links that lead to long form content it might be worthwhile to pull out the code bits from and add it on so the user can quickly grab the code that they need straight from docs the is a good one from this perspective because it provides the instructions on page and don t need to read long blogs to find the solution
0
42,015
10,741,402,326
IssuesEvent
2019-10-29 20:10:07
cakephp/bake
https://api.github.com/repos/cakephp/bake
closed
4.x Empty layout templates created for `bin/cake bake mailer Ticket`
Defect
* Using `cakephp/bake 4.x-dev 1d28d25` * Running `bin/cake bake mailer Ticket` produces ``` Creating file /var/virtual/cakefest2019/src/../templates/layout/email/html/ticket.php Could not write to `/var/virtual/cakefest2019/src/../templates/layout/email/html/ticket.php`. Creating file /var/virtual/cakefest2019/src/../templates/layout/email/text/ticket.php Could not write to `/var/virtual/cakefest2019/src/../templates/layout/email/text/ticket.php`. Creating file /var/virtual/cakefest2019/src/Mailer/TicketMailer.php Wrote `/var/virtual/cakefest2019/src/Mailer/TicketMailer.php` Baking test case for App\Mailer\TicketMailer ... Creating file /var/virtual/cakefest2019/tests/TestCase/Mailer/TicketMailerTest.php Wrote `/var/virtual/cakefest2019/tests/TestCase/Mailer/TicketMailerTest.php` ``` Note file path containing `..` causing issues writing the templates
1.0
4.x Empty layout templates created for `bin/cake bake mailer Ticket` - * Using `cakephp/bake 4.x-dev 1d28d25` * Running `bin/cake bake mailer Ticket` produces ``` Creating file /var/virtual/cakefest2019/src/../templates/layout/email/html/ticket.php Could not write to `/var/virtual/cakefest2019/src/../templates/layout/email/html/ticket.php`. Creating file /var/virtual/cakefest2019/src/../templates/layout/email/text/ticket.php Could not write to `/var/virtual/cakefest2019/src/../templates/layout/email/text/ticket.php`. Creating file /var/virtual/cakefest2019/src/Mailer/TicketMailer.php Wrote `/var/virtual/cakefest2019/src/Mailer/TicketMailer.php` Baking test case for App\Mailer\TicketMailer ... Creating file /var/virtual/cakefest2019/tests/TestCase/Mailer/TicketMailerTest.php Wrote `/var/virtual/cakefest2019/tests/TestCase/Mailer/TicketMailerTest.php` ``` Note file path containing `..` causing issues writing the templates
defect
x empty layout templates created for bin cake bake mailer ticket using cakephp bake x dev running bin cake bake mailer ticket produces creating file var virtual src templates layout email html ticket php could not write to var virtual src templates layout email html ticket php creating file var virtual src templates layout email text ticket php could not write to var virtual src templates layout email text ticket php creating file var virtual src mailer ticketmailer php wrote var virtual src mailer ticketmailer php baking test case for app mailer ticketmailer creating file var virtual tests testcase mailer ticketmailertest php wrote var virtual tests testcase mailer ticketmailertest php note file path containing causing issues writing the templates
1
4,017
2,544,716,145
IssuesEvent
2015-01-29 12:19:13
pychess/pychess
https://api.github.com/repos/pychess/pychess
closed
Analog clock in chessclock
Component-UI enhancement EyeCandy imported Milestone-Release1.0 Priority-Low
_From [lobais](https://code.google.com/u/lobais/) on August 13, 2006 16:30:38_ We could put a little analog clock, next to the digital time. Would look nice :) _Original issue: http://code.google.com/p/pychess/issues/detail?id=7_
1.0
Analog clock in chessclock - _From [lobais](https://code.google.com/u/lobais/) on August 13, 2006 16:30:38_ We could put a little analog clock, next to the digital time. Would look nice :) _Original issue: http://code.google.com/p/pychess/issues/detail?id=7_
non_defect
analog clock in chessclock from on august we could put a little analog clock next to the digital time would look nice original issue
0
69,141
22,203,474,009
IssuesEvent
2022-06-07 13:12:03
matrix-org/matrix-hookshot
https://api.github.com/repos/matrix-org/matrix-hookshot
closed
Use the RequiresClient widget capability to disable popout button in Element
T-Defect
The hookshot config window can't work without identity verification, so [set RequiresClient](https://github.com/matrix-org/matrix-react-sdk/pull/7005)
1.0
Use the RequiresClient widget capability to disable popout button in Element - The hookshot config window can't work without identity verification, so [set RequiresClient](https://github.com/matrix-org/matrix-react-sdk/pull/7005)
defect
use the requiresclient widget capability to disable popout button in element the hookshot config window can t work without identity verification so
1