Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
757
| labels
stringlengths 4
664
| body
stringlengths 3
261k
| index
stringclasses 10
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
232k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
19,230
| 3,155,728,719
|
IssuesEvent
|
2015-09-17 10:33:40
|
bigbluebutton/bigbluebutton
|
https://api.github.com/repos/bigbluebutton/bigbluebutton
|
closed
|
If try to change microphone when entering in BBB, it enters in a loop
|
Defect Normal Priority
|
Originally reported on Google Code with ID 1896
```
When you enter in BBB, you select No when you receive the question: have you heard your
voice? Then you select a different microphone and push Next. It never finishes. You
have to reload the page.
```
Reported by `smoral@adhoclearning.com` on 2015-03-06 14:44:53
|
1.0
|
If try to change microphone when entering in BBB, it enters in a loop - Originally reported on Google Code with ID 1896
```
When you enter in BBB, you select No when you receive the question: have you heard your
voice? Then you select a different microphone and push Next. It never finishes. You
have to reload the page.
```
Reported by `smoral@adhoclearning.com` on 2015-03-06 14:44:53
|
defect
|
if try to change microphone when entering in bbb it enters in a loop originally reported on google code with id when you enter in bbb you select no when you receive the question have you heard your voice then you select a different microphone and push next it never finishes you have to reload the page reported by smoral adhoclearning com on
| 1
|
89,638
| 25,862,949,281
|
IssuesEvent
|
2022-12-13 18:21:35
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
libtorch_python.so: undefined symbol: PyInstanceMethod_Type
|
module: build triaged topic: build
|
### ๐ Describe the bug
I run into this error vvv
```
C++ exception with description "/var/lib/jenkins/multipy/multipy/runtime/loader.cpp:780: libnvfuser_python.so: could not load library, dlopen says: /opt/conda/lib/python3.10/site-packages/torch/lib/libtorch_python.so: undefined symbol: PyInstanceMethod_Type" thrown in the test body.
```
link of failing CI: https://github.com/pytorch/pytorch/actions/runs/3586511045/jobs/6036274959
cmake lines related to the object: https://github.com/pytorch/pytorch/blob/d8335704a54f89e93f562cda7b49048e1eab1ebb/torch/csrc/jit/codegen/cuda/CMakeLists.txt#L165-L200
It's a little strange to me how the error was thrown from libtorch_python.so. This feels like a pytorch build issue.
I'm also a little bit uncertain how this is invoked.
In the setup, I have `libnvfuser_python.so` which depends on both `libnvfuser_codegen.so` and `libtorch_python.so`. While `libnvfuser_codegen.so` is dlopened by torch and does not have dependency on `libnvfuser_python.so`.
### Versions
This is failing on CI job:
[linux-bionic-cuda11.6-py3.10-gcc7 / test (deploy, 1, 1, linux.4xlarge.nvidia.gpu)](https://github.com/pytorch/pytorch/actions/runs/3590997095/jobs/6045485004)
cc @malfet @seemethere
|
2.0
|
libtorch_python.so: undefined symbol: PyInstanceMethod_Type - ### ๐ Describe the bug
I run into this error vvv
```
C++ exception with description "/var/lib/jenkins/multipy/multipy/runtime/loader.cpp:780: libnvfuser_python.so: could not load library, dlopen says: /opt/conda/lib/python3.10/site-packages/torch/lib/libtorch_python.so: undefined symbol: PyInstanceMethod_Type" thrown in the test body.
```
link of failing CI: https://github.com/pytorch/pytorch/actions/runs/3586511045/jobs/6036274959
cmake lines related to the object: https://github.com/pytorch/pytorch/blob/d8335704a54f89e93f562cda7b49048e1eab1ebb/torch/csrc/jit/codegen/cuda/CMakeLists.txt#L165-L200
It's a little strange to me how the error was thrown from libtorch_python.so. This feels like a pytorch build issue.
I'm also a little bit uncertain how this is invoked.
In the setup, I have `libnvfuser_python.so` which depends on both `libnvfuser_codegen.so` and `libtorch_python.so`. While `libnvfuser_codegen.so` is dlopened by torch and does not have dependency on `libnvfuser_python.so`.
### Versions
This is failing on CI job:
[linux-bionic-cuda11.6-py3.10-gcc7 / test (deploy, 1, 1, linux.4xlarge.nvidia.gpu)](https://github.com/pytorch/pytorch/actions/runs/3590997095/jobs/6045485004)
cc @malfet @seemethere
|
non_defect
|
libtorch python so undefined symbol pyinstancemethod type ๐ describe the bug i run into this error vvv c exception with description var lib jenkins multipy multipy runtime loader cpp libnvfuser python so could not load library dlopen says opt conda lib site packages torch lib libtorch python so undefined symbol pyinstancemethod type thrown in the test body link of failing ci cmake lines related to the object it s a little strange to me how the error was thrown from libtorch python so this feels like a pytorch build issue i m also a little bit uncertain how this is invoked in the setup i have libnvfuser python so which depends on both libnvfuser codegen so and libtorch python so while libnvfuser codegen so is dlopened by torch and does not have dependency on libnvfuser python so versions this is failing on ci job cc malfet seemethere
| 0
|
2,538
| 2,607,926,695
|
IssuesEvent
|
2015-02-26 00:24:56
|
chrsmithdemos/minify
|
https://api.github.com/repos/chrsmithdemos/minify
|
closed
|
Add a GET parameter for debugging
|
auto-migrated Priority-Medium Type-Defect
|
```
A way to see the original js without minifying it, mainly for debugging/
developing purposes:
In the .htaccess, we need the QSA flag:
RewriteRule ^(.*\.(css|js))$ /minify.php?files=$1 [L,NC,QSA]
And in the minify.php, simply:
if (isset($_GET['debug'])) {
echo $minify->combine(!($_GET['debug']));
} else {
echo $minify->combine();
}
So http://www.example.com/js/blabla.js?debug=1 should return the original
js, without minifying it.
Congrats for the software! It helps me a lot!
Victor
```
-----
Original issue reported on code.google.com by `espiga...@gmail.com` on 21 Jul 2007 at 3:24
|
1.0
|
Add a GET parameter for debugging - ```
A way to see the original js without minifying it, mainly for debugging/
developing purposes:
In the .htaccess, we need the QSA flag:
RewriteRule ^(.*\.(css|js))$ /minify.php?files=$1 [L,NC,QSA]
And in the minify.php, simply:
if (isset($_GET['debug'])) {
echo $minify->combine(!($_GET['debug']));
} else {
echo $minify->combine();
}
So http://www.example.com/js/blabla.js?debug=1 should return the original
js, without minifying it.
Congrats for the software! It helps me a lot!
Victor
```
-----
Original issue reported on code.google.com by `espiga...@gmail.com` on 21 Jul 2007 at 3:24
|
defect
|
add a get parameter for debugging a way to see the original js without minifying it mainly for debugging developing purposes in the htaccess we need the qsa flag rewriterule css js minify php files and in the minify php simply if isset get echo minify combine get else echo minify combine so should return the original js without minifying it congrats for the software it helps me a lot victor original issue reported on code google com by espiga gmail com on jul at
| 1
|
14,040
| 24,277,096,750
|
IssuesEvent
|
2022-09-28 14:34:26
|
CS3219-AY2223S1/cs3219-project-ay2223s1-g5
|
https://api.github.com/repos/CS3219-AY2223S1/cs3219-project-ay2223s1-g5
|
closed
|
[FR-EDITOR-1] The system should have a code editor which allows users to code together in real time.
|
functional requirement P1
|
- [x] #185
- [x] #186
|
1.0
|
[FR-EDITOR-1] The system should have a code editor which allows users to code together in real time. - - [x] #185
- [x] #186
|
non_defect
|
the system should have a code editor which allows users to code together in real time
| 0
|
118,714
| 4,752,110,756
|
IssuesEvent
|
2016-10-23 08:37:54
|
CS2103AUG2016-T13-C2/main
|
https://api.github.com/repos/CS2103AUG2016-T13-C2/main
|
closed
|
bugs in undo delete
|
priority.high type.bug
|
should not add at the bottom, and also carry the information of completion
|
1.0
|
bugs in undo delete - should not add at the bottom, and also carry the information of completion
|
non_defect
|
bugs in undo delete should not add at the bottom and also carry the information of completion
| 0
|
55,954
| 14,860,098,508
|
IssuesEvent
|
2021-01-18 19:48:48
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
SelectOneMenu: JS error calling disable()
|
defect
|
**Describe the defect**
Found during Integration Testing...
When calling the widget `disable()` method throws the following JS error using the Showcase.

This can happen whenver the items are rendered on the client.
```java
.attr("renderPanelContentOnClient", menu.getVar() == null, false);
```
**Environment:**
- PF Version: _10.0_
**Expected behavior**
The SelectOneMenu should be disabled.
**Example XHTML**
```html
<p:selectOneMenu id="hideNoselectionOption" value="#{selectOneMenuView.hideNoSelectOption}"
hideNoSelectionOption="#{not empty selectOneMenuView.hideNoSelectOption}">
<p:ajax update="@this" process="@this"/>
<f:selectItem itemLabel="Select One" itemValue="#{null}" noSelectionOption="true"/>
<f:selectItem itemLabel="Option1" itemValue="Option1"/>
<f:selectItem itemLabel="Option2" itemValue="Option2"/>
<f:selectItem itemLabel="Option3" itemValue="Option3"/>
</p:selectOneMenu>
```
|
1.0
|
SelectOneMenu: JS error calling disable() - **Describe the defect**
Found during Integration Testing...
When calling the widget `disable()` method throws the following JS error using the Showcase.

This can happen whenver the items are rendered on the client.
```java
.attr("renderPanelContentOnClient", menu.getVar() == null, false);
```
**Environment:**
- PF Version: _10.0_
**Expected behavior**
The SelectOneMenu should be disabled.
**Example XHTML**
```html
<p:selectOneMenu id="hideNoselectionOption" value="#{selectOneMenuView.hideNoSelectOption}"
hideNoSelectionOption="#{not empty selectOneMenuView.hideNoSelectOption}">
<p:ajax update="@this" process="@this"/>
<f:selectItem itemLabel="Select One" itemValue="#{null}" noSelectionOption="true"/>
<f:selectItem itemLabel="Option1" itemValue="Option1"/>
<f:selectItem itemLabel="Option2" itemValue="Option2"/>
<f:selectItem itemLabel="Option3" itemValue="Option3"/>
</p:selectOneMenu>
```
|
defect
|
selectonemenu js error calling disable describe the defect found during integration testing when calling the widget disable method throws the following js error using the showcase this can happen whenver the items are rendered on the client java attr renderpanelcontentonclient menu getvar null false environment pf version expected behavior the selectonemenu should be disabled example xhtml html p selectonemenu id hidenoselectionoption value selectonemenuview hidenoselectoption hidenoselectionoption not empty selectonemenuview hidenoselectoption
| 1
|
54,377
| 13,632,224,963
|
IssuesEvent
|
2020-09-24 19:17:43
|
department-of-veterans-affairs/va.gov-cms
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
|
opened
|
Rebuild taxonomy entity index programmatically on a regular basis
|
Content governance Dashboards and UI Defect Drupal engineering Good first task
|
## Description
Occasionally we're seeing cloned pages with edited Section taxonomy terms show up in the wrong /taxonomy/term/TID view
We've found that rebuilding the index seems to fix this.
## Acceptance Criteria
- [ ] Rebuild taxonomy entity index on a regular basis.
|
1.0
|
Rebuild taxonomy entity index programmatically on a regular basis - ## Description
Occasionally we're seeing cloned pages with edited Section taxonomy terms show up in the wrong /taxonomy/term/TID view
We've found that rebuilding the index seems to fix this.
## Acceptance Criteria
- [ ] Rebuild taxonomy entity index on a regular basis.
|
defect
|
rebuild taxonomy entity index programmatically on a regular basis description occasionally we re seeing cloned pages with edited section taxonomy terms show up in the wrong taxonomy term tid view we ve found that rebuilding the index seems to fix this acceptance criteria rebuild taxonomy entity index on a regular basis
| 1
|
237,125
| 18,153,028,726
|
IssuesEvent
|
2021-09-26 15:47:47
|
kimym56/42Project
|
https://api.github.com/repos/kimym56/42Project
|
opened
|
0927 PanResponder
|
documentation
|
## **Done**
- ์๋ ๋งํฌ ์ฝ๋(๋ ์ง ๋ฒ์๋ฅผ dragging์ผ๋ก ์ง์ ) ํ
์คํธ & ์๋๋ก์ด๋ location value ๋ฌธ์ ํด๊ฒฐ
https://blog.bam.tech/developer-news/how-to-handle-user-gestures-in-react-native-with-panresponder
## **To Do**
- ์๋ ๋งํฌ ์ฝ๋(๋ ์ง ๋ฒ์๋ฅผ touch(select)๋ก ์ง์ )๋ ์ ๋งํฌ ์ฝ๋ ๊ฒฐ์
https://react-day-picker.js.org/examples/selected-range/
|
1.0
|
0927 PanResponder - ## **Done**
- ์๋ ๋งํฌ ์ฝ๋(๋ ์ง ๋ฒ์๋ฅผ dragging์ผ๋ก ์ง์ ) ํ
์คํธ & ์๋๋ก์ด๋ location value ๋ฌธ์ ํด๊ฒฐ
https://blog.bam.tech/developer-news/how-to-handle-user-gestures-in-react-native-with-panresponder
## **To Do**
- ์๋ ๋งํฌ ์ฝ๋(๋ ์ง ๋ฒ์๋ฅผ touch(select)๋ก ์ง์ )๋ ์ ๋งํฌ ์ฝ๋ ๊ฒฐ์
https://react-day-picker.js.org/examples/selected-range/
|
non_defect
|
panresponder done ์๋ ๋งํฌ ์ฝ๋ ๋ ์ง ๋ฒ์๋ฅผ dragging์ผ๋ก ์ง์ ํ
์คํธ ์๋๋ก์ด๋ location value ๋ฌธ์ ํด๊ฒฐ to do ์๋ ๋งํฌ ์ฝ๋ ๋ ์ง ๋ฒ์๋ฅผ touch select ๋ก ์ง์ ๋ ์ ๋งํฌ ์ฝ๋ ๊ฒฐ์
| 0
|
77,218
| 26,855,486,153
|
IssuesEvent
|
2023-02-03 14:19:50
|
obophenotype/cell-ontology
|
https://api.github.com/repos/obophenotype/cell-ontology
|
closed
|
names that are too broad for their definitions (ciliary epithelial cells)
|
Priority-Medium Type-Defect auto-migrated typo text definition
|
```
TICKET: These terms have names that are too broad for their definitions, or
definitions that are two narrow for their names:
id: CL:0002304
name: non-pigmented ciliary epithelial cell
def: "A multi-ciliated cell of the retina that lacks visual pigment and
contributes to aqueous humor by secreting chloride ions. This cell type
maintains gap junctions with pigmented epithelial cells." [GOC:tfm,
PMID:15106942]
[Term]
id: CL:0002303
name: pigmented ciliary epithelial cell
def: "A ciliated epithelial cell of the retina, this cell type uptakes sodium
chloride and passes it to non-pigmented ciliary epithelial cells."
[PMID:15106942]
```
Original issue reported on code.google.com by `dosu...@gmail.com` on 27 Feb 2012 at 5:41
|
1.0
|
names that are too broad for their definitions (ciliary epithelial cells) - ```
TICKET: These terms have names that are too broad for their definitions, or
definitions that are two narrow for their names:
id: CL:0002304
name: non-pigmented ciliary epithelial cell
def: "A multi-ciliated cell of the retina that lacks visual pigment and
contributes to aqueous humor by secreting chloride ions. This cell type
maintains gap junctions with pigmented epithelial cells." [GOC:tfm,
PMID:15106942]
[Term]
id: CL:0002303
name: pigmented ciliary epithelial cell
def: "A ciliated epithelial cell of the retina, this cell type uptakes sodium
chloride and passes it to non-pigmented ciliary epithelial cells."
[PMID:15106942]
```
Original issue reported on code.google.com by `dosu...@gmail.com` on 27 Feb 2012 at 5:41
|
defect
|
names that are too broad for their definitions ciliary epithelial cells ticket these terms have names that are too broad for their definitions or definitions that are two narrow for their names id cl name non pigmented ciliary epithelial cell def a multi ciliated cell of the retina that lacks visual pigment and contributes to aqueous humor by secreting chloride ions this cell type maintains gap junctions with pigmented epithelial cells goc tfm pmid id cl name pigmented ciliary epithelial cell def a ciliated epithelial cell of the retina this cell type uptakes sodium chloride and passes it to non pigmented ciliary epithelial cells original issue reported on code google com by dosu gmail com on feb at
| 1
|
13,832
| 2,787,212,491
|
IssuesEvent
|
2015-05-08 02:51:48
|
mblanchette/maven-java-formatter-plugin
|
https://api.github.com/repos/mblanchette/maven-java-formatter-plugin
|
closed
|
Update embedded Eclipse to 4.4 to allow for Java 8 sources.
|
auto-migrated Priority-Medium Type-Defect
|
```
Version 0.4 embeds an older Eclipse which does not know about Java 8.
The recently released Eclipse 4.4 does, so it would be really nice if version
0.5 used this instead.
```
Original issue reported on code.google.com by `thorbjo...@gmail.com` on 27 Jun 2014 at 9:24
|
1.0
|
Update embedded Eclipse to 4.4 to allow for Java 8 sources. - ```
Version 0.4 embeds an older Eclipse which does not know about Java 8.
The recently released Eclipse 4.4 does, so it would be really nice if version
0.5 used this instead.
```
Original issue reported on code.google.com by `thorbjo...@gmail.com` on 27 Jun 2014 at 9:24
|
defect
|
update embedded eclipse to to allow for java sources version embeds an older eclipse which does not know about java the recently released eclipse does so it would be really nice if version used this instead original issue reported on code google com by thorbjo gmail com on jun at
| 1
|
84,817
| 10,418,983,553
|
IssuesEvent
|
2019-09-15 13:17:01
|
square/okhttp
|
https://api.github.com/repos/square/okhttp
|
closed
|
Improve/Fix visibility of sample recipes e.g. Certificate Pinning
|
documentation
|
The docs on https://square.github.io/okhttp/https/#certificate-pinning
say:
```
public SSLContext sslContextForTrustedCertificates(InputStream in) {
... // Full source omitted. See sample.
}
```
but there is no such sample.
https://github.com/square/okhttp/tree/master/samples/guide/src/main/java/okhttp3/recipes
|
1.0
|
Improve/Fix visibility of sample recipes e.g. Certificate Pinning - The docs on https://square.github.io/okhttp/https/#certificate-pinning
say:
```
public SSLContext sslContextForTrustedCertificates(InputStream in) {
... // Full source omitted. See sample.
}
```
but there is no such sample.
https://github.com/square/okhttp/tree/master/samples/guide/src/main/java/okhttp3/recipes
|
non_defect
|
improve fix visibility of sample recipes e g certificate pinning the docs on say public sslcontext sslcontextfortrustedcertificates inputstream in full source omitted see sample but there is no such sample
| 0
|
68,760
| 7,109,041,991
|
IssuesEvent
|
2018-01-17 03:35:14
|
bitcoinjs/bitcoinjs-lib
|
https://api.github.com/repos/bitcoinjs/bitcoinjs-lib
|
closed
|
BIP143 rejects uncompressed public keys in P2WPKH or P2WSH
|
bug testing
|
Not sure if I am misunderstanding something here but does this not only check if the output is p2wpkh and not p2wsh as well?
i.e. is correct?
```
if (kpPubKey.length !== 33 &&
(input.signType === scriptTypes.P2WPKH || input.signType === scriptTypes.P2WSH)) throw new Error('BIP143 rejects uncompressed public keys in P2WPKH or P2WSH')
```
https://github.com/bitcoinjs/bitcoinjs-lib/blob/86cd4a44a1524686d993f35b02a1ed331aaa551f/src/transaction_builder.js#L709
|
1.0
|
BIP143 rejects uncompressed public keys in P2WPKH or P2WSH - Not sure if I am misunderstanding something here but does this not only check if the output is p2wpkh and not p2wsh as well?
i.e. is correct?
```
if (kpPubKey.length !== 33 &&
(input.signType === scriptTypes.P2WPKH || input.signType === scriptTypes.P2WSH)) throw new Error('BIP143 rejects uncompressed public keys in P2WPKH or P2WSH')
```
https://github.com/bitcoinjs/bitcoinjs-lib/blob/86cd4a44a1524686d993f35b02a1ed331aaa551f/src/transaction_builder.js#L709
|
non_defect
|
rejects uncompressed public keys in or not sure if i am misunderstanding something here but does this not only check if the output is and not as well i e is correct if kppubkey length input signtype scripttypes input signtype scripttypes throw new error rejects uncompressed public keys in or
| 0
|
5,609
| 3,970,223,416
|
IssuesEvent
|
2016-05-04 05:44:51
|
kolliSuman/issues
|
https://api.github.com/repos/kolliSuman/issues
|
closed
|
QA_Introduction to Programmable Logic Controller_Self evaluation_smk
|
Category: Usability Developed By: VLEAD Release Number: Production Severity: S3 Status: Open
|
Defect Description :
In Self Evaluation page of "Introduction to Programmable Logic Controller and Introduction to Digital I/O Interface to PLCโ when we click on get start button the Content on the top of the page is not clear where Content should be present with no distortions
Actual Result :
In Self Evaluation page of "Introduction to Programmable Logic Controller and Introduction to Digital I/O Interface to PLCโ when we click on get start button the Content on the top of the page is not clear
Environment :
OS: Windows 7, Ubuntu-16.04,Centos-6
Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM ,
Processor:i5
Test Step Link:
https://github.com/Virtual-Labs/industrial-electrical-drives-nitk/blob/master/test-cases/integration_test-cases/Introduction%20to%20Programmable%20Logic%20Controller/Introduction%20to%20Programmable%20Logic%20Controller_04_Self%20evaluation_smk.org
|
True
|
QA_Introduction to Programmable Logic Controller_Self evaluation_smk - Defect Description :
In Self Evaluation page of "Introduction to Programmable Logic Controller and Introduction to Digital I/O Interface to PLCโ when we click on get start button the Content on the top of the page is not clear where Content should be present with no distortions
Actual Result :
In Self Evaluation page of "Introduction to Programmable Logic Controller and Introduction to Digital I/O Interface to PLCโ when we click on get start button the Content on the top of the page is not clear
Environment :
OS: Windows 7, Ubuntu-16.04,Centos-6
Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM ,
Processor:i5
Test Step Link:
https://github.com/Virtual-Labs/industrial-electrical-drives-nitk/blob/master/test-cases/integration_test-cases/Introduction%20to%20Programmable%20Logic%20Controller/Introduction%20to%20Programmable%20Logic%20Controller_04_Self%20evaluation_smk.org
|
non_defect
|
qa introduction to programmable logic controller self evaluation smk defect description in self evaluation page of introduction to programmable logic controller and introduction to digital i o interface to plcโ when we click on get start button the content on the top of the page is not clear where content should be present with no distortions actual result in self evaluation page of introduction to programmable logic controller and introduction to digital i o interface to plcโ when we click on get start button the content on the top of the page is not clear environment os windows ubuntu centos browsers firefox chrome chromium bandwidth hardware configuration processor test step link
| 0
|
63,182
| 17,409,860,851
|
IssuesEvent
|
2021-08-03 10:52:56
|
AtlasOfLivingAustralia/collectory
|
https://api.github.com/repos/AtlasOfLivingAustralia/collectory
|
closed
|
download button a dataset page doesn't work
|
priority-medium status-new type-defect
|
_From @mbohun on August 19, 2014 12:33_
_migrated from:_ https://code.google.com/p/ala/issues/detail?id=524
_date:_ Tue Jan 14 20:41:20 2014
_author:_ milo_nic...@hotmail.com
---
e.g. clicking the download records button on [http://collections.ala.org.au/public/show/dr365](http://collections.ala.org.au/public/show/dr365) produces an error - A reasonTypeId must be provided.
_Copied from original issue: AtlasOfLivingAustralia/biocache-hubs#55_
|
1.0
|
download button a dataset page doesn't work - _From @mbohun on August 19, 2014 12:33_
_migrated from:_ https://code.google.com/p/ala/issues/detail?id=524
_date:_ Tue Jan 14 20:41:20 2014
_author:_ milo_nic...@hotmail.com
---
e.g. clicking the download records button on [http://collections.ala.org.au/public/show/dr365](http://collections.ala.org.au/public/show/dr365) produces an error - A reasonTypeId must be provided.
_Copied from original issue: AtlasOfLivingAustralia/biocache-hubs#55_
|
defect
|
download button a dataset page doesn t work from mbohun on august migrated from date tue jan author milo nic hotmail com e g clicking the download records button on produces an error a reasontypeid must be provided copied from original issue atlasoflivingaustralia biocache hubs
| 1
|
203,692
| 15,380,550,885
|
IssuesEvent
|
2021-03-02 21:16:28
|
radicle-dev/radicle-link
|
https://api.github.com/repos/radicle-dev/radicle-link
|
closed
|
Split integration tests
|
brooming help wanted testing
|
`propagation_basic` was meant to test some, err, basic things.
Before it gets out of hand, we should split into a couple more modules. As those tests are becoming more heavy on the IOs, this also prepares for running them concurrently and/or behind a `slow` feature.
|
1.0
|
Split integration tests - `propagation_basic` was meant to test some, err, basic things.
Before it gets out of hand, we should split into a couple more modules. As those tests are becoming more heavy on the IOs, this also prepares for running them concurrently and/or behind a `slow` feature.
|
non_defect
|
split integration tests propagation basic was meant to test some err basic things before it gets out of hand we should split into a couple more modules as those tests are becoming more heavy on the ios this also prepares for running them concurrently and or behind a slow feature
| 0
|
2,772
| 2,607,944,887
|
IssuesEvent
|
2015-02-26 00:32:52
|
chrsmithdemos/switchlist
|
https://api.github.com/repos/chrsmithdemos/switchlist
|
opened
|
Add explicit checks for duplicate town names.
|
auto-migrated Priority-Medium Type-Defect
|
```
There's currently no validation that two towns don't have the same names. This
would mess up the "train stops" code that assumes only one town exists with a
given name.
```
-----
Original issue reported on code.google.com by `rwbowdi...@gmail.com` on 24 Apr 2011 at 5:25
|
1.0
|
Add explicit checks for duplicate town names. - ```
There's currently no validation that two towns don't have the same names. This
would mess up the "train stops" code that assumes only one town exists with a
given name.
```
-----
Original issue reported on code.google.com by `rwbowdi...@gmail.com` on 24 Apr 2011 at 5:25
|
defect
|
add explicit checks for duplicate town names there s currently no validation that two towns don t have the same names this would mess up the train stops code that assumes only one town exists with a given name original issue reported on code google com by rwbowdi gmail com on apr at
| 1
|
27,303
| 4,958,098,713
|
IssuesEvent
|
2016-12-02 08:27:19
|
TNGSB/eWallet
|
https://api.github.com/repos/TNGSB/eWallet
|
closed
|
eWallet_MobileApp(Reload) #105
|
Defect - High (Sev-2)
|
[Defect_Mobile App #105.xlsx](https://github.com/TNGSB/eWallet/files/593552/Defect_Mobile.App.105.xlsx)
Defect Description : System displayed a successful message and automatically redirect user to "Credit/Debit Card" page before the countdown counter time out
Test Description : To verify error message when user exceeded the countdown timer for credit card detail input
Refer attachment for POT
|
1.0
|
eWallet_MobileApp(Reload) #105 -
[Defect_Mobile App #105.xlsx](https://github.com/TNGSB/eWallet/files/593552/Defect_Mobile.App.105.xlsx)
Defect Description : System displayed a successful message and automatically redirect user to "Credit/Debit Card" page before the countdown counter time out
Test Description : To verify error message when user exceeded the countdown timer for credit card detail input
Refer attachment for POT
|
defect
|
ewallet mobileapp reload defect description system displayed a successful message and automatically redirect user to credit debit card page before the countdown counter time out test description to verify error message when user exceeded the countdown timer for credit card detail input refer attachment for pot
| 1
|
11,012
| 2,622,955,112
|
IssuesEvent
|
2015-03-04 09:03:13
|
folded/carve
|
https://api.github.com/repos/folded/carve
|
opened
|
Carve fails on (apparantly) valid input
|
auto-migrated Priority-Medium Type-Defect
|
```
Just run:
>intersect "(_0a.ply | _1a.ply | _2a.ply) A_MINUS_B _3s.ply"
with the attached files.
Files attached were created with carve by intersecting several simple
models. Viewing tool supplied with carve opens them without any problems.
```
Original issue reported on code.google.com by `ru.el...@gmail.com` on 17 May 2010 at 2:31
Attachments:
* [_0a.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-17/comment-0/_0a.ply)
* [_1a.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-17/comment-0/_1a.ply)
* [_2a.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-17/comment-0/_2a.ply)
* [_3s.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-17/comment-0/_3s.ply)
|
1.0
|
Carve fails on (apparantly) valid input - ```
Just run:
>intersect "(_0a.ply | _1a.ply | _2a.ply) A_MINUS_B _3s.ply"
with the attached files.
Files attached were created with carve by intersecting several simple
models. Viewing tool supplied with carve opens them without any problems.
```
Original issue reported on code.google.com by `ru.el...@gmail.com` on 17 May 2010 at 2:31
Attachments:
* [_0a.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-17/comment-0/_0a.ply)
* [_1a.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-17/comment-0/_1a.ply)
* [_2a.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-17/comment-0/_2a.ply)
* [_3s.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-17/comment-0/_3s.ply)
|
defect
|
carve fails on apparantly valid input just run intersect ply ply ply a minus b ply with the attached files files attached were created with carve by intersecting several simple models viewing tool supplied with carve opens them without any problems original issue reported on code google com by ru el gmail com on may at attachments
| 1
|
108,794
| 23,664,428,850
|
IssuesEvent
|
2022-08-26 19:07:06
|
unoplatform/uno
|
https://api.github.com/repos/unoplatform/uno
|
opened
|
Enable XAML Trimming for `net6.0-[ios|android|macos|maccatalyst]` and GTK targets
|
kind/enhancement area/code-generation
|
## What would you like to be added:
Support for [XAML Trimming](https://platform.uno/docs/articles/features/resources-trimming.html?) for `net6.0-[ios|android|macos|maccatalyst]` and GTK targets.
## Why is this needed:
Reduce the size of applications.
## Anything else we need to know?
WPF is explicitly excluded from this list as trimming is forcibly disabled by the SDK.
Work branch: https://github.com/unoplatform/uno/tree/dev/eb/xaml-trimming-net6
## Related issues
- https://github.com/dotnet/sdk/issues/27492
- https://github.com/xamarin/xamarin-android/issues/7301
|
1.0
|
Enable XAML Trimming for `net6.0-[ios|android|macos|maccatalyst]` and GTK targets - ## What would you like to be added:
Support for [XAML Trimming](https://platform.uno/docs/articles/features/resources-trimming.html?) for `net6.0-[ios|android|macos|maccatalyst]` and GTK targets.
## Why is this needed:
Reduce the size of applications.
## Anything else we need to know?
WPF is explicitly excluded from this list as trimming is forcibly disabled by the SDK.
Work branch: https://github.com/unoplatform/uno/tree/dev/eb/xaml-trimming-net6
## Related issues
- https://github.com/dotnet/sdk/issues/27492
- https://github.com/xamarin/xamarin-android/issues/7301
|
non_defect
|
enable xaml trimming for and gtk targets what would you like to be added support for for and gtk targets why is this needed reduce the size of applications anything else we need to know wpf is explicitly excluded from this list as trimming is forcibly disabled by the sdk work branch related issues
| 0
|
18,922
| 3,734,406,963
|
IssuesEvent
|
2016-03-08 06:41:42
|
kumulsoft/Fixed-Assets
|
https://api.github.com/repos/kumulsoft/Fixed-Assets
|
closed
|
Asset Transactions - Grid Page - Enhancement
|
enhancement Fixed HIGH Ready for testing
|
Remove the 'Trans Detail' column to give space to add two new columns 'Centre and Custodian)

|
1.0
|
Asset Transactions - Grid Page - Enhancement - Remove the 'Trans Detail' column to give space to add two new columns 'Centre and Custodian)

|
non_defect
|
asset transactions grid page enhancement remove the trans detail column to give space to add two new columns centre and custodian
| 0
|
62,085
| 17,023,847,476
|
IssuesEvent
|
2021-07-03 04:09:20
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
Only 1 of 2 streets with same name found
|
Component: nominatim Priority: major Resolution: duplicate Type: defect
|
**[Submitted to the original trac issue database at 11.18pm, Thursday, 3rd January 2013]**
http://nominatim.openstreetmap.org/search.php?q=tannenweg+lenzkirch&viewbox=8.19%2C47.88%2C8.21%2C47.87&polygon=1
Only one of the 2 residential streets named Tannenweg are found. The streets are merging in a Y shape in their western parts.
The error leads to a wrong bounding box for zooming. Thus depending on screen resolution and street geometry it might not be obvious to a human user that a second street with the same name exists (is only shown partially but the label off-screen).
|
1.0
|
Only 1 of 2 streets with same name found - **[Submitted to the original trac issue database at 11.18pm, Thursday, 3rd January 2013]**
http://nominatim.openstreetmap.org/search.php?q=tannenweg+lenzkirch&viewbox=8.19%2C47.88%2C8.21%2C47.87&polygon=1
Only one of the 2 residential streets named Tannenweg are found. The streets are merging in a Y shape in their western parts.
The error leads to a wrong bounding box for zooming. Thus depending on screen resolution and street geometry it might not be obvious to a human user that a second street with the same name exists (is only shown partially but the label off-screen).
|
defect
|
only of streets with same name found only one of the residential streets named tannenweg are found the streets are merging in a y shape in their western parts the error leads to a wrong bounding box for zooming thus depending on screen resolution and street geometry it might not be obvious to a human user that a second street with the same name exists is only shown partially but the label off screen
| 1
|
435,622
| 30,510,415,808
|
IssuesEvent
|
2023-07-18 20:21:19
|
Nnoemis/Test1
|
https://api.github.com/repos/Nnoemis/Test1
|
opened
|
How would you test a 5 kg capacity grocery shopping paper bag? Describe the tests that you could perform.
|
documentation
|
1. Insert 3 kg of oranges.
2. Insert 4.5 kg of oranges.
3.Insert 5 kg of oranges.
4. Insert + 100 g of oranges.
5. Insert + 200 g of oranges.
6. Every next time insert + 100 g of oranges and stop when paper bag is damaged. (Bounding testing)
7. Try the these steps with a wet paper bag.
8. Try the bag`s volume.
Sourse [https://simeon.svet-bg.com/index.php/forums/topic/think-testing-5-kg-bag/](url)
|
1.0
|
How would you test a 5 kg capacity grocery shopping paper bag? Describe the tests that you could perform. - 1. Insert 3 kg of oranges.
2. Insert 4.5 kg of oranges.
3.Insert 5 kg of oranges.
4. Insert + 100 g of oranges.
5. Insert + 200 g of oranges.
6. Every next time insert + 100 g of oranges and stop when paper bag is damaged. (Bounding testing)
7. Try the these steps with a wet paper bag.
8. Try the bag`s volume.
Sourse [https://simeon.svet-bg.com/index.php/forums/topic/think-testing-5-kg-bag/](url)
|
non_defect
|
how would you test a kg capacity grocery shopping paper bag describe the tests that you could perform insert kg of oranges insert kg of oranges insert kg of oranges insert g of oranges insert g of oranges every next time insert g of oranges and stop when paper bag is damaged bounding testing try the these steps with a wet paper bag try the bag s volume sourse url
| 0
|
41,834
| 10,679,235,358
|
IssuesEvent
|
2019-10-21 18:51:50
|
techo/voluntariado-eventual
|
https://api.github.com/repos/techo/voluntariado-eventual
|
closed
|
Error al mostrar puntos de encuentro en la inscripciรณn
|
Defecto
|
**Describรญ el error**
Como usuario al querer inscribirme en una actividad los puntos de encuentro se muestran incorrectamente.
**Para reproducirlo**
Pasos para reproducir el comportamiento:
1. Cargar una nueva actividad.
2. Cargar 2 puntos de encuentro en diferentes provincias
3. Como usuario ir a la actividad a inscribirme
4. Ver que las pronvincias de los puntos de encuentro se muestran incorrectamente.
**Comportamiento esperando**
Que las provincias se vean igual que en la carga de actividad.
**Capturas de pantalla**

|
1.0
|
Error al mostrar puntos de encuentro en la inscripciรณn - **Describรญ el error**
Como usuario al querer inscribirme en una actividad los puntos de encuentro se muestran incorrectamente.
**Para reproducirlo**
Pasos para reproducir el comportamiento:
1. Cargar una nueva actividad.
2. Cargar 2 puntos de encuentro en diferentes provincias
3. Como usuario ir a la actividad a inscribirme
4. Ver que las pronvincias de los puntos de encuentro se muestran incorrectamente.
**Comportamiento esperando**
Que las provincias se vean igual que en la carga de actividad.
**Capturas de pantalla**

|
defect
|
error al mostrar puntos de encuentro en la inscripciรณn describรญ el error como usuario al querer inscribirme en una actividad los puntos de encuentro se muestran incorrectamente para reproducirlo pasos para reproducir el comportamiento cargar una nueva actividad cargar puntos de encuentro en diferentes provincias como usuario ir a la actividad a inscribirme ver que las pronvincias de los puntos de encuentro se muestran incorrectamente comportamiento esperando que las provincias se vean igual que en la carga de actividad capturas de pantalla
| 1
|
74,877
| 25,379,761,270
|
IssuesEvent
|
2022-11-21 16:37:21
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
opened
|
Liquibase 4.17 compatibility
|
T: Defect
|
### Expected behavior
Jooq Liquibase extension to be compatible with recent version of Liquibase
### Actual behavior
The Jooq liquibase extension uses `FileSystemResourceAccessor` which was [removed](https://github.com/liquibase/liquibase/issues/3478) in Liquibase 4.17.
### Steps to reproduce the problem
Log output from an attempt to run the Gradle plugin `generateJooq` task:
```
matt @ ghost-s1 in ~/Workspace/github.com/mattupstate/acme (โk3d-acme-dev/default) on git:main+
$ ./gradlew acme-data:acme-data-sql:generateJooq
> Task :acme-data:acme-data-sql:generateJooq FAILED
Exception in thread "main" java.lang.NoClassDefFoundError: liquibase/resource/FileSystemResourceAccessor
at org.jooq.meta.extensions.liquibase.LiquibaseDatabase.export(LiquibaseDatabase.java:155)
at org.jooq.meta.extensions.AbstractInterpretingDatabase.connection(AbstractInterpretingDatabase.java:100)
at org.jooq.meta.extensions.AbstractInterpretingDatabase.create0(AbstractInterpretingDatabase.java:77)
at org.jooq.meta.AbstractDatabase.create(AbstractDatabase.java:369)
at org.jooq.meta.AbstractDatabase.create(AbstractDatabase.java:359)
at org.jooq.meta.AbstractDatabase.setConnection(AbstractDatabase.java:337)
at org.jooq.codegen.GenerationTool.run0(GenerationTool.java:553)
at org.jooq.codegen.GenerationTool.run(GenerationTool.java:240)
at org.jooq.codegen.GenerationTool.generate(GenerationTool.java:235)
at org.jooq.codegen.GenerationTool.main(GenerationTool.java:207)
Caused by: java.lang.ClassNotFoundException: liquibase.resource.FileSystemResourceAccessor
at java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:641)
at java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:188)
at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:520)
... 10 more
```
### jOOQ Version
3.17.4
### Database product and version
N/A
### Java Version
N/A
### OS Version
N/A
### JDBC driver name and version (include name if unofficial driver)
N/A
|
1.0
|
Liquibase 4.17 compatibility - ### Expected behavior
Jooq Liquibase extension to be compatible with recent version of Liquibase
### Actual behavior
The Jooq liquibase extension uses `FileSystemResourceAccessor` which was [removed](https://github.com/liquibase/liquibase/issues/3478) in Liquibase 4.17.
### Steps to reproduce the problem
Log output from an attempt to run the Gradle plugin `generateJooq` task:
```
matt @ ghost-s1 in ~/Workspace/github.com/mattupstate/acme (โk3d-acme-dev/default) on git:main+
$ ./gradlew acme-data:acme-data-sql:generateJooq
> Task :acme-data:acme-data-sql:generateJooq FAILED
Exception in thread "main" java.lang.NoClassDefFoundError: liquibase/resource/FileSystemResourceAccessor
at org.jooq.meta.extensions.liquibase.LiquibaseDatabase.export(LiquibaseDatabase.java:155)
at org.jooq.meta.extensions.AbstractInterpretingDatabase.connection(AbstractInterpretingDatabase.java:100)
at org.jooq.meta.extensions.AbstractInterpretingDatabase.create0(AbstractInterpretingDatabase.java:77)
at org.jooq.meta.AbstractDatabase.create(AbstractDatabase.java:369)
at org.jooq.meta.AbstractDatabase.create(AbstractDatabase.java:359)
at org.jooq.meta.AbstractDatabase.setConnection(AbstractDatabase.java:337)
at org.jooq.codegen.GenerationTool.run0(GenerationTool.java:553)
at org.jooq.codegen.GenerationTool.run(GenerationTool.java:240)
at org.jooq.codegen.GenerationTool.generate(GenerationTool.java:235)
at org.jooq.codegen.GenerationTool.main(GenerationTool.java:207)
Caused by: java.lang.ClassNotFoundException: liquibase.resource.FileSystemResourceAccessor
at java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:641)
at java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:188)
at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:520)
... 10 more
```
### jOOQ Version
3.17.4
### Database product and version
N/A
### Java Version
N/A
### OS Version
N/A
### JDBC driver name and version (include name if unofficial driver)
N/A
|
defect
|
liquibase compatibility expected behavior jooq liquibase extension to be compatible with recent version of liquibase actual behavior the jooq liquibase extension uses filesystemresourceaccessor which was in liquibase steps to reproduce the problem log output from an attempt to run the gradle plugin generatejooq task matt ghost in workspace github com mattupstate acme โ acme dev default on git main gradlew acme data acme data sql generatejooq task acme data acme data sql generatejooq failed exception in thread main java lang noclassdeffounderror liquibase resource filesystemresourceaccessor at org jooq meta extensions liquibase liquibasedatabase export liquibasedatabase java at org jooq meta extensions abstractinterpretingdatabase connection abstractinterpretingdatabase java at org jooq meta extensions abstractinterpretingdatabase abstractinterpretingdatabase java at org jooq meta abstractdatabase create abstractdatabase java at org jooq meta abstractdatabase create abstractdatabase java at org jooq meta abstractdatabase setconnection abstractdatabase java at org jooq codegen generationtool generationtool java at org jooq codegen generationtool run generationtool java at org jooq codegen generationtool generate generationtool java at org jooq codegen generationtool main generationtool java caused by java lang classnotfoundexception liquibase resource filesystemresourceaccessor at java base jdk internal loader builtinclassloader loadclass builtinclassloader java at java base jdk internal loader classloaders appclassloader loadclass classloaders java at java base java lang classloader loadclass classloader java more jooq version database product and version n a java version n a os version n a jdbc driver name and version include name if unofficial driver n a
| 1
|
15,881
| 2,869,088,191
|
IssuesEvent
|
2015-06-05 23:14:17
|
dart-lang/sdk
|
https://api.github.com/repos/dart-lang/sdk
|
closed
|
Collapse together errors of the same kind
|
Area-Pkg Pkg-PolymerDevExp PolymerMilestone-Next Priority-Medium Triaged Type-Defect
|
It's common to see one warning/error repeated in many places. For example the href vs \_href might happen a lot now that we introduce the warning for it. It would be great to combine all occurrences of the same error together.
Some ideas:
- for the log_injector UI, we can combine it graphically.
- for the pub-build/pub-serve output in stdout, we can delay reporting warnings until the end of the phase, collect all warnings together and report them as such. At least we can do that on a file-per-file basis.
|
1.0
|
Collapse together errors of the same kind - It's common to see one warning/error repeated in many places. For example the href vs \_href might happen a lot now that we introduce the warning for it. It would be great to combine all occurrences of the same error together.
Some ideas:
- for the log_injector UI, we can combine it graphically.
- for the pub-build/pub-serve output in stdout, we can delay reporting warnings until the end of the phase, collect all warnings together and report them as such. At least we can do that on a file-per-file basis.
|
defect
|
collapse together errors of the same kind it s common to see one warning error repeated in many places for example the href vs href might happen a lot now that we introduce the warning for it it would be great to combine all occurrences of the same error together some ideas for the log injector ui we can combine it graphically for the pub build pub serve output in stdout we can delay reporting warnings until the end of the phase collect all warnings together and report them as such at least we can do that on a file per file basis
| 1
|
4,357
| 10,965,734,316
|
IssuesEvent
|
2019-11-28 04:13:34
|
fga-eps-mds/2019.2-Over26
|
https://api.github.com/repos/fga-eps-mds/2019.2-Over26
|
closed
|
Atualizar diagramas do Documento de Arquitetura
|
Architecture Documentation EPS
|
## Descriรงรฃo da Mudanรงa *
<!--- Forneรงa um resumo geral da _issue_ -->
ร necessรกrio adequar os diagramas presentes no Documento de Arquitetura ร atual estrutura do projeto.
## Checklist *
<!-- Essa checklist propรตe a criaรงรฃo de uma boa issue -->
<!-- Se a issue รฉ sobre uma histรณria de usuรกrio, seu nome deve ser "USXX - Nome da histรณria-->
<!-- Se a issue รฉ sobre um bug, seu nome deve ser "BF - Nome curto do bug"-->
<!-- Se a issue รฉ sobre outra tarefa o nome deve ser uma simples descriรงรฃo da tarefa-->
- [x] Esta issue tem um nome significativo.
- [x] O nome da issue estรก no padrรฃo.
- [x] Esta issue tem uma descriรงรฃo de fรกcil entendimento.
- [x] Esta issue tem uma boa definiรงรฃo de critรฉrios de aceitaรงรฃo.
- [x] Esta issue tem labels associadas.
- [ ] Esta issue estรก associada ร uma milestone.
- [ ] Esta issue tem uma pontuaรงรฃo estimada.
## Tarefas *
<!-- Adicione aqui as tarefas necessรกrias para concluir a issue -->
- [x] Atualizar diagrama de classes
- [x] Atualizar diagrama lรณgico
- [x] Atualizar diagrama de pacotes
## Critรฉrios de Aceitaรงรฃo *
<!-- Liste aqui o conjunto de aspectos mecessรกrios para considerar a atividade como completa-->
<!-- Os itens serรฃo adicionados pelo Product Owner -->
- [x] Diagramas atualizados.
|
1.0
|
Atualizar diagramas do Documento de Arquitetura - ## Descriรงรฃo da Mudanรงa *
<!--- Forneรงa um resumo geral da _issue_ -->
ร necessรกrio adequar os diagramas presentes no Documento de Arquitetura ร atual estrutura do projeto.
## Checklist *
<!-- Essa checklist propรตe a criaรงรฃo de uma boa issue -->
<!-- Se a issue รฉ sobre uma histรณria de usuรกrio, seu nome deve ser "USXX - Nome da histรณria-->
<!-- Se a issue รฉ sobre um bug, seu nome deve ser "BF - Nome curto do bug"-->
<!-- Se a issue รฉ sobre outra tarefa o nome deve ser uma simples descriรงรฃo da tarefa-->
- [x] Esta issue tem um nome significativo.
- [x] O nome da issue estรก no padrรฃo.
- [x] Esta issue tem uma descriรงรฃo de fรกcil entendimento.
- [x] Esta issue tem uma boa definiรงรฃo de critรฉrios de aceitaรงรฃo.
- [x] Esta issue tem labels associadas.
- [ ] Esta issue estรก associada ร uma milestone.
- [ ] Esta issue tem uma pontuaรงรฃo estimada.
## Tarefas *
<!-- Adicione aqui as tarefas necessรกrias para concluir a issue -->
- [x] Atualizar diagrama de classes
- [x] Atualizar diagrama lรณgico
- [x] Atualizar diagrama de pacotes
## Critรฉrios de Aceitaรงรฃo *
<!-- Liste aqui o conjunto de aspectos mecessรกrios para considerar a atividade como completa-->
<!-- Os itens serรฃo adicionados pelo Product Owner -->
- [x] Diagramas atualizados.
|
non_defect
|
atualizar diagramas do documento de arquitetura descriรงรฃo da mudanรงa รฉ necessรกrio adequar os diagramas presentes no documento de arquitetura ร atual estrutura do projeto checklist esta issue tem um nome significativo o nome da issue estรก no padrรฃo esta issue tem uma descriรงรฃo de fรกcil entendimento esta issue tem uma boa definiรงรฃo de critรฉrios de aceitaรงรฃo esta issue tem labels associadas esta issue estรก associada ร uma milestone esta issue tem uma pontuaรงรฃo estimada tarefas atualizar diagrama de classes atualizar diagrama lรณgico atualizar diagrama de pacotes critรฉrios de aceitaรงรฃo diagramas atualizados
| 0
|
44,751
| 12,372,216,940
|
IssuesEvent
|
2020-05-18 19:59:26
|
FoldingAtHome/fah-issues
|
https://api.github.com/repos/FoldingAtHome/fah-issues
|
closed
|
Dialog Box opened from FahCore22.exe, entry point clReleaseDevice could not be located
|
FAHClient defect
|
## Environment
* OS: Windows 7 Enterprise 64 build 7601 SP1
* FAH version: April 28, 2020
## What were you trying to do
Folding on FAH for Covid-19
## What happened
NVIDIA GEOFORCE GT 610 as GPU:0 hangs, throws dialog box, FahCore_22.exe - Entry Point Not Found. White X in Red circle: The procedure entry point clReleaseDevice could not be located in the dynamic link library OpenCL.dll. "OK" the dialog box, and a second box is thrown of same type
OK that one and GPU process indicator turns to solid yellow. And this cycle continues over and over.
## To Reproduce
just keeps happening each time.
|
1.0
|
Dialog Box opened from FahCore22.exe, entry point clReleaseDevice could not be located - ## Environment
* OS: Windows 7 Enterprise 64 build 7601 SP1
* FAH version: April 28, 2020
## What were you trying to do
Folding on FAH for Covid-19
## What happened
NVIDIA GEOFORCE GT 610 as GPU:0 hangs, throws dialog box, FahCore_22.exe - Entry Point Not Found. White X in Red circle: The procedure entry point clReleaseDevice could not be located in the dynamic link library OpenCL.dll. "OK" the dialog box, and a second box is thrown of same type
OK that one and GPU process indicator turns to solid yellow. And this cycle continues over and over.
## To Reproduce
just keeps happening each time.
|
defect
|
dialog box opened from exe entry point clreleasedevice could not be located environment os windows enterprise build fah version april what were you trying to do folding on fah for covid what happened nvidia geoforce gt as gpu hangs throws dialog box fahcore exe entry point not found white x in red circle the procedure entry point clreleasedevice could not be located in the dynamic link library opencl dll ok the dialog box and a second box is thrown of same type ok that one and gpu process indicator turns to solid yellow and this cycle continues over and over to reproduce just keeps happening each time
| 1
|
46,293
| 24,466,746,399
|
IssuesEvent
|
2022-10-07 15:39:27
|
Kitware/vtk-js
|
https://api.github.com/repos/Kitware/vtk-js
|
closed
|
Setting a model variable as object always trigger modified()
|
type: bug ๐ type: performance โก๏ธ
|
### High-level description
If `macro.set(publicAPI, model, ['foo']);` and `foo` is an `Object` (e.g. `{a: 1, b: 2}`, then calling `setFoo({a: 1, b:2})` will always trigger `modified()`.
### Steps to reproduce
```
extend(...){
macro.set(publicAPI, model, ['foo']);
}
...
myobj.setFoo({a: 1, b: 2});
myobj.getMTime(); // =>1234
myobj.setFoo({a: 1, b: 2});
myobj.getMTime(); // =>1235
```
See [here](https://github.com/Kitware/vtk-js/blob/master/Sources/Widgets/Core/WidgetManager/index.js#L239) for a performance hit that modifies a WidgetRepresentation each time the mouse is moved which will always `requestData()` on the widget representation
### Detailed behavior
This is because `macro.setter()` does a shallow equal comparison `if (model[field] !== value) {` instead of a deep equal.
### Expected behavior
I can see 2 options:
- Add deep equal check if `value` is an object in `macro.setter()`
- Create `macro.setObject(publicAPI, model, ['foo']);` convenient setter similar to `macro.setArray()`
### Environment
- **vtk.js**: master
|
True
|
Setting a model variable as object always trigger modified() - ### High-level description
If `macro.set(publicAPI, model, ['foo']);` and `foo` is an `Object` (e.g. `{a: 1, b: 2}`, then calling `setFoo({a: 1, b:2})` will always trigger `modified()`.
### Steps to reproduce
```
extend(...){
macro.set(publicAPI, model, ['foo']);
}
...
myobj.setFoo({a: 1, b: 2});
myobj.getMTime(); // =>1234
myobj.setFoo({a: 1, b: 2});
myobj.getMTime(); // =>1235
```
See [here](https://github.com/Kitware/vtk-js/blob/master/Sources/Widgets/Core/WidgetManager/index.js#L239) for a performance hit that modifies a WidgetRepresentation each time the mouse is moved which will always `requestData()` on the widget representation
### Detailed behavior
This is because `macro.setter()` does a shallow equal comparison `if (model[field] !== value) {` instead of a deep equal.
### Expected behavior
I can see 2 options:
- Add deep equal check if `value` is an object in `macro.setter()`
- Create `macro.setObject(publicAPI, model, ['foo']);` convenient setter similar to `macro.setArray()`
### Environment
- **vtk.js**: master
|
non_defect
|
setting a model variable as object always trigger modified high level description if macro set publicapi model and foo is an object e g a b then calling setfoo a b will always trigger modified steps to reproduce extend macro set publicapi model myobj setfoo a b myobj getmtime myobj setfoo a b myobj getmtime see for a performance hit that modifies a widgetrepresentation each time the mouse is moved which will always requestdata on the widget representation detailed behavior this is because macro setter does a shallow equal comparison if model value instead of a deep equal expected behavior i can see options add deep equal check if value is an object in macro setter create macro setobject publicapi model convenient setter similar to macro setarray environment vtk js master
| 0
|
47,552
| 13,056,241,652
|
IssuesEvent
|
2020-07-30 04:05:51
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
closed
|
steamshovel file open slowness (Trac #735)
|
Migrated from Trac combo core defect
|
"After investigating this problem a bit more, we found out how to get around the slowness we've been experiencing when using steamshovel in the offline-software meta-project. When we just run steamshovel, without passing the files in the command line, but opening them from the GUI, things work fine. When loading the files from command line, it goes very slowly."
Possibly something different between cmd line loading and file menu loading.
Migrated from https://code.icecube.wisc.edu/ticket/735
```json
{
"status": "closed",
"changetime": "2015-07-10T14:08:08",
"description": "\"After investigating this problem a bit more, we found out how to get around the slowness we've been experiencing when using steamshovel in the offline-software meta-project. When we just run steamshovel, without passing the files in the command line, but opening them from the GUI, things work fine. When loading the files from command line, it goes very slowly.\"\n\nPossibly something different between cmd line loading and file menu loading.",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1436537288260585",
"component": "combo core",
"summary": "steamshovel file open slowness",
"priority": "normal",
"keywords": "",
"time": "2014-08-13T15:04:32",
"milestone": "",
"owner": "hdembinski",
"type": "defect"
}
```
|
1.0
|
steamshovel file open slowness (Trac #735) - "After investigating this problem a bit more, we found out how to get around the slowness we've been experiencing when using steamshovel in the offline-software meta-project. When we just run steamshovel, without passing the files in the command line, but opening them from the GUI, things work fine. When loading the files from command line, it goes very slowly."
Possibly something different between cmd line loading and file menu loading.
Migrated from https://code.icecube.wisc.edu/ticket/735
```json
{
"status": "closed",
"changetime": "2015-07-10T14:08:08",
"description": "\"After investigating this problem a bit more, we found out how to get around the slowness we've been experiencing when using steamshovel in the offline-software meta-project. When we just run steamshovel, without passing the files in the command line, but opening them from the GUI, things work fine. When loading the files from command line, it goes very slowly.\"\n\nPossibly something different between cmd line loading and file menu loading.",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1436537288260585",
"component": "combo core",
"summary": "steamshovel file open slowness",
"priority": "normal",
"keywords": "",
"time": "2014-08-13T15:04:32",
"milestone": "",
"owner": "hdembinski",
"type": "defect"
}
```
|
defect
|
steamshovel file open slowness trac after investigating this problem a bit more we found out how to get around the slowness we ve been experiencing when using steamshovel in the offline software meta project when we just run steamshovel without passing the files in the command line but opening them from the gui things work fine when loading the files from command line it goes very slowly possibly something different between cmd line loading and file menu loading migrated from json status closed changetime description after investigating this problem a bit more we found out how to get around the slowness we ve been experiencing when using steamshovel in the offline software meta project when we just run steamshovel without passing the files in the command line but opening them from the gui things work fine when loading the files from command line it goes very slowly n npossibly something different between cmd line loading and file menu loading reporter david schultz cc resolution fixed ts component combo core summary steamshovel file open slowness priority normal keywords time milestone owner hdembinski type defect
| 1
|
21,020
| 3,442,784,529
|
IssuesEvent
|
2015-12-15 00:20:22
|
prettydiff/prettydiff
|
https://api.github.com/repos/prettydiff/prettydiff
|
closed
|
Conditions in tags get rearranged in twig templates
|
Defect QA
|
I've been using atom-beautify on twig templates, only to realize that it messed up some logic parts of the templates.
When beautifying twig templates, conditions inside of tags get interpreted as attributes, changing **if [...] endif** to **endif [...] if**.
```twig
<html>
<body>
<a href="/linktarget.html" {% if active %}class="active"{% endif %}>linktext</a>
</body>
</html>
```
turns into
```twig
<html>
<body>
<a class="active" href="/linktarget.html" {% endif %} {% if active %}>linktext</a>
</body>
</html>
```
|
1.0
|
Conditions in tags get rearranged in twig templates - I've been using atom-beautify on twig templates, only to realize that it messed up some logic parts of the templates.
When beautifying twig templates, conditions inside of tags get interpreted as attributes, changing **if [...] endif** to **endif [...] if**.
```twig
<html>
<body>
<a href="/linktarget.html" {% if active %}class="active"{% endif %}>linktext</a>
</body>
</html>
```
turns into
```twig
<html>
<body>
<a class="active" href="/linktarget.html" {% endif %} {% if active %}>linktext</a>
</body>
</html>
```
|
defect
|
conditions in tags get rearranged in twig templates i ve been using atom beautify on twig templates only to realize that it messed up some logic parts of the templates when beautifying twig templates conditions inside of tags get interpreted as attributes changing if endif to endif if twig linktext turns into twig linktext
| 1
|
290,241
| 32,045,665,693
|
IssuesEvent
|
2023-09-23 01:34:13
|
Chiencc/asuswrt-gt-ac5300
|
https://api.github.com/repos/Chiencc/asuswrt-gt-ac5300
|
reopened
|
jquery.mobile-1.4.5.min.js: 1 vulnerabilities (highest severity is: 6.5)
|
Mend: dependency security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery.mobile-1.4.5.min.js</b></p></summary>
<p>Touch-Optimized Web Framework for Smartphones & Tablets</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery-mobile/1.4.5/jquery.mobile.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery-mobile/1.4.5/jquery.mobile.min.js</a></p>
<p>Path to dependency file: /release/src/router/www/sysdep/VZW-AC1300/www/QIS_wizard.htm</p>
<p>Path to vulnerable library: /release/src/router/www/sysdep/BLUECAVE/www/mobile/js/jquery.mobile.js,/release/src/router/www/sysdep/VZW-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC2200/www/js/jquery.mobile.js,/release/src/router/www/sysdep/BLUECAVE/www/./mobile/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1750/www/js/jquery.mobile.js</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (jquery.mobile version) | Remediation Possible** |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [WS-2019-0136](https://github.com/jquery/jquery-mobile/commit/b0d9cc758a48f13321750d7409fb7655dcdf2b50) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.5 | jquery.mobile-1.4.5.min.js | Direct | BMC.NET - 1.0.3;org.webjars:jquery-mobile - 1.3.0-1,1.4.3;jquery.mobile - 1.3.0;jQWidgets_Framework - 8.0.0,6.0.6 | ❌ |
<p>**In some cases, Remediation PR cannot be created automatically for a vulnerability despite the availability of remediation</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> WS-2019-0136</summary>
### Vulnerable Library - <b>jquery.mobile-1.4.5.min.js</b></p>
<p>Touch-Optimized Web Framework for Smartphones & Tablets</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery-mobile/1.4.5/jquery.mobile.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery-mobile/1.4.5/jquery.mobile.min.js</a></p>
<p>Path to dependency file: /release/src/router/www/sysdep/VZW-AC1300/www/QIS_wizard.htm</p>
<p>Path to vulnerable library: /release/src/router/www/sysdep/BLUECAVE/www/mobile/js/jquery.mobile.js,/release/src/router/www/sysdep/VZW-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC2200/www/js/jquery.mobile.js,/release/src/router/www/sysdep/BLUECAVE/www/./mobile/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1750/www/js/jquery.mobile.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery.mobile-1.4.5.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
All versions of Jquery mobile have an open redirect that leads to cross-site scripting when the endpoint reflects user input.
<p>Publish Date: 2019-06-13
<p>URL: <a href=https://github.com/jquery/jquery-mobile/commit/b0d9cc758a48f13321750d7409fb7655dcdf2b50>WS-2019-0136</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/WS-2019-0136">https://nvd.nist.gov/vuln/detail/WS-2019-0136</a></p>
<p>Release Date: 2019-06-13</p>
<p>Fix Resolution: BMC.NET - 1.0.3;org.webjars:jquery-mobile - 1.3.0-1,1.4.3;jquery.mobile - 1.3.0;jQWidgets_Framework - 8.0.0,6.0.6</p>
</p>
<p></p>
</details>
|
True
|
jquery.mobile-1.4.5.min.js: 1 vulnerabilities (highest severity is: 6.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery.mobile-1.4.5.min.js</b></p></summary>
<p>Touch-Optimized Web Framework for Smartphones & Tablets</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery-mobile/1.4.5/jquery.mobile.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery-mobile/1.4.5/jquery.mobile.min.js</a></p>
<p>Path to dependency file: /release/src/router/www/sysdep/VZW-AC1300/www/QIS_wizard.htm</p>
<p>Path to vulnerable library: /release/src/router/www/sysdep/BLUECAVE/www/mobile/js/jquery.mobile.js,/release/src/router/www/sysdep/VZW-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC2200/www/js/jquery.mobile.js,/release/src/router/www/sysdep/BLUECAVE/www/./mobile/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1750/www/js/jquery.mobile.js</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (jquery.mobile version) | Remediation Possible** |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [WS-2019-0136](https://github.com/jquery/jquery-mobile/commit/b0d9cc758a48f13321750d7409fb7655dcdf2b50) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.5 | jquery.mobile-1.4.5.min.js | Direct | BMC.NET - 1.0.3;org.webjars:jquery-mobile - 1.3.0-1,1.4.3;jquery.mobile - 1.3.0;jQWidgets_Framework - 8.0.0,6.0.6 | ❌ |
<p>**In some cases, Remediation PR cannot be created automatically for a vulnerability despite the availability of remediation</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> WS-2019-0136</summary>
### Vulnerable Library - <b>jquery.mobile-1.4.5.min.js</b></p>
<p>Touch-Optimized Web Framework for Smartphones & Tablets</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery-mobile/1.4.5/jquery.mobile.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery-mobile/1.4.5/jquery.mobile.min.js</a></p>
<p>Path to dependency file: /release/src/router/www/sysdep/VZW-AC1300/www/QIS_wizard.htm</p>
<p>Path to vulnerable library: /release/src/router/www/sysdep/BLUECAVE/www/mobile/js/jquery.mobile.js,/release/src/router/www/sysdep/VZW-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC2200/www/js/jquery.mobile.js,/release/src/router/www/sysdep/BLUECAVE/www/./mobile/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1300/www/js/jquery.mobile.js,/release/src/router/www/sysdep/MAP-AC1750/www/js/jquery.mobile.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery.mobile-1.4.5.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
All versions of Jquery mobile have an open redirect that leads to cross-site scripting when the endpoint reflects user input.
<p>Publish Date: 2019-06-13
<p>URL: <a href=https://github.com/jquery/jquery-mobile/commit/b0d9cc758a48f13321750d7409fb7655dcdf2b50>WS-2019-0136</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/WS-2019-0136">https://nvd.nist.gov/vuln/detail/WS-2019-0136</a></p>
<p>Release Date: 2019-06-13</p>
<p>Fix Resolution: BMC.NET - 1.0.3;org.webjars:jquery-mobile - 1.3.0-1,1.4.3;jquery.mobile - 1.3.0;jQWidgets_Framework - 8.0.0,6.0.6</p>
</p>
<p></p>
</details>
|
non_defect
|
jquery mobile min js vulnerabilities highest severity is vulnerable library jquery mobile min js touch optimized web framework for smartphones tablets library home page a href path to dependency file release src router www sysdep vzw www qis wizard htm path to vulnerable library release src router www sysdep bluecave www mobile js jquery mobile js release src router www sysdep vzw www js jquery mobile js release src router www sysdep map www js jquery mobile js release src router www sysdep map www js jquery mobile js release src router www sysdep bluecave www mobile js jquery mobile js release src router www sysdep map www js jquery mobile js release src router www sysdep map www js jquery mobile js found in head commit a href vulnerabilities cve severity cvss dependency type fixed in jquery mobile version remediation possible medium jquery mobile min js direct bmc net org webjars jquery mobile jquery mobile jqwidgets framework in some cases remediation pr cannot be created automatically for a vulnerability despite the availability of remediation details ws vulnerable library jquery mobile min js touch optimized web framework for smartphones tablets library home page a href path to dependency file release src router www sysdep vzw www qis wizard htm path to vulnerable library release src router www sysdep bluecave www mobile js jquery mobile js release src router www sysdep vzw www js jquery mobile js release src router www sysdep map www js jquery mobile js release src router www sysdep map www js jquery mobile js release src router www sysdep bluecave www mobile js jquery mobile js release src router www sysdep map www js jquery mobile js release src router www sysdep map www js jquery mobile js dependency hierarchy x jquery mobile min js vulnerable library found in head commit a href found in base branch master vulnerability details all versions of jquery mobile have an open redirect that leads to cross site scripting when the endpoint reflects user input publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bmc net org webjars jquery mobile jquery mobile jqwidgets framework
| 0
|
24,675
| 17,600,369,614
|
IssuesEvent
|
2021-08-17 11:04:21
|
google/web-stories-wp
|
https://api.github.com/repos/google/web-stories-wp
|
closed
|
Packages: Import React dependency from own React package
|
P2 Type: Infrastructure Pod: WP & Infra Package: React
|
<!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ -->
## Feature Description
<!-- A clear and concise description of what the problem is and what you want to happen. -->
We created our own dedicated `react` package in #8159.
Now, we want to re-export `react` and `react-dom/server` functions from this package and make any package depend on this one. Similar to `@wordpress/element`.
This makes it easier to maintain the React dependency and update it in the future.
## Alternatives Considered
<!-- A clear and concise description of any alternative solutions or features you've considered. -->
## Additional Context
<!-- Add any other context or screenshots about the feature request. -->
---
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance Criteria
<!-- One or more bullet points for acceptance criteria. -->
## Implementation Brief
<!-- One or more bullet points for how to technically implement the feature. -->
|
1.0
|
Packages: Import React dependency from own React package - <!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ -->
## Feature Description
<!-- A clear and concise description of what the problem is and what you want to happen. -->
We created our own dedicated `react` package in #8159.
Now, we want to re-export `react` and `react-dom/server` functions from this package and make any package depend on this one. Similar to `@wordpress/element`.
This makes it easier to maintain the React dependency and update it in the future.
## Alternatives Considered
<!-- A clear and concise description of any alternative solutions or features you've considered. -->
## Additional Context
<!-- Add any other context or screenshots about the feature request. -->
---
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance Criteria
<!-- One or more bullet points for acceptance criteria. -->
## Implementation Brief
<!-- One or more bullet points for how to technically implement the feature. -->
|
non_defect
|
packages import react dependency from own react package feature description we created our own dedicated react package in now we want to re export react and react dom server functions from this package and make any package depend on this one similar to wordpress element this makes it easier to maintain the react dependency and update it in the future alternatives considered additional context do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria implementation brief
| 0
|
72,881
| 31,769,581,331
|
IssuesEvent
|
2023-09-12 10:53:48
|
gauravrs18/issue_onboarding
|
https://api.github.com/repos/gauravrs18/issue_onboarding
|
closed
|
dev-angular-style-account-services-new-connection-component-approve-component
-consumer-details-component
-application-component
-map-component
|
CX-account-services
|
dev-angular-style-account-services-new-connection-component-approve-component
-consumer-details-component
-application-component
-map-component
|
1.0
|
dev-angular-style-account-services-new-connection-component-approve-component
-consumer-details-component
-application-component
-map-component - dev-angular-style-account-services-new-connection-component-approve-component
-consumer-details-component
-application-component
-map-component
|
non_defect
|
dev angular style account services new connection component approve component consumer details component application component map component dev angular style account services new connection component approve component consumer details component application component map component
| 0
|
15,552
| 5,143,603,622
|
IssuesEvent
|
2017-01-12 16:24:05
|
threadly/threadly
|
https://api.github.com/repos/threadly/threadly
|
closed
|
Perform actions to release version 4.9.0
|
non-code related
|
4.9.0 will be the last feature expansion release before 5.0.0 is released. This issue is to track the release process for 4.9.0. Once this issue is closed, it will indicate that 4.9.0 has been released, and is available.
Items to update:
- Homepage stable download
- Homepage javadocs
- Homepage news
- Wiki homepage for java docs
- Wiki page for complete list of javadocs links
- Add release information to github
Code operations:
- Deploy artifact to sonatype
- Tag git
- Keep build forever in jenkins
- rebase `unstable_staging` on top of `master`
|
1.0
|
Perform actions to release version 4.9.0 - 4.9.0 will be the last feature expansion release before 5.0.0 is released. This issue is to track the release process for 4.9.0. Once this issue is closed, it will indicate that 4.9.0 has been released, and is available.
Items to update:
- Homepage stable download
- Homepage javadocs
- Homepage news
- Wiki homepage for java docs
- Wiki page for complete list of javadocs links
- Add release information to github
Code operations:
- Deploy artifact to sonatype
- Tag git
- Keep build forever in jenkins
- rebase `unstable_staging` on top of `master`
|
non_defect
|
perform actions to release version will be the last feature expansion release before is released this issue is to track the release process for once this issue is closed it will indicate that has been released and is available items to update homepage stable download homepage javadocs homepage news wiki homepage for java docs wiki page for complete list of javadocs links add release information to github code operations deploy artifact to sonatype tag git keep build forever in jenkins rebase unstable staging on top of master
| 0
|
556,627
| 16,487,403,020
|
IssuesEvent
|
2021-05-24 20:13:02
|
mlflow/mlflow
|
https://api.github.com/repos/mlflow/mlflow
|
closed
|
[BUG] INTERNAL_SERVER_ERROR when sorting the runs table by metric or by date
|
area/tracking area/uiux bug priority/important-soon
|
### System information
- **Have I written custom code (as opposed to using a stock example script provided in MLflow)**: no
- **OS Platform and Distribution (e.g., Linux Ubuntu 16.04)**: Linux Debian
- **MLflow installed from (source or binary)**: binary
- **MLflow version (run ``mlflow --version``)**: 1.7.0
- **Python version**: 3.7
- **npm version, if running the dev UI**: -
- **Exact command to reproduce**: run the mlflow server and log parameters and metrics
### Describe the problem
When I try to sort the runs by any metric or date (parameters, versions and users work) I get an `INTERNAL_SERVER_ERROR` and the whole table disappears. Then, if I sort by a parameter, for example, the table comes back.
**UPDATE:**

### Code to reproduce issue
run the mlflow server and log parameters and metrics
### Other info / logs
This is the trace-back I got from the server:
```
2020/03/04 09:40:50 ERROR mlflow.server: Exception on /ajax-api/2.0/preview/mlflow/runs/search [POST]
Traceback (most recent call last):
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 1246, in _execute_context
cursor, statement, parameters, context
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/default.py", line 588, in do_execute
cursor.execute(statement, parameters)
pyodbc.ProgrammingError: ('42000', "[42000] [Microsoft][ODBC Driver 17 for SQL Server][SQL Server]Incorrect syntax near '1'. (102) (SQLExecDirectW)")
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.7/site-packages/mlflow/store/db/utils.py", line 73, in make_managed_session
yield session
File "/usr/local/lib/python3.7/site-packages/mlflow/store/tracking/sqlalchemy_store.py", line 651, in _search_runs
.offset(offset).limit(max_results).all()
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/orm/query.py", line 3233, in all
return list(self)
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/orm/query.py", line 3389, in __iter__
return self._execute_and_instances(context)
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/orm/query.py", line 3414, in _execute_and_instances
result = conn.execute(querycontext.statement, self._params)
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 982, in execute
return meth(self, multiparams, params)
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/sql/elements.py", line 293, in _execute_on_connection
return connection._execute_clauseelement(self, multiparams, params)
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 1101, in _execute_clauseelement
distilled_params,
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 1250, in _execute_context
e, statement, parameters, cursor, context
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 1476, in _handle_dbapi_exception
util.raise_from_cause(sqlalchemy_exception, exc_info)
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/util/compat.py", line 398, in raise_from_cause
reraise(type(exception), exception, tb=exc_tb, cause=cause)
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/util/compat.py", line 152, in reraise
raise value.with_traceback(tb)
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 1246, in _execute_context
cursor, statement, parameters, context
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/default.py", line 588, in do_execute
cursor.execute(statement, parameters)
sqlalchemy.exc.ProgrammingError: (pyodbc.ProgrammingError) ('42000', "[42000] [Microsoft][ODBC Driver 17 for SQL Server][SQL Server]Incorrect syntax near '1'. (102) (SQLExecDirectW)")
[SQL: SELECT DISTINCT TOP 100 runs.run_uuid AS runs_run_uuid, runs.name AS runs_name, runs.source_type AS runs_source_type, runs.source_name AS runs_source_name, runs.entry_point_name AS runs_entry_point_name, runs.user_id AS runs_user_id, runs.status AS runs_status, runs.start_time AS runs_start_time, runs.end_time AS runs_end_time, runs.source_version AS runs_source_version, runs.lifecycle_stage AS runs_lifecycle_stage, runs.artifact_uri AS runs_artifact_uri, runs.experiment_id AS runs_experiment_id, CASE WHEN (anon_1.is_nan IS 1) THEN ? WHEN (anon_1.value IS NULL) THEN ? ELSE ? END AS clause_1, anon_1.value AS anon_1_value
FROM runs LEFT OUTER JOIN (SELECT latest_metrics.[key] AS [key], latest_metrics.value AS value, latest_metrics.timestamp AS timestamp, latest_metrics.step AS step, latest_metrics.is_nan AS is_nan, latest_metrics.run_uuid AS run_uuid
FROM latest_metrics
WHERE latest_metrics.[key] = ?) AS anon_1 ON runs.run_uuid = anon_1.run_uuid
WHERE runs.experiment_id IN (?) AND runs.lifecycle_stage IN (?) ORDER BY clause_1, anon_1.value, runs.start_time DESC, runs.run_uuid]
[parameters: (1, 1, 0, 'auc', '4', 'active')]
(Background on this error at: http://sqlalche.me/e/f405)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.7/site-packages/mlflow/server/handlers.py", line 155, in wrapper
return func(*args, **kwargs)
File "/usr/local/lib/python3.7/site-packages/mlflow/server/handlers.py", line 394, in _search_runs
max_results, order_by, page_token)
File "/usr/local/lib/python3.7/site-packages/mlflow/store/tracking/abstract_store.py", line 230, in search_runs
order_by, page_token)
File "/usr/local/lib/python3.7/site-packages/mlflow/store/tracking/sqlalchemy_store.py", line 654, in _search_runs
next_page_token = compute_next_token(len(runs))
File "/usr/local/lib/python3.7/contextlib.py", line 130, in __exit__
self.gen.throw(type, value, traceback)
File "/usr/local/lib/python3.7/site-packages/mlflow/store/db/utils.py", line 80, in make_managed_session
raise MlflowException(message=e, error_code=INTERNAL_ERROR)
mlflow.exceptions.MlflowException: (pyodbc.ProgrammingError) ('42000', "[42000] [Microsoft][ODBC Driver 17 for SQL Server][SQL Server]Incorrect syntax near '1'. (102) (SQLExecDirectW)")
[SQL: SELECT DISTINCT TOP 100 runs.run_uuid AS runs_run_uuid, runs.name AS runs_name, runs.source_type AS runs_source_type, runs.source_name AS runs_source_name, runs.entry_point_name AS runs_entry_point_name, runs.user_id AS runs_user_id, runs.status AS runs_status, runs.start_time AS runs_start_time, runs.end_time AS runs_end_time, runs.source_version AS runs_source_version, runs.lifecycle_stage AS runs_lifecycle_stage, runs.artifact_uri AS runs_artifact_uri, runs.experiment_id AS runs_experiment_id, CASE WHEN (anon_1.is_nan IS 1) THEN ? WHEN (anon_1.value IS NULL) THEN ? ELSE ? END AS clause_1, anon_1.value AS anon_1_value
FROM runs LEFT OUTER JOIN (SELECT latest_metrics.[key] AS [key], latest_metrics.value AS value, latest_metrics.timestamp AS timestamp, latest_metrics.step AS step, latest_metrics.is_nan AS is_nan, latest_metrics.run_uuid AS run_uuid
FROM latest_metrics
WHERE latest_metrics.[key] = ?) AS anon_1 ON runs.run_uuid = anon_1.run_uuid
WHERE runs.experiment_id IN (?) AND runs.lifecycle_stage IN (?) ORDER BY clause_1, anon_1.value, runs.start_time DESC, runs.run_uuid]
[parameters: (1, 1, 0, 'auc', '4', 'active')]
(Background on this error at: http://sqlalche.me/e/f405)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.7/site-packages/flask/app.py", line 2446, in wsgi_app
response = self.full_dispatch_request()
File "/usr/local/lib/python3.7/site-packages/flask/app.py", line 1951, in full_dispatch_request
rv = self.handle_user_exception(e)
File "/usr/local/lib/python3.7/site-packages/flask/app.py", line 1820, in handle_user_exception
reraise(exc_type, exc_value, tb)
File "/usr/local/lib/python3.7/site-packages/flask/_compat.py", line 39, in reraise
raise value
File "/usr/local/lib/python3.7/site-packages/flask/app.py", line 1949, in full_dispatch_request
rv = self.dispatch_request()
File "/usr/local/lib/python3.7/site-packages/flask/app.py", line 1935, in dispatch_request
return self.view_functions[rule.endpoint](**req.view_args)
File "/usr/local/lib/python3.7/site-packages/mlflow/server/handlers.py", line 158, in wrapper
response.set_data(e.serialize_as_json())
File "/usr/local/lib/python3.7/site-packages/mlflow/exceptions.py", line 50, in serialize_as_json
return json.dumps(exception_dict)
File "/usr/local/lib/python3.7/json/__init__.py", line 231, in dumps
return _default_encoder.encode(obj)
File "/usr/local/lib/python3.7/json/encoder.py", line 199, in encode
chunks = self.iterencode(o, _one_shot=True)
File "/usr/local/lib/python3.7/json/encoder.py", line 257, in iterencode
return _iterencode(o, 0)
File "/usr/local/lib/python3.7/json/encoder.py", line 179, in default
raise TypeError(f'Object of type {o.__class__.__name__} '
TypeError: Object of type ProgrammingError is not JSON serializable
```
### **UPDATE**
I've updated the version to v1.7.2 and the bug is still there.
|
1.0
|
[BUG] INTERNAL_SERVER_ERROR when sorting the runs table by metric or by date - ### System information
- **Have I written custom code (as opposed to using a stock example script provided in MLflow)**: no
- **OS Platform and Distribution (e.g., Linux Ubuntu 16.04)**: Linux Debian
- **MLflow installed from (source or binary)**: binary
- **MLflow version (run ``mlflow --version``)**: 1.7.0
- **Python version**: 3.7
- **npm version, if running the dev UI**: -
- **Exact command to reproduce**: run the mlflow server and log parameters and metrics
### Describe the problem
When I try to sort the runs by any metric or date (parameters, versions and users work) I get an `INTERNAL_SERVER_ERROR` and the whole table disappears. Then, if I sort by a parameter, for example, the table comes back.
**UPDATE:**

### Code to reproduce issue
run the mlflow server and log parameters and metrics
### Other info / logs
This is the trace-back I got from the server:
```
2020/03/04 09:40:50 ERROR mlflow.server: Exception on /ajax-api/2.0/preview/mlflow/runs/search [POST]
Traceback (most recent call last):
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 1246, in _execute_context
cursor, statement, parameters, context
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/default.py", line 588, in do_execute
cursor.execute(statement, parameters)
pyodbc.ProgrammingError: ('42000', "[42000] [Microsoft][ODBC Driver 17 for SQL Server][SQL Server]Incorrect syntax near '1'. (102) (SQLExecDirectW)")
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.7/site-packages/mlflow/store/db/utils.py", line 73, in make_managed_session
yield session
File "/usr/local/lib/python3.7/site-packages/mlflow/store/tracking/sqlalchemy_store.py", line 651, in _search_runs
.offset(offset).limit(max_results).all()
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/orm/query.py", line 3233, in all
return list(self)
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/orm/query.py", line 3389, in __iter__
return self._execute_and_instances(context)
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/orm/query.py", line 3414, in _execute_and_instances
result = conn.execute(querycontext.statement, self._params)
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 982, in execute
return meth(self, multiparams, params)
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/sql/elements.py", line 293, in _execute_on_connection
return connection._execute_clauseelement(self, multiparams, params)
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 1101, in _execute_clauseelement
distilled_params,
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 1250, in _execute_context
e, statement, parameters, cursor, context
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 1476, in _handle_dbapi_exception
util.raise_from_cause(sqlalchemy_exception, exc_info)
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/util/compat.py", line 398, in raise_from_cause
reraise(type(exception), exception, tb=exc_tb, cause=cause)
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/util/compat.py", line 152, in reraise
raise value.with_traceback(tb)
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 1246, in _execute_context
cursor, statement, parameters, context
File "/usr/local/lib/python3.7/site-packages/sqlalchemy/engine/default.py", line 588, in do_execute
cursor.execute(statement, parameters)
sqlalchemy.exc.ProgrammingError: (pyodbc.ProgrammingError) ('42000', "[42000] [Microsoft][ODBC Driver 17 for SQL Server][SQL Server]Incorrect syntax near '1'. (102) (SQLExecDirectW)")
[SQL: SELECT DISTINCT TOP 100 runs.run_uuid AS runs_run_uuid, runs.name AS runs_name, runs.source_type AS runs_source_type, runs.source_name AS runs_source_name, runs.entry_point_name AS runs_entry_point_name, runs.user_id AS runs_user_id, runs.status AS runs_status, runs.start_time AS runs_start_time, runs.end_time AS runs_end_time, runs.source_version AS runs_source_version, runs.lifecycle_stage AS runs_lifecycle_stage, runs.artifact_uri AS runs_artifact_uri, runs.experiment_id AS runs_experiment_id, CASE WHEN (anon_1.is_nan IS 1) THEN ? WHEN (anon_1.value IS NULL) THEN ? ELSE ? END AS clause_1, anon_1.value AS anon_1_value
FROM runs LEFT OUTER JOIN (SELECT latest_metrics.[key] AS [key], latest_metrics.value AS value, latest_metrics.timestamp AS timestamp, latest_metrics.step AS step, latest_metrics.is_nan AS is_nan, latest_metrics.run_uuid AS run_uuid
FROM latest_metrics
WHERE latest_metrics.[key] = ?) AS anon_1 ON runs.run_uuid = anon_1.run_uuid
WHERE runs.experiment_id IN (?) AND runs.lifecycle_stage IN (?) ORDER BY clause_1, anon_1.value, runs.start_time DESC, runs.run_uuid]
[parameters: (1, 1, 0, 'auc', '4', 'active')]
(Background on this error at: http://sqlalche.me/e/f405)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.7/site-packages/mlflow/server/handlers.py", line 155, in wrapper
return func(*args, **kwargs)
File "/usr/local/lib/python3.7/site-packages/mlflow/server/handlers.py", line 394, in _search_runs
max_results, order_by, page_token)
File "/usr/local/lib/python3.7/site-packages/mlflow/store/tracking/abstract_store.py", line 230, in search_runs
order_by, page_token)
File "/usr/local/lib/python3.7/site-packages/mlflow/store/tracking/sqlalchemy_store.py", line 654, in _search_runs
next_page_token = compute_next_token(len(runs))
File "/usr/local/lib/python3.7/contextlib.py", line 130, in __exit__
self.gen.throw(type, value, traceback)
File "/usr/local/lib/python3.7/site-packages/mlflow/store/db/utils.py", line 80, in make_managed_session
raise MlflowException(message=e, error_code=INTERNAL_ERROR)
mlflow.exceptions.MlflowException: (pyodbc.ProgrammingError) ('42000', "[42000] [Microsoft][ODBC Driver 17 for SQL Server][SQL Server]Incorrect syntax near '1'. (102) (SQLExecDirectW)")
[SQL: SELECT DISTINCT TOP 100 runs.run_uuid AS runs_run_uuid, runs.name AS runs_name, runs.source_type AS runs_source_type, runs.source_name AS runs_source_name, runs.entry_point_name AS runs_entry_point_name, runs.user_id AS runs_user_id, runs.status AS runs_status, runs.start_time AS runs_start_time, runs.end_time AS runs_end_time, runs.source_version AS runs_source_version, runs.lifecycle_stage AS runs_lifecycle_stage, runs.artifact_uri AS runs_artifact_uri, runs.experiment_id AS runs_experiment_id, CASE WHEN (anon_1.is_nan IS 1) THEN ? WHEN (anon_1.value IS NULL) THEN ? ELSE ? END AS clause_1, anon_1.value AS anon_1_value
FROM runs LEFT OUTER JOIN (SELECT latest_metrics.[key] AS [key], latest_metrics.value AS value, latest_metrics.timestamp AS timestamp, latest_metrics.step AS step, latest_metrics.is_nan AS is_nan, latest_metrics.run_uuid AS run_uuid
FROM latest_metrics
WHERE latest_metrics.[key] = ?) AS anon_1 ON runs.run_uuid = anon_1.run_uuid
WHERE runs.experiment_id IN (?) AND runs.lifecycle_stage IN (?) ORDER BY clause_1, anon_1.value, runs.start_time DESC, runs.run_uuid]
[parameters: (1, 1, 0, 'auc', '4', 'active')]
(Background on this error at: http://sqlalche.me/e/f405)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/lib/python3.7/site-packages/flask/app.py", line 2446, in wsgi_app
response = self.full_dispatch_request()
File "/usr/local/lib/python3.7/site-packages/flask/app.py", line 1951, in full_dispatch_request
rv = self.handle_user_exception(e)
File "/usr/local/lib/python3.7/site-packages/flask/app.py", line 1820, in handle_user_exception
reraise(exc_type, exc_value, tb)
File "/usr/local/lib/python3.7/site-packages/flask/_compat.py", line 39, in reraise
raise value
File "/usr/local/lib/python3.7/site-packages/flask/app.py", line 1949, in full_dispatch_request
rv = self.dispatch_request()
File "/usr/local/lib/python3.7/site-packages/flask/app.py", line 1935, in dispatch_request
return self.view_functions[rule.endpoint](**req.view_args)
File "/usr/local/lib/python3.7/site-packages/mlflow/server/handlers.py", line 158, in wrapper
response.set_data(e.serialize_as_json())
File "/usr/local/lib/python3.7/site-packages/mlflow/exceptions.py", line 50, in serialize_as_json
return json.dumps(exception_dict)
File "/usr/local/lib/python3.7/json/__init__.py", line 231, in dumps
return _default_encoder.encode(obj)
File "/usr/local/lib/python3.7/json/encoder.py", line 199, in encode
chunks = self.iterencode(o, _one_shot=True)
File "/usr/local/lib/python3.7/json/encoder.py", line 257, in iterencode
return _iterencode(o, 0)
File "/usr/local/lib/python3.7/json/encoder.py", line 179, in default
raise TypeError(f'Object of type {o.__class__.__name__} '
TypeError: Object of type ProgrammingError is not JSON serializable
```
### **UPDATE**
I've updated the version to v1.7.2 and the bug is still there.
|
non_defect
|
internal server error when sorting the runs table by metric or by date system information have i written custom code as opposed to using a stock example script provided in mlflow no os platform and distribution e g linux ubuntu linux debian mlflow installed from source or binary binary mlflow version run mlflow version python version npm version if running the dev ui exact command to reproduce run the mlflow server and log parameters and metrics describe the problem when i try to sort the runs by any metric or date parameters versions and users work i get an internal server error and the whole table disappears then if i sort by a parameter for example the table comes back update code to reproduce issue run the mlflow server and log parameters and metrics other info logs this is the trace back i got from the server error mlflow server exception on ajax api preview mlflow runs search traceback most recent call last file usr local lib site packages sqlalchemy engine base py line in execute context cursor statement parameters context file usr local lib site packages sqlalchemy engine default py line in do execute cursor execute statement parameters pyodbc programmingerror incorrect syntax near sqlexecdirectw the above exception was the direct cause of the following exception traceback most recent call last file usr local lib site packages mlflow store db utils py line in make managed session yield session file usr local lib site packages mlflow store tracking sqlalchemy store py line in search runs offset offset limit max results all file usr local lib site packages sqlalchemy orm query py line in all return list self file usr local lib site packages sqlalchemy orm query py line in iter return self execute and instances context file usr local lib site packages sqlalchemy orm query py line in execute and instances result conn execute querycontext statement self params file usr local lib site packages sqlalchemy engine base py line in execute return meth self multiparams params file usr local lib site packages sqlalchemy sql elements py line in execute on connection return connection execute clauseelement self multiparams params file usr local lib site packages sqlalchemy engine base py line in execute clauseelement distilled params file usr local lib site packages sqlalchemy engine base py line in execute context e statement parameters cursor context file usr local lib site packages sqlalchemy engine base py line in handle dbapi exception util raise from cause sqlalchemy exception exc info file usr local lib site packages sqlalchemy util compat py line in raise from cause reraise type exception exception tb exc tb cause cause file usr local lib site packages sqlalchemy util compat py line in reraise raise value with traceback tb file usr local lib site packages sqlalchemy engine base py line in execute context cursor statement parameters context file usr local lib site packages sqlalchemy engine default py line in do execute cursor execute statement parameters sqlalchemy exc programmingerror pyodbc programmingerror incorrect syntax near sqlexecdirectw sql select distinct top runs run uuid as runs run uuid runs name as runs name runs source type as runs source type runs source name as runs source name runs entry point name as runs entry point name runs user id as runs user id runs status as runs status runs start time as runs start time runs end time as runs end time runs source version as runs source version runs lifecycle stage as runs lifecycle stage runs artifact uri as runs artifact uri runs experiment id as runs experiment id case when anon is nan is then when anon value is null then else end as clause anon value as anon value from runs left outer join select latest metrics as latest metrics value as value latest metrics timestamp as timestamp latest metrics step as step latest metrics is nan as is nan latest metrics run uuid as run uuid from latest metrics where latest metrics as anon on runs run uuid anon run uuid where runs experiment id in and runs lifecycle stage in order by clause anon value runs start time desc runs run uuid background on this error at during handling of the above exception another exception occurred traceback most recent call last file usr local lib site packages mlflow server handlers py line in wrapper return func args kwargs file usr local lib site packages mlflow server handlers py line in search runs max results order by page token file usr local lib site packages mlflow store tracking abstract store py line in search runs order by page token file usr local lib site packages mlflow store tracking sqlalchemy store py line in search runs next page token compute next token len runs file usr local lib contextlib py line in exit self gen throw type value traceback file usr local lib site packages mlflow store db utils py line in make managed session raise mlflowexception message e error code internal error mlflow exceptions mlflowexception pyodbc programmingerror incorrect syntax near sqlexecdirectw sql select distinct top runs run uuid as runs run uuid runs name as runs name runs source type as runs source type runs source name as runs source name runs entry point name as runs entry point name runs user id as runs user id runs status as runs status runs start time as runs start time runs end time as runs end time runs source version as runs source version runs lifecycle stage as runs lifecycle stage runs artifact uri as runs artifact uri runs experiment id as runs experiment id case when anon is nan is then when anon value is null then else end as clause anon value as anon value from runs left outer join select latest metrics as latest metrics value as value latest metrics timestamp as timestamp latest metrics step as step latest metrics is nan as is nan latest metrics run uuid as run uuid from latest metrics where latest metrics as anon on runs run uuid anon run uuid where runs experiment id in and runs lifecycle stage in order by clause anon value runs start time desc runs run uuid background on this error at during handling of the above exception another exception occurred traceback most recent call last file usr local lib site packages flask app py line in wsgi app response self full dispatch request file usr local lib site packages flask app py line in full dispatch request rv self handle user exception e file usr local lib site packages flask app py line in handle user exception reraise exc type exc value tb file usr local lib site packages flask compat py line in reraise raise value file usr local lib site packages flask app py line in full dispatch request rv self dispatch request file usr local lib site packages flask app py line in dispatch request return self view functions req view args file usr local lib site packages mlflow server handlers py line in wrapper response set data e serialize as json file usr local lib site packages mlflow exceptions py line in serialize as json return json dumps exception dict file usr local lib json init py line in dumps return default encoder encode obj file usr local lib json encoder py line in encode chunks self iterencode o one shot true file usr local lib json encoder py line in iterencode return iterencode o file usr local lib json encoder py line in default raise typeerror f object of type o class name typeerror object of type programmingerror is not json serializable update i ve updated the version to and the bug is still there
| 0
|
95,111
| 3,934,256,902
|
IssuesEvent
|
2016-04-25 21:56:59
|
ceylon/ceylon-ide-eclipse
|
https://api.github.com/repos/ceylon/ceylon-ide-eclipse
|
closed
|
initializer annotation and highlight range are completely broken
|
bug on last release bug on master high priority
|
`AdditionalAnnotationCreator` is busted :-(
|
1.0
|
initializer annotation and highlight range are completely broken - `AdditionalAnnotationCreator` is busted :-(
|
non_defect
|
initializer annotation and highlight range are completely broken additionalannotationcreator is busted
| 0
|
9,812
| 2,615,175,313
|
IssuesEvent
|
2015-03-01 06:58:36
|
chrsmith/reaver-wps
|
https://api.github.com/repos/chrsmith/reaver-wps
|
opened
|
1.4 reaver issue with AWUS036H.
|
auto-migrated Priority-Triage Type-Defect
|
```
I have gone through just about ever post out there about wireless not working
after 2 minutes etc. Here are my findings from the past week of fiddling around
with this.
I have installed two AWUS036H usb cards on the same laptop running backtrack 5
r3 (i have also tested this with Kali linux 1.0.7 with same results) Hard drive
and live cd. all the findings are exactly the same.
all of the operating systems have 1.4 reaver on them.
My problem is, when i start using reaver 1.4 within couple minutes my awus036h
card rtl8781L along with all the wireless cards stop working. I know its
difficult to imagine. The only real work around i have found is when reaver
comes up with error Warning! Received timeout occured (then reaver hangs and
stops working) is to unplug the card from usb, plug it back in and put it in
monitor mode then continue the attack until it locks up again.
Things that i have tried:
1) rule out all software conflicts, did airmon-ng start (wlan0 and wlan1 cards
which are both my awus036h cards).. kill all possible conflicts.. run
airodump-ng on one of the cards while keeping the other card on another
terminal window so i have two terminals running airdump while on one of those
runs reaver as well. when reaver locks up, both of the monitored interfaces go
down and by going down i mean all input like packet, data, and all that goes
flat dead. there is no software conflict. i have ran iwconfig before and after
reaver takes a shit and no configuration has been changed with interfaces.
nothing got switched between managed or monitored modes, there is plenty of
power to both cards, i have used the Y adapter cables to input additional power
to the cards so thats not the issue as well. I should also mention i have an
internal wireless card that is intel something.. which still continued to run
airdump after reaver has taken out my awus036h card driver. (while both of the
cards are seased, i switched reaver to my intel wireless card mon3 interface
(actual adapter is wlan3) it locked that up too! within 2 mins use. I have
pretty much done all i can but it seems reaver 1.4 triggers some kind of glitch
when used with awus036h card which brings down all the wireless interfaces no
matter that they are if you point reaver to the divice.. it will bring it down
until you physically unplug it and put it back in OR reboot. then you have
another 1-2 minutes before it locks up again. I have not had any issues using
awus036hA the aerthos chipset that keeps running just fine. but rtl8717l is
another story. oh and also i have completely uninstalled dhcp client, disabled
network monitor, suplements, configured /network/interface so none of any
software causes any kind of conflit or even looks at the wireless interfaces.
is there anything else i've missed? and no i have not been able to find an
actual solution to this problem although i have read a lot of posts and
problems may sound similar but not in their entirety. yes i have tried every
single one of the suggested fixes and ran out of options but to stock up on new
wireless cards. any suggestions?
```
Original issue reported on code.google.com by `berezini...@gmail.com` on 9 Jul 2014 at 9:50
|
1.0
|
1.4 reaver issue with AWUS036H. - ```
I have gone through just about ever post out there about wireless not working
after 2 minutes etc. Here are my findings from the past week of fiddling around
with this.
I have installed two AWUS036H usb cards on the same laptop running backtrack 5
r3 (i have also tested this with Kali linux 1.0.7 with same results) Hard drive
and live cd. all the findings are exactly the same.
all of the operating systems have 1.4 reaver on them.
My problem is, when i start using reaver 1.4 within couple minutes my awus036h
card rtl8781L along with all the wireless cards stop working. I know its
difficult to imagine. The only real work around i have found is when reaver
comes up with error Warning! Received timeout occured (then reaver hangs and
stops working) is to unplug the card from usb, plug it back in and put it in
monitor mode then continue the attack until it locks up again.
Things that i have tried:
1) rule out all software conflicts, did airmon-ng start (wlan0 and wlan1 cards
which are both my awus036h cards).. kill all possible conflicts.. run
airodump-ng on one of the cards while keeping the other card on another
terminal window so i have two terminals running airdump while on one of those
runs reaver as well. when reaver locks up, both of the monitored interfaces go
down and by going down i mean all input like packet, data, and all that goes
flat dead. there is no software conflict. i have ran iwconfig before and after
reaver takes a shit and no configuration has been changed with interfaces.
nothing got switched between managed or monitored modes, there is plenty of
power to both cards, i have used the Y adapter cables to input additional power
to the cards so thats not the issue as well. I should also mention i have an
internal wireless card that is intel something.. which still continued to run
airdump after reaver has taken out my awus036h card driver. (while both of the
cards are seased, i switched reaver to my intel wireless card mon3 interface
(actual adapter is wlan3) it locked that up too! within 2 mins use. I have
pretty much done all i can but it seems reaver 1.4 triggers some kind of glitch
when used with awus036h card which brings down all the wireless interfaces no
matter that they are if you point reaver to the divice.. it will bring it down
until you physically unplug it and put it back in OR reboot. then you have
another 1-2 minutes before it locks up again. I have not had any issues using
awus036hA the aerthos chipset that keeps running just fine. but rtl8717l is
another story. oh and also i have completely uninstalled dhcp client, disabled
network monitor, suplements, configured /network/interface so none of any
software causes any kind of conflit or even looks at the wireless interfaces.
is there anything else i've missed? and no i have not been able to find an
actual solution to this problem although i have read a lot of posts and
problems may sound similar but not in their entirety. yes i have tried every
single one of the suggested fixes and ran out of options but to stock up on new
wireless cards. any suggestions?
```
Original issue reported on code.google.com by `berezini...@gmail.com` on 9 Jul 2014 at 9:50
|
defect
|
reaver issue with i have gone through just about ever post out there about wireless not working after minutes etc here are my findings from the past week of fiddling around with this i have installed two usb cards on the same laptop running backtrack i have also tested this with kali linux with same results hard drive and live cd all the findings are exactly the same all of the operating systems have reaver on them my problem is when i start using reaver within couple minutes my card along with all the wireless cards stop working i know its difficult to imagine the only real work around i have found is when reaver comes up with error warning received timeout occured then reaver hangs and stops working is to unplug the card from usb plug it back in and put it in monitor mode then continue the attack until it locks up again things that i have tried rule out all software conflicts did airmon ng start and cards which are both my cards kill all possible conflicts run airodump ng on one of the cards while keeping the other card on another terminal window so i have two terminals running airdump while on one of those runs reaver as well when reaver locks up both of the monitored interfaces go down and by going down i mean all input like packet data and all that goes flat dead there is no software conflict i have ran iwconfig before and after reaver takes a shit and no configuration has been changed with interfaces nothing got switched between managed or monitored modes there is plenty of power to both cards i have used the y adapter cables to input additional power to the cards so thats not the issue as well i should also mention i have an internal wireless card that is intel something which still continued to run airdump after reaver has taken out my card driver while both of the cards are seased i switched reaver to my intel wireless card interface actual adapter is it locked that up too within mins use i have pretty much done all i can but it seems reaver triggers some kind of glitch when used with card which brings down all the wireless interfaces no matter that they are if you point reaver to the divice it will bring it down until you physically unplug it and put it back in or reboot then you have another minutes before it locks up again i have not had any issues using the aerthos chipset that keeps running just fine but is another story oh and also i have completely uninstalled dhcp client disabled network monitor suplements configured network interface so none of any software causes any kind of conflit or even looks at the wireless interfaces is there anything else i ve missed and no i have not been able to find an actual solution to this problem although i have read a lot of posts and problems may sound similar but not in their entirety yes i have tried every single one of the suggested fixes and ran out of options but to stock up on new wireless cards any suggestions original issue reported on code google com by berezini gmail com on jul at
| 1
|
48,476
| 13,094,936,504
|
IssuesEvent
|
2020-08-03 13:18:59
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
closed
|
CacheSimpleConfig should accept cache name in its constructors
|
Module: ICache Source: Community Team: Core Type: Defect
|
Other configs - such as `MapConfig` or `QueueConfig` have a constructor with a name.
This allow to use constructs such as: `config.addMapConfig(new MapConfig("myMap").setFoo(foo));`
CacheSimpleConfig does not have this constructor. It should have it for the same of consistency.
|
1.0
|
CacheSimpleConfig should accept cache name in its constructors - Other configs - such as `MapConfig` or `QueueConfig` have a constructor with a name.
This allow to use constructs such as: `config.addMapConfig(new MapConfig("myMap").setFoo(foo));`
CacheSimpleConfig does not have this constructor. It should have it for the same of consistency.
|
defect
|
cachesimpleconfig should accept cache name in its constructors other configs such as mapconfig or queueconfig have a constructor with a name this allow to use constructs such as config addmapconfig new mapconfig mymap setfoo foo cachesimpleconfig does not have this constructor it should have it for the same of consistency
| 1
|
18,118
| 3,024,309,808
|
IssuesEvent
|
2015-08-02 13:46:35
|
aayush93/xbt
|
https://api.github.com/repos/aayush93/xbt
|
closed
|
IPv6 time
|
auto-migrated Priority-Medium Type-Defect
|
```
How about IPv6 support ?
```
Original issue reported on code.google.com by `allfe...@yandex.ru` on 21 Nov 2012 at 6:47
|
1.0
|
IPv6 time - ```
How about IPv6 support ?
```
Original issue reported on code.google.com by `allfe...@yandex.ru` on 21 Nov 2012 at 6:47
|
defect
|
time how about support original issue reported on code google com by allfe yandex ru on nov at
| 1
|
64,294
| 18,397,282,207
|
IssuesEvent
|
2021-10-12 12:50:02
|
vector-im/element-android
|
https://api.github.com/repos/vector-im/element-android
|
closed
|
Emails are not listed in the screen notification by email
|
T-Defect feature:settings
|
At least an email is associated to the account
- Clear cache from Settings/General clear cache
- Go to settings/notifications
- Observe that the section "Email notification" displays the "no email" placeholder
- Click on this placeholder
- The emails are loaded
- Go back
- The emails are display and user can enable Email notification
I think we have to force a refresh of the 3Pid as per https://github.com/vector-im/element-android/blob/develop/vector/src/main/java/im/vector/app/features/settings/threepids/ThreePidsSettingsViewModel.kt#L106 and this is not done in https://github.com/vector-im/element-android/blob/develop/vector/src/main/java/im/vector/app/features/settings/notifications/VectorSettingsNotificationPreferenceFragment.kt#L424
- [ ] So I think this is just about changing the parameter in https://github.com/vector-im/element-android/blob/develop/vector/src/main/java/im/vector/app/features/settings/notifications/VectorSettingsNotificationPreferenceFragment.kt#L424
- [ ] Also the pushers has to be refreshed, if configured on another session of the same user, the displayed information are not correct when entering the screen (user has to go to the debug screen about pushers to get the latest pushers, and then the information will be correct.
|
1.0
|
Emails are not listed in the screen notification by email - At least an email is associated to the account
- Clear cache from Settings/General clear cache
- Go to settings/notifications
- Observe that the section "Email notification" displays the "no email" placeholder
- Click on this placeholder
- The emails are loaded
- Go back
- The emails are display and user can enable Email notification
I think we have to force a refresh of the 3Pid as per https://github.com/vector-im/element-android/blob/develop/vector/src/main/java/im/vector/app/features/settings/threepids/ThreePidsSettingsViewModel.kt#L106 and this is not done in https://github.com/vector-im/element-android/blob/develop/vector/src/main/java/im/vector/app/features/settings/notifications/VectorSettingsNotificationPreferenceFragment.kt#L424
- [ ] So I think this is just about changing the parameter in https://github.com/vector-im/element-android/blob/develop/vector/src/main/java/im/vector/app/features/settings/notifications/VectorSettingsNotificationPreferenceFragment.kt#L424
- [ ] Also the pushers has to be refreshed, if configured on another session of the same user, the displayed information are not correct when entering the screen (user has to go to the debug screen about pushers to get the latest pushers, and then the information will be correct.
|
defect
|
emails are not listed in the screen notification by email at least an email is associated to the account clear cache from settings general clear cache go to settings notifications observe that the section email notification displays the no email placeholder click on this placeholder the emails are loaded go back the emails are display and user can enable email notification i think we have to force a refresh of the as per and this is not done in so i think this is just about changing the parameter in also the pushers has to be refreshed if configured on another session of the same user the displayed information are not correct when entering the screen user has to go to the debug screen about pushers to get the latest pushers and then the information will be correct
| 1
|
223,087
| 17,103,014,843
|
IssuesEvent
|
2021-07-09 13:55:02
|
OpenSlides/OpenSlides
|
https://api.github.com/repos/OpenSlides/OpenSlides
|
opened
|
Provide info about used environment variables
|
OpenSlides4 documentation
|
Each service should provide information about all used/recognized environment variables and its defaults so we do not have to guess during deployment.
Proposal: We use a uniform format for this definitions and every service should additionally mention it in its README.
E. g. every service should have a `default.env` file with content like this:
```
# Environment variables used in this service
# This variable is used for something more or less important.
FOO_BAR=42
# This variable can be used in a nice way ...
BAZ_BAR=foo
...
```
|
1.0
|
Provide info about used environment variables - Each service should provide information about all used/recognized environment variables and its defaults so we do not have to guess during deployment.
Proposal: We use a uniform format for this definitions and every service should additionally mention it in its README.
E. g. every service should have a `default.env` file with content like this:
```
# Environment variables used in this service
# This variable is used for something more or less important.
FOO_BAR=42
# This variable can be used in a nice way ...
BAZ_BAR=foo
...
```
|
non_defect
|
provide info about used environment variables each service should provide information about all used recognized environment variables and its defaults so we do not have to guess during deployment proposal we use a uniform format for this definitions and every service should additionally mention it in its readme e g every service should have a default env file with content like this environment variables used in this service this variable is used for something more or less important foo bar this variable can be used in a nice way baz bar foo
| 0
|
38,652
| 8,951,492,190
|
IssuesEvent
|
2019-01-25 14:07:13
|
hazelcast/hazelcast-cpp-client
|
https://api.github.com/repos/hazelcast/hazelcast-cpp-client
|
closed
|
MixedMapAPITestInstance/MixedMapAPITest.testGetAllPutAll/0 sometimes fail
|
Estimation: M Priority: High Type: Defect
|
During nightly test(https://hazelcast-l337.ci.cloudbees.com/job/cpp-linux-nightly-64-SHARED-Debug/737/#), we observe test failure:
```
[0;31m[ FAILED ] [mMixedMapAPITestInstance/MixedMapAPITest.testGetAllPutAll/0, where GetParam() = 0x28ac7f0
[0;31m[ FAILED ] [mMixedMapAPITestInstance/MixedMapAPITest.testGetAllPutAll/1, where GetParam() = 0x28ac480
```
The error is:
```C++ exception with description "ExceptionMessage {Connection closed.} at InvocationService::cleanResources" thrown in the test body.```
|
1.0
|
MixedMapAPITestInstance/MixedMapAPITest.testGetAllPutAll/0 sometimes fail - During nightly test(https://hazelcast-l337.ci.cloudbees.com/job/cpp-linux-nightly-64-SHARED-Debug/737/#), we observe test failure:
```
[0;31m[ FAILED ] [mMixedMapAPITestInstance/MixedMapAPITest.testGetAllPutAll/0, where GetParam() = 0x28ac7f0
[0;31m[ FAILED ] [mMixedMapAPITestInstance/MixedMapAPITest.testGetAllPutAll/1, where GetParam() = 0x28ac480
```
The error is:
```C++ exception with description "ExceptionMessage {Connection closed.} at InvocationService::cleanResources" thrown in the test body.```
|
defect
|
mixedmapapitestinstance mixedmapapitest testgetallputall sometimes fail during nightly test we observe test failure mmixedmapapitestinstance mixedmapapitest testgetallputall where getparam mmixedmapapitestinstance mixedmapapitest testgetallputall where getparam the error is c exception with description exceptionmessage connection closed at invocationservice cleanresources thrown in the test body
| 1
|
22,352
| 3,640,405,441
|
IssuesEvent
|
2016-02-13 00:01:48
|
netty/netty
|
https://api.github.com/repos/netty/netty
|
closed
|
HTTP/2 codec may not always call Http2Connection$Listener.onStreamRemoved
|
defect
|
If `Http2Connection$Listener.onStreamAdded` is called it is not always the case that `Http2Connection$Listener.onStreamRemoved` will be called. Some use cases (including `InboundHttp2ToHttpAdapter`) depend upon the `onStreamRemoved` method to be called in order to clean up allocated buffers.
[Http2ConnectionHandler.channelInactive](https://github.com/netty/netty/blob/4.1/codec-http2/src/main/java/io/netty/handler/codec/http2/Http2ConnectionHandler.java#L180) currently only iterates over active streams, and this does not account for streams which may still exist but are not active.
|
1.0
|
HTTP/2 codec may not always call Http2Connection$Listener.onStreamRemoved - If `Http2Connection$Listener.onStreamAdded` is called it is not always the case that `Http2Connection$Listener.onStreamRemoved` will be called. Some use cases (including `InboundHttp2ToHttpAdapter`) depend upon the `onStreamRemoved` method to be called in order to clean up allocated buffers.
[Http2ConnectionHandler.channelInactive](https://github.com/netty/netty/blob/4.1/codec-http2/src/main/java/io/netty/handler/codec/http2/Http2ConnectionHandler.java#L180) currently only iterates over active streams, and this does not account for streams which may still exist but are not active.
|
defect
|
http codec may not always call listener onstreamremoved if listener onstreamadded is called it is not always the case that listener onstreamremoved will be called some use cases including depend upon the onstreamremoved method to be called in order to clean up allocated buffers currently only iterates over active streams and this does not account for streams which may still exist but are not active
| 1
|
102,664
| 4,157,996,019
|
IssuesEvent
|
2016-06-16 23:43:26
|
expressjs/serve-index
|
https://api.github.com/repos/expressjs/serve-index
|
closed
|
nsp security error on 1.7.3
|
deps low priority
|
Looking at [master](https://github.com/expressjs/serve-index/blob/master/package.json#L9), it looks like all that needs to be done is cut a new version, but we're seeing this error in https://github.com/redfin/react-server/issues/291
```
dougwade packages/react-server-cli โนreact-server-cli-nsp*โบ ยป nsp check
(+) 1 vulnerabilities found
โโโโโโโโโโโโโโโโโฌโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ โ Regular Expression Denial of Service โ
โโโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค
โ Name โ negotiator โ
โโโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค
โ Installed โ 0.5.3 โ
โโโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค
โ Vulnerable โ <= 0.6.0 โ
โโโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค
โ Patched โ >= 0.6.1 โ
โโโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค
โ Path โ react-server-cli@0.3.2 > webpack-dev-server@1.14.1 > serve-index@1.7.3 > accepts@1.2.13 > negotiator@0.5.3 โ
โโโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค
โ More Info โ https://nodesecurity.io/advisories/106 โ
โโโโโโโโโโโโโโโโโดโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
```
|
1.0
|
nsp security error on 1.7.3 - Looking at [master](https://github.com/expressjs/serve-index/blob/master/package.json#L9), it looks like all that needs to be done is cut a new version, but we're seeing this error in https://github.com/redfin/react-server/issues/291
```
dougwade packages/react-server-cli โนreact-server-cli-nsp*โบ ยป nsp check
(+) 1 vulnerabilities found
โโโโโโโโโโโโโโโโโฌโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ โ Regular Expression Denial of Service โ
โโโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค
โ Name โ negotiator โ
โโโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค
โ Installed โ 0.5.3 โ
โโโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค
โ Vulnerable โ <= 0.6.0 โ
โโโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค
โ Patched โ >= 0.6.1 โ
โโโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค
โ Path โ react-server-cli@0.3.2 > webpack-dev-server@1.14.1 > serve-index@1.7.3 > accepts@1.2.13 > negotiator@0.5.3 โ
โโโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค
โ More Info โ https://nodesecurity.io/advisories/106 โ
โโโโโโโโโโโโโโโโโดโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
```
|
non_defect
|
nsp security error on looking at it looks like all that needs to be done is cut a new version but we re seeing this error in dougwade packages react server cli โนreact server cli nsp โบ ยป nsp check vulnerabilities found โโโโโโโโโโโโโโโโโฌโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ โ โ regular expression denial of service โ โโโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค โ name โ negotiator โ โโโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค โ installed โ โ โโโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค โ vulnerable โ โ โโโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค โ patched โ โ โโโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค โ path โ react server cli webpack dev server serve index accepts negotiator โ โโโโโโโโโโโโโโโโโผโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโค โ more info โ โ โโโโโโโโโโโโโโโโโดโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
| 0
|
51,818
| 13,211,319,155
|
IssuesEvent
|
2020-08-15 22:16:52
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
example scripts for credo fail (Trac #1118)
|
Incomplete Migration Migrated from Trac combo reconstruction defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1118">https://code.icecube.wisc.edu/projects/icecube/ticket/1118</a>, reported by kjmeagherand owned by jtatar</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"_ts": "1550067117911749",
"description": "credol3.py and nugen2107l2b.py require particleforge\ntest1.py and test2.py use old I3Units import\n\nno meta-project level documentation either",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"time": "2015-08-17T13:34:45",
"component": "combo reconstruction",
"summary": "example scripts for credo fail",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "jtatar",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
example scripts for credo fail (Trac #1118) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1118">https://code.icecube.wisc.edu/projects/icecube/ticket/1118</a>, reported by kjmeagherand owned by jtatar</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"_ts": "1550067117911749",
"description": "credol3.py and nugen2107l2b.py require particleforge\ntest1.py and test2.py use old I3Units import\n\nno meta-project level documentation either",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"time": "2015-08-17T13:34:45",
"component": "combo reconstruction",
"summary": "example scripts for credo fail",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "jtatar",
"type": "defect"
}
```
</p>
</details>
|
defect
|
example scripts for credo fail trac migrated from json status closed changetime ts description py and py require particleforge py and py use old import n nno meta project level documentation either reporter kjmeagher cc resolution fixed time component combo reconstruction summary example scripts for credo fail priority normal keywords milestone owner jtatar type defect
| 1
|
255,089
| 27,484,737,617
|
IssuesEvent
|
2023-03-04 01:13:46
|
panasalap/linux-4.1.15
|
https://api.github.com/repos/panasalap/linux-4.1.15
|
opened
|
CVE-2016-4581 (Medium) detected in linux-yocto-devv4.2.8
|
security vulnerability
|
## CVE-2016-4581 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-devv4.2.8</b></p></summary>
<p>
<p>Linux Embedded Kernel - tracks the next mainline release</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-dev>https://git.yoctoproject.org/git/linux-yocto-dev</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/pnode.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/pnode.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
fs/pnode.c in the Linux kernel before 4.5.4 does not properly traverse a mount propagation tree in a certain case involving a slave mount, which allows local users to cause a denial of service (NULL pointer dereference and OOPS) via a crafted series of mount system calls.
<p>Publish Date: 2016-05-23
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-4581>CVE-2016-4581</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-4581">https://nvd.nist.gov/vuln/detail/CVE-2016-4581</a></p>
<p>Release Date: 2016-05-23</p>
<p>Fix Resolution: 4.5.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2016-4581 (Medium) detected in linux-yocto-devv4.2.8 - ## CVE-2016-4581 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-devv4.2.8</b></p></summary>
<p>
<p>Linux Embedded Kernel - tracks the next mainline release</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-dev>https://git.yoctoproject.org/git/linux-yocto-dev</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/pnode.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/pnode.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
fs/pnode.c in the Linux kernel before 4.5.4 does not properly traverse a mount propagation tree in a certain case involving a slave mount, which allows local users to cause a denial of service (NULL pointer dereference and OOPS) via a crafted series of mount system calls.
<p>Publish Date: 2016-05-23
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-4581>CVE-2016-4581</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-4581">https://nvd.nist.gov/vuln/detail/CVE-2016-4581</a></p>
<p>Release Date: 2016-05-23</p>
<p>Fix Resolution: 4.5.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in linux yocto cve medium severity vulnerability vulnerable library linux yocto linux embedded kernel tracks the next mainline release library home page a href found in base branch master vulnerable source files fs pnode c fs pnode c vulnerability details fs pnode c in the linux kernel before does not properly traverse a mount propagation tree in a certain case involving a slave mount which allows local users to cause a denial of service null pointer dereference and oops via a crafted series of mount system calls publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
126,064
| 26,775,795,044
|
IssuesEvent
|
2023-01-31 17:03:09
|
Clueless-Community/seamless-ui
|
https://api.github.com/repos/Clueless-Community/seamless-ui
|
opened
|
Make Accordion
|
codepeak 22
|
Need to improve this component on `\Accordion\src\accordion-3.html`
## Improvements required :
Make an accordion which opens on click and only one can remain open at a time, rest should be closed

|
1.0
|
Make Accordion - Need to improve this component on `\Accordion\src\accordion-3.html`
## Improvements required :
Make an accordion which opens on click and only one can remain open at a time, rest should be closed

|
non_defect
|
make accordion need to improve this component on accordion src accordion html improvements required make an accordion which opens on click and only one can remain open at a time rest should be closed
| 0
|
65,760
| 3,240,202,109
|
IssuesEvent
|
2015-10-15 01:29:08
|
BYU-ARCLITE/Ayamel-Examples
|
https://api.github.com/repos/BYU-ARCLITE/Ayamel-Examples
|
closed
|
Annotation Editor Fails to Register Change
|
Annotations Bug duplicate Priority 2
|
When making changes to annotations, they are saved as shown below, but the editor fails to register that the "save" took place:

After seeing that screen and clicking on the arrow to return to the content, the system pops up this message:

|
1.0
|
Annotation Editor Fails to Register Change - When making changes to annotations, they are saved as shown below, but the editor fails to register that the "save" took place:

After seeing that screen and clicking on the arrow to return to the content, the system pops up this message:

|
non_defect
|
annotation editor fails to register change when making changes to annotations they are saved as shown below but the editor fails to register that the save took place after seeing that screen and clicking on the arrow to return to the content the system pops up this message
| 0
|
762
| 2,587,972,319
|
IssuesEvent
|
2015-02-17 21:49:28
|
chrsmith/codesearch
|
https://api.github.com/repos/chrsmith/codesearch
|
opened
|
csearch: print statistics of -f fregexp hits if -verbose [PATCH]
|
auto-migrated Priority-Medium Type-Defect
|
```
csearch already prints the trigrams and number of files matched from the index
if -verbose mode is on.
This patch preprocesses the list of fileids so we can print how many were
matched by the filename regex.
It does double the number of calls to ix.Name(fileid), but that call looks
cheap and any slowdown would be dwarfed by the disk access during the grep
stage anyway.
```
-----
Original issue reported on code.google.com by `dgryski` on 17 Feb 2012 at 1:50
Attachments:
* [filename-regex-stats.patch](https://storage.googleapis.com/google-code-attachments/codesearch/issue-14/comment-0/filename-regex-stats.patch)
|
1.0
|
csearch: print statistics of -f fregexp hits if -verbose [PATCH] - ```
csearch already prints the trigrams and number of files matched from the index
if -verbose mode is on.
This patch preprocesses the list of fileids so we can print how many were
matched by the filename regex.
It does double the number of calls to ix.Name(fileid), but that call looks
cheap and any slowdown would be dwarfed by the disk access during the grep
stage anyway.
```
-----
Original issue reported on code.google.com by `dgryski` on 17 Feb 2012 at 1:50
Attachments:
* [filename-regex-stats.patch](https://storage.googleapis.com/google-code-attachments/codesearch/issue-14/comment-0/filename-regex-stats.patch)
|
defect
|
csearch print statistics of f fregexp hits if verbose csearch already prints the trigrams and number of files matched from the index if verbose mode is on this patch preprocesses the list of fileids so we can print how many were matched by the filename regex it does double the number of calls to ix name fileid but that call looks cheap and any slowdown would be dwarfed by the disk access during the grep stage anyway original issue reported on code google com by dgryski on feb at attachments
| 1
|
55,519
| 14,530,821,787
|
IssuesEvent
|
2020-12-14 19:50:26
|
mozilla-extensions/normandy-devtools
|
https://api.github.com/repos/mozilla-extensions/normandy-devtools
|
closed
|
The dropdowns from the Normandy Devtools are not fixed in their position when scrolling through the page
|
[QA]:Normal issue defect
|
[Affected versions]:
- Firefox Release 83.0 (Build ID: 20201112153044)
[Affected Platforms]:
- Windows 10 x64
[Prerequisites]:
- Have the Firefox browser installed.
- Have the [Normandy Devtools V2.2.0](https://github.com/mozilla-extensions/normandy-devtools/releases) addon installed.
- Have VPN access to the staging environment.
[Steps to reproduce]:
1. Open the Normandy Devtools and choose the Stage environment.
2. Click on the โCreate Recipeโ button.
3. Click on any dropdown field from the page.
4. Scroll the page and observe the behavior.
[Expected result]:
- The dropdown remains fixed on its area.
[Actual result]:
- The dropdown is floating on the page while scrolling.
[Notes]:
- This issue is reproducible with the โActionโ, โSampling Typeโ, โCountriesโ, and โLocalesโ dropdowns.
- Attached a screen recording of the issue.

|
1.0
|
The dropdowns from the Normandy Devtools are not fixed in their position when scrolling through the page - [Affected versions]:
- Firefox Release 83.0 (Build ID: 20201112153044)
[Affected Platforms]:
- Windows 10 x64
[Prerequisites]:
- Have the Firefox browser installed.
- Have the [Normandy Devtools V2.2.0](https://github.com/mozilla-extensions/normandy-devtools/releases) addon installed.
- Have VPN access to the staging environment.
[Steps to reproduce]:
1. Open the Normandy Devtools and choose the Stage environment.
2. Click on the โCreate Recipeโ button.
3. Click on any dropdown field from the page.
4. Scroll the page and observe the behavior.
[Expected result]:
- The dropdown remains fixed on its area.
[Actual result]:
- The dropdown is floating on the page while scrolling.
[Notes]:
- This issue is reproducible with the โActionโ, โSampling Typeโ, โCountriesโ, and โLocalesโ dropdowns.
- Attached a screen recording of the issue.

|
defect
|
the dropdowns from the normandy devtools are not fixed in their position when scrolling through the page firefox release build id windows have the firefox browser installed have the addon installed have vpn access to the staging environment open the normandy devtools and choose the stage environment click on the โcreate recipeโ button click on any dropdown field from the page scroll the page and observe the behavior the dropdown remains fixed on its area the dropdown is floating on the page while scrolling this issue is reproducible with the โactionโ โsampling typeโ โcountriesโ and โlocalesโ dropdowns attached a screen recording of the issue
| 1
|
58,864
| 16,836,181,954
|
IssuesEvent
|
2021-06-18 12:24:10
|
BOINC/boinc
|
https://api.github.com/repos/BOINC/boinc
|
closed
|
Right clicking the BOINC Manager logo on the titlebar on Windows causes it to instantly crash
|
C: Manager R: duplicate T: Defect
|
**Describe the bug**
A clear and concise description of what the bug is.
The BOINC Manager will instantly crash if the titlebar logo is right clicked.
**Steps To Reproduce**
1. Open the BOINC Manager
2. Right click the BOINC Manager logo in the titlebar
**Expected behavior**
A clear and concise description of what you expected to happen.
Standard titlebar context menu items
**Screenshots**
If applicable, add screenshots to help explain your problem.
**System Information**
- OS: Windows 10
- BOINC Version: 7.16.11
**Additional context**
Add any other context about the problem here.
|
1.0
|
Right clicking the BOINC Manager logo on the titlebar on Windows causes it to instantly crash - **Describe the bug**
A clear and concise description of what the bug is.
The BOINC Manager will instantly crash if the titlebar logo is right clicked.
**Steps To Reproduce**
1. Open the BOINC Manager
2. Right click the BOINC Manager logo in the titlebar
**Expected behavior**
A clear and concise description of what you expected to happen.
Standard titlebar context menu items
**Screenshots**
If applicable, add screenshots to help explain your problem.
**System Information**
- OS: Windows 10
- BOINC Version: 7.16.11
**Additional context**
Add any other context about the problem here.
|
defect
|
right clicking the boinc manager logo on the titlebar on windows causes it to instantly crash describe the bug a clear and concise description of what the bug is the boinc manager will instantly crash if the titlebar logo is right clicked steps to reproduce open the boinc manager right click the boinc manager logo in the titlebar expected behavior a clear and concise description of what you expected to happen standard titlebar context menu items screenshots if applicable add screenshots to help explain your problem system information os windows boinc version additional context add any other context about the problem here
| 1
|
25,190
| 2,677,853,450
|
IssuesEvent
|
2015-03-26 04:43:01
|
JukkaL/mypy
|
https://api.github.com/repos/JukkaL/mypy
|
closed
|
Runtime access to type annotations
|
feature pep484 priority
|
The current implementation erases type information from __annotations__ and friends at runtime:
In [2]: def f(x: List[int]) -> int:
...: pass
In [3]: f.__annotations__
Out[3]: {'return': int, 'x': list} ## lost x: List[int]
Is this by design, and if so, why? Is there some reason to not keep an accurate run-time representation of type annotations? Could that information be made available uniformly on every typeable thing?
In [4]: class A:
...: i = Undefined(int)
In [5]: A.__annotations__ ## lost i: int
Perhaps even include a run-time representation of some inferred types e.g. all types of all instance attributes might be made available via class.__annotations__ (or something like it), regardless of how those types were declared (on __init__ or via class attributes), or even inferred (by mypy from body of __init__ or otherwise).
Thoughts?
|
1.0
|
Runtime access to type annotations - The current implementation erases type information from __annotations__ and friends at runtime:
In [2]: def f(x: List[int]) -> int:
...: pass
In [3]: f.__annotations__
Out[3]: {'return': int, 'x': list} ## lost x: List[int]
Is this by design, and if so, why? Is there some reason to not keep an accurate run-time representation of type annotations? Could that information be made available uniformly on every typeable thing?
In [4]: class A:
...: i = Undefined(int)
In [5]: A.__annotations__ ## lost i: int
Perhaps even include a run-time representation of some inferred types e.g. all types of all instance attributes might be made available via class.__annotations__ (or something like it), regardless of how those types were declared (on __init__ or via class attributes), or even inferred (by mypy from body of __init__ or otherwise).
Thoughts?
|
non_defect
|
runtime access to type annotations the current implementation erases type information from annotations and friends at runtime in def f x list int pass in f annotations out return int x list lost x list is this by design and if so why is there some reason to not keep an accurate run time representation of type annotations could that information be made available uniformly on every typeable thing in class a i undefined int in a annotations lost i int perhaps even include a run time representation of some inferred types e g all types of all instance attributes might be made available via class annotations or something like it regardless of how those types were declared on init or via class attributes or even inferred by mypy from body of init or otherwise thoughts
| 0
|
142,719
| 21,875,074,689
|
IssuesEvent
|
2022-05-19 09:24:20
|
flutter/flutter
|
https://api.github.com/repos/flutter/flutter
|
closed
|
Time picker with incorrect time separator format for locale `id`
|
framework f: material design f: date/time picker a: internationalization
|
Should be `.`, not `:`.
b/229057815
b/231987861
|
1.0
|
Time picker with incorrect time separator format for locale `id` - Should be `.`, not `:`.
b/229057815
b/231987861
|
non_defect
|
time picker with incorrect time separator format for locale id should be not b b
| 0
|
67,459
| 20,961,613,639
|
IssuesEvent
|
2022-03-27 21:49:42
|
abedmaatalla/sipdroid
|
https://api.github.com/repos/abedmaatalla/sipdroid
|
closed
|
C2DM
|
Priority-Medium Type-Defect auto-migrated
|
```
Feature request:
Have you considered adding C2DM (cloud-to-device-messaging) to siproid and
pbxes.org to replace the TCP standby connection that is currently optimal?
```
Original issue reported on code.google.com by `ezelspin...@gmail.com` on 17 Dec 2011 at 9:00
|
1.0
|
C2DM - ```
Feature request:
Have you considered adding C2DM (cloud-to-device-messaging) to siproid and
pbxes.org to replace the TCP standby connection that is currently optimal?
```
Original issue reported on code.google.com by `ezelspin...@gmail.com` on 17 Dec 2011 at 9:00
|
defect
|
feature request have you considered adding cloud to device messaging to siproid and pbxes org to replace the tcp standby connection that is currently optimal original issue reported on code google com by ezelspin gmail com on dec at
| 1
|
64,480
| 18,684,729,310
|
IssuesEvent
|
2021-11-01 10:56:18
|
obophenotype/cell-ontology
|
https://api.github.com/repos/obophenotype/cell-ontology
|
closed
|
Project: Review all cases where cell types have multiple develops_from relatiionships
|
Priority-Medium Type-Defect auto-migrated autoclosed-unfixed
|
```
Assuming develops_from indicates lineage then the following is problematic
cell1 subclassOf develops_from some cell2
cell1 subClassOf develops_from some cell3
e.g.
id: CL:0000140
name: odontocyte
relationship: develops_from CL:0000060 ! odontoblast
relationship: develops_from CL:0000134 ! mesenchymal stem cell
unless cell2 subClassOf cell3 OR cell3 SubClassOf cell2 (in which case the
pattern reflects redundancy and one axioms should be deleted) or cell2 and
cell3 overlap and all cell1 develop from some cell in that is in a class
defined by the intersection of cell2 and cell3.
I suspect that this is rarely the case where this pattern appears in CL. More
likely these are attempts at recording some/some relationships.
All such cases should therefore be reviewed.
```
Original issue reported on code.google.com by `dosu...@gmail.com` on 16 Jul 2014 at 12:13
|
1.0
|
Project: Review all cases where cell types have multiple develops_from relatiionships - ```
Assuming develops_from indicates lineage then the following is problematic
cell1 subclassOf develops_from some cell2
cell1 subClassOf develops_from some cell3
e.g.
id: CL:0000140
name: odontocyte
relationship: develops_from CL:0000060 ! odontoblast
relationship: develops_from CL:0000134 ! mesenchymal stem cell
unless cell2 subClassOf cell3 OR cell3 SubClassOf cell2 (in which case the
pattern reflects redundancy and one axioms should be deleted) or cell2 and
cell3 overlap and all cell1 develop from some cell in that is in a class
defined by the intersection of cell2 and cell3.
I suspect that this is rarely the case where this pattern appears in CL. More
likely these are attempts at recording some/some relationships.
All such cases should therefore be reviewed.
```
Original issue reported on code.google.com by `dosu...@gmail.com` on 16 Jul 2014 at 12:13
|
defect
|
project review all cases where cell types have multiple develops from relatiionships assuming develops from indicates lineage then the following is problematic subclassof develops from some subclassof develops from some e g id cl name odontocyte relationship develops from cl odontoblast relationship develops from cl mesenchymal stem cell unless subclassof or subclassof in which case the pattern reflects redundancy and one axioms should be deleted or and overlap and all develop from some cell in that is in a class defined by the intersection of and i suspect that this is rarely the case where this pattern appears in cl more likely these are attempts at recording some some relationships all such cases should therefore be reviewed original issue reported on code google com by dosu gmail com on jul at
| 1
|
11,392
| 2,649,862,610
|
IssuesEvent
|
2015-03-15 11:09:50
|
dakkanner/ra-duty-scheduler
|
https://api.github.com/repos/dakkanner/ra-duty-scheduler
|
closed
|
Weekend EOT additional assigned days
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
Create schedule where the end of term lands on a Friday or saturday.
What is the expected output? What do you see instead?
It schedules whoever should be on for the whole weekend, making them have an inflated count of days.
```
Original issue reported on code.google.com by `dakkan...@gmail.com` on 21 Nov 2014 at 9:20
|
1.0
|
Weekend EOT additional assigned days - ```
What steps will reproduce the problem?
Create schedule where the end of term lands on a Friday or saturday.
What is the expected output? What do you see instead?
It schedules whoever should be on for the whole weekend, making them have an inflated count of days.
```
Original issue reported on code.google.com by `dakkan...@gmail.com` on 21 Nov 2014 at 9:20
|
defect
|
weekend eot additional assigned days what steps will reproduce the problem create schedule where the end of term lands on a friday or saturday what is the expected output what do you see instead it schedules whoever should be on for the whole weekend making them have an inflated count of days original issue reported on code google com by dakkan gmail com on nov at
| 1
|
603,507
| 18,668,490,905
|
IssuesEvent
|
2021-10-30 08:35:48
|
renovatebot/renovate
|
https://api.github.com/repos/renovatebot/renovate
|
closed
|
Renovate GitHub-native automerge debug message when `merge` method is not allowed
|
type:bug priority-4-low platform:github status:ready
|
### How are you running Renovate?
WhiteSource Renovate hosted app on github.com
### Please select which platform you are using if self-hosting.
_No response_
### If you're self-hosting Renovate, tell us what version of Renovate you run.
28.8.3
### Describe the bug
## Intro
I'm trying out the new automerge on my `honkinggoose/gitgosling` repo, with Renovate version `28.8.3`.
## Steps to reproduce:
1. The repository had no pending PRs.
1. I requested a lockfile update via the Dashboard.
1. I wondered what the behavior is now, so I checked the logs from this run (pasted below in log section of bug report).
1. I was expecting Renovate to automerge the lockfile right away, as it's the only PR and it's fully up to date with `main` right when its created.
1. Instead I got a debug message: `"message": "Merge method merge commits are not allowed on this repository"`
## Repo settings
On my repository I only allow `squash` merges:

I also have a branch protection rule stating that any PR branch must be up-to-date with `main` before it can be merged.
## Further context
Link to PR that had the debug logs: https://github.com/HonkingGoose/git-gosling/pull/313
Note that Renovate did manage to automerge the PR in the end, but I think it used the Renovate fallback to do that.
If GitHub-native automerge _requires_ that the normal merge is available we should document that, because we don't mention this right now.
@rarkins thought:
> Seems we need a way to detect merge type
### Relevant debug logs
<details><summary>Relevant logs for initial run</summary>
```
DEBUG: getBranchStatus(renovate/lock-file-maintenance)(branch="renovate/lock-file-maintenance")
DEBUG: branch status check result(branch="renovate/lock-file-maintenance")
{
"state": "pending",
"statuses": []
}
DEBUG: No check runs found(branch="renovate/lock-file-maintenance")
{
"result": {
"total_count": 0,
"check_runs": []
}
}
DEBUG: Branch status yellow(branch="renovate/lock-file-maintenance")
DEBUG: Creating PR(branch="renovate/lock-file-maintenance")
{
"prTitle": "chore(deps): lock file maintenance"
}
DEBUG: Creating PR(branch="renovate/lock-file-maintenance")
{
"title": "chore(deps): lock file maintenance",
"head": "HonkingGoose:renovate/lock-file-maintenance",
"base": "main",
"draft": false
}
DEBUG: PR created(branch="renovate/lock-file-maintenance")
{
"pr": 313,
"draft": false
}
DEBUG: Adding labels '' to #313(branch="renovate/lock-file-maintenance")
DEBUG: GitHub automerge unknown error(branch="renovate/lock-file-maintenance")
{
"prNumber": 313,
"errors": [
{
"type": "UNPROCESSABLE",
"path": [
"enablePullRequestAutoMerge"
],
"locations": [
{
"line": 5,
"column": 3
}
],
"message": "Merge method merge commits are not allowed on this repository"
}
]
}
INFO: PR created(branch="renovate/lock-file-maintenance")
{
"pr": 313,
"prTitle": "chore(deps): lock file maintenance"
}
DEBUG: Skipping assignees and reviewers as automerge=true(branch="renovate/lock-file-maintenance")
DEBUG: Created Pull Request #313(branch="renovate/lock-file-maintenance")
DEBUG: PR is configured for automerge(branch="renovate/lock-file-maintenance")
DEBUG: PR is not ready for merge(branch="renovate/lock-file-maintenance")
DEBUG: getBranchPr(renovate/lock-file-maintenance)
DEBUG: findPr(renovate/lock-file-maintenance, undefined, open)
DEBUG: Found PR #313
DEBUG: PR not found in open or closed PRs list - trying to fetch it directly
{
"prNo": 313
}
DEBUG: Ensuring Dependency Dashboard
DEBUG: ensureIssue(Dependency Dashboard)
DEBUG: Patching issue
DEBUG: Issue updated
DEBUG: Removing any stale branches
DEBUG: config.repoIsOnboarded=true
DEBUG: Branch lists
{
"branchList": [
"renovate/lock-file-maintenance",
"renovate/npm-8.x"
],
"renovateBranches": [
"renovate/lock-file-maintenance"
]
}
DEBUG: remainingBranches=
DEBUG: No branches to clean up
DEBUG: Repository timing splits (milliseconds)
{
"splits": {
"init": 2411,
"extract": 2967,
"lookup": 2479,
"update": 130426
},
"total": 139426
}
```
</details>
<details><summary>Relevant logs for run that automerged the lockfile PR</summary>
```
DEBUG: Branch status green(branch="renovate/lock-file-maintenance")
DEBUG: Processing existing PR(branch="renovate/lock-file-maintenance")
DEBUG: Pull Request #313 does not need updating(branch="renovate/lock-file-maintenance")
DEBUG: PR is configured for automerge(branch="renovate/lock-file-maintenance")
DEBUG: resolveBranchStatus(branchName=renovate/lock-file-maintenance, ignoreTests=false)(branch="renovate/lock-file-maintenance")
DEBUG: getBranchStatus(renovate/lock-file-maintenance)(branch="renovate/lock-file-maintenance")
DEBUG: branch status check result(branch="renovate/lock-file-maintenance")
{
"state": "success",
"statuses": [
{
"url": "https://api.github.com/repos/HonkingGoose/git-gosling/statuses/2b9cedd52417d823105802e402932d23f354d9fb",
"avatar_url": "https://avatars.githubusercontent.com/in/8329?v=4",
"id": 14944469753,
"node_id": "SC_kwDOEbfyB88AAAADesKC-Q",
"state": "success",
"description": "Deployment has completed",
"target_url": "https://vercel.com/honkinggoose/git-gosling/EBFgt4sMCXh7utZs9nu4zRzTQLYM",
"context": "Vercel",
"created_at": "2021-10-23T07:14:57Z",
"updated_at": "2021-10-23T07:14:57Z"
}
]
}
DEBUG: check runs result(branch="renovate/lock-file-maintenance")
{
"checkRuns": [
{
"name": "build (14.x)",
"status": "completed",
"conclusion": "success"
},
{
"name": "lint (14.x)",
"status": "completed",
"conclusion": "success"
}
]
}
DEBUG: Branch status green(branch="renovate/lock-file-maintenance")
DEBUG: Automerging #313 with strategy auto(branch="renovate/lock-file-maintenance")
DEBUG: mergePr(313, renovate/lock-file-maintenance)(branch="renovate/lock-file-maintenance")
DEBUG: mergePr(branch="renovate/lock-file-maintenance")
{
"options": {
"body": {
"merge_method": "squash"
}
},
"url": "repos/HonkingGoose/git-gosling/pulls/313/merge"
}
DEBUG: PR merged(branch="renovate/lock-file-maintenance")
{
"automergeResult": {
"sha": "ec34406e5fa1722a1a1efa8a9add78593a5f9bc6",
"merged": true,
"message": "Pull Request successfully merged"
},
"pr": 313
}
INFO: PR automerged(branch="renovate/lock-file-maintenance")
{
"pr": 313,
"prTitle": "chore(deps): lock file maintenance"
}
DEBUG: No remote branch to delete(branch="renovate/lock-file-maintenance")
{
"branchName": "renovate/lock-file-maintenance"
}
DEBUG: No local branch to delete(branch="renovate/lock-file-maintenance")
{
"branchName": "renovate/lock-file-maintenance"
}
INFO: Renovating repository again after automerge result
```
</details>
### Have you created a minimal reproduction repository?
No reproduction, but I have linked to a public repo where it occurs
|
1.0
|
Renovate GitHub-native automerge debug message when `merge` method is not allowed - ### How are you running Renovate?
WhiteSource Renovate hosted app on github.com
### Please select which platform you are using if self-hosting.
_No response_
### If you're self-hosting Renovate, tell us what version of Renovate you run.
28.8.3
### Describe the bug
## Intro
I'm trying out the new automerge on my `honkinggoose/gitgosling` repo, with Renovate version `28.8.3`.
## Steps to reproduce:
1. The repository had no pending PRs.
1. I requested a lockfile update via the Dashboard.
1. I wondered what the behavior is now, so I checked the logs from this run (pasted below in log section of bug report).
1. I was expecting Renovate to automerge the lockfile right away, as it's the only PR and it's fully up to date with `main` right when its created.
1. Instead I got a debug message: `"message": "Merge method merge commits are not allowed on this repository"`
## Repo settings
On my repository I only allow `squash` merges:

I also have a branch protection rule stating that any PR branch must be up-to-date with `main` before it can be merged.
## Further context
Link to PR that had the debug logs: https://github.com/HonkingGoose/git-gosling/pull/313
Note that Renovate did manage to automerge the PR in the end, but I think it used the Renovate fallback to do that.
If GitHub-native automerge _requires_ that the normal merge is available we should document that, because we don't mention this right now.
@rarkins thought:
> Seems we need a way to detect merge type
### Relevant debug logs
<details><summary>Relevant logs for initial run</summary>
```
DEBUG: getBranchStatus(renovate/lock-file-maintenance)(branch="renovate/lock-file-maintenance")
DEBUG: branch status check result(branch="renovate/lock-file-maintenance")
{
"state": "pending",
"statuses": []
}
DEBUG: No check runs found(branch="renovate/lock-file-maintenance")
{
"result": {
"total_count": 0,
"check_runs": []
}
}
DEBUG: Branch status yellow(branch="renovate/lock-file-maintenance")
DEBUG: Creating PR(branch="renovate/lock-file-maintenance")
{
"prTitle": "chore(deps): lock file maintenance"
}
DEBUG: Creating PR(branch="renovate/lock-file-maintenance")
{
"title": "chore(deps): lock file maintenance",
"head": "HonkingGoose:renovate/lock-file-maintenance",
"base": "main",
"draft": false
}
DEBUG: PR created(branch="renovate/lock-file-maintenance")
{
"pr": 313,
"draft": false
}
DEBUG: Adding labels '' to #313(branch="renovate/lock-file-maintenance")
DEBUG: GitHub automerge unknown error(branch="renovate/lock-file-maintenance")
{
"prNumber": 313,
"errors": [
{
"type": "UNPROCESSABLE",
"path": [
"enablePullRequestAutoMerge"
],
"locations": [
{
"line": 5,
"column": 3
}
],
"message": "Merge method merge commits are not allowed on this repository"
}
]
}
INFO: PR created(branch="renovate/lock-file-maintenance")
{
"pr": 313,
"prTitle": "chore(deps): lock file maintenance"
}
DEBUG: Skipping assignees and reviewers as automerge=true(branch="renovate/lock-file-maintenance")
DEBUG: Created Pull Request #313(branch="renovate/lock-file-maintenance")
DEBUG: PR is configured for automerge(branch="renovate/lock-file-maintenance")
DEBUG: PR is not ready for merge(branch="renovate/lock-file-maintenance")
DEBUG: getBranchPr(renovate/lock-file-maintenance)
DEBUG: findPr(renovate/lock-file-maintenance, undefined, open)
DEBUG: Found PR #313
DEBUG: PR not found in open or closed PRs list - trying to fetch it directly
{
"prNo": 313
}
DEBUG: Ensuring Dependency Dashboard
DEBUG: ensureIssue(Dependency Dashboard)
DEBUG: Patching issue
DEBUG: Issue updated
DEBUG: Removing any stale branches
DEBUG: config.repoIsOnboarded=true
DEBUG: Branch lists
{
"branchList": [
"renovate/lock-file-maintenance",
"renovate/npm-8.x"
],
"renovateBranches": [
"renovate/lock-file-maintenance"
]
}
DEBUG: remainingBranches=
DEBUG: No branches to clean up
DEBUG: Repository timing splits (milliseconds)
{
"splits": {
"init": 2411,
"extract": 2967,
"lookup": 2479,
"update": 130426
},
"total": 139426
}
```
</details>
<details><summary>Relevant logs for run that automerged the lockfile PR</summary>
```
DEBUG: Branch status green(branch="renovate/lock-file-maintenance")
DEBUG: Processing existing PR(branch="renovate/lock-file-maintenance")
DEBUG: Pull Request #313 does not need updating(branch="renovate/lock-file-maintenance")
DEBUG: PR is configured for automerge(branch="renovate/lock-file-maintenance")
DEBUG: resolveBranchStatus(branchName=renovate/lock-file-maintenance, ignoreTests=false)(branch="renovate/lock-file-maintenance")
DEBUG: getBranchStatus(renovate/lock-file-maintenance)(branch="renovate/lock-file-maintenance")
DEBUG: branch status check result(branch="renovate/lock-file-maintenance")
{
"state": "success",
"statuses": [
{
"url": "https://api.github.com/repos/HonkingGoose/git-gosling/statuses/2b9cedd52417d823105802e402932d23f354d9fb",
"avatar_url": "https://avatars.githubusercontent.com/in/8329?v=4",
"id": 14944469753,
"node_id": "SC_kwDOEbfyB88AAAADesKC-Q",
"state": "success",
"description": "Deployment has completed",
"target_url": "https://vercel.com/honkinggoose/git-gosling/EBFgt4sMCXh7utZs9nu4zRzTQLYM",
"context": "Vercel",
"created_at": "2021-10-23T07:14:57Z",
"updated_at": "2021-10-23T07:14:57Z"
}
]
}
DEBUG: check runs result(branch="renovate/lock-file-maintenance")
{
"checkRuns": [
{
"name": "build (14.x)",
"status": "completed",
"conclusion": "success"
},
{
"name": "lint (14.x)",
"status": "completed",
"conclusion": "success"
}
]
}
DEBUG: Branch status green(branch="renovate/lock-file-maintenance")
DEBUG: Automerging #313 with strategy auto(branch="renovate/lock-file-maintenance")
DEBUG: mergePr(313, renovate/lock-file-maintenance)(branch="renovate/lock-file-maintenance")
DEBUG: mergePr(branch="renovate/lock-file-maintenance")
{
"options": {
"body": {
"merge_method": "squash"
}
},
"url": "repos/HonkingGoose/git-gosling/pulls/313/merge"
}
DEBUG: PR merged(branch="renovate/lock-file-maintenance")
{
"automergeResult": {
"sha": "ec34406e5fa1722a1a1efa8a9add78593a5f9bc6",
"merged": true,
"message": "Pull Request successfully merged"
},
"pr": 313
}
INFO: PR automerged(branch="renovate/lock-file-maintenance")
{
"pr": 313,
"prTitle": "chore(deps): lock file maintenance"
}
DEBUG: No remote branch to delete(branch="renovate/lock-file-maintenance")
{
"branchName": "renovate/lock-file-maintenance"
}
DEBUG: No local branch to delete(branch="renovate/lock-file-maintenance")
{
"branchName": "renovate/lock-file-maintenance"
}
INFO: Renovating repository again after automerge result
```
</details>
### Have you created a minimal reproduction repository?
No reproduction, but I have linked to a public repo where it occurs
|
non_defect
|
renovate github native automerge debug message when merge method is not allowed how are you running renovate whitesource renovate hosted app on github com please select which platform you are using if self hosting no response if you re self hosting renovate tell us what version of renovate you run describe the bug intro i m trying out the new automerge on my honkinggoose gitgosling repo with renovate version steps to reproduce the repository had no pending prs i requested a lockfile update via the dashboard i wondered what the behavior is now so i checked the logs from this run pasted below in log section of bug report i was expecting renovate to automerge the lockfile right away as it s the only pr and it s fully up to date with main right when its created instead i got a debug message message merge method merge commits are not allowed on this repository repo settings on my repository i only allow squash merges i also have a branch protection rule stating that any pr branch must be up to date with main before it can be merged further context link to pr that had the debug logs note that renovate did manage to automerge the pr in the end but i think it used the renovate fallback to do that if github native automerge requires that the normal merge is available we should document that because we don t mention this right now rarkins thought seems we need a way to detect merge type relevant debug logs relevant logs for initial run debug getbranchstatus renovate lock file maintenance branch renovate lock file maintenance debug branch status check result branch renovate lock file maintenance state pending statuses debug no check runs found branch renovate lock file maintenance result total count check runs debug branch status yellow branch renovate lock file maintenance debug creating pr branch renovate lock file maintenance prtitle chore deps lock file maintenance debug creating pr branch renovate lock file maintenance title chore deps lock file maintenance head honkinggoose renovate lock file maintenance base main draft false debug pr created branch renovate lock file maintenance pr draft false debug adding labels to branch renovate lock file maintenance debug github automerge unknown error branch renovate lock file maintenance prnumber errors type unprocessable path enablepullrequestautomerge locations line column message merge method merge commits are not allowed on this repository info pr created branch renovate lock file maintenance pr prtitle chore deps lock file maintenance debug skipping assignees and reviewers as automerge true branch renovate lock file maintenance debug created pull request branch renovate lock file maintenance debug pr is configured for automerge branch renovate lock file maintenance debug pr is not ready for merge branch renovate lock file maintenance debug getbranchpr renovate lock file maintenance debug findpr renovate lock file maintenance undefined open debug found pr debug pr not found in open or closed prs list trying to fetch it directly prno debug ensuring dependency dashboard debug ensureissue dependency dashboard debug patching issue debug issue updated debug removing any stale branches debug config repoisonboarded true debug branch lists branchlist renovate lock file maintenance renovate npm x renovatebranches renovate lock file maintenance debug remainingbranches debug no branches to clean up debug repository timing splits milliseconds splits init extract lookup update total relevant logs for run that automerged the lockfile pr debug branch status green branch renovate lock file maintenance debug processing existing pr branch renovate lock file maintenance debug pull request does not need updating branch renovate lock file maintenance debug pr is configured for automerge branch renovate lock file maintenance debug resolvebranchstatus branchname renovate lock file maintenance ignoretests false branch renovate lock file maintenance debug getbranchstatus renovate lock file maintenance branch renovate lock file maintenance debug branch status check result branch renovate lock file maintenance state success statuses url avatar url id node id sc q state success description deployment has completed target url context vercel created at updated at debug check runs result branch renovate lock file maintenance checkruns name build x status completed conclusion success name lint x status completed conclusion success debug branch status green branch renovate lock file maintenance debug automerging with strategy auto branch renovate lock file maintenance debug mergepr renovate lock file maintenance branch renovate lock file maintenance debug mergepr branch renovate lock file maintenance options body merge method squash url repos honkinggoose git gosling pulls merge debug pr merged branch renovate lock file maintenance automergeresult sha merged true message pull request successfully merged pr info pr automerged branch renovate lock file maintenance pr prtitle chore deps lock file maintenance debug no remote branch to delete branch renovate lock file maintenance branchname renovate lock file maintenance debug no local branch to delete branch renovate lock file maintenance branchname renovate lock file maintenance info renovating repository again after automerge result have you created a minimal reproduction repository no reproduction but i have linked to a public repo where it occurs
| 0
|
344,437
| 30,746,789,385
|
IssuesEvent
|
2023-07-28 15:37:49
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
Failing test: Jest Integration Tests.src/core/server/integration_tests/saved_objects/migrations/group3 - SO type registrations does not remove types from registrations without updating excludeOnUpgradeQuery
|
Team:Core failed-test
|
A test failed on a tracked branch
```
Error: Missing version for public endpoint GET /api/observability_onboarding/custom_logs/step/{name}
at parseEndpoint (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/kbn-server-route-repository/src/parse_endpoint.ts:23:11)
at /var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/x-pack/plugins/observability_onboarding/server/routes/register_routes.ts:39:47
at Array.forEach (<anonymous>)
at forEach (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/x-pack/plugins/observability_onboarding/server/routes/register_routes.ts:37:10)
at ObservabilityOnboardingPlugin.setup (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/x-pack/plugins/observability_onboarding/server/plugin.ts:66:19)
at PluginWrapper.setup (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/core/plugins/core-plugins-server-internal/src/plugin.ts:105:26)
at PluginsSystem.setup [as setupPlugins] (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/core/plugins/core-plugins-server-internal/src/plugins_system.ts:131:40)
at PluginsService.setupPlugins [as setup] (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/core/plugins/core-plugins-server-internal/src/plugins_service.ts:166:52)
at Server.setup (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/core/root/core-root-server-internal/src/server.ts:348:26)
at Root.setup (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/core/root/core-root-server-internal/src/root/index.ts:66:14)
at Object.<anonymous> (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/src/core/server/integration_tests/saved_objects/migrations/group3/type_registrations.test.ts:151:19)
```
First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/30745#01884a02-35b0-4943-a777-bd69fc053f9a)
<!-- kibanaCiData = {"failed-test":{"test.class":"Jest Integration Tests.src/core/server/integration_tests/saved_objects/migrations/group3","test.name":"SO type registrations does not remove types from registrations without updating excludeOnUpgradeQuery","test.failCount":3}} -->
|
1.0
|
Failing test: Jest Integration Tests.src/core/server/integration_tests/saved_objects/migrations/group3 - SO type registrations does not remove types from registrations without updating excludeOnUpgradeQuery - A test failed on a tracked branch
```
Error: Missing version for public endpoint GET /api/observability_onboarding/custom_logs/step/{name}
at parseEndpoint (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/kbn-server-route-repository/src/parse_endpoint.ts:23:11)
at /var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/x-pack/plugins/observability_onboarding/server/routes/register_routes.ts:39:47
at Array.forEach (<anonymous>)
at forEach (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/x-pack/plugins/observability_onboarding/server/routes/register_routes.ts:37:10)
at ObservabilityOnboardingPlugin.setup (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/x-pack/plugins/observability_onboarding/server/plugin.ts:66:19)
at PluginWrapper.setup (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/core/plugins/core-plugins-server-internal/src/plugin.ts:105:26)
at PluginsSystem.setup [as setupPlugins] (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/core/plugins/core-plugins-server-internal/src/plugins_system.ts:131:40)
at PluginsService.setupPlugins [as setup] (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/core/plugins/core-plugins-server-internal/src/plugins_service.ts:166:52)
at Server.setup (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/core/root/core-root-server-internal/src/server.ts:348:26)
at Root.setup (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/packages/core/root/core-root-server-internal/src/root/index.ts:66:14)
at Object.<anonymous> (/var/lib/buildkite-agent/builds/kb-n2-4-spot-05ef58b5d43908c8/elastic/kibana-on-merge/kibana/src/core/server/integration_tests/saved_objects/migrations/group3/type_registrations.test.ts:151:19)
```
First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/30745#01884a02-35b0-4943-a777-bd69fc053f9a)
<!-- kibanaCiData = {"failed-test":{"test.class":"Jest Integration Tests.src/core/server/integration_tests/saved_objects/migrations/group3","test.name":"SO type registrations does not remove types from registrations without updating excludeOnUpgradeQuery","test.failCount":3}} -->
|
non_defect
|
failing test jest integration tests src core server integration tests saved objects migrations so type registrations does not remove types from registrations without updating excludeonupgradequery a test failed on a tracked branch error missing version for public endpoint get api observability onboarding custom logs step name at parseendpoint var lib buildkite agent builds kb spot elastic kibana on merge kibana packages kbn server route repository src parse endpoint ts at var lib buildkite agent builds kb spot elastic kibana on merge kibana x pack plugins observability onboarding server routes register routes ts at array foreach at foreach var lib buildkite agent builds kb spot elastic kibana on merge kibana x pack plugins observability onboarding server routes register routes ts at observabilityonboardingplugin setup var lib buildkite agent builds kb spot elastic kibana on merge kibana x pack plugins observability onboarding server plugin ts at pluginwrapper setup var lib buildkite agent builds kb spot elastic kibana on merge kibana packages core plugins core plugins server internal src plugin ts at pluginssystem setup var lib buildkite agent builds kb spot elastic kibana on merge kibana packages core plugins core plugins server internal src plugins system ts at pluginsservice setupplugins var lib buildkite agent builds kb spot elastic kibana on merge kibana packages core plugins core plugins server internal src plugins service ts at server setup var lib buildkite agent builds kb spot elastic kibana on merge kibana packages core root core root server internal src server ts at root setup var lib buildkite agent builds kb spot elastic kibana on merge kibana packages core root core root server internal src root index ts at object var lib buildkite agent builds kb spot elastic kibana on merge kibana src core server integration tests saved objects migrations type registrations test ts first failure
| 0
|
721
| 2,587,731,740
|
IssuesEvent
|
2015-02-17 20:18:55
|
chrsmith/codesearch
|
https://api.github.com/repos/chrsmith/codesearch
|
opened
|
cindex fails on the first run: mmap() returns EINVAL
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. build all utils (or download binaries from this site)
2. be sure that there is no .csearchindex (or it has zero size)
3. run `cindex /usr/include` or any other path
`cindex` fails, because mmap returns error with EINVAL code.
This happens inside mmap_linux.go, on line 24.
That function (mmapFile) calls mmap with zero size, which fails.
I had manually copy ~/.csearchindex\~ to ~/.csearchindex.
That fixed the issue.
I expected cindex to work right out of the box, even with no .csearchindex.
I'm using:
Gentoo Base System release 2.0.3, Linux 3.1.6, amd64.
Go repository is the newest (built a few minutes ago from head).
```
-----
Original issue reported on code.google.com by zaitsev....@gmail.com on 19 Jan 2012 at 8:10
|
1.0
|
cindex fails on the first run: mmap() returns EINVAL - ```
What steps will reproduce the problem?
1. build all utils (or download binaries from this site)
2. be sure that there is no .csearchindex (or it has zero size)
3. run `cindex /usr/include` or any other path
`cindex` fails, because mmap returns error with EINVAL code.
This happens inside mmap_linux.go, on line 24.
That function (mmapFile) calls mmap with zero size, which fails.
I had manually copy ~/.csearchindex\~ to ~/.csearchindex.
That fixed the issue.
I expected cindex to work right out of the box, even with no .csearchindex.
I'm using:
Gentoo Base System release 2.0.3, Linux 3.1.6, amd64.
Go repository is the newest (built a few minutes ago from head).
```
-----
Original issue reported on code.google.com by zaitsev....@gmail.com on 19 Jan 2012 at 8:10
|
defect
|
cindex fails on the first run mmap returns einval what steps will reproduce the problem build all utils or download binaries from this site be sure that there is no csearchindex or it has zero size run cindex usr include or any other path cindex fails because mmap returns error with einval code this happens inside mmap linux go on line that function mmapfile calls mmap with zero size which fails i had manually copy csearchindex to csearchindex that fixed the issue i expected cindex to work right out of the box even with no csearchindex i m using gentoo base system release linux go repository is the newest built a few minutes ago from head original issue reported on code google com by zaitsev gmail com on jan at
| 1
|
75,119
| 9,819,146,358
|
IssuesEvent
|
2019-06-13 21:08:15
|
bradyrx/climpred
|
https://api.github.com/repos/bradyrx/climpred
|
closed
|
Consolidate bootstrap API in sphinx
|
documentation
|
As a start, I just dropped all public functions from `bootstrap` into the API. I imagine only a few of these are supposed to be called by the user. @aaronspring, if you could slim it down to only what is necessary, that would be great. Unless all of them should be on there.
|
1.0
|
Consolidate bootstrap API in sphinx - As a start, I just dropped all public functions from `bootstrap` into the API. I imagine only a few of these are supposed to be called by the user. @aaronspring, if you could slim it down to only what is necessary, that would be great. Unless all of them should be on there.
|
non_defect
|
consolidate bootstrap api in sphinx as a start i just dropped all public functions from bootstrap into the api i imagine only a few of these are supposed to be called by the user aaronspring if you could slim it down to only what is necessary that would be great unless all of them should be on there
| 0
|
36,654
| 8,049,079,421
|
IssuesEvent
|
2018-08-01 08:59:58
|
zotonic/zotonic
|
https://api.github.com/repos/zotonic/zotonic
|
opened
|
Concurrent category hierarchy save problem
|
core defect
|
It is possible that multiple concurrent category hierarchy saves lead to sql constraint problems. Normally this does not happen when you enable new modules manually, but when you install new sites via scripts it can happen that multiple modules with new categories in their datamodel save a new hierarchy concurrently.
The problem can be solved by making sure `hierarchy:save/3` can't run concurrently by either running it from a gen_server or as a job.
```erlang
5:59:39.716 [error] channelwww z_db:300 z_db error {error,error,<<"23505">>,<<"duplicate key value violates unique constraint \"hierarchy_pkey\"">>,[{detail,<<"Key (name, id)=($category, 345) already exists.">>}]} in query "\n insert into hierarchy\n (name, id, parent_id, lvl, nr, lft, rght)\n values\n ($1, $2, $3, $4, $5, $6, $7)" with ['$category',345,104,2,4998046,4998046,4998046]
15:59:39.717 [error] gen_server <0.1364.0> terminated with reason: no match of right hand value {rollback,{{error,{error,error,<<"23505">>,<<"duplicate key value violates unique constraint \"hierarchy_pkey\"">>,[{detail,<<"Key (name, id)=($category, 345) already exists.">>}]}},[{z_db,'-q/4-fun-2-',5,[{file,"src/db/z_db.erl"},{line,301}]},{timer,tc,2,[{file,"timer.erl"},{line,181}]},{z_db,with_connection,3,[{file,"src/db/z_db.erl"},{line,211}]},{lists,foreach,2,[{file,"lists.erl"},{line,1338}]},{m_hierarchy,save_nocheck_trans,3,[{file,"src/models/m_hierarchy.erl"},{line,295}]},{z_db,'-transaction1/2-fun-0-',...},...]}} in m_hierarchy:save_nocheck/3 line 261
```
|
1.0
|
Concurrent category hierarchy save problem - It is possible that multiple concurrent category hierarchy saves lead to sql constraint problems. Normally this does not happen when you enable new modules manually, but when you install new sites via scripts it can happen that multiple modules with new categories in their datamodel save a new hierarchy concurrently.
The problem can be solved by making sure `hierarchy:save/3` can't run concurrently by either running it from a gen_server or as a job.
```erlang
5:59:39.716 [error] channelwww z_db:300 z_db error {error,error,<<"23505">>,<<"duplicate key value violates unique constraint \"hierarchy_pkey\"">>,[{detail,<<"Key (name, id)=($category, 345) already exists.">>}]} in query "\n insert into hierarchy\n (name, id, parent_id, lvl, nr, lft, rght)\n values\n ($1, $2, $3, $4, $5, $6, $7)" with ['$category',345,104,2,4998046,4998046,4998046]
15:59:39.717 [error] gen_server <0.1364.0> terminated with reason: no match of right hand value {rollback,{{error,{error,error,<<"23505">>,<<"duplicate key value violates unique constraint \"hierarchy_pkey\"">>,[{detail,<<"Key (name, id)=($category, 345) already exists.">>}]}},[{z_db,'-q/4-fun-2-',5,[{file,"src/db/z_db.erl"},{line,301}]},{timer,tc,2,[{file,"timer.erl"},{line,181}]},{z_db,with_connection,3,[{file,"src/db/z_db.erl"},{line,211}]},{lists,foreach,2,[{file,"lists.erl"},{line,1338}]},{m_hierarchy,save_nocheck_trans,3,[{file,"src/models/m_hierarchy.erl"},{line,295}]},{z_db,'-transaction1/2-fun-0-',...},...]}} in m_hierarchy:save_nocheck/3 line 261
```
|
defect
|
concurrent category hierarchy save problem it is possible that multiple concurrent category hierarchy saves lead to sql constraint problems normally this does not happen when you enable new modules manually but when you install new sites via scripts it can happen that multiple modules with new categories in their datamodel save a new hierarchy concurrently the problem can be solved by making sure hierarchy save can t run concurrently by either running it from a gen server or as a job erlang channelwww z db z db error error error in query n insert into hierarchy n name id parent id lvl nr lft rght n values n with gen server terminated with reason no match of right hand value rollback error error error timer tc z db with connection lists foreach m hierarchy save nocheck trans z db fun in m hierarchy save nocheck line
| 1
|
67,230
| 27,754,555,272
|
IssuesEvent
|
2023-03-16 00:38:44
|
devssa/onde-codar-em-salvador
|
https://api.github.com/repos/devssa/onde-codar-em-salvador
|
closed
|
[OPORTUNIDADE] [REMOTO] Pessoa Desenvolvedora Back-end na Enjoei
|
BACK-END JAVASCRIPT JQUERY GIT POSTGRESQL RUBY AWS REMOTO BACKEND GITHUB RUBY ON RAILS MICROSERVICES ELASTICSEARCH AWS CLOUD HELP WANTED GOLANG Stale
|
### Sobre a empresa e desafio
Criada em 2009 para ser um blog para conectar pessoas que queriam se desfazer de peรงas de roupas que nรฃo usavam mais. Em meses, a Enjoei se transformou em um social commerce que oferece um espaรงo exclusivo para cada cliente criar a sua โlojinhaโ. Hoje รฉ um marketplace, um app, um site, uma plataforma de tecnologia com foco em moda. Investida por Monashees+, Bessemer Venture Partners, Dynamo e Grupo Globo e que em 2020 levantou R$ 1,13 bilhรฃo em IPO. Histรณria incrรญvel, nรฃo รฉ?
Pois bem, queremos conectar vocรช pessoa desenvolvedora ao prรณximo passo na sua carreira. Estamos com uma oportunidade, para fazer parte do time Enjoei e construir uma histรณria de sucesso, como Back-end Developer. Buscamos um perfil que tenha interesse genuรญno em produzir software de qualidade, que se preocupe com o produto, goste de desafios diรกrios e de solucionar questรตes de performance e de arquitetura para orientar e compartilhar conhecimento com pessoas desenvolvedoras menos experientes.
### Requerimentos
Para este desafio, รฉ necessรกrio ter experiรชncia em desenvolvimento de preferรชncia em Ruby โ Ruby on Rails, Javascript (Jquery) ou Golang, Git e Github, PostgresSQL, desenvolvimento em microservices, Elasticsearch e AWS
### Modelo de contraรงรฃo CLT
### Link
Vamos nesse desafio juntos? Saiba mais -> https://byintera.in/21y
|
1.0
|
[OPORTUNIDADE] [REMOTO] Pessoa Desenvolvedora Back-end na Enjoei - ### Sobre a empresa e desafio
Criada em 2009 para ser um blog para conectar pessoas que queriam se desfazer de peรงas de roupas que nรฃo usavam mais. Em meses, a Enjoei se transformou em um social commerce que oferece um espaรงo exclusivo para cada cliente criar a sua โlojinhaโ. Hoje รฉ um marketplace, um app, um site, uma plataforma de tecnologia com foco em moda. Investida por Monashees+, Bessemer Venture Partners, Dynamo e Grupo Globo e que em 2020 levantou R$ 1,13 bilhรฃo em IPO. Histรณria incrรญvel, nรฃo รฉ?
Pois bem, queremos conectar vocรช pessoa desenvolvedora ao prรณximo passo na sua carreira. Estamos com uma oportunidade, para fazer parte do time Enjoei e construir uma histรณria de sucesso, como Back-end Developer. Buscamos um perfil que tenha interesse genuรญno em produzir software de qualidade, que se preocupe com o produto, goste de desafios diรกrios e de solucionar questรตes de performance e de arquitetura para orientar e compartilhar conhecimento com pessoas desenvolvedoras menos experientes.
### Requerimentos
Para este desafio, รฉ necessรกrio ter experiรชncia em desenvolvimento de preferรชncia em Ruby โ Ruby on Rails, Javascript (Jquery) ou Golang, Git e Github, PostgresSQL, desenvolvimento em microservices, Elasticsearch e AWS
### Modelo de contraรงรฃo CLT
### Link
Vamos nesse desafio juntos? Saiba mais -> https://byintera.in/21y
|
non_defect
|
pessoa desenvolvedora back end na enjoei sobre a empresa e desafio criada em para ser um blog para conectar pessoas que queriam se desfazer de peรงas de roupas que nรฃo usavam mais em meses a enjoei se transformou em um social commerce que oferece um espaรงo exclusivo para cada cliente criar a sua โlojinhaโ hoje รฉ um marketplace um app um site uma plataforma de tecnologia com foco em moda investida por monashees bessemer venture partners dynamo e grupo globo e que em levantou r bilhรฃo em ipo histรณria incrรญvel nรฃo รฉ pois bem queremos conectar vocรช pessoa desenvolvedora ao prรณximo passo na sua carreira estamos com uma oportunidade para fazer parte do time enjoei e construir uma histรณria de sucesso como back end developer buscamos um perfil que tenha interesse genuรญno em produzir software de qualidade que se preocupe com o produto goste de desafios diรกrios e de solucionar questรตes de performance e de arquitetura para orientar e compartilhar conhecimento com pessoas desenvolvedoras menos experientes requerimentos para este desafio รฉ necessรกrio ter experiรชncia em desenvolvimento de preferรชncia em ruby โ ruby on rails javascript jquery ou golang git e github postgressql desenvolvimento em microservices elasticsearch e aws modelo de contraรงรฃo clt link vamos nesse desafio juntos saiba mais
| 0
|
28,384
| 4,389,865,150
|
IssuesEvent
|
2016-08-09 00:00:35
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
stress: failed test in cockroach/storage/storage.test: TestReplicateAddAndRemove
|
Robot test-failure
|
Binary: cockroach/static-tests.tar.gz sha: https://github.com/cockroachdb/cockroach/commits/c1114f920f13baf1df10cb555fd7516088249ae8
Stress build found a failed test:
```
=== RUN TestReplicateAddAndRemove
I160726 05:24:33.209159 storage/engine/rocksdb.go:353 opening in memory rocksdb instance
W160726 05:24:33.210050 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node
W160726 05:24:33.211342 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node
I160726 05:24:33.211509 storage/store.go:2671 system config not yet available
I160726 05:24:33.212090 storage/replica_command.go:1637 store=1:1 ([]=) range=1 [/Min-/Max): new range lease replica {1 1 1} 1970-01-01 00:00:00 +0000 UTC 900ms following replica {0 0 0} 1970-01-01 00:00:00 +0000 UTC 0 [physicalTime=1970-01-01 00:00:00 +0000 UTC]
I160726 05:24:33.212416 storage/engine/rocksdb.go:353 opening in memory rocksdb instance
W160726 05:24:33.213239 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node
I160726 05:24:33.213720 storage/store.go:2671 system config not yet available
I160726 05:24:33.213840 gossip/client.go:71 starting client to 127.0.0.1:38919
I160726 05:24:33.213844 storage/engine/rocksdb.go:353 opening in memory rocksdb instance
W160726 05:24:33.214625 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node
I160726 05:24:33.215180 storage/store.go:2671 system config not yet available
I160726 05:24:33.215304 gossip/client.go:71 starting client to 127.0.0.1:38919
I160726 05:24:33.215322 storage/engine/rocksdb.go:353 opening in memory rocksdb instance
W160726 05:24:33.216338 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node
I160726 05:24:33.217256 storage/store.go:2671 system config not yet available
I160726 05:24:33.217397 gossip/client.go:71 starting client to 127.0.0.1:38919
I160726 05:24:33.218819 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 12 in 243.521ยตs. encoded size=2580, 26 KV pairs, 2 log entries
I160726 05:24:33.219148 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > next_replica_id:2
I160726 05:24:33.219734 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.220086 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:4 StoreID:4 ReplicaID:2} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2}]
I160726 05:24:33.220386 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.220413 storage/replica_raftstorage.go:587 store=4:4 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 12 (encoded size=2580, 26 KV pairs, 2 log entries)
I160726 05:24:33.220963 storage/replica_raftstorage.go:590 store=4:4 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 492.712ยตs
I160726 05:24:33.221150 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 14 in 161.195ยตs. encoded size=3431, 29 KV pairs, 4 log entries
I160726 05:24:33.221452 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > next_replica_id:3
I160726 05:24:33.222290 storage/replica_raftstorage.go:587 store=2:2 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 14 (encoded size=3431, 29 KV pairs, 4 log entries)
I160726 05:24:33.222730 storage/replica_raftstorage.go:590 store=2:2 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 401.576ยตs
I160726 05:24:33.223154 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.223391 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.223721 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
I160726 05:24:33.224071 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.224404 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:2 StoreID:2 ReplicaID:3} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2} {NodeID:2 StoreID:2 ReplicaID:3}]
I160726 05:24:33.224628 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.225384 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.225797 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.226119 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
I160726 05:24:33.227423 storage/replica.go:1945 store=2:2 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.227676 storage/replica.go:1945 store=2:2 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.228066 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
I160726 05:24:33.229166 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 17 in 200.114ยตs. encoded size=4442, 33 KV pairs, 7 log entries
I160726 05:24:33.229475 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > replicas:<node_id:2 store_id:2 replica_id:3 > next_replica_id:4
I160726 05:24:33.230504 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.230881 storage/replica_raftstorage.go:587 store=3:3 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 17 (encoded size=4442, 33 KV pairs, 7 log entries)
I160726 05:24:33.230906 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:3 StoreID:3 ReplicaID:4} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2} {NodeID:2 StoreID:2 ReplicaID:3} {NodeID:3 StoreID:3 ReplicaID:4}]
W160726 05:24:33.231050 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1}
I160726 05:24:33.231186 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.231622 storage/replica_raftstorage.go:590 store=3:3 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 694.304ยตs
I160726 05:24:33.231941 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
W160726 05:24:33.232215 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1}
I160726 05:24:33.232549 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.232733 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > replicas:<node_id:2 store_id:2 replica_id:3 > replicas:<node_id:3 store_id:3 replica_id:4 > next_replica_id:5
I160726 05:24:33.232931 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
W160726 05:24:33.233887 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1}
I160726 05:24:33.235119 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.235346 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.235687 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
I160726 05:24:33.235916 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.236198 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing REMOVE_REPLICA {NodeID:2 StoreID:2 ReplicaID:3} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2} {NodeID:3 StoreID:3 ReplicaID:4}]
I160726 05:24:33.237944 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
W160726 05:24:33.237957 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1}
I160726 05:24:33.238396 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.239552 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.240543 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
W160726 05:24:33.240997 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1}
I160726 05:24:33.241153 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
I160726 05:24:33.241324 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.241659 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
W160726 05:24:33.266682 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: <nil>
W160726 05:24:33.266805 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 0} closed by the remote: <nil>
W160726 05:24:33.266829 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 0} closed by the remote: <nil>
W160726 05:24:33.266852 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: <nil>
W160726 05:24:33.266874 storage/replica.go:337 range 1: outgoing raft transport stream to {3 3 0} closed by the remote: <nil>
W160726 05:24:33.266907 storage/replica.go:337 range 1: outgoing raft transport stream to {3 3 4} closed by the remote: <nil>
W160726 05:24:33.266921 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: <nil>
W160726 05:24:33.266941 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:1 StoreID:1 ReplicaID:1}: no store registered for {NodeID:3 StoreID:3 ReplicaID:4}
I160726 05:24:33.267618 storage/engine/rocksdb.go:353 opening in memory rocksdb instance
W160726 05:24:33.268403 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node
I160726 05:24:33.269801 storage/store.go:2671 system config not yet available
W160726 05:24:33.269940 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node
I160726 05:24:33.270410 storage/replica_command.go:1637 store=1:1 ([]=) range=1 [/Min-/Max): new range lease replica {1 1 1} 1970-01-01 00:00:00 +0000 UTC 900ms following replica {0 0 0} 1970-01-01 00:00:00 +0000 UTC 0 [physicalTime=1970-01-01 00:00:00 +0000 UTC]
I160726 05:24:33.270697 storage/engine/rocksdb.go:353 opening in memory rocksdb instance
W160726 05:24:33.271184 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node
I160726 05:24:33.271629 storage/store.go:2671 system config not yet available
I160726 05:24:33.271813 gossip/client.go:71 starting client to 127.0.0.1:37700
I160726 05:24:33.271900 storage/engine/rocksdb.go:353 opening in memory rocksdb instance
W160726 05:24:33.272505 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node
I160726 05:24:33.273221 storage/store.go:2671 system config not yet available
I160726 05:24:33.273356 gossip/client.go:71 starting client to 127.0.0.1:37700
I160726 05:24:33.273382 storage/engine/rocksdb.go:353 opening in memory rocksdb instance
W160726 05:24:33.274134 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node
I160726 05:24:33.274729 storage/store.go:2671 system config not yet available
I160726 05:24:33.274752 gossip/client.go:71 starting client to 127.0.0.1:37700
I160726 05:24:33.276138 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 12 in 142.306ยตs. encoded size=2580, 26 KV pairs, 2 log entries
I160726 05:24:33.276432 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > next_replica_id:2
I160726 05:24:33.277000 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.277368 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:4 StoreID:4 ReplicaID:2} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2}]
I160726 05:24:33.277783 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.277899 storage/replica_raftstorage.go:587 store=4:4 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 12 (encoded size=2580, 26 KV pairs, 2 log entries)
I160726 05:24:33.278272 storage/replica_raftstorage.go:590 store=4:4 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 332.578ยตs
I160726 05:24:33.278606 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 14 in 131.608ยตs. encoded size=3431, 29 KV pairs, 4 log entries
I160726 05:24:33.278815 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > next_replica_id:3
I160726 05:24:33.279760 storage/replica_raftstorage.go:587 store=2:2 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 14 (encoded size=3431, 29 KV pairs, 4 log entries)
I160726 05:24:33.280204 storage/replica_raftstorage.go:590 store=2:2 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 399.401ยตs
I160726 05:24:33.280443 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.280728 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.280936 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.281222 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
I160726 05:24:33.281321 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:2 StoreID:2 ReplicaID:3} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2} {NodeID:2 StoreID:2 ReplicaID:3}]
I160726 05:24:33.281495 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.282152 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.282741 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.283186 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
I160726 05:24:33.284773 storage/replica.go:1945 store=2:2 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.285021 storage/replica.go:1945 store=2:2 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.285465 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
I160726 05:24:33.289996 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > replicas:<node_id:2 store_id:2 replica_id:3 > next_replica_id:4
W160726 05:24:33.291631 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1}
I160726 05:24:33.291894 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.292323 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.292666 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing REMOVE_REPLICA {NodeID:2 StoreID:2 ReplicaID:3} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2}]
W160726 05:24:33.293532 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1}
I160726 05:24:33.294206 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
W160726 05:24:33.294991 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1}
I160726 05:24:33.295056 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 19 in 184.365ยตs. encoded size=5306, 36 KV pairs, 9 log entries
I160726 05:24:33.295408 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > next_replica_id:4
I160726 05:24:33.296003 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.296205 storage/replica_raftstorage.go:587 store=3:3 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 19 (encoded size=5306, 36 KV pairs, 9 log entries)
I160726 05:24:33.296535 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
I160726 05:24:33.296757 storage/replica_raftstorage.go:590 store=3:3 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 506.141ยตs
I160726 05:24:33.297024 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.297498 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:3 StoreID:3 ReplicaID:4} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2} {NodeID:3 StoreID:3 ReplicaID:4}]
I160726 05:24:33.297547 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.298421 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.299146 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.299590 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
I160726 05:24:33.300598 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.300760 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.301169 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
W160726 05:24:33.704343 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:33.704394 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:34.204542 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:34.204557 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
I160726 05:24:34.270157 gossip/client.go:71 starting client to 127.0.0.1:37298
I160726 05:24:34.270721 gossip/client.go:95 closing client to node 3 (127.0.0.1:37298): stopping outgoing client to node 3 (127.0.0.1:37298); already have incoming
W160726 05:24:34.504339 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:34.504431 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:34.804438 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:34.804451 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:35.304476 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:35.304622 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:35.604507 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:35.604522 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:35.904329 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:35.904397 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:36.204411 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:36.204476 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:36.504400 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:36.504468 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:36.804397 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:36.804411 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:37.304381 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:37.304395 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:37.604380 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:37.604402 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:38.104389 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:38.104405 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:38.504301 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:38.504315 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:38.804308 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:38.804363 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:39.304551 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:39.304564 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:39.604313 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:39.604324 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:40.104258 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:40.104336 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:40.604440 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:40.604453 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:41.004364 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:41.005420 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:41.304399 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:41.304471 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:41.704448 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:41.704472 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:42.204429 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:42.204482 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:42.604358 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:42.604371 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:43.004396 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:43.004418 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:43.504343 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:43.504373 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:43.804438 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:43.804511 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:44.304409 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:44.304424 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:44.604345 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:44.604361 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:45.004281 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:45.004314 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:45.504305 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:45.504352 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:45.904459 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:45.904486 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:46.204439 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:46.204465 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:46.704299 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:46.704315 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:47.204330 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:47.204471 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:47.604430 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:47.604445 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:48.004317 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:48.004330 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
I160726 05:24:48.388402 stopper.go:408 quiesceing; tasks left:
1 storage/replica_range_lease.go:167
1 storage/replica.go:2395
I160726 05:24:48.388490 stopper.go:408 quiesceing; tasks left:
1 storage/replica.go:2395
W160726 05:24:48.388557 storage/store.go:1282 error gossiping system config: range 1: replica node_id:2 store_id:2 replica_id:3 not lease holder; <nil> is
W160726 05:24:48.388805 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: <nil>
W160726 05:24:48.388860 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 0} closed by the remote: <nil>
W160726 05:24:48.388891 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: <nil>
W160726 05:24:48.388898 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 0} closed by the remote: <nil>
W160726 05:24:48.388911 storage/replica.go:337 range 1: outgoing raft transport stream to {3 3 0} closed by the remote: rpc error: code = 13 desc = transport is closing
W160726 05:24:48.388915 storage/replica.go:337 range 1: outgoing raft transport stream to {3 3 4} closed by the remote: <nil>
W160726 05:24:48.388932 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: <nil>
--- FAIL: TestReplicateAddAndRemove (15.18s)
testing.go:115: condition failed to evaluate within 15s: addFirst: false, expected [39 0 39 39], got [39 5 39 39]
```
Run Details:
```
0 runs so far, 0 failures, over 5s
0 runs so far, 0 failures, over 10s
0 runs so far, 0 failures, over 15s
0 runs so far, 0 failures, over 20s
0 runs so far, 0 failures, over 25s
0 runs so far, 0 failures, over 30s
0 runs so far, 0 failures, over 35s
0 runs so far, 0 failures, over 40s
0 runs so far, 0 failures, over 45s
0 runs so far, 0 failures, over 50s
0 runs so far, 0 failures, over 55s
0 runs so far, 0 failures, over 1m0s
0 runs so far, 0 failures, over 1m5s
0 runs so far, 0 failures, over 1m10s
0 runs so far, 0 failures, over 1m15s
6 runs so far, 0 failures, over 1m20s
16 runs so far, 0 failures, over 1m25s
16 runs so far, 0 failures, over 1m30s
16 runs so far, 0 failures, over 1m35s
16 runs so far, 0 failures, over 1m40s
16 runs so far, 0 failures, over 1m45s
16 runs so far, 0 failures, over 1m50s
16 runs so far, 0 failures, over 1m55s
16 runs so far, 0 failures, over 2m0s
16 runs so far, 0 failures, over 2m5s
16 runs so far, 0 failures, over 2m10s
16 runs so far, 0 failures, over 2m15s
16 runs so far, 0 failures, over 2m20s
16 runs so far, 0 failures, over 2m25s
16 runs so far, 0 failures, over 2m30s
16 runs so far, 0 failures, over 2m35s
20 runs so far, 0 failures, over 2m40s
30 runs so far, 0 failures, over 2m45s
32 runs so far, 0 failures, over 2m50s
32 runs so far, 0 failures, over 2m55s
32 runs so far, 0 failures, over 3m0s
32 runs so far, 0 failures, over 3m5s
32 runs so far, 0 failures, over 3m10s
32 runs so far, 0 failures, over 3m15s
32 runs so far, 0 failures, over 3m20s
32 runs so far, 0 failures, over 3m25s
32 runs so far, 0 failures, over 3m30s
32 runs so far, 0 failures, over 3m35s
32 runs so far, 0 failures, over 3m40s
32 runs so far, 0 failures, over 3m45s
32 runs so far, 0 failures, over 3m50s
32 runs so far, 0 failures, over 3m55s
35 runs so far, 0 failures, over 4m0s
44 runs so far, 0 failures, over 4m5s
48 runs so far, 0 failures, over 4m10s
48 runs so far, 0 failures, over 4m15s
48 runs so far, 0 failures, over 4m20s
48 runs so far, 0 failures, over 4m25s
48 runs so far, 0 failures, over 4m30s
48 runs so far, 0 failures, over 4m35s
48 runs so far, 0 failures, over 4m40s
48 runs so far, 0 failures, over 4m45s
48 runs so far, 0 failures, over 4m50s
48 runs so far, 0 failures, over 4m55s
48 runs so far, 0 failures, over 5m0s
48 runs so far, 0 failures, over 5m5s
48 runs so far, 0 failures, over 5m10s
49 runs so far, 0 failures, over 5m15s
52 runs so far, 0 failures, over 5m20s
60 runs so far, 0 failures, over 5m25s
64 runs so far, 0 failures, over 5m30s
64 runs so far, 0 failures, over 5m35s
64 runs so far, 0 failures, over 5m40s
64 runs so far, 0 failures, over 5m45s
64 runs so far, 0 failures, over 5m50s
64 runs so far, 0 failures, over 5m55s
64 runs so far, 0 failures, over 6m0s
64 runs so far, 0 failures, over 6m5s
64 runs so far, 0 failures, over 6m10s
64 runs so far, 0 failures, over 6m15s
64 runs so far, 0 failures, over 6m20s
64 runs so far, 0 failures, over 6m25s
64 runs so far, 0 failures, over 6m30s
65 runs so far, 0 failures, over 6m35s
68 runs so far, 0 failures, over 6m40s
77 runs so far, 0 failures, over 6m45s
80 runs completed, 1 failures, over 6m50s
FAIL
```
Please assign, take a look and update the issue accordingly.
|
1.0
|
stress: failed test in cockroach/storage/storage.test: TestReplicateAddAndRemove - Binary: cockroach/static-tests.tar.gz sha: https://github.com/cockroachdb/cockroach/commits/c1114f920f13baf1df10cb555fd7516088249ae8
Stress build found a failed test:
```
=== RUN TestReplicateAddAndRemove
I160726 05:24:33.209159 storage/engine/rocksdb.go:353 opening in memory rocksdb instance
W160726 05:24:33.210050 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node
W160726 05:24:33.211342 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node
I160726 05:24:33.211509 storage/store.go:2671 system config not yet available
I160726 05:24:33.212090 storage/replica_command.go:1637 store=1:1 ([]=) range=1 [/Min-/Max): new range lease replica {1 1 1} 1970-01-01 00:00:00 +0000 UTC 900ms following replica {0 0 0} 1970-01-01 00:00:00 +0000 UTC 0 [physicalTime=1970-01-01 00:00:00 +0000 UTC]
I160726 05:24:33.212416 storage/engine/rocksdb.go:353 opening in memory rocksdb instance
W160726 05:24:33.213239 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node
I160726 05:24:33.213720 storage/store.go:2671 system config not yet available
I160726 05:24:33.213840 gossip/client.go:71 starting client to 127.0.0.1:38919
I160726 05:24:33.213844 storage/engine/rocksdb.go:353 opening in memory rocksdb instance
W160726 05:24:33.214625 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node
I160726 05:24:33.215180 storage/store.go:2671 system config not yet available
I160726 05:24:33.215304 gossip/client.go:71 starting client to 127.0.0.1:38919
I160726 05:24:33.215322 storage/engine/rocksdb.go:353 opening in memory rocksdb instance
W160726 05:24:33.216338 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node
I160726 05:24:33.217256 storage/store.go:2671 system config not yet available
I160726 05:24:33.217397 gossip/client.go:71 starting client to 127.0.0.1:38919
I160726 05:24:33.218819 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 12 in 243.521ยตs. encoded size=2580, 26 KV pairs, 2 log entries
I160726 05:24:33.219148 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > next_replica_id:2
I160726 05:24:33.219734 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.220086 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:4 StoreID:4 ReplicaID:2} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2}]
I160726 05:24:33.220386 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.220413 storage/replica_raftstorage.go:587 store=4:4 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 12 (encoded size=2580, 26 KV pairs, 2 log entries)
I160726 05:24:33.220963 storage/replica_raftstorage.go:590 store=4:4 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 492.712ยตs
I160726 05:24:33.221150 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 14 in 161.195ยตs. encoded size=3431, 29 KV pairs, 4 log entries
I160726 05:24:33.221452 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > next_replica_id:3
I160726 05:24:33.222290 storage/replica_raftstorage.go:587 store=2:2 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 14 (encoded size=3431, 29 KV pairs, 4 log entries)
I160726 05:24:33.222730 storage/replica_raftstorage.go:590 store=2:2 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 401.576ยตs
I160726 05:24:33.223154 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.223391 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.223721 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
I160726 05:24:33.224071 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.224404 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:2 StoreID:2 ReplicaID:3} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2} {NodeID:2 StoreID:2 ReplicaID:3}]
I160726 05:24:33.224628 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.225384 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.225797 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.226119 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
I160726 05:24:33.227423 storage/replica.go:1945 store=2:2 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.227676 storage/replica.go:1945 store=2:2 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.228066 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
I160726 05:24:33.229166 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 17 in 200.114ยตs. encoded size=4442, 33 KV pairs, 7 log entries
I160726 05:24:33.229475 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > replicas:<node_id:2 store_id:2 replica_id:3 > next_replica_id:4
I160726 05:24:33.230504 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.230881 storage/replica_raftstorage.go:587 store=3:3 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 17 (encoded size=4442, 33 KV pairs, 7 log entries)
I160726 05:24:33.230906 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:3 StoreID:3 ReplicaID:4} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2} {NodeID:2 StoreID:2 ReplicaID:3} {NodeID:3 StoreID:3 ReplicaID:4}]
W160726 05:24:33.231050 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1}
I160726 05:24:33.231186 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.231622 storage/replica_raftstorage.go:590 store=3:3 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 694.304ยตs
I160726 05:24:33.231941 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
W160726 05:24:33.232215 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1}
I160726 05:24:33.232549 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.232733 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > replicas:<node_id:2 store_id:2 replica_id:3 > replicas:<node_id:3 store_id:3 replica_id:4 > next_replica_id:5
I160726 05:24:33.232931 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
W160726 05:24:33.233887 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1}
I160726 05:24:33.235119 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.235346 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.235687 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
I160726 05:24:33.235916 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.236198 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing REMOVE_REPLICA {NodeID:2 StoreID:2 ReplicaID:3} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2} {NodeID:3 StoreID:3 ReplicaID:4}]
I160726 05:24:33.237944 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
W160726 05:24:33.237957 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1}
I160726 05:24:33.238396 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.239552 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.240543 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
W160726 05:24:33.240997 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1}
I160726 05:24:33.241153 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
I160726 05:24:33.241324 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.241659 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
W160726 05:24:33.266682 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: <nil>
W160726 05:24:33.266805 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 0} closed by the remote: <nil>
W160726 05:24:33.266829 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 0} closed by the remote: <nil>
W160726 05:24:33.266852 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: <nil>
W160726 05:24:33.266874 storage/replica.go:337 range 1: outgoing raft transport stream to {3 3 0} closed by the remote: <nil>
W160726 05:24:33.266907 storage/replica.go:337 range 1: outgoing raft transport stream to {3 3 4} closed by the remote: <nil>
W160726 05:24:33.266921 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: <nil>
W160726 05:24:33.266941 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:1 StoreID:1 ReplicaID:1}: no store registered for {NodeID:3 StoreID:3 ReplicaID:4}
I160726 05:24:33.267618 storage/engine/rocksdb.go:353 opening in memory rocksdb instance
W160726 05:24:33.268403 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node
I160726 05:24:33.269801 storage/store.go:2671 system config not yet available
W160726 05:24:33.269940 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node
I160726 05:24:33.270410 storage/replica_command.go:1637 store=1:1 ([]=) range=1 [/Min-/Max): new range lease replica {1 1 1} 1970-01-01 00:00:00 +0000 UTC 900ms following replica {0 0 0} 1970-01-01 00:00:00 +0000 UTC 0 [physicalTime=1970-01-01 00:00:00 +0000 UTC]
I160726 05:24:33.270697 storage/engine/rocksdb.go:353 opening in memory rocksdb instance
W160726 05:24:33.271184 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node
I160726 05:24:33.271629 storage/store.go:2671 system config not yet available
I160726 05:24:33.271813 gossip/client.go:71 starting client to 127.0.0.1:37700
I160726 05:24:33.271900 storage/engine/rocksdb.go:353 opening in memory rocksdb instance
W160726 05:24:33.272505 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node
I160726 05:24:33.273221 storage/store.go:2671 system config not yet available
I160726 05:24:33.273356 gossip/client.go:71 starting client to 127.0.0.1:37700
I160726 05:24:33.273382 storage/engine/rocksdb.go:353 opening in memory rocksdb instance
W160726 05:24:33.274134 gossip/gossip.go:942 not connected to cluster; use --join to specify a connected node
I160726 05:24:33.274729 storage/store.go:2671 system config not yet available
I160726 05:24:33.274752 gossip/client.go:71 starting client to 127.0.0.1:37700
I160726 05:24:33.276138 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 12 in 142.306ยตs. encoded size=2580, 26 KV pairs, 2 log entries
I160726 05:24:33.276432 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > next_replica_id:2
I160726 05:24:33.277000 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.277368 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:4 StoreID:4 ReplicaID:2} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2}]
I160726 05:24:33.277783 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.277899 storage/replica_raftstorage.go:587 store=4:4 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 12 (encoded size=2580, 26 KV pairs, 2 log entries)
I160726 05:24:33.278272 storage/replica_raftstorage.go:590 store=4:4 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 332.578ยตs
I160726 05:24:33.278606 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 14 in 131.608ยตs. encoded size=3431, 29 KV pairs, 4 log entries
I160726 05:24:33.278815 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > next_replica_id:3
I160726 05:24:33.279760 storage/replica_raftstorage.go:587 store=2:2 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 14 (encoded size=3431, 29 KV pairs, 4 log entries)
I160726 05:24:33.280204 storage/replica_raftstorage.go:590 store=2:2 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 399.401ยตs
I160726 05:24:33.280443 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.280728 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.280936 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.281222 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
I160726 05:24:33.281321 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:2 StoreID:2 ReplicaID:3} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2} {NodeID:2 StoreID:2 ReplicaID:3}]
I160726 05:24:33.281495 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.282152 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.282741 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.283186 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
I160726 05:24:33.284773 storage/replica.go:1945 store=2:2 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.285021 storage/replica.go:1945 store=2:2 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.285465 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
I160726 05:24:33.289996 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > replicas:<node_id:2 store_id:2 replica_id:3 > next_replica_id:4
W160726 05:24:33.291631 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1}
I160726 05:24:33.291894 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.292323 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.292666 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing REMOVE_REPLICA {NodeID:2 StoreID:2 ReplicaID:3} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2}]
W160726 05:24:33.293532 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1}
I160726 05:24:33.294206 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
W160726 05:24:33.294991 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:3}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1}
I160726 05:24:33.295056 storage/replica_raftstorage.go:466 generated snapshot for range 1 at index 19 in 184.365ยตs. encoded size=5306, 36 KV pairs, 9 log entries
I160726 05:24:33.295408 storage/replica_command.go:3145 store=1:1 ([]=) range=1 [/Min-/Max): change replicas of 1: read existing descriptor range_id:1 start_key:"" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:4 store_id:4 replica_id:2 > next_replica_id:4
I160726 05:24:33.296003 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.296205 storage/replica_raftstorage.go:587 store=3:3 ([]=) range=1 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 1 at index 19 (encoded size=5306, 36 KV pairs, 9 log entries)
I160726 05:24:33.296535 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
I160726 05:24:33.296757 storage/replica_raftstorage.go:590 store=3:3 ([]=) range=1 [/Min-/Max): with replicaID [?], applied preemptive snapshot for range 1 in 506.141ยตs
I160726 05:24:33.297024 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.297498 storage/replica.go:1414 store=1:1 ([]=) range=1 [/Min-/Max): proposing ADD_REPLICA {NodeID:3 StoreID:3 ReplicaID:4} for range 1: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:4 StoreID:4 ReplicaID:2} {NodeID:3 StoreID:3 ReplicaID:4}]
I160726 05:24:33.297547 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.298421 storage/replica.go:1945 store=1:1 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.299146 storage/replica.go:1945 store=4:4 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.299590 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
I160726 05:24:33.300598 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/""/RangeDescriptor,/Min), ConditionalPut [/Local/Range/""/RangeDescriptor,/Min), Put [/Meta2/Max,/Min), Put [/Meta1/Max,/Min), pErr=<nil>
I160726 05:24:33.300760 storage/replica.go:1945 store=3:3 ([]=) range=1 [/Min-/Max): applied part of replica change txn: EndTransaction [/Local/Range/""/RangeDescriptor,/Min), pErr=<nil>
I160726 05:24:33.301169 storage/replica_command.go:2934 unable to gossip first range; hasLease=false, err=<nil>
W160726 05:24:33.704343 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:33.704394 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:34.204542 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:34.204557 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
I160726 05:24:34.270157 gossip/client.go:71 starting client to 127.0.0.1:37298
I160726 05:24:34.270721 gossip/client.go:95 closing client to node 3 (127.0.0.1:37298): stopping outgoing client to node 3 (127.0.0.1:37298); already have incoming
W160726 05:24:34.504339 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:34.504431 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:34.804438 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:34.804451 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:35.304476 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:35.304622 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:35.604507 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:35.604522 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:35.904329 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:35.904397 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:36.204411 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:36.204476 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:36.504400 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:36.504468 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:36.804397 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:36.804411 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:37.304381 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:37.304395 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:37.604380 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:37.604402 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:38.104389 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:38.104405 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:38.504301 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:38.504315 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:38.804308 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:38.804363 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:39.304551 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:39.304564 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:39.604313 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:39.604324 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:40.104258 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:40.104336 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:40.604440 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:40.604453 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:41.004364 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:41.005420 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:41.304399 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:41.304471 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:41.704448 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:41.704472 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:42.204429 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:42.204482 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:42.604358 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:42.604371 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:43.004396 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:43.004418 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:43.504343 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:43.504373 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:43.804438 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:43.804511 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:44.304409 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:44.304424 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:44.604345 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:44.604361 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:45.004281 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:45.004314 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:45.504305 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:45.504352 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:45.904459 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:45.904486 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:46.204439 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:46.204465 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:46.704299 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:46.704315 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:47.204330 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:47.204471 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:47.604430 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:47.604445 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:48.004317 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
W160726 05:24:48.004330 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: rpc error: code = 2 desc = range 1: discarding message from {NodeID:2 StoreID:2 ReplicaID:3}, older than NextReplicaID 5
I160726 05:24:48.388402 stopper.go:408 quiesceing; tasks left:
1 storage/replica_range_lease.go:167
1 storage/replica.go:2395
I160726 05:24:48.388490 stopper.go:408 quiesceing; tasks left:
1 storage/replica.go:2395
W160726 05:24:48.388557 storage/store.go:1282 error gossiping system config: range 1: replica node_id:2 store_id:2 replica_id:3 not lease holder; <nil> is
W160726 05:24:48.388805 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: <nil>
W160726 05:24:48.388860 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 0} closed by the remote: <nil>
W160726 05:24:48.388891 storage/replica.go:337 range 1: outgoing raft transport stream to {4 4 2} closed by the remote: <nil>
W160726 05:24:48.388898 storage/replica.go:337 range 1: outgoing raft transport stream to {2 2 0} closed by the remote: <nil>
W160726 05:24:48.388911 storage/replica.go:337 range 1: outgoing raft transport stream to {3 3 0} closed by the remote: rpc error: code = 13 desc = transport is closing
W160726 05:24:48.388915 storage/replica.go:337 range 1: outgoing raft transport stream to {3 3 4} closed by the remote: <nil>
W160726 05:24:48.388932 storage/replica.go:337 range 1: outgoing raft transport stream to {1 1 1} closed by the remote: <nil>
--- FAIL: TestReplicateAddAndRemove (15.18s)
testing.go:115: condition failed to evaluate within 15s: addFirst: false, expected [39 0 39 39], got [39 5 39 39]
```
Run Details:
```
0 runs so far, 0 failures, over 5s
0 runs so far, 0 failures, over 10s
0 runs so far, 0 failures, over 15s
0 runs so far, 0 failures, over 20s
0 runs so far, 0 failures, over 25s
0 runs so far, 0 failures, over 30s
0 runs so far, 0 failures, over 35s
0 runs so far, 0 failures, over 40s
0 runs so far, 0 failures, over 45s
0 runs so far, 0 failures, over 50s
0 runs so far, 0 failures, over 55s
0 runs so far, 0 failures, over 1m0s
0 runs so far, 0 failures, over 1m5s
0 runs so far, 0 failures, over 1m10s
0 runs so far, 0 failures, over 1m15s
6 runs so far, 0 failures, over 1m20s
16 runs so far, 0 failures, over 1m25s
16 runs so far, 0 failures, over 1m30s
16 runs so far, 0 failures, over 1m35s
16 runs so far, 0 failures, over 1m40s
16 runs so far, 0 failures, over 1m45s
16 runs so far, 0 failures, over 1m50s
16 runs so far, 0 failures, over 1m55s
16 runs so far, 0 failures, over 2m0s
16 runs so far, 0 failures, over 2m5s
16 runs so far, 0 failures, over 2m10s
16 runs so far, 0 failures, over 2m15s
16 runs so far, 0 failures, over 2m20s
16 runs so far, 0 failures, over 2m25s
16 runs so far, 0 failures, over 2m30s
16 runs so far, 0 failures, over 2m35s
20 runs so far, 0 failures, over 2m40s
30 runs so far, 0 failures, over 2m45s
32 runs so far, 0 failures, over 2m50s
32 runs so far, 0 failures, over 2m55s
32 runs so far, 0 failures, over 3m0s
32 runs so far, 0 failures, over 3m5s
32 runs so far, 0 failures, over 3m10s
32 runs so far, 0 failures, over 3m15s
32 runs so far, 0 failures, over 3m20s
32 runs so far, 0 failures, over 3m25s
32 runs so far, 0 failures, over 3m30s
32 runs so far, 0 failures, over 3m35s
32 runs so far, 0 failures, over 3m40s
32 runs so far, 0 failures, over 3m45s
32 runs so far, 0 failures, over 3m50s
32 runs so far, 0 failures, over 3m55s
35 runs so far, 0 failures, over 4m0s
44 runs so far, 0 failures, over 4m5s
48 runs so far, 0 failures, over 4m10s
48 runs so far, 0 failures, over 4m15s
48 runs so far, 0 failures, over 4m20s
48 runs so far, 0 failures, over 4m25s
48 runs so far, 0 failures, over 4m30s
48 runs so far, 0 failures, over 4m35s
48 runs so far, 0 failures, over 4m40s
48 runs so far, 0 failures, over 4m45s
48 runs so far, 0 failures, over 4m50s
48 runs so far, 0 failures, over 4m55s
48 runs so far, 0 failures, over 5m0s
48 runs so far, 0 failures, over 5m5s
48 runs so far, 0 failures, over 5m10s
49 runs so far, 0 failures, over 5m15s
52 runs so far, 0 failures, over 5m20s
60 runs so far, 0 failures, over 5m25s
64 runs so far, 0 failures, over 5m30s
64 runs so far, 0 failures, over 5m35s
64 runs so far, 0 failures, over 5m40s
64 runs so far, 0 failures, over 5m45s
64 runs so far, 0 failures, over 5m50s
64 runs so far, 0 failures, over 5m55s
64 runs so far, 0 failures, over 6m0s
64 runs so far, 0 failures, over 6m5s
64 runs so far, 0 failures, over 6m10s
64 runs so far, 0 failures, over 6m15s
64 runs so far, 0 failures, over 6m20s
64 runs so far, 0 failures, over 6m25s
64 runs so far, 0 failures, over 6m30s
65 runs so far, 0 failures, over 6m35s
68 runs so far, 0 failures, over 6m40s
77 runs so far, 0 failures, over 6m45s
80 runs completed, 1 failures, over 6m50s
FAIL
```
Please assign, take a look and update the issue accordingly.
|
non_defect
|
stress failed test in cockroach storage storage test testreplicateaddandremove binary cockroach static tests tar gz sha stress build found a failed test run testreplicateaddandremove storage engine rocksdb go opening in memory rocksdb instance gossip gossip go not connected to cluster use join to specify a connected node gossip gossip go not connected to cluster use join to specify a connected node storage store go system config not yet available storage replica command go store range storage engine rocksdb go opening in memory rocksdb instance gossip gossip go not connected to cluster use join to specify a connected node storage store go system config not yet available gossip client go starting client to storage engine rocksdb go opening in memory rocksdb instance gossip gossip go not connected to cluster use join to specify a connected node storage store go system config not yet available gossip client go starting client to storage engine rocksdb go opening in memory rocksdb instance gossip gossip go not connected to cluster use join to specify a connected node storage store go system config not yet available gossip client go starting client to storage replica raftstorage go generated snapshot for range at index in encoded size kv pairs log entries storage replica command go store range min max change replicas of read existing descriptor range id start key end key replicas next replica id storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica raftstorage go store range applying preemptive snapshot for range at index encoded size kv pairs log entries storage replica raftstorage go store range applied preemptive snapshot for range in storage replica raftstorage go generated snapshot for range at index in encoded size kv pairs log entries storage replica command go store range min max change replicas of read existing descriptor range id start key end key replicas replicas next replica id storage replica raftstorage go store range applying preemptive snapshot for range at index encoded size kv pairs log entries storage replica raftstorage go store range applied preemptive snapshot for range in storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica command go unable to gossip first range haslease false err storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica command go unable to gossip first range haslease false err storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica command go unable to gossip first range haslease false err storage replica raftstorage go generated snapshot for range at index in encoded size kv pairs log entries storage replica command go store range min max change replicas of read existing descriptor range id start key end key replicas replicas replicas next replica id storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica raftstorage go store range applying preemptive snapshot for range at index encoded size kv pairs log entries storage replica go store range storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc unable to accept raft message from nodeid storeid replicaid no store registered for nodeid storeid replicaid storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica raftstorage go store range applied preemptive snapshot for range in storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc unable to accept raft message from nodeid storeid replicaid no store registered for nodeid storeid replicaid storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica command go store range min max change replicas of read existing descriptor range id start key end key replicas replicas replicas replicas next replica id storage replica command go unable to gossip first range haslease false err storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc unable to accept raft message from nodeid storeid replicaid no store registered for nodeid storeid replicaid storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica command go unable to gossip first range haslease false err storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc unable to accept raft message from nodeid storeid replicaid no store registered for nodeid storeid replicaid storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc unable to accept raft message from nodeid storeid replicaid no store registered for nodeid storeid replicaid storage replica command go unable to gossip first range haslease false err storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica command go unable to gossip first range haslease false err storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc unable to accept raft message from nodeid storeid replicaid no store registered for nodeid storeid replicaid storage engine rocksdb go opening in memory rocksdb instance gossip gossip go not connected to cluster use join to specify a connected node storage store go system config not yet available gossip gossip go not connected to cluster use join to specify a connected node storage replica command go store range storage engine rocksdb go opening in memory rocksdb instance gossip gossip go not connected to cluster use join to specify a connected node storage store go system config not yet available gossip client go starting client to storage engine rocksdb go opening in memory rocksdb instance gossip gossip go not connected to cluster use join to specify a connected node storage store go system config not yet available gossip client go starting client to storage engine rocksdb go opening in memory rocksdb instance gossip gossip go not connected to cluster use join to specify a connected node storage store go system config not yet available gossip client go starting client to storage replica raftstorage go generated snapshot for range at index in encoded size kv pairs log entries storage replica command go store range min max change replicas of read existing descriptor range id start key end key replicas next replica id storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica raftstorage go store range applying preemptive snapshot for range at index encoded size kv pairs log entries storage replica raftstorage go store range applied preemptive snapshot for range in storage replica raftstorage go generated snapshot for range at index in encoded size kv pairs log entries storage replica command go store range min max change replicas of read existing descriptor range id start key end key replicas replicas next replica id storage replica raftstorage go store range applying preemptive snapshot for range at index encoded size kv pairs log entries storage replica raftstorage go store range applied preemptive snapshot for range in storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica command go unable to gossip first range haslease false err storage replica go store range storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica command go unable to gossip first range haslease false err storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica command go unable to gossip first range haslease false err storage replica command go store range min max change replicas of read existing descriptor range id start key end key replicas replicas replicas next replica id storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc unable to accept raft message from nodeid storeid replicaid no store registered for nodeid storeid replicaid storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc unable to accept raft message from nodeid storeid replicaid no store registered for nodeid storeid replicaid storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc unable to accept raft message from nodeid storeid replicaid no store registered for nodeid storeid replicaid storage replica raftstorage go generated snapshot for range at index in encoded size kv pairs log entries storage replica command go store range min max change replicas of read existing descriptor range id start key end key replicas replicas next replica id storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica raftstorage go store range applying preemptive snapshot for range at index encoded size kv pairs log entries storage replica command go unable to gossip first range haslease false err storage replica raftstorage go store range applied preemptive snapshot for range in storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica command go unable to gossip first range haslease false err storage replica go store range min max applied part of replica change txn begintransaction local range rangedescriptor min conditionalput local range rangedescriptor min put max min put max min perr storage replica go store range min max applied part of replica change txn endtransaction local range rangedescriptor min perr storage replica command go unable to gossip first range haslease false err storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid gossip client go starting client to gossip client go closing client to node stopping outgoing client to node already have incoming storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc range discarding message from nodeid storeid replicaid older than nextreplicaid stopper go quiesceing tasks left storage replica range lease go storage replica go stopper go quiesceing tasks left storage replica go storage store go error gossiping system config range replica node id store id replica id not lease holder is storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote rpc error code desc transport is closing storage replica go range outgoing raft transport stream to closed by the remote storage replica go range outgoing raft transport stream to closed by the remote fail testreplicateaddandremove testing go condition failed to evaluate within addfirst false expected got run details runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs completed failures over fail please assign take a look and update the issue accordingly
| 0
|
161,932
| 12,600,462,146
|
IssuesEvent
|
2020-06-11 08:11:50
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
[Failing Test] [sig-scheduling] SchedulerPreemption [Serial] PreemptionExecutionPath runs ReplicaSets to verify preemption running path [Conformance]
|
kind/failing-test priority/critical-urgent sig/scheduling
|
<!-- Please only use this template for submitting reports about continuously failing tests or jobs in Kubernetes CI -->
**Which jobs are failing**:
`Conformance - GCE - master (ci-kubernetes-gce-conformance-latest)`
**Which test(s) are failing**:
`[sig-scheduling] SchedulerPreemption [Serial] PreemptionExecutionPath runs ReplicaSets to verify preemption running path [Conformance] `
**Since when has it been failing**:
`06-09-20 22:27 PDT`
**Testgrid link**:
https://testgrid.k8s.io/sig-release-master-blocking#Conformance%20-%20GCE%20-%20master
**Reason for failure**:
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597
Jun 9 15:35:40.256: failed pod observation expectations: rs1 had more than 2 pods created: 3
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/preemption.go:655
```
**Anything else we need to know**:
/sig scheduling
/cc @kubernetes/ci-signal
/priority critical-urgent
/milestone v1.19
|
1.0
|
[Failing Test] [sig-scheduling] SchedulerPreemption [Serial] PreemptionExecutionPath runs ReplicaSets to verify preemption running path [Conformance] - <!-- Please only use this template for submitting reports about continuously failing tests or jobs in Kubernetes CI -->
**Which jobs are failing**:
`Conformance - GCE - master (ci-kubernetes-gce-conformance-latest)`
**Which test(s) are failing**:
`[sig-scheduling] SchedulerPreemption [Serial] PreemptionExecutionPath runs ReplicaSets to verify preemption running path [Conformance] `
**Since when has it been failing**:
`06-09-20 22:27 PDT`
**Testgrid link**:
https://testgrid.k8s.io/sig-release-master-blocking#Conformance%20-%20GCE%20-%20master
**Reason for failure**:
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597
Jun 9 15:35:40.256: failed pod observation expectations: rs1 had more than 2 pods created: 3
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/preemption.go:655
```
**Anything else we need to know**:
/sig scheduling
/cc @kubernetes/ci-signal
/priority critical-urgent
/milestone v1.19
|
non_defect
|
schedulerpreemption preemptionexecutionpath runs replicasets to verify preemption running path which jobs are failing conformance gce master ci kubernetes gce conformance latest which test s are failing schedulerpreemption preemptionexecutionpath runs replicasets to verify preemption running path since when has it been failing pdt testgrid link reason for failure go src io kubernetes output dockerized go src io kubernetes test framework framework go jun failed pod observation expectations had more than pods created go src io kubernetes output dockerized go src io kubernetes test scheduling preemption go anything else we need to know sig scheduling cc kubernetes ci signal priority critical urgent milestone
| 0
|
376,400
| 26,198,767,739
|
IssuesEvent
|
2023-01-03 15:39:12
|
taikoxyz/taiko-mono
|
https://api.github.com/repos/taikoxyz/taiko-mono
|
closed
|
feat(docs): add documentation on how to generate a private key
|
documentation good first issue dx
|
### Describe the feature request
There are a lot of questions about how to generate a private key for this documentation: https://taiko.xyz/docs/alpha-1-testnet/deploy-a-contract#steps
Let's add some steps on how to do it. There are two decent ways to do this, I would highly suggest to use a dummy account and not the private key of an account that has any real funds.
You can create a dummy account keypair manually with openssl (something like this https://gist.github.com/miguelmota/3793b160992b4ea0b616497b8e5aee2f) , or you can just create an account with metamask and export the private key.
We should keep in mind that not everyone has metamask. Perhaps we can even suggest usage of a useless private key if they don't care about the contract deployer (however doing this means that address will not show up when we scan the network for the POAP).
### Describe alternatives you've considered
None
### Additional context
None
|
1.0
|
feat(docs): add documentation on how to generate a private key - ### Describe the feature request
There are a lot of questions about how to generate a private key for this documentation: https://taiko.xyz/docs/alpha-1-testnet/deploy-a-contract#steps
Let's add some steps on how to do it. There are two decent ways to do this, I would highly suggest to use a dummy account and not the private key of an account that has any real funds.
You can create a dummy account keypair manually with openssl (something like this https://gist.github.com/miguelmota/3793b160992b4ea0b616497b8e5aee2f) , or you can just create an account with metamask and export the private key.
We should keep in mind that not everyone has metamask. Perhaps we can even suggest usage of a useless private key if they don't care about the contract deployer (however doing this means that address will not show up when we scan the network for the POAP).
### Describe alternatives you've considered
None
### Additional context
None
|
non_defect
|
feat docs add documentation on how to generate a private key describe the feature request there are a lot of questions about how to generate a private key for this documentation let s add some steps on how to do it there are two decent ways to do this i would highly suggest to use a dummy account and not the private key of an account that has any real funds you can create a dummy account keypair manually with openssl something like this or you can just create an account with metamask and export the private key we should keep in mind that not everyone has metamask perhaps we can even suggest usage of a useless private key if they don t care about the contract deployer however doing this means that address will not show up when we scan the network for the poap describe alternatives you ve considered none additional context none
| 0
|
41,328
| 10,417,616,108
|
IssuesEvent
|
2019-09-14 23:34:01
|
avereon/xenon
|
https://api.github.com/repos/avereon/xenon
|
closed
|
Guide tool icons not transparent
|
bug / error / defect
|
The guide tool icons do not appear to be using the alpha channel when rendering. This may be due to CSS configuration, guide renderer configuration or some other rendering configuration that does not allow those pixels to be translucent.
|
1.0
|
Guide tool icons not transparent - The guide tool icons do not appear to be using the alpha channel when rendering. This may be due to CSS configuration, guide renderer configuration or some other rendering configuration that does not allow those pixels to be translucent.
|
defect
|
guide tool icons not transparent the guide tool icons do not appear to be using the alpha channel when rendering this may be due to css configuration guide renderer configuration or some other rendering configuration that does not allow those pixels to be translucent
| 1
|
23,673
| 3,851,865,309
|
IssuesEvent
|
2016-04-06 05:27:47
|
GPF/imame4all
|
https://api.github.com/repos/GPF/imame4all
|
closed
|
Steelseries Free Controler support.
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. Buy a SteelSeries Free Controller
2. Connect it in either Gamepad or Arcade (icade) mode
3. Basic buttons are mapped to coin and start and other buttons dont map right
due to the way mame4ios seems to map buttons to functions and then those
functions to game buttons and actions. When I remap the buttons, I have to
assign "button 1" to "coin" (and other weird combinations) which messes up
other things.
What is the expected output? What do you see instead?
I was hoping to remap a key to "coin" instead of the key being mapped to "coin"
then me having to map "coin" to something else. Or if there was a way to
support the weird order of keys the Steelseries Free spits out.
What version of the product are you using? On what operating system?
Latest as of Feb 27 (the Feb9release of mame4ios reloaded)
Please provide any additional information below.
I will donate a SteelSeries Free Controller if there is interest in support.
Supporting the Free's Arcade mode should be supper simple as it sends key
down/key up just like any other iCade device but the Free sends different
combinations for buttons than the others making the orientations funny.
The Free also can be a full analog gamepad controller with dual analog sticks.
This is what would be awesome to have supported as well.
Again, let me know if interest in support and I will donate a controller.
```
Original issue reported on code.google.com by `RyanStep...@gmail.com` on 28 Feb 2013 at 12:58
|
1.0
|
Steelseries Free Controler support. - ```
What steps will reproduce the problem?
1. Buy a SteelSeries Free Controller
2. Connect it in either Gamepad or Arcade (icade) mode
3. Basic buttons are mapped to coin and start and other buttons dont map right
due to the way mame4ios seems to map buttons to functions and then those
functions to game buttons and actions. When I remap the buttons, I have to
assign "button 1" to "coin" (and other weird combinations) which messes up
other things.
What is the expected output? What do you see instead?
I was hoping to remap a key to "coin" instead of the key being mapped to "coin"
then me having to map "coin" to something else. Or if there was a way to
support the weird order of keys the Steelseries Free spits out.
What version of the product are you using? On what operating system?
Latest as of Feb 27 (the Feb9release of mame4ios reloaded)
Please provide any additional information below.
I will donate a SteelSeries Free Controller if there is interest in support.
Supporting the Free's Arcade mode should be supper simple as it sends key
down/key up just like any other iCade device but the Free sends different
combinations for buttons than the others making the orientations funny.
The Free also can be a full analog gamepad controller with dual analog sticks.
This is what would be awesome to have supported as well.
Again, let me know if interest in support and I will donate a controller.
```
Original issue reported on code.google.com by `RyanStep...@gmail.com` on 28 Feb 2013 at 12:58
|
defect
|
steelseries free controler support what steps will reproduce the problem buy a steelseries free controller connect it in either gamepad or arcade icade mode basic buttons are mapped to coin and start and other buttons dont map right due to the way seems to map buttons to functions and then those functions to game buttons and actions when i remap the buttons i have to assign button to coin and other weird combinations which messes up other things what is the expected output what do you see instead i was hoping to remap a key to coin instead of the key being mapped to coin then me having to map coin to something else or if there was a way to support the weird order of keys the steelseries free spits out what version of the product are you using on what operating system latest as of feb the of reloaded please provide any additional information below i will donate a steelseries free controller if there is interest in support supporting the free s arcade mode should be supper simple as it sends key down key up just like any other icade device but the free sends different combinations for buttons than the others making the orientations funny the free also can be a full analog gamepad controller with dual analog sticks this is what would be awesome to have supported as well again let me know if interest in support and i will donate a controller original issue reported on code google com by ryanstep gmail com on feb at
| 1
|
51,719
| 13,211,296,750
|
IssuesEvent
|
2020-08-15 22:07:29
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
[photospline] - divzero - loop preconditions aren't checked (Trac #922)
|
Incomplete Migration Migrated from Trac combo reconstruction defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/922">https://code.icecube.wisc.edu/projects/icecube/ticket/922</a>, reported by negaand owned by jvansanten</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-04-12T17:41:42",
"_ts": "1428860502423619",
"description": "http://goo.gl/FFVVbF\n\nloop preconditions aren't checked allowing a potential divide-by-zero error to occur in 7 steps.\n\nfix: pre-check and hard fail if loop pre-conditions suck",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"time": "2015-04-10T04:17:37",
"component": "combo reconstruction",
"summary": "[photospline] - divzero - loop preconditions aren't checked",
"priority": "normal",
"keywords": "photospline divzero",
"milestone": "",
"owner": "jvansanten",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[photospline] - divzero - loop preconditions aren't checked (Trac #922) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/922">https://code.icecube.wisc.edu/projects/icecube/ticket/922</a>, reported by negaand owned by jvansanten</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-04-12T17:41:42",
"_ts": "1428860502423619",
"description": "http://goo.gl/FFVVbF\n\nloop preconditions aren't checked allowing a potential divide-by-zero error to occur in 7 steps.\n\nfix: pre-check and hard fail if loop pre-conditions suck",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"time": "2015-04-10T04:17:37",
"component": "combo reconstruction",
"summary": "[photospline] - divzero - loop preconditions aren't checked",
"priority": "normal",
"keywords": "photospline divzero",
"milestone": "",
"owner": "jvansanten",
"type": "defect"
}
```
</p>
</details>
|
defect
|
divzero loop preconditions aren t checked trac migrated from json status closed changetime ts description preconditions aren t checked allowing a potential divide by zero error to occur in steps n nfix pre check and hard fail if loop pre conditions suck reporter nega cc resolution fixed time component combo reconstruction summary divzero loop preconditions aren t checked priority normal keywords photospline divzero milestone owner jvansanten type defect
| 1
|
65,453
| 6,964,020,386
|
IssuesEvent
|
2017-12-08 19:49:09
|
legchikov/coding_interview
|
https://api.github.com/repos/legchikov/coding_interview
|
closed
|
[Arrays and Strings][Tests] Is Unique
|
test
|
Implement an algorithm to determine if a string has all unique characters.
What if you cannot use additional data structures?
|
1.0
|
[Arrays and Strings][Tests] Is Unique - Implement an algorithm to determine if a string has all unique characters.
What if you cannot use additional data structures?
|
non_defect
|
is unique implement an algorithm to determine if a string has all unique characters what if you cannot use additional data structures
| 0
|
32,834
| 6,953,397,182
|
IssuesEvent
|
2017-12-06 20:52:48
|
Dzhuneyt/jquery-tubular
|
https://api.github.com/repos/Dzhuneyt/jquery-tubular
|
closed
|
Video is not playing in Android device
|
auto-migrated Priority-Medium Type-Defect
|
```
I am using your tubular js for playing the video in background. it's working
well few days ago. but now when i opened the website in my android phone the
video doesn't show. and also the folllowing url doesn't show video in
background in android device.
http://www.seanmccambridge.com/tubular/
```
Original issue reported on code.google.com by `jitendra...@technosoftwares.in` on 26 May 2014 at 2:21
|
1.0
|
Video is not playing in Android device - ```
I am using your tubular js for playing the video in background. it's working
well few days ago. but now when i opened the website in my android phone the
video doesn't show. and also the folllowing url doesn't show video in
background in android device.
http://www.seanmccambridge.com/tubular/
```
Original issue reported on code.google.com by `jitendra...@technosoftwares.in` on 26 May 2014 at 2:21
|
defect
|
video is not playing in android device i am using your tubular js for playing the video in background it s working well few days ago but now when i opened the website in my android phone the video doesn t show and also the folllowing url doesn t show video in background in android device original issue reported on code google com by jitendra technosoftwares in on may at
| 1
|
160,836
| 6,103,449,580
|
IssuesEvent
|
2017-06-20 18:44:08
|
craftercms/craftercms
|
https://api.github.com/repos/craftercms/craftercms
|
opened
|
[studio] Bulk publish tries to publish items that area already published
|
enhancement Priority: Medium
|
We need the bulk publish algorithm to the following if possible:
- Query the database for all items that are not published
- For that list, add item to publish queue
Ping me if you have any questions.
|
1.0
|
[studio] Bulk publish tries to publish items that area already published - We need the bulk publish algorithm to the following if possible:
- Query the database for all items that are not published
- For that list, add item to publish queue
Ping me if you have any questions.
|
non_defect
|
bulk publish tries to publish items that area already published we need the bulk publish algorithm to the following if possible query the database for all items that are not published for that list add item to publish queue ping me if you have any questions
| 0
|
71,792
| 23,804,601,000
|
IssuesEvent
|
2022-09-03 21:02:48
|
ascott18/TellMeWhen
|
https://api.github.com/repos/ascott18/TellMeWhen
|
closed
|
[Bug]: Swing Timer bug with flurry
|
T: defect S: resolved
|
### WoW Version
TBC Classic
### TellMeWhen Version
9.2.4 r10
### Describe the bug
Swing timer is sometimes inaccurate, possibly due to when it checks for flurry being up or not.
[Video of the behavior](https://www.youtube.com/watch?v=GHjEt8qJBGM)
[screenshot](https://imgur.com/oZGoDgs)
### Export Strings
```text
N/A
```
|
1.0
|
[Bug]: Swing Timer bug with flurry - ### WoW Version
TBC Classic
### TellMeWhen Version
9.2.4 r10
### Describe the bug
Swing timer is sometimes inaccurate, possibly due to when it checks for flurry being up or not.
[Video of the behavior](https://www.youtube.com/watch?v=GHjEt8qJBGM)
[screenshot](https://imgur.com/oZGoDgs)
### Export Strings
```text
N/A
```
|
defect
|
swing timer bug with flurry wow version tbc classic tellmewhen version describe the bug swing timer is sometimes inaccurate possibly due to when it checks for flurry being up or not export strings text n a
| 1
|
46,645
| 5,825,420,871
|
IssuesEvent
|
2017-05-07 21:18:14
|
mnc123/uapp
|
https://api.github.com/repos/mnc123/uapp
|
closed
|
Variaciรณn de pantalla de espera
|
Test
|
Cuando hay 5 usuarios en cola cambia la pรกgina de espera, esto me parece correcto pero hay que cambiarlo de vista cuando ya este adentro de los primeros 5 (o 4 no recuerdo), tambiรฉn ver en detalle cรณmo varรญa la pantalla de espera en estos casos.
|
1.0
|
Variaciรณn de pantalla de espera - Cuando hay 5 usuarios en cola cambia la pรกgina de espera, esto me parece correcto pero hay que cambiarlo de vista cuando ya este adentro de los primeros 5 (o 4 no recuerdo), tambiรฉn ver en detalle cรณmo varรญa la pantalla de espera en estos casos.
|
non_defect
|
variaciรณn de pantalla de espera cuando hay usuarios en cola cambia la pรกgina de espera esto me parece correcto pero hay que cambiarlo de vista cuando ya este adentro de los primeros o no recuerdo tambiรฉn ver en detalle cรณmo varรญa la pantalla de espera en estos casos
| 0
|
42,832
| 11,299,257,540
|
IssuesEvent
|
2020-01-17 10:47:23
|
vector-im/riot-web
|
https://api.github.com/repos/vector-im/riot-web
|
opened
|
Complete security incorrectly sends you through key backup restore
|
bug defect p1 story:1
|
Tom tested the complete security flow on his account with key backup, and when clicking start, he was sent into a restore key backup flow, but that's not intended.
The expected thing is you are prompted for the SSSS passphrase and the device is verified.
|
1.0
|
Complete security incorrectly sends you through key backup restore - Tom tested the complete security flow on his account with key backup, and when clicking start, he was sent into a restore key backup flow, but that's not intended.
The expected thing is you are prompted for the SSSS passphrase and the device is verified.
|
defect
|
complete security incorrectly sends you through key backup restore tom tested the complete security flow on his account with key backup and when clicking start he was sent into a restore key backup flow but that s not intended the expected thing is you are prompted for the ssss passphrase and the device is verified
| 1
|
357,524
| 25,176,399,412
|
IssuesEvent
|
2022-11-11 09:38:46
|
rama-pang/pe
|
https://api.github.com/repos/rama-pang/pe
|
opened
|
Numbering error in Use Case
|
severity.VeryLow type.DocumentationBug
|
In Use Case UC6 - Sorting the tutor list, Extensions section, number 1b; the next command after this has number "1a1" instead of "1b1".
<!--session: 1668152096750-d2fd9f28-c252-473b-8fc4-fdb88cf3f718-->
<!--Version: Web v3.4.4-->
|
1.0
|
Numbering error in Use Case - In Use Case UC6 - Sorting the tutor list, Extensions section, number 1b; the next command after this has number "1a1" instead of "1b1".
<!--session: 1668152096750-d2fd9f28-c252-473b-8fc4-fdb88cf3f718-->
<!--Version: Web v3.4.4-->
|
non_defect
|
numbering error in use case in use case sorting the tutor list extensions section number the next command after this has number instead of
| 0
|
126
| 2,511,982,053
|
IssuesEvent
|
2015-01-14 13:11:56
|
enthought/chaco
|
https://api.github.com/repos/enthought/chaco
|
opened
|
cntr.c could use an update
|
defect
|
The code in `cntr.c` is outdated and a little buggy (as I discovered when I tried to pass the optional mask to the `Cntr` type). It appears to be a clone of [this code](https://github.com/matplotlib/matplotlib/blob/master/src/cntr.c) in matplotlib, which looks a little more up-to-date.
|
1.0
|
cntr.c could use an update - The code in `cntr.c` is outdated and a little buggy (as I discovered when I tried to pass the optional mask to the `Cntr` type). It appears to be a clone of [this code](https://github.com/matplotlib/matplotlib/blob/master/src/cntr.c) in matplotlib, which looks a little more up-to-date.
|
defect
|
cntr c could use an update the code in cntr c is outdated and a little buggy as i discovered when i tried to pass the optional mask to the cntr type it appears to be a clone of in matplotlib which looks a little more up to date
| 1
|
55,814
| 23,609,640,486
|
IssuesEvent
|
2022-08-24 11:16:52
|
microsoft/vscode-cpptools
|
https://api.github.com/repos/microsoft/vscode-cpptools
|
closed
|
Language service is idle, "Invalid opened file instance. Ignoring IntelliSense message for file" for multiroot workspace
|
bug Language Service more info needed Feature: Multiroot Not Repro - Internal
|
Bug type: Language Service
<!-- Prior to creating a bug report, please review:
๐ Existing issues at https://github.com/Microsoft/vscode-cpptools/issues
๐ Our documentation at https://code.visualstudio.com/docs/languages/cpp
๐ FAQs at https://code.visualstudio.com/docs/cpp/faq-cpp
-->
**Describe the bug**
- OS and Version: Windows 10 21H2
- VS Code Version: 1.66
- C/C++ Extension Version: 1.9.7
- Other extensions you installed (and if the issue persists after disabling them): C/C++ Extension pack v1.1.0 (without CMake-related extensions). Issue still persists.
- If using SSH remote, specify OS of remote machine: Windows Server 2019
- A clear and concise description of what the bug is, including information about the workspace (i.e. is the workspace a single project or multiple projects, size of the project, etc).
**Steps to reproduce**
1. Connect via SSH to remote Windows OS;
2. Add the directory that is adjacent to the `${workspaceFolder}` in step 4 (it should not have `.vscode` folder unlike the next two);
3. Add several sub-directories of a `${workspaceFolder}`;
4. Add the `${workspaceFolder}` itself;
5. Switch "C_Cpp: Logging Level" to "Debug"
6. Hovering over C++ code will reveal `Unsupported Markup content received. Kind is:`
7. In "C/C++" Output window there should be `Invalid opened file instance. Ignoring IntelliSense message for file X`
**Expected behavior**
IntelliSense is expected to work on hover.
**Code sample and logs**
<!-- Please provide code sample, your c_cpp_properties.json and logs. -->
- Configurations in `c_cpp_properties.json`
```json
{
"configurations": [
{
"name": "Win32",
"includePath": [
"${workspaceFolder}/subdir1/**",
"${workspaceFolder}/subdir2/**"
],
"defines": [
"_DEBUG",
"UNICODE",
"_UNICODE"
],
"cStandard": "c17",
"cppStandard": "c++17",
}
],
"version": 4
}
```
- Logs from running `C/C++: Log Diagnostics` from the VS Code command palette
```
-------- Diagnostics - 3/31/2022, 2:50:23 PM
Version: 1.9.7
Current Configuration:
{
"name": "Win32",
"includePath": [
"${workspaceFolder}/**",
"C:/Program Files (x86)/Intel/oneAPI/compiler/latest/windows/include/sycl"
],
"defines": [
"_DEBUG",
"UNICODE",
"_UNICODE"
],
"windowsSdkVersion": "10.0.17763.0",
"compilerPath": "cl.exe",
"compilerPathIsExplicit": true,
"cStandardIsExplicit": true,
"cppStandardIsExplicit": true,
"intelliSenseModeIsExplicit": false,
"compilerArgs": [],
"cStandard": "c17",
"cppStandard": "c++17",
"mergeConfigurations": false,
"browse": {
"path": [
"${workspaceFolder}/**",
"C:/Program Files (x86)/Intel/oneAPI/compiler/latest/windows/include/sycl"
],
"limitSymbolsToIncludedHeaders": true
}
}
No active translation units.
```
- Logs from [the language server logging](https://code.visualstudio.com/docs/cpp/enable-logging-cpp#_enable-logging-for-the-language-server)
```
loggingLevel: Debug
cpptools/didChangeCppProperties
Attempting to get defaults from C++ compiler in "compilerPath" property: 'cl.exe'
Attempting to get defaults from C compiler in "compilerPath" property: 'cl.exe'
Folder: C:/PROGRAM FILES (X86)/MICROSOFT VISUAL STUDIO/2017/PROFESSIONAL/VC/TOOLS/MSVC/14.16.27023/ATLMFC/INCLUDE/* will be indexed
Folder: C:/PROGRAM FILES (X86)/MICROSOFT VISUAL STUDIO/2017/PROFESSIONAL/VC/TOOLS/MSVC/14.16.27023/INCLUDE/* will be indexed
Folder: C:/PROGRAM FILES (X86)/WINDOWS KITS/10/INCLUDE/10.0.17763.0/CPPWINRT/ will be indexed
Folder: C:/PROGRAM FILES (X86)/WINDOWS KITS/10/INCLUDE/10.0.17763.0/SHARED/ will be indexed
Folder: C:/PROGRAM FILES (X86)/WINDOWS KITS/10/INCLUDE/10.0.17763.0/UCRT/ will be indexed
Folder: C:/PROGRAM FILES (X86)/WINDOWS KITS/10/INCLUDE/10.0.17763.0/UM/ will be indexed
Folder: C:/PROGRAM FILES (X86)/WINDOWS KITS/10/INCLUDE/10.0.17763.0/WINRT/ will be indexed
Folder: C:/<WORKSPACE FOLDER>/ will be indexed
Folder: C:/PROGRAM FILES (X86)/INTEL/ONEAPI/COMPILER/LATEST/WINDOWS/INCLUDE/SYCL/ will be indexed
textDocument/didOpen: <FILE1>
cpptools/textEditorSelectionChange
cpptools/activeDocumentChange: <FILE1>
cpptools/getDocumentSymbols: <FILE1> (id: 2)
cpptools/getCodeActions: <FILE1> (id: 3)
cpptools/getFoldingRanges: <FILE1> (id: 4)
cpptools/getSemanticTokens: <FILE1> (id: 5)
textDocument/hover: <FILE1> (id: 6)
Invalid opened file instance. Ignoring IntelliSense message for file <FILE1>.
cpptools/getDiagnostics (id: 10)
```
**Additional context**
This bug might be related to #8762, so steps to reproduce should be similar.
**Possible workarounds**
- exclude step 2;
|
1.0
|
Language service is idle, "Invalid opened file instance. Ignoring IntelliSense message for file" for multiroot workspace - Bug type: Language Service
<!-- Prior to creating a bug report, please review:
๐ Existing issues at https://github.com/Microsoft/vscode-cpptools/issues
๐ Our documentation at https://code.visualstudio.com/docs/languages/cpp
๐ FAQs at https://code.visualstudio.com/docs/cpp/faq-cpp
-->
**Describe the bug**
- OS and Version: Windows 10 21H2
- VS Code Version: 1.66
- C/C++ Extension Version: 1.9.7
- Other extensions you installed (and if the issue persists after disabling them): C/C++ Extension pack v1.1.0 (without CMake-related extensions). Issue still persists.
- If using SSH remote, specify OS of remote machine: Windows Server 2019
- A clear and concise description of what the bug is, including information about the workspace (i.e. is the workspace a single project or multiple projects, size of the project, etc).
**Steps to reproduce**
1. Connect via SSH to remote Windows OS;
2. Add the directory that is adjacent to the `${workspaceFolder}` in step 4 (it should not have `.vscode` folder unlike the next two);
3. Add several sub-directories of a `${workspaceFolder}`;
4. Add the `${workspaceFolder}` itself;
5. Switch "C_Cpp: Logging Level" to "Debug"
6. Hovering over C++ code will reveal `Unsupported Markup content received. Kind is:`
7. In "C/C++" Output window there should be `Invalid opened file instance. Ignoring IntelliSense message for file X`
**Expected behavior**
IntelliSense is expected to work on hover.
**Code sample and logs**
<!-- Please provide code sample, your c_cpp_properties.json and logs. -->
- Configurations in `c_cpp_properties.json`
```json
{
"configurations": [
{
"name": "Win32",
"includePath": [
"${workspaceFolder}/subdir1/**",
"${workspaceFolder}/subdir2/**"
],
"defines": [
"_DEBUG",
"UNICODE",
"_UNICODE"
],
"cStandard": "c17",
"cppStandard": "c++17",
}
],
"version": 4
}
```
- Logs from running `C/C++: Log Diagnostics` from the VS Code command palette
```
-------- Diagnostics - 3/31/2022, 2:50:23 PM
Version: 1.9.7
Current Configuration:
{
"name": "Win32",
"includePath": [
"${workspaceFolder}/**",
"C:/Program Files (x86)/Intel/oneAPI/compiler/latest/windows/include/sycl"
],
"defines": [
"_DEBUG",
"UNICODE",
"_UNICODE"
],
"windowsSdkVersion": "10.0.17763.0",
"compilerPath": "cl.exe",
"compilerPathIsExplicit": true,
"cStandardIsExplicit": true,
"cppStandardIsExplicit": true,
"intelliSenseModeIsExplicit": false,
"compilerArgs": [],
"cStandard": "c17",
"cppStandard": "c++17",
"mergeConfigurations": false,
"browse": {
"path": [
"${workspaceFolder}/**",
"C:/Program Files (x86)/Intel/oneAPI/compiler/latest/windows/include/sycl"
],
"limitSymbolsToIncludedHeaders": true
}
}
No active translation units.
```
- Logs from [the language server logging](https://code.visualstudio.com/docs/cpp/enable-logging-cpp#_enable-logging-for-the-language-server)
```
loggingLevel: Debug
cpptools/didChangeCppProperties
Attempting to get defaults from C++ compiler in "compilerPath" property: 'cl.exe'
Attempting to get defaults from C compiler in "compilerPath" property: 'cl.exe'
Folder: C:/PROGRAM FILES (X86)/MICROSOFT VISUAL STUDIO/2017/PROFESSIONAL/VC/TOOLS/MSVC/14.16.27023/ATLMFC/INCLUDE/* will be indexed
Folder: C:/PROGRAM FILES (X86)/MICROSOFT VISUAL STUDIO/2017/PROFESSIONAL/VC/TOOLS/MSVC/14.16.27023/INCLUDE/* will be indexed
Folder: C:/PROGRAM FILES (X86)/WINDOWS KITS/10/INCLUDE/10.0.17763.0/CPPWINRT/ will be indexed
Folder: C:/PROGRAM FILES (X86)/WINDOWS KITS/10/INCLUDE/10.0.17763.0/SHARED/ will be indexed
Folder: C:/PROGRAM FILES (X86)/WINDOWS KITS/10/INCLUDE/10.0.17763.0/UCRT/ will be indexed
Folder: C:/PROGRAM FILES (X86)/WINDOWS KITS/10/INCLUDE/10.0.17763.0/UM/ will be indexed
Folder: C:/PROGRAM FILES (X86)/WINDOWS KITS/10/INCLUDE/10.0.17763.0/WINRT/ will be indexed
Folder: C:/<WORKSPACE FOLDER>/ will be indexed
Folder: C:/PROGRAM FILES (X86)/INTEL/ONEAPI/COMPILER/LATEST/WINDOWS/INCLUDE/SYCL/ will be indexed
textDocument/didOpen: <FILE1>
cpptools/textEditorSelectionChange
cpptools/activeDocumentChange: <FILE1>
cpptools/getDocumentSymbols: <FILE1> (id: 2)
cpptools/getCodeActions: <FILE1> (id: 3)
cpptools/getFoldingRanges: <FILE1> (id: 4)
cpptools/getSemanticTokens: <FILE1> (id: 5)
textDocument/hover: <FILE1> (id: 6)
Invalid opened file instance. Ignoring IntelliSense message for file <FILE1>.
cpptools/getDiagnostics (id: 10)
```
**Additional context**
This bug might be related to #8762, so steps to reproduce should be similar.
**Possible workarounds**
- exclude step 2;
|
non_defect
|
language service is idle invalid opened file instance ignoring intellisense message for file for multiroot workspace bug type language service prior to creating a bug report please review ๐ existing issues at ๐ our documentation at ๐ faqs at describe the bug os and version windows vs code version c c extension version other extensions you installed and if the issue persists after disabling them c c extension pack without cmake related extensions issue still persists if using ssh remote specify os of remote machine windows server a clear and concise description of what the bug is including information about the workspace i e is the workspace a single project or multiple projects size of the project etc steps to reproduce connect via ssh to remote windows os add the directory that is adjacent to the workspacefolder in step it should not have vscode folder unlike the next two add several sub directories of a workspacefolder add the workspacefolder itself switch c cpp logging level to debug hovering over c code will reveal unsupported markup content received kind is in c c output window there should be invalid opened file instance ignoring intellisense message for file x expected behavior intellisense is expected to work on hover code sample and logs configurations in c cpp properties json json configurations name includepath workspacefolder workspacefolder defines debug unicode unicode cstandard cppstandard c version logs from running c c log diagnostics from the vs code command palette diagnostics pm version current configuration name includepath workspacefolder c program files intel oneapi compiler latest windows include sycl defines debug unicode unicode windowssdkversion compilerpath cl exe compilerpathisexplicit true cstandardisexplicit true cppstandardisexplicit true intellisensemodeisexplicit false compilerargs cstandard cppstandard c mergeconfigurations false browse path workspacefolder c program files intel oneapi compiler latest windows include sycl limitsymbolstoincludedheaders true no active translation units logs from logginglevel debug cpptools didchangecppproperties attempting to get defaults from c compiler in compilerpath property cl exe attempting to get defaults from c compiler in compilerpath property cl exe folder c program files microsoft visual studio professional vc tools msvc atlmfc include will be indexed folder c program files microsoft visual studio professional vc tools msvc include will be indexed folder c program files windows kits include cppwinrt will be indexed folder c program files windows kits include shared will be indexed folder c program files windows kits include ucrt will be indexed folder c program files windows kits include um will be indexed folder c program files windows kits include winrt will be indexed folder c will be indexed folder c program files intel oneapi compiler latest windows include sycl will be indexed textdocument didopen cpptools texteditorselectionchange cpptools activedocumentchange cpptools getdocumentsymbols id cpptools getcodeactions id cpptools getfoldingranges id cpptools getsemantictokens id textdocument hover id invalid opened file instance ignoring intellisense message for file cpptools getdiagnostics id additional context this bug might be related to so steps to reproduce should be similar possible workarounds exclude step
| 0
|
309,736
| 23,304,228,421
|
IssuesEvent
|
2022-08-07 19:34:44
|
bitwes/Gut
|
https://api.github.com/repos/bitwes/Gut
|
closed
|
In the wiki, anchors for assert_typeof and assert_nottypeof are wrong
|
documentation
|
They link to https://github.com/bitwes/Gut/wiki/%24assert_typeof instead of https://github.com/bitwes/Gut/wiki/Asserts-and-Methods#assert_typeof
|
1.0
|
In the wiki, anchors for assert_typeof and assert_nottypeof are wrong - They link to https://github.com/bitwes/Gut/wiki/%24assert_typeof instead of https://github.com/bitwes/Gut/wiki/Asserts-and-Methods#assert_typeof
|
non_defect
|
in the wiki anchors for assert typeof and assert nottypeof are wrong they link to instead of
| 0
|
16,836
| 2,948,324,068
|
IssuesEvent
|
2015-07-06 01:32:22
|
Winetricks/winetricks
|
https://api.github.com/repos/Winetricks/winetricks
|
closed
|
New Steam Verb: "Magic: The Gathering - Duels of the Planeswalkers 2012"
|
auto-migrated Priority-Medium Type-Defect
|
```
Attached verb will install this game through Steam as per the instructions in
AddingSteamGames, with no workarounds used.
As of now however, sound does not work in-game, but this is an upstream bug in
Wine 1.3.29 itself. There is also occasional crashes, but again I couldn't find
a workaround in AppDB. (I'll report these when I get an opportunity to do so)
Tested on Mac OS X 10.7.1 (Lion) under Wine 1.3.29 and Winetricks revision
20110629.
```
Original issue reported on code.google.com by `aersunst...@googlemail.com` on 9 Oct 2011 at 11:35
Attachments:
* [mtg_dotp2012_steam.verb](https://storage.googleapis.com/google-code-attachments/winetricks/issue-135/comment-0/mtg_dotp2012_steam.verb)
|
1.0
|
New Steam Verb: "Magic: The Gathering - Duels of the Planeswalkers 2012" - ```
Attached verb will install this game through Steam as per the instructions in
AddingSteamGames, with no workarounds used.
As of now however, sound does not work in-game, but this is an upstream bug in
Wine 1.3.29 itself. There is also occasional crashes, but again I couldn't find
a workaround in AppDB. (I'll report these when I get an opportunity to do so)
Tested on Mac OS X 10.7.1 (Lion) under Wine 1.3.29 and Winetricks revision
20110629.
```
Original issue reported on code.google.com by `aersunst...@googlemail.com` on 9 Oct 2011 at 11:35
Attachments:
* [mtg_dotp2012_steam.verb](https://storage.googleapis.com/google-code-attachments/winetricks/issue-135/comment-0/mtg_dotp2012_steam.verb)
|
defect
|
new steam verb magic the gathering duels of the planeswalkers attached verb will install this game through steam as per the instructions in addingsteamgames with no workarounds used as of now however sound does not work in game but this is an upstream bug in wine itself there is also occasional crashes but again i couldn t find a workaround in appdb i ll report these when i get an opportunity to do so tested on mac os x lion under wine and winetricks revision original issue reported on code google com by aersunst googlemail com on oct at attachments
| 1
|
214,790
| 7,276,786,175
|
IssuesEvent
|
2018-02-21 17:21:12
|
TylerConlee/slab
|
https://api.github.com/repos/TylerConlee/slab
|
closed
|
Display CSAT as part of ticket notification
|
enhancement priority:normal
|
On the More Info screen, previous ticket CSATs can be displayed to give some context as to the attitude of the customer towards the support given. With a history of how the customer responds, an agent can then tailor the messaging even further to ensure the customer gets the best service possible.
|
1.0
|
Display CSAT as part of ticket notification - On the More Info screen, previous ticket CSATs can be displayed to give some context as to the attitude of the customer towards the support given. With a history of how the customer responds, an agent can then tailor the messaging even further to ensure the customer gets the best service possible.
|
non_defect
|
display csat as part of ticket notification on the more info screen previous ticket csats can be displayed to give some context as to the attitude of the customer towards the support given with a history of how the customer responds an agent can then tailor the messaging even further to ensure the customer gets the best service possible
| 0
|
31,148
| 4,691,651,642
|
IssuesEvent
|
2016-10-11 11:22:56
|
skaut/SRS
|
https://api.github.com/repos/skaut/SRS
|
closed
|
Hromadnรฉ pลiลazovรกnรญ rolรญ
|
Testovรกnรญ
|
Tak jako se nynรญ dรก hromadnฤ uลพivatelลฏm zadat platba a tisk dokladลฏ v evidenci รบฤastnรญkลฏ, prosรญm o รบpravu, aby se hromadnฤ daly zadรกvat i role.
|
1.0
|
Hromadnรฉ pลiลazovรกnรญ rolรญ - Tak jako se nynรญ dรก hromadnฤ uลพivatelลฏm zadat platba a tisk dokladลฏ v evidenci รบฤastnรญkลฏ, prosรญm o รบpravu, aby se hromadnฤ daly zadรกvat i role.
|
non_defect
|
hromadnรฉ pลiลazovรกnรญ rolรญ tak jako se nynรญ dรก hromadnฤ uลพivatelลฏm zadat platba a tisk dokladลฏ v evidenci รบฤastnรญkลฏ prosรญm o รบpravu aby se hromadnฤ daly zadรกvat i role
| 0
|
121,867
| 4,822,453,788
|
IssuesEvent
|
2016-11-05 21:22:28
|
ReactiveX/rxjs
|
https://api.github.com/repos/ReactiveX/rxjs
|
closed
|
Materialize: Notification property incorrectly is `exception` rather than `error` for `kind: "E"`
|
priority: critical type: bug
|
<!--
Thank you for raising your concerns, we appreciate your feedback and contributions to this repository.
Before you continue, consider the following:
If you have a "How do I do ...?" question, it is better for you and for us that this question is placed in [StackOverflow](http://stackoverflow.com/questions/tagged/rxjs5) or some chat channel. This way, you are making it easier for others to learn from your experiences too.
These "Issues" are meant only for technical problems, bugs, and proposals related to the library.
If your issue is a bug, please follow the format below:
-->
**RxJS version:** 5.0.0-rc.1
**Code to reproduce:**
```js
Rx.Observable.throw(new Error('sad'))
.materialize()
.map(n => n.error.message)
.subscribe(x => console.log(x))
```
**Expected behavior:**
logs `"sad"`
**Actual behavior:**
Throws cannot read "message" of undefined
**Additional information:**
Currently `error` is in the `exception` field.
Required to upgrade falcor-router to Rx5
|
1.0
|
Materialize: Notification property incorrectly is `exception` rather than `error` for `kind: "E"` - <!--
Thank you for raising your concerns, we appreciate your feedback and contributions to this repository.
Before you continue, consider the following:
If you have a "How do I do ...?" question, it is better for you and for us that this question is placed in [StackOverflow](http://stackoverflow.com/questions/tagged/rxjs5) or some chat channel. This way, you are making it easier for others to learn from your experiences too.
These "Issues" are meant only for technical problems, bugs, and proposals related to the library.
If your issue is a bug, please follow the format below:
-->
**RxJS version:** 5.0.0-rc.1
**Code to reproduce:**
```js
Rx.Observable.throw(new Error('sad'))
.materialize()
.map(n => n.error.message)
.subscribe(x => console.log(x))
```
**Expected behavior:**
logs `"sad"`
**Actual behavior:**
Throws cannot read "message" of undefined
**Additional information:**
Currently `error` is in the `exception` field.
Required to upgrade falcor-router to Rx5
|
non_defect
|
materialize notification property incorrectly is exception rather than error for kind e thank you for raising your concerns we appreciate your feedback and contributions to this repository before you continue consider the following if you have a how do i do question it is better for you and for us that this question is placed in or some chat channel this way you are making it easier for others to learn from your experiences too these issues are meant only for technical problems bugs and proposals related to the library if your issue is a bug please follow the format below rxjs version rc code to reproduce js rx observable throw new error sad materialize map n n error message subscribe x console log x expected behavior logs sad actual behavior throws cannot read message of undefined additional information currently error is in the exception field required to upgrade falcor router to
| 0
|
33,304
| 27,370,658,707
|
IssuesEvent
|
2023-02-27 23:10:10
|
pypa/packaging
|
https://api.github.com/repos/pypa/packaging
|
closed
|
Set up permissions to github workflows
|
infrastructure
|
I have suggested a PR to setuptools repo to harden permissions to the github workflows https://github.com/pypa/setuptools/pull/3833, if that's ok, I also would like to do the same to the packaging repo in order to increase supply-chain security by limiting the impact of an eventual compromised workflow.
Let me know if a PR is also welcome in this repo and I'll submit it ASAP.
|
1.0
|
Set up permissions to github workflows - I have suggested a PR to setuptools repo to harden permissions to the github workflows https://github.com/pypa/setuptools/pull/3833, if that's ok, I also would like to do the same to the packaging repo in order to increase supply-chain security by limiting the impact of an eventual compromised workflow.
Let me know if a PR is also welcome in this repo and I'll submit it ASAP.
|
non_defect
|
set up permissions to github workflows i have suggested a pr to setuptools repo to harden permissions to the github workflows if that s ok i also would like to do the same to the packaging repo in order to increase supply chain security by limiting the impact of an eventual compromised workflow let me know if a pr is also welcome in this repo and i ll submit it asap
| 0
|
64,420
| 18,670,712,558
|
IssuesEvent
|
2021-10-30 16:55:27
|
theseion/Fuel
|
https://api.github.com/repos/theseion/Fuel
|
closed
|
Fully implement DiskProxy
|
Priority-Medium auto-migrated Type-Defect stale
|
```
Fully implement DiskProxy (with any object as argument). it could be necessary
to change our format to "headerStep [instStep refStep]+ trailerStep"
```
Original issue reported on code.google.com by `marianopeck` on 1 Dec 2011 at 12:17
|
1.0
|
Fully implement DiskProxy - ```
Fully implement DiskProxy (with any object as argument). it could be necessary
to change our format to "headerStep [instStep refStep]+ trailerStep"
```
Original issue reported on code.google.com by `marianopeck` on 1 Dec 2011 at 12:17
|
defect
|
fully implement diskproxy fully implement diskproxy with any object as argument it could be necessary to change our format to headerstep trailerstep original issue reported on code google com by marianopeck on dec at
| 1
|
346,011
| 10,383,015,832
|
IssuesEvent
|
2019-09-10 08:46:50
|
acidanthera/bugtracker
|
https://api.github.com/repos/acidanthera/bugtracker
|
closed
|
ART Frequency, CPU Frequency and Clock Drift
|
priority:high project:oc
|
OCCPU is misreporting my CPU's frequency. I believe it's also causing the macOS system clock to run fast:
```
00:226 00:113 OCCPU: Found Intel(R) Xeon(R) Gold 6136 CPU @ 3.00GHz
00:345 00:118 OCCPU: Signature 50654 Stepping 4 Model 55 Family 6 Type 0 ExtModel 5 ExtFamily 0
00:461 00:116 OCCPU: Detected Apple Processor Type: 0F -> 0F01
00:575 00:114 OCCPU: Ratio Min 12 Max 30 Current 36 Turbo 37 36 36 36
00:695 00:119 OCCPU: ART Frequency 2880000000 2880MHz 24000000 * 240 / 2 = 2880000000
00:821 00:126 OCCPU: TSC Frequency 2992972839 2992MHz
00:944 00:122 OCCPU: CPU Frequency 2880000000 2880MHz
01:068 00:123 OCCPU: FSB Frequency 96000000 96MHz
01:206 00:138 OCCPU: Pkg 1 Cores 12 Threads 24
```
Looking at `OcCpu.c`, the following lines look suspect to me:
```
AsmCpuid (CPUID_TIME_STAMP_COUNTER, &CpuidEax, &CpuidEbx, NULL, NULL);
if (CpuidEax > 0 && CpuidEbx > 0) {
Cpu->CPUFrequency = MultU64x32 (BASE_ART_CLOCK_SOURCE, (UINT32) DivU64x32 (CpuidEbx, CpuidEax));
```
https://github.com/acidanthera/OcSupportPkg/blob/3e98de1f89a9e946f026ecc24e4856e2096dc338/Library/OcCpuLib/OcCpuLib.c#L667
Should that be storing to `Cpu->ARTFrequency` instead? I tested by making this change and setting `Cpu->CPUFrequency` from `Cpu->TSCFrequency`. `About this Mac` now reports the correct frequency (3Ghz instead of 2.88Ghz). Also, the clock no longer drifts whereas it used to run a few seconds per minute faster than real-time.
|
1.0
|
ART Frequency, CPU Frequency and Clock Drift - OCCPU is misreporting my CPU's frequency. I believe it's also causing the macOS system clock to run fast:
```
00:226 00:113 OCCPU: Found Intel(R) Xeon(R) Gold 6136 CPU @ 3.00GHz
00:345 00:118 OCCPU: Signature 50654 Stepping 4 Model 55 Family 6 Type 0 ExtModel 5 ExtFamily 0
00:461 00:116 OCCPU: Detected Apple Processor Type: 0F -> 0F01
00:575 00:114 OCCPU: Ratio Min 12 Max 30 Current 36 Turbo 37 36 36 36
00:695 00:119 OCCPU: ART Frequency 2880000000 2880MHz 24000000 * 240 / 2 = 2880000000
00:821 00:126 OCCPU: TSC Frequency 2992972839 2992MHz
00:944 00:122 OCCPU: CPU Frequency 2880000000 2880MHz
01:068 00:123 OCCPU: FSB Frequency 96000000 96MHz
01:206 00:138 OCCPU: Pkg 1 Cores 12 Threads 24
```
Looking at `OcCpu.c`, the following lines look suspect to me:
```
AsmCpuid (CPUID_TIME_STAMP_COUNTER, &CpuidEax, &CpuidEbx, NULL, NULL);
if (CpuidEax > 0 && CpuidEbx > 0) {
Cpu->CPUFrequency = MultU64x32 (BASE_ART_CLOCK_SOURCE, (UINT32) DivU64x32 (CpuidEbx, CpuidEax));
```
https://github.com/acidanthera/OcSupportPkg/blob/3e98de1f89a9e946f026ecc24e4856e2096dc338/Library/OcCpuLib/OcCpuLib.c#L667
Should that be storing to `Cpu->ARTFrequency` instead? I tested by making this change and setting `Cpu->CPUFrequency` from `Cpu->TSCFrequency`. `About this Mac` now reports the correct frequency (3Ghz instead of 2.88Ghz). Also, the clock no longer drifts whereas it used to run a few seconds per minute faster than real-time.
|
non_defect
|
art frequency cpu frequency and clock drift occpu is misreporting my cpu s frequency i believe it s also causing the macos system clock to run fast occpu found intel r xeon r gold cpu occpu signature stepping model family type extmodel extfamily occpu detected apple processor type occpu ratio min max current turbo occpu art frequency occpu tsc frequency occpu cpu frequency occpu fsb frequency occpu pkg cores threads looking at occpu c the following lines look suspect to me asmcpuid cpuid time stamp counter cpuideax cpuidebx null null if cpuideax cpuidebx cpu cpufrequency base art clock source cpuidebx cpuideax should that be storing to cpu artfrequency instead i tested by making this change and setting cpu cpufrequency from cpu tscfrequency about this mac now reports the correct frequency instead of also the clock no longer drifts whereas it used to run a few seconds per minute faster than real time
| 0
|
31,669
| 6,583,368,914
|
IssuesEvent
|
2017-09-13 05:15:18
|
kuenzign/dyknow-panel-extractor
|
https://api.github.com/repos/kuenzign/dyknow-panel-extractor
|
closed
|
Links in the about box do not load a web page
|
auto-migrated Priority-Medium Project-DPXCommon Type-Defect
|
_From @GoogleCodeExporter on March 15, 2015 13:14_
```
When you click on a link in the about box, it does not load the intended web
page in the default browser.
```
Original issue reported on code.google.com by `jjhatf02` on 28 Apr 2010 at 3:32
_Copied from original issue: JaredHatfield/dyknow-panel-extractor#11_
|
1.0
|
Links in the about box do not load a web page - _From @GoogleCodeExporter on March 15, 2015 13:14_
```
When you click on a link in the about box, it does not load the intended web
page in the default browser.
```
Original issue reported on code.google.com by `jjhatf02` on 28 Apr 2010 at 3:32
_Copied from original issue: JaredHatfield/dyknow-panel-extractor#11_
|
defect
|
links in the about box do not load a web page from googlecodeexporter on march when you click on a link in the about box it does not load the intended web page in the default browser original issue reported on code google com by on apr at copied from original issue jaredhatfield dyknow panel extractor
| 1
|
52,447
| 10,863,340,405
|
IssuesEvent
|
2019-11-14 14:57:50
|
USGCRP/gcis
|
https://api.github.com/repos/USGCRP/gcis
|
closed
|
Error building GCIS under Perl 5.22+
|
a bug context Core Code context IT priority low type technical
|
```
Can't load 'blib/arch/auto/Algorithm/Permute/Permute.so' for module Algorithm::Permute: blib/arch/auto/Algorithm/Permute/Permute.so: undefined symbol: PUSHBLOCK at /usr/local/lib/perl5/5.24.0/x86_64-linux/DynaLoader.pm line 193.
at test.pl line 13.
Compilation failed in require at test.pl line 13.
BEGIN failed--compilation aborted at test.pl line 13.
```
Possibly related to using Perl 5.24. See [this issue](https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=825012).
|
1.0
|
Error building GCIS under Perl 5.22+ - ```
Can't load 'blib/arch/auto/Algorithm/Permute/Permute.so' for module Algorithm::Permute: blib/arch/auto/Algorithm/Permute/Permute.so: undefined symbol: PUSHBLOCK at /usr/local/lib/perl5/5.24.0/x86_64-linux/DynaLoader.pm line 193.
at test.pl line 13.
Compilation failed in require at test.pl line 13.
BEGIN failed--compilation aborted at test.pl line 13.
```
Possibly related to using Perl 5.24. See [this issue](https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=825012).
|
non_defect
|
error building gcis under perl can t load blib arch auto algorithm permute permute so for module algorithm permute blib arch auto algorithm permute permute so undefined symbol pushblock at usr local lib linux dynaloader pm line at test pl line compilation failed in require at test pl line begin failed compilation aborted at test pl line possibly related to using perl see
| 0
|
20,690
| 3,407,672,115
|
IssuesEvent
|
2015-12-04 05:03:12
|
jarrodek/ChromeRestClient
|
https://api.github.com/repos/jarrodek/ChromeRestClient
|
closed
|
Unable to install the rest client
|
OpSys-Windows Priority-High Status-New Type-Defect
|
Originally reported on Google Code with ID 296
```
What steps will reproduce the problem?
1. Access below url
https://chrome.google.com/webstore/detail/advanced-rest-client/hgmloofddffdnphfgcellkdfbfbjeloo
2.click on FREE and proceed with installation
3.Installation fails with error "An error has occurred
Advanced REST client (extension ID "hgmloofddffdnphfgcellkdfbfbjeloo") is blocked by
the administrator."
What is the expected output? What do you see instead?
Installation of Advanced Rest client
On what operating system, browser and browser version?
Windows 2007, 64 bit, Chrome Version 37.0.2062.102 m
Please provide any additional information below.
The error is "An error has occurred
Advanced REST client (extension ID "hgmloofddffdnphfgcellkdfbfbjeloo") is blocked by
the administrator."
```
Reported by `VishalVShroff` on 2014-12-29 09:34:58
|
1.0
|
Unable to install the rest client - Originally reported on Google Code with ID 296
```
What steps will reproduce the problem?
1. Access below url
https://chrome.google.com/webstore/detail/advanced-rest-client/hgmloofddffdnphfgcellkdfbfbjeloo
2.click on FREE and proceed with installation
3.Installation fails with error "An error has occurred
Advanced REST client (extension ID "hgmloofddffdnphfgcellkdfbfbjeloo") is blocked by
the administrator."
What is the expected output? What do you see instead?
Installation of Advanced Rest client
On what operating system, browser and browser version?
Windows 2007, 64 bit, Chrome Version 37.0.2062.102 m
Please provide any additional information below.
The error is "An error has occurred
Advanced REST client (extension ID "hgmloofddffdnphfgcellkdfbfbjeloo") is blocked by
the administrator."
```
Reported by `VishalVShroff` on 2014-12-29 09:34:58
|
defect
|
unable to install the rest client originally reported on google code with id what steps will reproduce the problem access below url click on free and proceed with installation installation fails with error an error has occurred advanced rest client extension id hgmloofddffdnphfgcellkdfbfbjeloo is blocked by the administrator what is the expected output what do you see instead installation of advanced rest client on what operating system browser and browser version windows bit chrome version m please provide any additional information below the error is an error has occurred advanced rest client extension id hgmloofddffdnphfgcellkdfbfbjeloo is blocked by the administrator reported by vishalvshroff on
| 1
|
40,446
| 9,998,937,114
|
IssuesEvent
|
2019-07-12 09:24:31
|
contao/contao
|
https://api.github.com/repos/contao/contao
|
closed
|
new version of leafo/scssphp
|
defect
|
**Affected version(s)**
all 4.4 and above
**Description**
there is a new version for leafo/scssphp v0.8.2 which fixes some major issues...
could you please update the requirements
|
1.0
|
new version of leafo/scssphp - **Affected version(s)**
all 4.4 and above
**Description**
there is a new version for leafo/scssphp v0.8.2 which fixes some major issues...
could you please update the requirements
|
defect
|
new version of leafo scssphp affected version s all and above description there is a new version for leafo scssphp which fixes some major issues could you please update the requirements
| 1
|
374,756
| 26,130,450,486
|
IssuesEvent
|
2022-12-29 03:36:25
|
CameronRedmore/memory-deck
|
https://api.github.com/repos/CameronRedmore/memory-deck
|
opened
|
Create wiki pages
|
documentation enhancement
|
- [ ] How the plugin works
- [ ] List dependencies
- [ ] Instructions on how to run main.py manually for debugging purposes, potentially enable that functionality by default
|
1.0
|
Create wiki pages - - [ ] How the plugin works
- [ ] List dependencies
- [ ] Instructions on how to run main.py manually for debugging purposes, potentially enable that functionality by default
|
non_defect
|
create wiki pages how the plugin works list dependencies instructions on how to run main py manually for debugging purposes potentially enable that functionality by default
| 0
|
76,788
| 26,598,842,628
|
IssuesEvent
|
2023-01-23 14:26:12
|
vector-im/element-ios
|
https://api.github.com/repos/vector-im/element-ios
|
closed
|
1 translation error was found.
|
T-Defect A-Timeline S-Major O-Frequent
|
### Steps to reproduce
n/a
### Outcome

**The target language is Simplified Chinese, for Traditional Chinese, the translation here is correct.**
### Your phone model
iPhone 7 Plus
### Operating system version
iOS 15.6.1
### Application version
Element iOS 1.9.14
### Homeserver
Synapse 1.74.0
### Will you send logs?
No
|
1.0
|
1 translation error was found. - ### Steps to reproduce
n/a
### Outcome

**The target language is Simplified Chinese, for Traditional Chinese, the translation here is correct.**
### Your phone model
iPhone 7 Plus
### Operating system version
iOS 15.6.1
### Application version
Element iOS 1.9.14
### Homeserver
Synapse 1.74.0
### Will you send logs?
No
|
defect
|
translation error was found steps to reproduce n a outcome the target language is simplified chinese for traditional chinese the translation here is correct your phone model iphone plus operating system version ios application version element ios homeserver synapse will you send logs no
| 1
|
811,963
| 30,308,066,395
|
IssuesEvent
|
2023-07-10 10:52:20
|
GoogleCloudPlatform/python-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
|
closed
|
securitycenter.snippets.snippets_mute_config_test: test_bulk_mute_findings failed
|
priority: p2 type: bug api: securitycenter samples flakybot: issue flakybot: flaky
|
This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: d07634990df61b990c3a6908ce1502c2a6d69752
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/0f52d7e2-8f21-43d7-8582-2c6e0c4df6be), [Sponge](http://sponge2/0f52d7e2-8f21-43d7-8582-2c6e0c4df6be)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 72, in error_remapped_callable
return callable_(*args, **kwargs)
File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 1030, in __call__
return _end_unary_response_blocking(state, call, False, None)
File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 910, in _end_unary_response_blocking
raise _InactiveRpcError(state) # pytype: disable=not-instantiable
grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
status = StatusCode.INTERNAL
details = "Internal error encountered."
debug_error_string = "UNKNOWN:Error received from peer ipv4:74.125.195.95:443 {created_time:"2023-06-27T11:09:30.860723791+00:00", grpc_status:13, grpc_message:"Internal error encountered."}"
>
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/workspace/securitycenter/snippets/snippets_mute_config_test.py", line 121, in test_bulk_mute_findings
f"projects/{PROJECT_ID}", f'resource.project_display_name="{PROJECT_ID}"'
File "/workspace/securitycenter/snippets/snippets_mute_config.py", line 260, in bulk_mute_findings
response = client.bulk_mute_findings(request)
File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/cloud/securitycenter_v1/services/security_center/client.py", line 875, in bulk_mute_findings
metadata=metadata,
File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/gapic_v1/method.py", line 113, in __call__
return wrapped_func(*args, **kwargs)
File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 74, in error_remapped_callable
raise exceptions.from_grpc_error(exc) from exc
google.api_core.exceptions.InternalServerError: 500 Internal error encountered.</pre></details>
|
1.0
|
securitycenter.snippets.snippets_mute_config_test: test_bulk_mute_findings failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: d07634990df61b990c3a6908ce1502c2a6d69752
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/0f52d7e2-8f21-43d7-8582-2c6e0c4df6be), [Sponge](http://sponge2/0f52d7e2-8f21-43d7-8582-2c6e0c4df6be)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 72, in error_remapped_callable
return callable_(*args, **kwargs)
File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 1030, in __call__
return _end_unary_response_blocking(state, call, False, None)
File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 910, in _end_unary_response_blocking
raise _InactiveRpcError(state) # pytype: disable=not-instantiable
grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
status = StatusCode.INTERNAL
details = "Internal error encountered."
debug_error_string = "UNKNOWN:Error received from peer ipv4:74.125.195.95:443 {created_time:"2023-06-27T11:09:30.860723791+00:00", grpc_status:13, grpc_message:"Internal error encountered."}"
>
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/workspace/securitycenter/snippets/snippets_mute_config_test.py", line 121, in test_bulk_mute_findings
f"projects/{PROJECT_ID}", f'resource.project_display_name="{PROJECT_ID}"'
File "/workspace/securitycenter/snippets/snippets_mute_config.py", line 260, in bulk_mute_findings
response = client.bulk_mute_findings(request)
File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/cloud/securitycenter_v1/services/security_center/client.py", line 875, in bulk_mute_findings
metadata=metadata,
File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/gapic_v1/method.py", line 113, in __call__
return wrapped_func(*args, **kwargs)
File "/workspace/securitycenter/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 74, in error_remapped_callable
raise exceptions.from_grpc_error(exc) from exc
google.api_core.exceptions.InternalServerError: 500 Internal error encountered.</pre></details>
|
non_defect
|
securitycenter snippets snippets mute config test test bulk mute findings failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output traceback most recent call last file workspace securitycenter snippets nox py lib site packages google api core grpc helpers py line in error remapped callable return callable args kwargs file workspace securitycenter snippets nox py lib site packages grpc channel py line in call return end unary response blocking state call false none file workspace securitycenter snippets nox py lib site packages grpc channel py line in end unary response blocking raise inactiverpcerror state pytype disable not instantiable grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with status statuscode internal details internal error encountered debug error string unknown error received from peer created time grpc status grpc message internal error encountered the above exception was the direct cause of the following exception traceback most recent call last file workspace securitycenter snippets snippets mute config test py line in test bulk mute findings f projects project id f resource project display name project id file workspace securitycenter snippets snippets mute config py line in bulk mute findings response client bulk mute findings request file workspace securitycenter snippets nox py lib site packages google cloud securitycenter services security center client py line in bulk mute findings metadata metadata file workspace securitycenter snippets nox py lib site packages google api core gapic method py line in call return wrapped func args kwargs file workspace securitycenter snippets nox py lib site packages google api core grpc helpers py line in error remapped callable raise exceptions from grpc error exc from exc google api core exceptions internalservererror internal error encountered
| 0
|
55,852
| 8,032,543,746
|
IssuesEvent
|
2018-07-28 16:38:53
|
nf-core/cookiecutter
|
https://api.github.com/repos/nf-core/cookiecutter
|
opened
|
Documentation redundancy
|
documentation
|
The files `local.md` and `adding_your_own.md` have similar (if not identical) sections for Docker and Singularity - we should remove these in one location at least :-)
|
1.0
|
Documentation redundancy - The files `local.md` and `adding_your_own.md` have similar (if not identical) sections for Docker and Singularity - we should remove these in one location at least :-)
|
non_defect
|
documentation redundancy the files local md and adding your own md have similar if not identical sections for docker and singularity we should remove these in one location at least
| 0
|
45,197
| 2,920,859,454
|
IssuesEvent
|
2015-06-24 21:08:28
|
vickychijwani/quill
|
https://api.github.com/repos/vickychijwani/quill
|
reopened
|
Memory leak in WebViewFragment
|
bug help wanted high-priority
|
Steps to reproduce:
1. Open a post preview (the new WebView-based one)
2. Select some text
3. Tap back twice to come back to the post list
4. Open up the post again and watch LeakCanary dump a heap
Note:
- [x] Test fix on older Android versions (webkit vs chromium)
- [x] Merge BrowserActivity and WebViewFragment (or fix and test *both*)
----
Unfortunately the heap trace (below) is unhelpful:
```
06-23 03:53:53.152 22607-25907/me.vickychijwani.spectre D/LeakCanary๏น In me.vickychijwani.spectre:1.0:1.
* me.vickychijwani.spectre.view.PostViewActivity has leaked:
* GC ROOT org.chromium.content.browser.input.PopupTouchHandleDrawable.mContext
* leaks me.vickychijwani.spectre.view.PostViewActivity instance
* Reference Key: c5347412-4c57-49a2-9df7-9760955cedeb
* Device: LGE google Nexus 4 occam
* Android Version: 5.1.1 API: 22
* Durations: watch=5124ms, gc=162ms, heap dump=11630ms, analysis=43875ms
```
Maybe analyzing the [actual heap dump](https://www.dropbox.com/s/zodudiuacbctjcr/quill_issue_75.hprof?dl=0) will help.
|
1.0
|
Memory leak in WebViewFragment - Steps to reproduce:
1. Open a post preview (the new WebView-based one)
2. Select some text
3. Tap back twice to come back to the post list
4. Open up the post again and watch LeakCanary dump a heap
Note:
- [x] Test fix on older Android versions (webkit vs chromium)
- [x] Merge BrowserActivity and WebViewFragment (or fix and test *both*)
----
Unfortunately the heap trace (below) is unhelpful:
```
06-23 03:53:53.152 22607-25907/me.vickychijwani.spectre D/LeakCanary๏น In me.vickychijwani.spectre:1.0:1.
* me.vickychijwani.spectre.view.PostViewActivity has leaked:
* GC ROOT org.chromium.content.browser.input.PopupTouchHandleDrawable.mContext
* leaks me.vickychijwani.spectre.view.PostViewActivity instance
* Reference Key: c5347412-4c57-49a2-9df7-9760955cedeb
* Device: LGE google Nexus 4 occam
* Android Version: 5.1.1 API: 22
* Durations: watch=5124ms, gc=162ms, heap dump=11630ms, analysis=43875ms
```
Maybe analyzing the [actual heap dump](https://www.dropbox.com/s/zodudiuacbctjcr/quill_issue_75.hprof?dl=0) will help.
|
non_defect
|
memory leak in webviewfragment steps to reproduce open a post preview the new webview based one select some text tap back twice to come back to the post list open up the post again and watch leakcanary dump a heap note test fix on older android versions webkit vs chromium merge browseractivity and webviewfragment or fix and test both unfortunately the heap trace below is unhelpful me vickychijwani spectre d leakcanary๏น in me vickychijwani spectre me vickychijwani spectre view postviewactivity has leaked gc root org chromium content browser input popuptouchhandledrawable mcontext leaks me vickychijwani spectre view postviewactivity instance reference key device lge google nexus occam android version api durations watch gc heap dump analysis maybe analyzing the will help
| 0
|
144,864
| 5,547,156,747
|
IssuesEvent
|
2017-03-23 04:11:23
|
ClaytonPassmore/ProjectOrange
|
https://api.github.com/repos/ClaytonPassmore/ProjectOrange
|
closed
|
Develop test case to validate NFP
|
Priority: Medium Server
|
We claimed that we should be able to handle 100 decks with 500 cards each.
Generate test data so that we can verify this claim.
|
1.0
|
Develop test case to validate NFP - We claimed that we should be able to handle 100 decks with 500 cards each.
Generate test data so that we can verify this claim.
|
non_defect
|
develop test case to validate nfp we claimed that we should be able to handle decks with cards each generate test data so that we can verify this claim
| 0
|
49,852
| 13,187,281,185
|
IssuesEvent
|
2020-08-13 02:55:03
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
Neutrino simulation event has a pulse width of zero (Trac #2180)
|
Incomplete Migration Migrated from Trac combo simulation defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2180">https://code.icecube.wisc.edu/ticket/2180</a>, reported by lwille and owned by juancarlos</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-03-27T16:25:07",
"description": "I'm running into a Monopod error with about 10% of the files in Nancy's NuGen_new simulation set. The error is \"FATAL (millipede): Assertion failed: p->GetWidth() > 0 (MillipedeDOMCacheMap.cxx:379 in void MillipedeDOMCacheMap::UpdateData(const I3TimeWindow&, const I3RecoPulseSeriesMap&, const I3TimeWindowSeriesMap&, double, double, double, bool))\".\n\nLooking into this error, it appears that about 10% of the time, an event will have a single pulse with a width reported as 0. I'm uncertain if this is an issue with the simulation file processing or something else. An example of an event with a zero pulse width is here `/data/ana/Cscd/StartingEvents/NuGen_new/NuTau/medium_energy/IC86_flasher_p1=0.3_p2=0.0_domeff_081/l2/1/l2_00000248.i3.zst` Event 450.",
"reporter": "lwille",
"cc": "jvansanten, nwhitehorn",
"resolution": "duplicate",
"_ts": "1553703907985721",
"component": "combo simulation",
"summary": "Neutrino simulation event has a pulse width of zero",
"priority": "normal",
"keywords": "",
"time": "2018-08-08T18:33:10",
"milestone": "Vernal Equinox 2019",
"owner": "juancarlos",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
Neutrino simulation event has a pulse width of zero (Trac #2180) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2180">https://code.icecube.wisc.edu/ticket/2180</a>, reported by lwille and owned by juancarlos</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-03-27T16:25:07",
"description": "I'm running into a Monopod error with about 10% of the files in Nancy's NuGen_new simulation set. The error is \"FATAL (millipede): Assertion failed: p->GetWidth() > 0 (MillipedeDOMCacheMap.cxx:379 in void MillipedeDOMCacheMap::UpdateData(const I3TimeWindow&, const I3RecoPulseSeriesMap&, const I3TimeWindowSeriesMap&, double, double, double, bool))\".\n\nLooking into this error, it appears that about 10% of the time, an event will have a single pulse with a width reported as 0. I'm uncertain if this is an issue with the simulation file processing or something else. An example of an event with a zero pulse width is here `/data/ana/Cscd/StartingEvents/NuGen_new/NuTau/medium_energy/IC86_flasher_p1=0.3_p2=0.0_domeff_081/l2/1/l2_00000248.i3.zst` Event 450.",
"reporter": "lwille",
"cc": "jvansanten, nwhitehorn",
"resolution": "duplicate",
"_ts": "1553703907985721",
"component": "combo simulation",
"summary": "Neutrino simulation event has a pulse width of zero",
"priority": "normal",
"keywords": "",
"time": "2018-08-08T18:33:10",
"milestone": "Vernal Equinox 2019",
"owner": "juancarlos",
"type": "defect"
}
```
</p>
</details>
|
defect
|
neutrino simulation event has a pulse width of zero trac migrated from json status closed changetime description i m running into a monopod error with about of the files in nancy s nugen new simulation set the error is fatal millipede assertion failed p getwidth millipededomcachemap cxx in void millipededomcachemap updatedata const const const double double double bool n nlooking into this error it appears that about of the time an event will have a single pulse with a width reported as i m uncertain if this is an issue with the simulation file processing or something else an example of an event with a zero pulse width is here data ana cscd startingevents nugen new nutau medium energy flasher domeff zst event reporter lwille cc jvansanten nwhitehorn resolution duplicate ts component combo simulation summary neutrino simulation event has a pulse width of zero priority normal keywords time milestone vernal equinox owner juancarlos type defect
| 1
|
430,940
| 30,208,283,102
|
IssuesEvent
|
2023-07-05 10:58:20
|
scylladb/scylladb
|
https://api.github.com/repos/scylladb/scylladb
|
opened
|
doc: improve the docs for integration with Elasticsearch
|
Documentation
|
https://opensource.docs.scylladb.com/stable/using-scylla/integrations/integration-elasticsearch.html
Feedback from @zseta :
> My guess would be that the user might prefer a short and to-the-point page and not a list of links that lead to long-form content. It might be worthwhile to pull out the code bits from [this blog](https://www.scylladb.com/2019/03/07/scylla-and-elasticsearch-part-two-practical-examples-to-support-full-text-search-workloads/), and add it on [this page](https://opensource.docs.scylladb.com/stable/using-scylla/integrations/integration-elasticsearch.html) so the user can quickly grab the code that they need straight from docs.
>
> The [Databricks example](https://opensource.docs.scylladb.com/stable/using-scylla/integrations/integration-databricks.html) is a good one from this perspective, because it provides the instructions on-page and don't need to read long blogs to find the solution.
|
1.0
|
doc: improve the docs for integration with Elasticsearch - https://opensource.docs.scylladb.com/stable/using-scylla/integrations/integration-elasticsearch.html
Feedback from @zseta :
> My guess would be that the user might prefer a short and to-the-point page and not a list of links that lead to long-form content. It might be worthwhile to pull out the code bits from [this blog](https://www.scylladb.com/2019/03/07/scylla-and-elasticsearch-part-two-practical-examples-to-support-full-text-search-workloads/), and add it on [this page](https://opensource.docs.scylladb.com/stable/using-scylla/integrations/integration-elasticsearch.html) so the user can quickly grab the code that they need straight from docs.
>
> The [Databricks example](https://opensource.docs.scylladb.com/stable/using-scylla/integrations/integration-databricks.html) is a good one from this perspective, because it provides the instructions on-page and don't need to read long blogs to find the solution.
|
non_defect
|
doc improve the docs for integration with elasticsearch feedback from zseta my guess would be that the user might prefer a short and to the point page and not a list of links that lead to long form content it might be worthwhile to pull out the code bits from and add it on so the user can quickly grab the code that they need straight from docs the is a good one from this perspective because it provides the instructions on page and don t need to read long blogs to find the solution
| 0
|
42,015
| 10,741,402,326
|
IssuesEvent
|
2019-10-29 20:10:07
|
cakephp/bake
|
https://api.github.com/repos/cakephp/bake
|
closed
|
4.x Empty layout templates created for `bin/cake bake mailer Ticket`
|
Defect
|
* Using `cakephp/bake 4.x-dev 1d28d25`
* Running `bin/cake bake mailer Ticket` produces
```
Creating file /var/virtual/cakefest2019/src/../templates/layout/email/html/ticket.php
Could not write to `/var/virtual/cakefest2019/src/../templates/layout/email/html/ticket.php`.
Creating file /var/virtual/cakefest2019/src/../templates/layout/email/text/ticket.php
Could not write to `/var/virtual/cakefest2019/src/../templates/layout/email/text/ticket.php`.
Creating file /var/virtual/cakefest2019/src/Mailer/TicketMailer.php
Wrote `/var/virtual/cakefest2019/src/Mailer/TicketMailer.php`
Baking test case for App\Mailer\TicketMailer ...
Creating file /var/virtual/cakefest2019/tests/TestCase/Mailer/TicketMailerTest.php
Wrote `/var/virtual/cakefest2019/tests/TestCase/Mailer/TicketMailerTest.php`
```
Note file path containing `..` causing issues writing the templates
|
1.0
|
4.x Empty layout templates created for `bin/cake bake mailer Ticket` - * Using `cakephp/bake 4.x-dev 1d28d25`
* Running `bin/cake bake mailer Ticket` produces
```
Creating file /var/virtual/cakefest2019/src/../templates/layout/email/html/ticket.php
Could not write to `/var/virtual/cakefest2019/src/../templates/layout/email/html/ticket.php`.
Creating file /var/virtual/cakefest2019/src/../templates/layout/email/text/ticket.php
Could not write to `/var/virtual/cakefest2019/src/../templates/layout/email/text/ticket.php`.
Creating file /var/virtual/cakefest2019/src/Mailer/TicketMailer.php
Wrote `/var/virtual/cakefest2019/src/Mailer/TicketMailer.php`
Baking test case for App\Mailer\TicketMailer ...
Creating file /var/virtual/cakefest2019/tests/TestCase/Mailer/TicketMailerTest.php
Wrote `/var/virtual/cakefest2019/tests/TestCase/Mailer/TicketMailerTest.php`
```
Note file path containing `..` causing issues writing the templates
|
defect
|
x empty layout templates created for bin cake bake mailer ticket using cakephp bake x dev running bin cake bake mailer ticket produces creating file var virtual src templates layout email html ticket php could not write to var virtual src templates layout email html ticket php creating file var virtual src templates layout email text ticket php could not write to var virtual src templates layout email text ticket php creating file var virtual src mailer ticketmailer php wrote var virtual src mailer ticketmailer php baking test case for app mailer ticketmailer creating file var virtual tests testcase mailer ticketmailertest php wrote var virtual tests testcase mailer ticketmailertest php note file path containing causing issues writing the templates
| 1
|
4,017
| 2,544,716,145
|
IssuesEvent
|
2015-01-29 12:19:13
|
pychess/pychess
|
https://api.github.com/repos/pychess/pychess
|
closed
|
Analog clock in chessclock
|
Component-UI enhancement EyeCandy imported Milestone-Release1.0 Priority-Low
|
_From [lobais](https://code.google.com/u/lobais/) on August 13, 2006 16:30:38_
We could put a little analog clock, next to the digital time. Would look
nice :)
_Original issue: http://code.google.com/p/pychess/issues/detail?id=7_
|
1.0
|
Analog clock in chessclock - _From [lobais](https://code.google.com/u/lobais/) on August 13, 2006 16:30:38_
We could put a little analog clock, next to the digital time. Would look
nice :)
_Original issue: http://code.google.com/p/pychess/issues/detail?id=7_
|
non_defect
|
analog clock in chessclock from on august we could put a little analog clock next to the digital time would look nice original issue
| 0
|
69,141
| 22,203,474,009
|
IssuesEvent
|
2022-06-07 13:12:03
|
matrix-org/matrix-hookshot
|
https://api.github.com/repos/matrix-org/matrix-hookshot
|
closed
|
Use the RequiresClient widget capability to disable popout button in Element
|
T-Defect
|
The hookshot config window can't work without identity verification, so [set RequiresClient](https://github.com/matrix-org/matrix-react-sdk/pull/7005)
|
1.0
|
Use the RequiresClient widget capability to disable popout button in Element - The hookshot config window can't work without identity verification, so [set RequiresClient](https://github.com/matrix-org/matrix-react-sdk/pull/7005)
|
defect
|
use the requiresclient widget capability to disable popout button in element the hookshot config window can t work without identity verification so
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.