Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
258,439
22,319,807,671
IssuesEvent
2022-06-14 04:38:52
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: tpccbench/nodes=9/cpu=4/chaos/partition failed
C-test-failure O-robot O-roachtest release-blocker T-kv branch-release-22.1
roachtest.tpccbench/nodes=9/cpu=4/chaos/partition [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=5377296&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=5377296&tab=artifacts#/tpccbench/nodes=9/cpu=4/chaos/partition) on release-22.1 @ [380f15c04826b432273cbd7ffe96b335ffc90396](https://github.com/cockroachdb/cockroach/commits/380f15c04826b432273cbd7ffe96b335ffc90396): ``` The test failed on branch=release-22.1, cloud=gce: test artifacts and logs in: /artifacts/tpccbench/nodes=9/cpu=4/chaos/partition/run_1 cluster.go:1845,tpcc.go:1153,tpcc.go:1163,search.go:43,search.go:173,tpcc.go:1159,tpcc.go:931,test_runner.go:883: one or more parallel execution failure (1) attached stack trace -- stack trace: | github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).ParallelE | github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:2042 | github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Parallel | github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:1923 | github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Start | github.com/cockroachdb/cockroach/pkg/roachprod/install/cockroach.go:167 | github.com/cockroachdb/cockroach/pkg/roachprod.Start | github.com/cockroachdb/cockroach/pkg/roachprod/roachprod.go:660 | main.(*clusterImpl).StartE | main/pkg/cmd/roachtest/cluster.go:1803 | main.(*clusterImpl).Start | main/pkg/cmd/roachtest/cluster.go:1844 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runTPCCBench.func3 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/tpcc.go:1153 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runTPCCBench.func4 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/tpcc.go:1163 | github.com/cockroachdb/cockroach/pkg/util/search.searchWithSearcher | github.com/cockroachdb/cockroach/pkg/util/search/search.go:43 | github.com/cockroachdb/cockroach/pkg/util/search.(*lineSearcher).Search | github.com/cockroachdb/cockroach/pkg/util/search/search.go:173 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runTPCCBench | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/tpcc.go:1159 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerTPCCBenchSpec.func1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/tpcc.go:931 | main.(*testRunner).runTest.func2 | main/pkg/cmd/roachtest/test_runner.go:883 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1581 Wraps: (2) one or more parallel execution failure Error types: (1) *withstack.withStack (2) *errutil.leafError ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*tpccbench/nodes=9/cpu=4/chaos/partition.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-16400
2.0
roachtest: tpccbench/nodes=9/cpu=4/chaos/partition failed - roachtest.tpccbench/nodes=9/cpu=4/chaos/partition [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=5377296&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=5377296&tab=artifacts#/tpccbench/nodes=9/cpu=4/chaos/partition) on release-22.1 @ [380f15c04826b432273cbd7ffe96b335ffc90396](https://github.com/cockroachdb/cockroach/commits/380f15c04826b432273cbd7ffe96b335ffc90396): ``` The test failed on branch=release-22.1, cloud=gce: test artifacts and logs in: /artifacts/tpccbench/nodes=9/cpu=4/chaos/partition/run_1 cluster.go:1845,tpcc.go:1153,tpcc.go:1163,search.go:43,search.go:173,tpcc.go:1159,tpcc.go:931,test_runner.go:883: one or more parallel execution failure (1) attached stack trace -- stack trace: | github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).ParallelE | github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:2042 | github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Parallel | github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:1923 | github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Start | github.com/cockroachdb/cockroach/pkg/roachprod/install/cockroach.go:167 | github.com/cockroachdb/cockroach/pkg/roachprod.Start | github.com/cockroachdb/cockroach/pkg/roachprod/roachprod.go:660 | main.(*clusterImpl).StartE | main/pkg/cmd/roachtest/cluster.go:1803 | main.(*clusterImpl).Start | main/pkg/cmd/roachtest/cluster.go:1844 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runTPCCBench.func3 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/tpcc.go:1153 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runTPCCBench.func4 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/tpcc.go:1163 | github.com/cockroachdb/cockroach/pkg/util/search.searchWithSearcher | github.com/cockroachdb/cockroach/pkg/util/search/search.go:43 | github.com/cockroachdb/cockroach/pkg/util/search.(*lineSearcher).Search | github.com/cockroachdb/cockroach/pkg/util/search/search.go:173 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runTPCCBench | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/tpcc.go:1159 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerTPCCBenchSpec.func1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/tpcc.go:931 | main.(*testRunner).runTest.func2 | main/pkg/cmd/roachtest/test_runner.go:883 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1581 Wraps: (2) one or more parallel execution failure Error types: (1) *withstack.withStack (2) *errutil.leafError ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*tpccbench/nodes=9/cpu=4/chaos/partition.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-16400
test
roachtest tpccbench nodes cpu chaos partition failed roachtest tpccbench nodes cpu chaos partition with on release the test failed on branch release cloud gce test artifacts and logs in artifacts tpccbench nodes cpu chaos partition run cluster go tpcc go tpcc go search go search go tpcc go tpcc go test runner go one or more parallel execution failure attached stack trace stack trace github com cockroachdb cockroach pkg roachprod install syncedcluster parallele github com cockroachdb cockroach pkg roachprod install cluster synced go github com cockroachdb cockroach pkg roachprod install syncedcluster parallel github com cockroachdb cockroach pkg roachprod install cluster synced go github com cockroachdb cockroach pkg roachprod install syncedcluster start github com cockroachdb cockroach pkg roachprod install cockroach go github com cockroachdb cockroach pkg roachprod start github com cockroachdb cockroach pkg roachprod roachprod go main clusterimpl starte main pkg cmd roachtest cluster go main clusterimpl start main pkg cmd roachtest cluster go github com cockroachdb cockroach pkg cmd roachtest tests runtpccbench github com cockroachdb cockroach pkg cmd roachtest tests tpcc go github com cockroachdb cockroach pkg cmd roachtest tests runtpccbench github com cockroachdb cockroach pkg cmd roachtest tests tpcc go github com cockroachdb cockroach pkg util search searchwithsearcher github com cockroachdb cockroach pkg util search search go github com cockroachdb cockroach pkg util search linesearcher search github com cockroachdb cockroach pkg util search search go github com cockroachdb cockroach pkg cmd roachtest tests runtpccbench github com cockroachdb cockroach pkg cmd roachtest tests tpcc go github com cockroachdb cockroach pkg cmd roachtest tests registertpccbenchspec github com cockroachdb cockroach pkg cmd roachtest tests tpcc go main testrunner runtest main pkg cmd roachtest test runner go runtime goexit goroot src runtime asm s wraps one or more parallel execution failure error types withstack withstack errutil leaferror help see see cc cockroachdb kv triage jira issue crdb
1
43,063
12,965,177,247
IssuesEvent
2020-07-20 21:50:13
jtimberlake/griffin
https://api.github.com/repos/jtimberlake/griffin
opened
WS-2019-0379 (Medium) detected in commons-codec-1.11.jar
security vulnerability
## WS-2019-0379 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-codec-1.11.jar</b></p></summary> <p>The Apache Commons Codec package contains simple encoder and decoders for various formats such as Base64 and Hexadecimal. In addition to these widely used encoders and decoders, the codec package also maintains a collection of phonetic encoding utilities.</p> <p>Path to dependency file: /tmp/ws-scm/griffin/service/hibernate_mysql_pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.11/commons-codec-1.11.jar</p> <p> Dependency Hierarchy: - elasticsearch-rest-client-6.2.4.jar (Root Library) - :x: **commons-codec-1.11.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jtimberlake/griffin/commit/7b8d4cb53c4eab239eecb18da5b2a6048b2fce60">7b8d4cb53c4eab239eecb18da5b2a6048b2fce60</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache commons-codec before version โ€œcommons-codec-1.13-RC1โ€ is vulnerable to information disclosure due to Improper Input validation. <p>Publish Date: 2019-05-20 <p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113">https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113</a></p> <p>Release Date: 2019-05-12</p> <p>Fix Resolution: 1.13-RC1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-codec","packageName":"commons-codec","packageVersion":"1.11","isTransitiveDependency":true,"dependencyTree":"org.elasticsearch.client:elasticsearch-rest-client:6.2.4;commons-codec:commons-codec:1.11","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.13-RC1"}],"vulnerabilityIdentifier":"WS-2019-0379","vulnerabilityDetails":"Apache commons-codec before version โ€œcommons-codec-1.13-RC1โ€ is vulnerable to information disclosure due to Improper Input validation.","vulnerabilityUrl":"https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
WS-2019-0379 (Medium) detected in commons-codec-1.11.jar - ## WS-2019-0379 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-codec-1.11.jar</b></p></summary> <p>The Apache Commons Codec package contains simple encoder and decoders for various formats such as Base64 and Hexadecimal. In addition to these widely used encoders and decoders, the codec package also maintains a collection of phonetic encoding utilities.</p> <p>Path to dependency file: /tmp/ws-scm/griffin/service/hibernate_mysql_pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.11/commons-codec-1.11.jar</p> <p> Dependency Hierarchy: - elasticsearch-rest-client-6.2.4.jar (Root Library) - :x: **commons-codec-1.11.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jtimberlake/griffin/commit/7b8d4cb53c4eab239eecb18da5b2a6048b2fce60">7b8d4cb53c4eab239eecb18da5b2a6048b2fce60</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache commons-codec before version โ€œcommons-codec-1.13-RC1โ€ is vulnerable to information disclosure due to Improper Input validation. <p>Publish Date: 2019-05-20 <p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113">https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113</a></p> <p>Release Date: 2019-05-12</p> <p>Fix Resolution: 1.13-RC1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-codec","packageName":"commons-codec","packageVersion":"1.11","isTransitiveDependency":true,"dependencyTree":"org.elasticsearch.client:elasticsearch-rest-client:6.2.4;commons-codec:commons-codec:1.11","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.13-RC1"}],"vulnerabilityIdentifier":"WS-2019-0379","vulnerabilityDetails":"Apache commons-codec before version โ€œcommons-codec-1.13-RC1โ€ is vulnerable to information disclosure due to Improper Input validation.","vulnerabilityUrl":"https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_test
ws medium detected in commons codec jar ws medium severity vulnerability vulnerable library commons codec jar the apache commons codec package contains simple encoder and decoders for various formats such as and hexadecimal in addition to these widely used encoders and decoders the codec package also maintains a collection of phonetic encoding utilities path to dependency file tmp ws scm griffin service hibernate mysql pom xml path to vulnerable library home wss scanner repository commons codec commons codec commons codec jar dependency hierarchy elasticsearch rest client jar root library x commons codec jar vulnerable library found in head commit a href vulnerability details apache commons codec before version โ€œcommons codec โ€ is vulnerable to information disclosure due to improper input validation publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails apache commons codec before version โ€œcommons codec โ€ is vulnerable to information disclosure due to improper input validation vulnerabilityurl
0
105,617
13,200,745,662
IssuesEvent
2020-08-14 08:49:49
boogle-lab/ssho-issue-board
https://api.github.com/repos/boogle-lab/ssho-issue-board
closed
[2์ฐจ ํ”„๋กœํ† ํƒ€์ž…][4์ฃผ์ฐจ] ์ƒ์ž UI ์Šคํƒ€์ผ๋ง ์ ์šฉ
CLIENT DESIGN NEW-FEATURE
- ์ƒ์ž UI ์Šคํƒ€์ผ๋ง ์ ์šฉ - ๊ฐ„๋‹จํ•œ UI ๋””์ž์ธ @hesed98 - UI ๋ฐ˜์˜ ์ปดํฌ๋„ŒํŠธ ์Šคํƒ€์ผ๋ง @Zinyon - ์Šคํƒ€์ผ๋ง ๊ณผ์ • ์ •๋ฆฌํ•ด์„œ(์ถ”ํ›„ ์Šคํƒ€์ผ๋ง ๊ด€๋ จ ๋ฐฉ๋ฒ• ๊ตฌ์ฒดํ™”) ์—…๋ฐ์ดํŠธ @Zinyon
1.0
[2์ฐจ ํ”„๋กœํ† ํƒ€์ž…][4์ฃผ์ฐจ] ์ƒ์ž UI ์Šคํƒ€์ผ๋ง ์ ์šฉ - - ์ƒ์ž UI ์Šคํƒ€์ผ๋ง ์ ์šฉ - ๊ฐ„๋‹จํ•œ UI ๋””์ž์ธ @hesed98 - UI ๋ฐ˜์˜ ์ปดํฌ๋„ŒํŠธ ์Šคํƒ€์ผ๋ง @Zinyon - ์Šคํƒ€์ผ๋ง ๊ณผ์ • ์ •๋ฆฌํ•ด์„œ(์ถ”ํ›„ ์Šคํƒ€์ผ๋ง ๊ด€๋ จ ๋ฐฉ๋ฒ• ๊ตฌ์ฒดํ™”) ์—…๋ฐ์ดํŠธ @Zinyon
non_test
์ƒ์ž ui ์Šคํƒ€์ผ๋ง ์ ์šฉ ์ƒ์ž ui ์Šคํƒ€์ผ๋ง ์ ์šฉ ๊ฐ„๋‹จํ•œ ui ๋””์ž์ธ ui ๋ฐ˜์˜ ์ปดํฌ๋„ŒํŠธ ์Šคํƒ€์ผ๋ง zinyon ์Šคํƒ€์ผ๋ง ๊ณผ์ • ์ •๋ฆฌํ•ด์„œ ์ถ”ํ›„ ์Šคํƒ€์ผ๋ง ๊ด€๋ จ ๋ฐฉ๋ฒ• ๊ตฌ์ฒดํ™” ์—…๋ฐ์ดํŠธ zinyon
0
168,882
26,709,672,702
IssuesEvent
2023-01-27 21:58:04
ethereum/solidity
https://api.github.com/repos/ethereum/solidity
closed
Unable to disambiguate shadowed function pointers
language design :rage4:
## Description Attempting to pass a function pointer where the name of the function being passed is shadowed results in `Error: No matching declaration found after variable lookup`. In theory, the compiler should be able to disambiguate which version of the function is being passed based on the type definition where the function pointer is used. ## Environment - Compiler version: 0.8.10 ## Steps to Reproduce Pointer.sol ```solidity // SPDX-License-Identifier: MIT pragma solidity 0.8.10; contract Pointer { function shadowed(uint256 a) internal {} function shadowed(uint256 a, uint256 b) internal {} function passFn() internal { takeFn(shadowed); } function takeFn(function(uint256) internal fnPtr) private { fnPtr(0); } } ``` ```bash $ solc --version Version: 0.8.10+commit.fc410830.Linux.g++ $ solc Pointer.sol Error: No matching declaration found after variable lookup. --> Pointer.sol:8:16: | 8 | takeFn(shadowed); | ^^^^^^^^ ``` Removing the shadowing function (`shadowed(uint256, uint256)`) results in a compile-able contract. Because `takeFn()` explicitly accepts a function pointer of type `shadowed(uint256)`, it should be possible to determine which version of the method is being passed. In cases where this isn't possible, it could be nice to have some syntax at the callsite to differentiate which version of a function is expected, though this may be a significant syntactic change. To steal from rust's turbofish operator as an example: ```rust fnPtr::<&shadowed(uint256)>(100); ```
1.0
Unable to disambiguate shadowed function pointers - ## Description Attempting to pass a function pointer where the name of the function being passed is shadowed results in `Error: No matching declaration found after variable lookup`. In theory, the compiler should be able to disambiguate which version of the function is being passed based on the type definition where the function pointer is used. ## Environment - Compiler version: 0.8.10 ## Steps to Reproduce Pointer.sol ```solidity // SPDX-License-Identifier: MIT pragma solidity 0.8.10; contract Pointer { function shadowed(uint256 a) internal {} function shadowed(uint256 a, uint256 b) internal {} function passFn() internal { takeFn(shadowed); } function takeFn(function(uint256) internal fnPtr) private { fnPtr(0); } } ``` ```bash $ solc --version Version: 0.8.10+commit.fc410830.Linux.g++ $ solc Pointer.sol Error: No matching declaration found after variable lookup. --> Pointer.sol:8:16: | 8 | takeFn(shadowed); | ^^^^^^^^ ``` Removing the shadowing function (`shadowed(uint256, uint256)`) results in a compile-able contract. Because `takeFn()` explicitly accepts a function pointer of type `shadowed(uint256)`, it should be possible to determine which version of the method is being passed. In cases where this isn't possible, it could be nice to have some syntax at the callsite to differentiate which version of a function is expected, though this may be a significant syntactic change. To steal from rust's turbofish operator as an example: ```rust fnPtr::<&shadowed(uint256)>(100); ```
non_test
unable to disambiguate shadowed function pointers description attempting to pass a function pointer where the name of the function being passed is shadowed results in error no matching declaration found after variable lookup in theory the compiler should be able to disambiguate which version of the function is being passed based on the type definition where the function pointer is used environment compiler version steps to reproduce pointer sol solidity spdx license identifier mit pragma solidity contract pointer function shadowed a internal function shadowed a b internal function passfn internal takefn shadowed function takefn function internal fnptr private fnptr bash solc version version commit linux g solc pointer sol error no matching declaration found after variable lookup pointer sol takefn shadowed removing the shadowing function shadowed results in a compile able contract because takefn explicitly accepts a function pointer of type shadowed it should be possible to determine which version of the method is being passed in cases where this isn t possible it could be nice to have some syntax at the callsite to differentiate which version of a function is expected though this may be a significant syntactic change to steal from rust s turbofish operator as an example rust fnptr
0
108,348
4,337,593,143
IssuesEvent
2016-07-28 01:16:55
pombase/pombase-chado
https://api.github.com/repos/pombase/pombase-chado
opened
Load all publication details from Canto into Chado
medium priority
For #67 will we need all the publication triage information from Canto. See pombase/canto#1207
1.0
Load all publication details from Canto into Chado - For #67 will we need all the publication triage information from Canto. See pombase/canto#1207
non_test
load all publication details from canto into chado for will we need all the publication triage information from canto see pombase canto
0
55,164
11,396,614,902
IssuesEvent
2020-01-30 13:53:44
eclipse-theia/theia
https://api.github.com/repos/eclipse-theia/theia
closed
workbench.action.closeEditorsInGroup doesn't close it self (the current Widget)
bug vscode
Description workbench.action.closeEditorsInGroup command doesn't close it self (the current Widget) Reproduction Steps Open couple of files in agroup execute workbench.action.closeEditorsInGroup command. Can also be tested with this vscode ext - https://github.com/tomer-epstein/vscode-close-editor.git Diagnostics: plugin-vscode-commands-contribution.ts ({ owner }) => CodeEditorWidget.is(owner) && owner !== editor
1.0
workbench.action.closeEditorsInGroup doesn't close it self (the current Widget) - Description workbench.action.closeEditorsInGroup command doesn't close it self (the current Widget) Reproduction Steps Open couple of files in agroup execute workbench.action.closeEditorsInGroup command. Can also be tested with this vscode ext - https://github.com/tomer-epstein/vscode-close-editor.git Diagnostics: plugin-vscode-commands-contribution.ts ({ owner }) => CodeEditorWidget.is(owner) && owner !== editor
non_test
workbench action closeeditorsingroup doesn t close it self the current widget description workbench action closeeditorsingroup command doesn t close it self the current widget reproduction steps open couple of files in agroup execute workbench action closeeditorsingroup command can also be tested with this vscode ext diagnostics plugin vscode commands contribution ts owner codeeditorwidget is owner owner editor
0
99,340
4,053,291,905
IssuesEvent
2016-05-24 08:04:23
nvs/gem
https://api.github.com/repos/nvs/gem
opened
Is the 'Intro' needed?
Priority: Later Type: Discussion
I can probably count on my hand the number of people I've seen watch it. If the information in game was better, it may not be needed at all.
1.0
Is the 'Intro' needed? - I can probably count on my hand the number of people I've seen watch it. If the information in game was better, it may not be needed at all.
non_test
is the intro needed i can probably count on my hand the number of people i ve seen watch it if the information in game was better it may not be needed at all
0
338,045
30,278,158,223
IssuesEvent
2023-07-07 22:02:30
microsoft/vscode-python
https://api.github.com/repos/microsoft/vscode-python
closed
Group test explorer test by workspace in multi-root workspace
feature-request area-testing
I am using the multi-root workspace in vs code to open multiple python packages, each with their own `tests` folder containing the tests for that package. The test explorer displays all tests, but does not group them by the workspace they come from. For a simple multi-root workspace with just two packages, this looks like the following screenshot: ![image](https://user-images.githubusercontent.com/25877269/206271562-0aae8e28-c672-4150-a838-55b8527cb23e.png) Since all folders are named in the same way, its cumbersome to determine which entry is from which workspace. I also searched through the extension's settings, but didn't find anyting related. I think the best solution would be to group the tests simililarly as in the repositories in the source control view. What do you think?
1.0
Group test explorer test by workspace in multi-root workspace - I am using the multi-root workspace in vs code to open multiple python packages, each with their own `tests` folder containing the tests for that package. The test explorer displays all tests, but does not group them by the workspace they come from. For a simple multi-root workspace with just two packages, this looks like the following screenshot: ![image](https://user-images.githubusercontent.com/25877269/206271562-0aae8e28-c672-4150-a838-55b8527cb23e.png) Since all folders are named in the same way, its cumbersome to determine which entry is from which workspace. I also searched through the extension's settings, but didn't find anyting related. I think the best solution would be to group the tests simililarly as in the repositories in the source control view. What do you think?
test
group test explorer test by workspace in multi root workspace i am using the multi root workspace in vs code to open multiple python packages each with their own tests folder containing the tests for that package the test explorer displays all tests but does not group them by the workspace they come from for a simple multi root workspace with just two packages this looks like the following screenshot since all folders are named in the same way its cumbersome to determine which entry is from which workspace i also searched through the extension s settings but didn t find anyting related i think the best solution would be to group the tests simililarly as in the repositories in the source control view what do you think
1
438,904
30,668,507,502
IssuesEvent
2023-07-25 20:16:28
storybookjs/storybook
https://api.github.com/repos/storybookjs/storybook
closed
[Documentation]: Extend the Environment variables documentation for `VITE_` environment variables docs
documentation
### Describe the problem Currently, the documentation lacks information on how to use certain types of environment variables (e.g., `VITE_`) and which ones are currently supported. Mentioned in the following [discussion](https://github.com/storybookjs/storybook/discussions/22703#discussioncomment-6454120) ### Additional context N/A
1.0
[Documentation]: Extend the Environment variables documentation for `VITE_` environment variables docs - ### Describe the problem Currently, the documentation lacks information on how to use certain types of environment variables (e.g., `VITE_`) and which ones are currently supported. Mentioned in the following [discussion](https://github.com/storybookjs/storybook/discussions/22703#discussioncomment-6454120) ### Additional context N/A
non_test
extend the environment variables documentation for vite environment variables docs describe the problem currently the documentation lacks information on how to use certain types of environment variables e g vite and which ones are currently supported mentioned in the following additional context n a
0
442,748
30,854,703,607
IssuesEvent
2023-08-02 19:34:56
dogboy21/serializationisbad
https://api.github.com/repos/dogboy21/serializationisbad
closed
LanteaCraft listed, likely shouldn't be?
documentation
LanteaCraft is linked along with SGCraft, but LanteaCraft forked SGCraft in 1.6.4, and was totally rewritten very soon after, the classes you are patching do not exist in LanteaCraft. Our repo is here: https://github.com/PC-Logix/LanteaCraft I've checked and can not find any references to "ObjectInputStream" in any of LanteaCraft code. It's possible this existed long long ago, but any builds based on the original SGCraft before our full rewrite were removed from CurseForce ~9 years ago. If I am missing anything in the current LanteaCraft source please let me know, and I'll re-work the code as needed.
1.0
LanteaCraft listed, likely shouldn't be? - LanteaCraft is linked along with SGCraft, but LanteaCraft forked SGCraft in 1.6.4, and was totally rewritten very soon after, the classes you are patching do not exist in LanteaCraft. Our repo is here: https://github.com/PC-Logix/LanteaCraft I've checked and can not find any references to "ObjectInputStream" in any of LanteaCraft code. It's possible this existed long long ago, but any builds based on the original SGCraft before our full rewrite were removed from CurseForce ~9 years ago. If I am missing anything in the current LanteaCraft source please let me know, and I'll re-work the code as needed.
non_test
lanteacraft listed likely shouldn t be lanteacraft is linked along with sgcraft but lanteacraft forked sgcraft in and was totally rewritten very soon after the classes you are patching do not exist in lanteacraft our repo is here i ve checked and can not find any references to objectinputstream in any of lanteacraft code it s possible this existed long long ago but any builds based on the original sgcraft before our full rewrite were removed from curseforce years ago if i am missing anything in the current lanteacraft source please let me know and i ll re work the code as needed
0
247,829
7,924,113,181
IssuesEvent
2018-07-05 15:54:05
weglot/weglot-php
https://api.github.com/repos/weglot/weglot-php
closed
Text node condition have been removed
priority: high status: have to reproduce type: bug
<!-- This form is for bug reports and feature requests ONLY! --> **Is this a BUG REPORT or FEATURE REQUEST?**: BUG **What happened**: When you pass`<script>This is translated but should not be</script>` it get's translated. Same for `<noscript>` **What you expected to happen**: We should have the same behavior as `0.5.8` where we had a filter function that check the text like this: ```php protected function check() { return ($this->node->parent()->tag != 'script' && $this->node->parent()->tag != 'style' && $this->node->parent()->tag != 'noscript' && $this->node->parent()->tag != 'code' && !is_numeric(TextUtil::fullTrim($this->node->outertext)) && !preg_match('/^\d+%$/', TextUtil::fullTrim($this->node->outertext)) && strpos($this->node->outertext, '[vc_') === false && strpos($this->node->outertext, '<?php') === false); } ``` **How to reproduce it (some code is appreciated)**: ```php <?php require_once __DIR__. '/weglot-php/vendor/autoload.php'; use Weglot\Client\Client; use Weglot\Parser\Parser; use Weglot\Parser\ConfigProvider\ManualConfigProvider; use Weglot\Client\Api\Enum\BotType; // Url to parse $url = 'https://weglot.com/documentation/getting-started'; // Config manually $config = new ManualConfigProvider($url, BotType::HUMAN); // Client $client = new Client('wg_XXXXX'); $parser = new Parser($client, $config); $contents = <<<EOT <script>This is translated but should not be</script> <style>This is translated but should not be</style> <code>This is translated but should not be</code> <noscript>This is translated but should not be</noscript> <p>This is translated but should not be [vc_ok]</p> EOT; // Run the Parser $translatedContent = $parser->translate($contents, 'en', 'fr'); // dumping returned object echo $translatedContent; ```
1.0
Text node condition have been removed - <!-- This form is for bug reports and feature requests ONLY! --> **Is this a BUG REPORT or FEATURE REQUEST?**: BUG **What happened**: When you pass`<script>This is translated but should not be</script>` it get's translated. Same for `<noscript>` **What you expected to happen**: We should have the same behavior as `0.5.8` where we had a filter function that check the text like this: ```php protected function check() { return ($this->node->parent()->tag != 'script' && $this->node->parent()->tag != 'style' && $this->node->parent()->tag != 'noscript' && $this->node->parent()->tag != 'code' && !is_numeric(TextUtil::fullTrim($this->node->outertext)) && !preg_match('/^\d+%$/', TextUtil::fullTrim($this->node->outertext)) && strpos($this->node->outertext, '[vc_') === false && strpos($this->node->outertext, '<?php') === false); } ``` **How to reproduce it (some code is appreciated)**: ```php <?php require_once __DIR__. '/weglot-php/vendor/autoload.php'; use Weglot\Client\Client; use Weglot\Parser\Parser; use Weglot\Parser\ConfigProvider\ManualConfigProvider; use Weglot\Client\Api\Enum\BotType; // Url to parse $url = 'https://weglot.com/documentation/getting-started'; // Config manually $config = new ManualConfigProvider($url, BotType::HUMAN); // Client $client = new Client('wg_XXXXX'); $parser = new Parser($client, $config); $contents = <<<EOT <script>This is translated but should not be</script> <style>This is translated but should not be</style> <code>This is translated but should not be</code> <noscript>This is translated but should not be</noscript> <p>This is translated but should not be [vc_ok]</p> EOT; // Run the Parser $translatedContent = $parser->translate($contents, 'en', 'fr'); // dumping returned object echo $translatedContent; ```
non_test
text node condition have been removed is this a bug report or feature request bug what happened when you pass this is translated but should not be it get s translated same for what you expected to happen we should have the same behavior as where we had a filter function that check the text like this php protected function check return this node parent tag script this node parent tag style this node parent tag noscript this node parent tag code is numeric textutil fulltrim this node outertext preg match d textutil fulltrim this node outertext strpos this node outertext vc false strpos this node outertext php false how to reproduce it some code is appreciated php php require once dir weglot php vendor autoload php use weglot client client use weglot parser parser use weglot parser configprovider manualconfigprovider use weglot client api enum bottype url to parse url config manually config new manualconfigprovider url bottype human client client new client wg xxxxx parser new parser client config contents eot this is translated but should not be this is translated but should not be this is translated but should not be this is translated but should not be this is translated but should not be eot run the parser translatedcontent parser translate contents en fr dumping returned object echo translatedcontent
0
23,388
4,933,748,914
IssuesEvent
2016-11-28 17:08:56
antoinecarme/pyaf
https://api.github.com/repos/antoinecarme/pyaf
opened
Add a Jupyter Notebook to demonstrate the use of Hierarchical Forecasting
Documentation priority:high
need a similar doc with hierarchical and grouped time series detailed examples.
1.0
Add a Jupyter Notebook to demonstrate the use of Hierarchical Forecasting - need a similar doc with hierarchical and grouped time series detailed examples.
non_test
add a jupyter notebook to demonstrate the use of hierarchical forecasting need a similar doc with hierarchical and grouped time series detailed examples
0
76,828
7,546,207,557
IssuesEvent
2018-04-18 01:36:39
khughitt/eda
https://api.github.com/repos/khughitt/eda
closed
Add test code
testing
Write unit tests using [testthat](https://cran.r-project.org/web/packages/testthat/index.html): 1. Use a small dataset (e.g. `iris`) 2. Test out each of public-facing EDA classes 3. Write tests for datasets with and without metadata 4. Pass in inappropriate inputs / edge cases (e.g. 1 x n data frames) 5. Add Travis config and link to Github
1.0
Add test code - Write unit tests using [testthat](https://cran.r-project.org/web/packages/testthat/index.html): 1. Use a small dataset (e.g. `iris`) 2. Test out each of public-facing EDA classes 3. Write tests for datasets with and without metadata 4. Pass in inappropriate inputs / edge cases (e.g. 1 x n data frames) 5. Add Travis config and link to Github
test
add test code write unit tests using use a small dataset e g iris test out each of public facing eda classes write tests for datasets with and without metadata pass in inappropriate inputs edge cases e g x n data frames add travis config and link to github
1
89,958
25,935,749,363
IssuesEvent
2022-12-16 14:01:22
docker/docs
https://api.github.com/repos/docker/docs
closed
Higher level explanations + simple examples
area/Build lifecycle/stale
File: [compose/reference/build.md](https://docs.docker.com/compose/reference/build/) Higher level summaries of what CLI options do + a simple example or two = everyone understands docs infinitely better.
1.0
Higher level explanations + simple examples - File: [compose/reference/build.md](https://docs.docker.com/compose/reference/build/) Higher level summaries of what CLI options do + a simple example or two = everyone understands docs infinitely better.
non_test
higher level explanations simple examples file higher level summaries of what cli options do a simple example or two everyone understands docs infinitely better
0
264,201
20,010,223,996
IssuesEvent
2022-02-01 04:55:45
personalrobotics/ada_feeding
https://api.github.com/repos/personalrobotics/ada_feeding
opened
Add consistency and improve documentation for collision checking
documentation
We should have more consistency (and more documentation/comments) about where we follow what collision constraints. It took me a long time to understand that the collision constraints in Feeding Demo (e.g., [this](https://github.com/personalrobotics/ada_feeding/blob/b0b0575423d6ee9d19e3e8918908da0a8c78e63c/src/action/FeedFoodToPerson.cpp#L30)) are actually empty as far as I can tell, and to get the world collision constraints you have to call a function in aikido (e.g., [this call](https://github.com/personalrobotics/ada_feeding/blob/b0b0575423d6ee9d19e3e8918908da0a8c78e63c/src/action/MoveAbovePlate.cpp#L24) to [this aikido function](https://github.com/personalrobotics/aikido/blob/c015cea62017b6253264460c70b28df9119c69c0/src/robot/Robot.cpp#L250)). Before I realized the different collision constraints that are being used, I'd freely copy/paste a call to `planToConfiguration` without considering which collision constraints are being used. This makes it more challenging for new developers to onboard onto ADA. See #11 for the problem that led me to realize the confusion and inconsistencies in how we treat collisions. As a solution, I'd propose the following: 1. All collision constraints should come from calls to FeedingDemo. So we should add a FeedingDemo function (e.g., `getWorldCollisionConstraint()`) for getting the collision constraints that are currently gotten from the aikido function. 2. Remove and/or consolidate the empty `mCollisionFreeConstraint` and `mCollisionFreeConstraintWithWallFurtherBack`. The new function that gets a blank collision constraint should be called something like `getEmptyCollisionConstraint()`.
1.0
Add consistency and improve documentation for collision checking - We should have more consistency (and more documentation/comments) about where we follow what collision constraints. It took me a long time to understand that the collision constraints in Feeding Demo (e.g., [this](https://github.com/personalrobotics/ada_feeding/blob/b0b0575423d6ee9d19e3e8918908da0a8c78e63c/src/action/FeedFoodToPerson.cpp#L30)) are actually empty as far as I can tell, and to get the world collision constraints you have to call a function in aikido (e.g., [this call](https://github.com/personalrobotics/ada_feeding/blob/b0b0575423d6ee9d19e3e8918908da0a8c78e63c/src/action/MoveAbovePlate.cpp#L24) to [this aikido function](https://github.com/personalrobotics/aikido/blob/c015cea62017b6253264460c70b28df9119c69c0/src/robot/Robot.cpp#L250)). Before I realized the different collision constraints that are being used, I'd freely copy/paste a call to `planToConfiguration` without considering which collision constraints are being used. This makes it more challenging for new developers to onboard onto ADA. See #11 for the problem that led me to realize the confusion and inconsistencies in how we treat collisions. As a solution, I'd propose the following: 1. All collision constraints should come from calls to FeedingDemo. So we should add a FeedingDemo function (e.g., `getWorldCollisionConstraint()`) for getting the collision constraints that are currently gotten from the aikido function. 2. Remove and/or consolidate the empty `mCollisionFreeConstraint` and `mCollisionFreeConstraintWithWallFurtherBack`. The new function that gets a blank collision constraint should be called something like `getEmptyCollisionConstraint()`.
non_test
add consistency and improve documentation for collision checking we should have more consistency and more documentation comments about where we follow what collision constraints it took me a long time to understand that the collision constraints in feeding demo e g are actually empty as far as i can tell and to get the world collision constraints you have to call a function in aikido e g to before i realized the different collision constraints that are being used i d freely copy paste a call to plantoconfiguration without considering which collision constraints are being used this makes it more challenging for new developers to onboard onto ada see for the problem that led me to realize the confusion and inconsistencies in how we treat collisions as a solution i d propose the following all collision constraints should come from calls to feedingdemo so we should add a feedingdemo function e g getworldcollisionconstraint for getting the collision constraints that are currently gotten from the aikido function remove and or consolidate the empty mcollisionfreeconstraint and mcollisionfreeconstraintwithwallfurtherback the new function that gets a blank collision constraint should be called something like getemptycollisionconstraint
0
131,369
18,244,865,905
IssuesEvent
2021-10-01 17:00:17
ibm-skills-network/olx-cli
https://api.github.com/repos/ibm-skills-network/olx-cli
opened
CVE-2021-3807 (High) detected in ansi-regex-3.0.0.tgz
security vulnerability
## CVE-2021-3807 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ansi-regex-3.0.0.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-3.0.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-3.0.0.tgz</a></p> <p>Path to dependency file: olx-cli/package.json</p> <p>Path to vulnerable library: olx-cli/node_modules/ansi-regex/package.json</p> <p> Dependency Hierarchy: - plugin-help-3.2.2.tgz (Root Library) - wrap-ansi-4.0.0.tgz - strip-ansi-4.0.0.tgz - :x: **ansi-regex-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ibm-skills-network/olx-cli/commit/0eb81ebf6f4ad8012837f17100271c5496a2794c">0eb81ebf6f4ad8012837f17100271c5496a2794c</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ansi-regex is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807>CVE-2021-3807</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/">https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/</a></p> <p>Release Date: 2021-09-17</p> <p>Fix Resolution: ansi-regex - 5.0.1,6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-3807 (High) detected in ansi-regex-3.0.0.tgz - ## CVE-2021-3807 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ansi-regex-3.0.0.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-3.0.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-3.0.0.tgz</a></p> <p>Path to dependency file: olx-cli/package.json</p> <p>Path to vulnerable library: olx-cli/node_modules/ansi-regex/package.json</p> <p> Dependency Hierarchy: - plugin-help-3.2.2.tgz (Root Library) - wrap-ansi-4.0.0.tgz - strip-ansi-4.0.0.tgz - :x: **ansi-regex-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ibm-skills-network/olx-cli/commit/0eb81ebf6f4ad8012837f17100271c5496a2794c">0eb81ebf6f4ad8012837f17100271c5496a2794c</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ansi-regex is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807>CVE-2021-3807</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/">https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/</a></p> <p>Release Date: 2021-09-17</p> <p>Fix Resolution: ansi-regex - 5.0.1,6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in ansi regex tgz cve high severity vulnerability vulnerable library ansi regex tgz regular expression for matching ansi escape codes library home page a href path to dependency file olx cli package json path to vulnerable library olx cli node modules ansi regex package json dependency hierarchy plugin help tgz root library wrap ansi tgz strip ansi tgz x ansi regex tgz vulnerable library found in head commit a href found in base branch master vulnerability details ansi regex is vulnerable to inefficient regular expression complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ansi regex step up your open source security game with whitesource
0
40,480
8,793,954,529
IssuesEvent
2018-12-21 22:22:01
MicrosoftDocs/live-share
https://api.github.com/repos/MicrosoftDocs/live-share
closed
Feature request: Allow AL debugging
area: co-debug area: extensions bug vscode
<!-- For Visual Studio problems/feedback, please use the "Report a Problem..." feature built into the tool. See https://aka.ms/vsls-vsproblem. For VS Code issues, attach verbose logs as follows: 1. Press F1 (or Ctrl-Shift-P), type "export logs" and run the "Live Share: Export Logs" command. 2. Drag and drop the zip to the issue on this screen and wait for it to upload before creating the issue. For feature requests, please include enough of this same info so we know if the request is tool or language/platform specific. --> **Product and Version** [VSCode]: 1.21.0 **OS Version** [Windows]: Server 2016 1607 (14393.2007) **Live Share Extension Version**: 0.2.206 **Target Platform or Language** [e.g. Node.js]: AL **Steps to Reproduce / Scenario:** 1a. Go through https://docs.microsoft.com/en-us/dynamics-nav/developer/devenv-get-started to get the AL environment up and running. 1b. Alternatively, if you have docker running, you can do docker run -e accept_eula=y microsoft/dynamics-nav:devpreview. The docker log will tell you the ip, user and password as well as where you find the correct vsix. With the vsix installed you can call "AL: Go!" to get an initial project and deploy it with F5 2. If that works, stop the debugging session 3. Activate a live share 4. The following error appears for me One or more errors occured. (Exception has been thrown by the target of an invocation.) The logs are here: [vsls-logs.zip](https://github.com/MicrosoftDocs/live-share/files/1814382/vsls-logs.zip)
1.0
Feature request: Allow AL debugging - <!-- For Visual Studio problems/feedback, please use the "Report a Problem..." feature built into the tool. See https://aka.ms/vsls-vsproblem. For VS Code issues, attach verbose logs as follows: 1. Press F1 (or Ctrl-Shift-P), type "export logs" and run the "Live Share: Export Logs" command. 2. Drag and drop the zip to the issue on this screen and wait for it to upload before creating the issue. For feature requests, please include enough of this same info so we know if the request is tool or language/platform specific. --> **Product and Version** [VSCode]: 1.21.0 **OS Version** [Windows]: Server 2016 1607 (14393.2007) **Live Share Extension Version**: 0.2.206 **Target Platform or Language** [e.g. Node.js]: AL **Steps to Reproduce / Scenario:** 1a. Go through https://docs.microsoft.com/en-us/dynamics-nav/developer/devenv-get-started to get the AL environment up and running. 1b. Alternatively, if you have docker running, you can do docker run -e accept_eula=y microsoft/dynamics-nav:devpreview. The docker log will tell you the ip, user and password as well as where you find the correct vsix. With the vsix installed you can call "AL: Go!" to get an initial project and deploy it with F5 2. If that works, stop the debugging session 3. Activate a live share 4. The following error appears for me One or more errors occured. (Exception has been thrown by the target of an invocation.) The logs are here: [vsls-logs.zip](https://github.com/MicrosoftDocs/live-share/files/1814382/vsls-logs.zip)
non_test
feature request allow al debugging for visual studio problems feedback please use the report a problem feature built into the tool see for vs code issues attach verbose logs as follows press or ctrl shift p type export logs and run the live share export logs command drag and drop the zip to the issue on this screen and wait for it to upload before creating the issue for feature requests please include enough of this same info so we know if the request is tool or language platform specific product and version os version server live share extension version target platform or language al steps to reproduce scenario go through to get the al environment up and running alternatively if you have docker running you can do docker run e accept eula y microsoft dynamics nav devpreview the docker log will tell you the ip user and password as well as where you find the correct vsix with the vsix installed you can call al go to get an initial project and deploy it with if that works stop the debugging session activate a live share the following error appears for me one or more errors occured exception has been thrown by the target of an invocation the logs are here
0
12,733
3,088,830,916
IssuesEvent
2015-08-25 18:30:54
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
closed
#if preprocessor directive incorrectly handles enclosed C# verbatim string
Area-Language Design
To reproduce, just use this code and make sure that DEBUG is set: #if !DEBUG [Description(@" # text")] class Test { } #endif Actually, this produces a CS1024 error because the compiler seems to interpret the "# text" portion of the string as a preprocessor directive.
1.0
#if preprocessor directive incorrectly handles enclosed C# verbatim string - To reproduce, just use this code and make sure that DEBUG is set: #if !DEBUG [Description(@" # text")] class Test { } #endif Actually, this produces a CS1024 error because the compiler seems to interpret the "# text" portion of the string as a preprocessor directive.
non_test
if preprocessor directive incorrectly handles enclosed c verbatim string to reproduce just use this code and make sure that debug is set if debug description text class test endif actually this produces a error because the compiler seems to interpret the text portion of the string as a preprocessor directive
0
260,089
27,755,058,832
IssuesEvent
2023-03-16 01:18:32
andygonzalez2010/store
https://api.github.com/repos/andygonzalez2010/store
opened
CVE-2023-1108 (High) detected in undertow-core-2.0.19.Final.jar
Mend: dependency security vulnerability
## CVE-2023-1108 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>undertow-core-2.0.19.Final.jar</b></p></summary> <p>Undertow</p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/undertow/undertow-core/2.0.19.Final/undertow-core-2.0.19.Final.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-undertow-2.1.4.RELEASE.jar (Root Library) - :x: **undertow-core-2.0.19.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/andygonzalez2010/store/commit/3f6d614029f4d6cfdddfcef8468949cb7822503c">3f6d614029f4d6cfdddfcef8468949cb7822503c</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Undertow: Infinite loop in SslConduit during close. A flaw was found in undertow. This issue makes achieving a denial of service possible due to an unexpected handshake status updated in SslConduit, where the loop never terminates. <p>Publish Date: 2023-03-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1108>CVE-2023-1108</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2023-1108">https://nvd.nist.gov/vuln/detail/CVE-2023-1108</a></p> <p>Release Date: 2023-03-01</p> <p>Fix Resolution: org.teiid:vdb-base-builder - 1.6.0;io.syndesis.server:server-runtime - 1.3.5,1.13.1;io.syndesis.meta:meta - 1.13.1,1.3.5,1.13.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-1108 (High) detected in undertow-core-2.0.19.Final.jar - ## CVE-2023-1108 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>undertow-core-2.0.19.Final.jar</b></p></summary> <p>Undertow</p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/undertow/undertow-core/2.0.19.Final/undertow-core-2.0.19.Final.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-undertow-2.1.4.RELEASE.jar (Root Library) - :x: **undertow-core-2.0.19.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/andygonzalez2010/store/commit/3f6d614029f4d6cfdddfcef8468949cb7822503c">3f6d614029f4d6cfdddfcef8468949cb7822503c</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Undertow: Infinite loop in SslConduit during close. A flaw was found in undertow. This issue makes achieving a denial of service possible due to an unexpected handshake status updated in SslConduit, where the loop never terminates. <p>Publish Date: 2023-03-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1108>CVE-2023-1108</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2023-1108">https://nvd.nist.gov/vuln/detail/CVE-2023-1108</a></p> <p>Release Date: 2023-03-01</p> <p>Fix Resolution: org.teiid:vdb-base-builder - 1.6.0;io.syndesis.server:server-runtime - 1.3.5,1.13.1;io.syndesis.meta:meta - 1.13.1,1.3.5,1.13.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in undertow core final jar cve high severity vulnerability vulnerable library undertow core final jar undertow path to dependency file pom xml path to vulnerable library home wss scanner repository io undertow undertow core final undertow core final jar dependency hierarchy spring boot starter undertow release jar root library x undertow core final jar vulnerable library found in head commit a href found in base branch master vulnerability details undertow infinite loop in sslconduit during close a flaw was found in undertow this issue makes achieving a denial of service possible due to an unexpected handshake status updated in sslconduit where the loop never terminates publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org teiid vdb base builder io syndesis server server runtime io syndesis meta meta step up your open source security game with mend
0
710
12,867,031,996
IssuesEvent
2020-07-10 05:47:03
shioyama/hakodate-wikipedia
https://api.github.com/repos/shioyama/hakodate-wikipedia
opened
Merriman Colbert Harris
people
Founded Aoyama Gakuin University. His wife founded Iai Yochien (in Hakodate), was very active in girls education. References: - [English Wikipedia page](https://en.wikipedia.org/wiki/Merriman_Colbert_Harris) (mentions Hakodate briefly) - [Japanese Wikipedia page](https://ja.wikipedia.org/wiki/%E3%83%A1%E3%83%AA%E3%83%9E%E3%83%B3%E3%83%BB%E3%83%8F%E3%83%AA%E3%82%B9)
1.0
Merriman Colbert Harris - Founded Aoyama Gakuin University. His wife founded Iai Yochien (in Hakodate), was very active in girls education. References: - [English Wikipedia page](https://en.wikipedia.org/wiki/Merriman_Colbert_Harris) (mentions Hakodate briefly) - [Japanese Wikipedia page](https://ja.wikipedia.org/wiki/%E3%83%A1%E3%83%AA%E3%83%9E%E3%83%B3%E3%83%BB%E3%83%8F%E3%83%AA%E3%82%B9)
non_test
merriman colbert harris founded aoyama gakuin university his wife founded iai yochien in hakodate was very active in girls education references mentions hakodate briefly
0
451,001
32,002,648,252
IssuesEvent
2023-09-21 13:11:07
typescript-eslint/typescript-eslint
https://api.github.com/repos/typescript-eslint/typescript-eslint
closed
Docs: Out-of-date TS error codes in eslint-recommended.ts
good first issue documentation accepting prs
### Before You File a Documentation Request Please Confirm You Have Done The Following... - [X] I have looked for existing [open or closed documentation requests](https://github.com/typescript-eslint/typescript-eslint/issues?q=is%3Aissue+label%3Adocumentation) that match my proposal. - [X] I have [read the FAQ](https://typescript-eslint.io/linting/troubleshooting) and my problem is not listed. ### Suggested Changes https://github.com/typescript-eslint/typescript-eslint/blob/939e42f05afa9536e03923b8829f060cdf35ada0/packages/eslint-plugin/src/configs/eslint-recommended.ts includes comments referring to corresponding TS error codes for rules. Example: https://github.com/typescript-eslint/typescript-eslint/blob/939e42f05afa9536e03923b8829f060cdf35ada0/packages/eslint-plugin/src/configs/eslint-recommended.ts#L11C17-L11C17 Some of those error codes are out of date: https://github.com/eslint/eslint/pull/17500#discussion_r1311685415. Someone should go through and update the TS codes to whatever they are now as of the latest stable version of Typescript. ### Affected URL(s) https://github.com/typescript-eslint/typescript-eslint/blob/939e42f05afa9536e03923b8829f060cdf35ada0/packages/eslint-plugin/src/configs/eslint-recommended.ts (Technically comments rather than website, but they're used by some as docs too)
1.0
Docs: Out-of-date TS error codes in eslint-recommended.ts - ### Before You File a Documentation Request Please Confirm You Have Done The Following... - [X] I have looked for existing [open or closed documentation requests](https://github.com/typescript-eslint/typescript-eslint/issues?q=is%3Aissue+label%3Adocumentation) that match my proposal. - [X] I have [read the FAQ](https://typescript-eslint.io/linting/troubleshooting) and my problem is not listed. ### Suggested Changes https://github.com/typescript-eslint/typescript-eslint/blob/939e42f05afa9536e03923b8829f060cdf35ada0/packages/eslint-plugin/src/configs/eslint-recommended.ts includes comments referring to corresponding TS error codes for rules. Example: https://github.com/typescript-eslint/typescript-eslint/blob/939e42f05afa9536e03923b8829f060cdf35ada0/packages/eslint-plugin/src/configs/eslint-recommended.ts#L11C17-L11C17 Some of those error codes are out of date: https://github.com/eslint/eslint/pull/17500#discussion_r1311685415. Someone should go through and update the TS codes to whatever they are now as of the latest stable version of Typescript. ### Affected URL(s) https://github.com/typescript-eslint/typescript-eslint/blob/939e42f05afa9536e03923b8829f060cdf35ada0/packages/eslint-plugin/src/configs/eslint-recommended.ts (Technically comments rather than website, but they're used by some as docs too)
non_test
docs out of date ts error codes in eslint recommended ts before you file a documentation request please confirm you have done the following i have looked for existing that match my proposal i have and my problem is not listed suggested changes includes comments referring to corresponding ts error codes for rules example some of those error codes are out of date someone should go through and update the ts codes to whatever they are now as of the latest stable version of typescript affected url s technically comments rather than website but they re used by some as docs too
0
14,864
25,926,522,061
IssuesEvent
2022-12-16 05:14:29
jackcmh1/todo_app_test
https://api.github.com/repos/jackcmh1/todo_app_test
closed
Github Project ์นธ๋ฐ˜
Bare Minimum Requirement
- [x] Todo App์„ ์ง์ ‘ ๋งŒ๋“ ๋‹ค๊ณ  ์ƒ๊ฐํ•˜๊ณ  ์ด์Šˆ, ๋งˆ์ผ์Šคํ†ค์„ ์Šค์Šค๋กœ ์ •ํ•ด๋ด…๋‹ˆ๋‹ค. - [x] Github Project ์นธ๋ฐ˜์„ ๋งŒ๋“ค๊ณ , ์ด์Šˆ์™€ ๋งˆ์ผ์Šคํ†ค์„ ์—ฐ๊ฒฐํ•˜๊ณ  ์นธ๋ฐ˜ ๋ทฐ๋กœ ๋ณ€๊ฒฝํ•ฉ๋‹ˆ๋‹ค
1.0
Github Project ์นธ๋ฐ˜ - - [x] Todo App์„ ์ง์ ‘ ๋งŒ๋“ ๋‹ค๊ณ  ์ƒ๊ฐํ•˜๊ณ  ์ด์Šˆ, ๋งˆ์ผ์Šคํ†ค์„ ์Šค์Šค๋กœ ์ •ํ•ด๋ด…๋‹ˆ๋‹ค. - [x] Github Project ์นธ๋ฐ˜์„ ๋งŒ๋“ค๊ณ , ์ด์Šˆ์™€ ๋งˆ์ผ์Šคํ†ค์„ ์—ฐ๊ฒฐํ•˜๊ณ  ์นธ๋ฐ˜ ๋ทฐ๋กœ ๋ณ€๊ฒฝํ•ฉ๋‹ˆ๋‹ค
non_test
github project ์นธ๋ฐ˜ todo app์„ ์ง์ ‘ ๋งŒ๋“ ๋‹ค๊ณ  ์ƒ๊ฐํ•˜๊ณ  ์ด์Šˆ ๋งˆ์ผ์Šคํ†ค์„ ์Šค์Šค๋กœ ์ •ํ•ด๋ด…๋‹ˆ๋‹ค github project ์นธ๋ฐ˜์„ ๋งŒ๋“ค๊ณ  ์ด์Šˆ์™€ ๋งˆ์ผ์Šคํ†ค์„ ์—ฐ๊ฒฐํ•˜๊ณ  ์นธ๋ฐ˜ ๋ทฐ๋กœ ๋ณ€๊ฒฝํ•ฉ๋‹ˆ๋‹ค
0
255,868
21,963,029,271
IssuesEvent
2022-05-24 17:24:05
pulumi/pulumi-kubernetes
https://api.github.com/repos/pulumi/pulumi-kubernetes
closed
Use dedicated GKE cluster for each CI run
kind/engineering area/tests resolution/fixed
We manage most tests in dedicated namespaces. Colliding tests have been a massive pain lately. We already boot a GKE cluster for one of the tests, so we might as well pull that out, and have every test for a given CI invocation run inside it. This would also remove our need to have a perpetually-idle CI cluster, which overall would reduce costs. It also has no impact on test time, since one test already boots GKE.
1.0
Use dedicated GKE cluster for each CI run - We manage most tests in dedicated namespaces. Colliding tests have been a massive pain lately. We already boot a GKE cluster for one of the tests, so we might as well pull that out, and have every test for a given CI invocation run inside it. This would also remove our need to have a perpetually-idle CI cluster, which overall would reduce costs. It also has no impact on test time, since one test already boots GKE.
test
use dedicated gke cluster for each ci run we manage most tests in dedicated namespaces colliding tests have been a massive pain lately we already boot a gke cluster for one of the tests so we might as well pull that out and have every test for a given ci invocation run inside it this would also remove our need to have a perpetually idle ci cluster which overall would reduce costs it also has no impact on test time since one test already boots gke
1
244,925
26,490,450,109
IssuesEvent
2023-01-17 22:05:49
brave/brave-ios
https://api.github.com/repos/brave/brave-ios
closed
Brave iOS 1.33 "Block Scripts" doesn't block js on first page load
sec-high Epic: Security
### Description: I have Brave configured to always be in Private Browsing Only mode and with all shields enabled, including "Block Scripts". I have a Twitter URL bookmarked (not logged in due to Private Browsing always being enabled), https://mobile.twitter.com/thegrugq - previously I would have to tap the shield and disable "Block Scripts". As of the most recent update, when I tap the bookmark for thegrugq Twitter page, it loads without JS being blocked! Then if I tap refresh the JS is blocked as expected. ### Steps to Reproduce 1. Private Browsing Only mode enabled 2. Block Scripts enabled 3. https://mobile.twitter.com/thegrugq bookmarked 4. Go to the bookmark 5. Site loads normally, unexpected 6. Reload, JS is blocked as expected **Actual result:** Site loads normally although scripts should be blocked. **Expected result:** Site should not load (as it is JS based) until Block Scripts is disabled. **Reproduces how often:** Easily Reproduced **Brave Version:** 1.33 (21.12.14.16) **Device details:** iPhone SE 1 iOS 15.2 **Website problems only:** - did you check with Brave Shields down? - did you check in Safari/Firefox (WkWebView-based browsers)? ### Additional Information Doesn't seem to affect other bookmarks, e.g. I have https://www.bloomberg.com/quote/XAUAUD:CUR bookmarked and JS is blocked as expected there.
True
Brave iOS 1.33 "Block Scripts" doesn't block js on first page load - ### Description: I have Brave configured to always be in Private Browsing Only mode and with all shields enabled, including "Block Scripts". I have a Twitter URL bookmarked (not logged in due to Private Browsing always being enabled), https://mobile.twitter.com/thegrugq - previously I would have to tap the shield and disable "Block Scripts". As of the most recent update, when I tap the bookmark for thegrugq Twitter page, it loads without JS being blocked! Then if I tap refresh the JS is blocked as expected. ### Steps to Reproduce 1. Private Browsing Only mode enabled 2. Block Scripts enabled 3. https://mobile.twitter.com/thegrugq bookmarked 4. Go to the bookmark 5. Site loads normally, unexpected 6. Reload, JS is blocked as expected **Actual result:** Site loads normally although scripts should be blocked. **Expected result:** Site should not load (as it is JS based) until Block Scripts is disabled. **Reproduces how often:** Easily Reproduced **Brave Version:** 1.33 (21.12.14.16) **Device details:** iPhone SE 1 iOS 15.2 **Website problems only:** - did you check with Brave Shields down? - did you check in Safari/Firefox (WkWebView-based browsers)? ### Additional Information Doesn't seem to affect other bookmarks, e.g. I have https://www.bloomberg.com/quote/XAUAUD:CUR bookmarked and JS is blocked as expected there.
non_test
brave ios block scripts doesn t block js on first page load description i have brave configured to always be in private browsing only mode and with all shields enabled including block scripts i have a twitter url bookmarked not logged in due to private browsing always being enabled previously i would have to tap the shield and disable block scripts as of the most recent update when i tap the bookmark for thegrugq twitter page it loads without js being blocked then if i tap refresh the js is blocked as expected steps to reproduce private browsing only mode enabled block scripts enabled bookmarked go to the bookmark site loads normally unexpected reload js is blocked as expected actual result site loads normally although scripts should be blocked expected result site should not load as it is js based until block scripts is disabled reproduces how often easily reproduced brave version device details iphone se ios website problems only did you check with brave shields down did you check in safari firefox wkwebview based browsers additional information doesn t seem to affect other bookmarks e g i have bookmarked and js is blocked as expected there
0
332,328
29,239,788,811
IssuesEvent
2023-05-23 01:02:20
stakwork/sphinx-nav-fiber
https://api.github.com/repos/stakwork/sphinx-nav-fiber
closed
Write unit/e2e tests on the add source button
Automated Testing
We want to unit test this button on the bottom right ![Screen Shot 2023-05-03 at 1 30 39 PM](https://user-images.githubusercontent.com/15950706/236010963-ce459d2c-11d2-40c2-bf69-36ee64161046.png)
1.0
Write unit/e2e tests on the add source button - We want to unit test this button on the bottom right ![Screen Shot 2023-05-03 at 1 30 39 PM](https://user-images.githubusercontent.com/15950706/236010963-ce459d2c-11d2-40c2-bf69-36ee64161046.png)
test
write unit tests on the add source button we want to unit test this button on the bottom right
1
338,489
30,301,594,666
IssuesEvent
2023-07-10 06:28:39
quarkusio/quarkus
https://api.github.com/repos/quarkusio/quarkus
closed
Support Constructor Injection in @QuarkusTest
kind/enhancement area/testing
### Description Currently I tried to inject beans via constructor like declaring fields in a general Kotlin class, it does not work as expected. I have tried the following, both are failed. ```kotlin @QuarkusTest class MyTest(val posts: PostRepository) {} ``` ```kotlin @QuarkusTest class MyTest @Inject constructor(val posts: PostRepository) {} ``` ### Implementation ideas I hope it is working like a simple class with constructor injection automaticially. ```kotlin @QuarkusTest class MyTest(val posts: PostRepository) {} ```
1.0
Support Constructor Injection in @QuarkusTest - ### Description Currently I tried to inject beans via constructor like declaring fields in a general Kotlin class, it does not work as expected. I have tried the following, both are failed. ```kotlin @QuarkusTest class MyTest(val posts: PostRepository) {} ``` ```kotlin @QuarkusTest class MyTest @Inject constructor(val posts: PostRepository) {} ``` ### Implementation ideas I hope it is working like a simple class with constructor injection automaticially. ```kotlin @QuarkusTest class MyTest(val posts: PostRepository) {} ```
test
support constructor injection in quarkustest description currently i tried to inject beans via constructor like declaring fields in a general kotlin class it does not work as expected i have tried the following both are failed kotlin quarkustest class mytest val posts postrepository kotlin quarkustest class mytest inject constructor val posts postrepository implementation ideas i hope it is working like a simple class with constructor injection automaticially kotlin quarkustest class mytest val posts postrepository
1
178,948
30,039,129,624
IssuesEvent
2023-06-27 14:29:25
FRONTENDSCHOOL5/final-25-would-you
https://api.github.com/repos/FRONTENDSCHOOL5/final-25-would-you
closed
[๐Ÿ’„design] ๊ณตํ†ตUI(header) ๋งˆํฌ์—…
design
## ์ž‘์—…ํ•  ๋‚ด์šฉ - ๊ณตํ†ตUI(header) ๋งˆํฌ์—… ## ๊ตฌํ˜„ ๋ฐฉ๋ฒ• ๋ฐ ์˜ˆ์ƒ ๊ฒฐ๊ณผ - ๋งˆํฌ์—… ์ถ”๊ฐ€ ## ํŠน์ด์‚ฌํ•ญ - feat_#9_header ๋ธŒ๋žœ์น˜ ์ƒ์„ฑํ•ด์„œ ๋”ฐ๋กœ ์ž‘์—…
1.0
[๐Ÿ’„design] ๊ณตํ†ตUI(header) ๋งˆํฌ์—… - ## ์ž‘์—…ํ•  ๋‚ด์šฉ - ๊ณตํ†ตUI(header) ๋งˆํฌ์—… ## ๊ตฌํ˜„ ๋ฐฉ๋ฒ• ๋ฐ ์˜ˆ์ƒ ๊ฒฐ๊ณผ - ๋งˆํฌ์—… ์ถ”๊ฐ€ ## ํŠน์ด์‚ฌํ•ญ - feat_#9_header ๋ธŒ๋žœ์น˜ ์ƒ์„ฑํ•ด์„œ ๋”ฐ๋กœ ์ž‘์—…
non_test
๊ณตํ†ตui header ๋งˆํฌ์—… ์ž‘์—…ํ•  ๋‚ด์šฉ ๊ณตํ†ตui header ๋งˆํฌ์—… ๊ตฌํ˜„ ๋ฐฉ๋ฒ• ๋ฐ ์˜ˆ์ƒ ๊ฒฐ๊ณผ ๋งˆํฌ์—… ์ถ”๊ฐ€ ํŠน์ด์‚ฌํ•ญ feat header ๋ธŒ๋žœ์น˜ ์ƒ์„ฑํ•ด์„œ ๋”ฐ๋กœ ์ž‘์—…
0
258,359
22,306,546,833
IssuesEvent
2022-06-13 13:30:42
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
[pkg/stanza] helper/multiline: add "integration" tests
help wanted test coverage
Multiline helper started to be logically complicated and it would be nice to have tests which helps to understand usecase with force flushing logs. This comes from the open-telemetry/opentelemetry-log-collection#434 [link to comment](https://github.com/open-telemetry/opentelemetry-log-collection/pull/434#pullrequestreview-916099415) > I think there are probably two things that will be necessary: > > 1. Create a custom implementation of io.Reader that behaves predictably in our test cases. Probably it takes a slice of timestamps and corresponding []byte that are written at those times. > 1. Make Flusher use time.Now and/or time.Since as variables so the test Reader can manipulate them in test code. ([Similar example](https://github.com/open-telemetry/opentelemetry-log-collection/blob/01ffa4f428908ff5e7e4b4d79d5dcac54ad643a2/operator/helper/time_test.go#L33)) How unit tests work now ([link to comment](https://github.com/open-telemetry/opentelemetry-log-collection/pull/434#discussion_r830792913)): > This should be considered in file input operator point of view. File input operator reads data from file and run it against splitFunc. If there is some tokens to emit, it emits, otherwise it ends with some data unflushed. Filelog input operator is able to keep track pointer to last data being flushed. So in next iteration it takes data from that point (meantime can be more data added to the end of the file) and run against splitFunc. If there is no new data added to the end of the file for forcePeriod, flusher force to flush it. > > This is hardly for me to explain, so I will add list just in case: > > Filelog input operator: > > 1. Read file from previous pointer position (0 for start_at: beginning, EOF for start_at: end, pointer position if already in loop > 1. Run against splitFunc and update pointer position > a. return tokens if any (will update pointer position to the end of last token) > b. force flush if needed (no change since forcePeriod) (will update pointer position at the EOF) > c. wait for more data (will do not update pointer position) > 1. go to 1 I don't see a possibility to test full flow (1-3) in one unit test. That's why kind of "integration" test is needed
1.0
[pkg/stanza] helper/multiline: add "integration" tests - Multiline helper started to be logically complicated and it would be nice to have tests which helps to understand usecase with force flushing logs. This comes from the open-telemetry/opentelemetry-log-collection#434 [link to comment](https://github.com/open-telemetry/opentelemetry-log-collection/pull/434#pullrequestreview-916099415) > I think there are probably two things that will be necessary: > > 1. Create a custom implementation of io.Reader that behaves predictably in our test cases. Probably it takes a slice of timestamps and corresponding []byte that are written at those times. > 1. Make Flusher use time.Now and/or time.Since as variables so the test Reader can manipulate them in test code. ([Similar example](https://github.com/open-telemetry/opentelemetry-log-collection/blob/01ffa4f428908ff5e7e4b4d79d5dcac54ad643a2/operator/helper/time_test.go#L33)) How unit tests work now ([link to comment](https://github.com/open-telemetry/opentelemetry-log-collection/pull/434#discussion_r830792913)): > This should be considered in file input operator point of view. File input operator reads data from file and run it against splitFunc. If there is some tokens to emit, it emits, otherwise it ends with some data unflushed. Filelog input operator is able to keep track pointer to last data being flushed. So in next iteration it takes data from that point (meantime can be more data added to the end of the file) and run against splitFunc. If there is no new data added to the end of the file for forcePeriod, flusher force to flush it. > > This is hardly for me to explain, so I will add list just in case: > > Filelog input operator: > > 1. Read file from previous pointer position (0 for start_at: beginning, EOF for start_at: end, pointer position if already in loop > 1. Run against splitFunc and update pointer position > a. return tokens if any (will update pointer position to the end of last token) > b. force flush if needed (no change since forcePeriod) (will update pointer position at the EOF) > c. wait for more data (will do not update pointer position) > 1. go to 1 I don't see a possibility to test full flow (1-3) in one unit test. That's why kind of "integration" test is needed
test
helper multiline add integration tests multiline helper started to be logically complicated and it would be nice to have tests which helps to understand usecase with force flushing logs this comes from the open telemetry opentelemetry log collection i think there are probably two things that will be necessary create a custom implementation of io reader that behaves predictably in our test cases probably it takes a slice of timestamps and corresponding byte that are written at those times make flusher use time now and or time since as variables so the test reader can manipulate them in test code how unit tests work now this should be considered in file input operator point of view file input operator reads data from file and run it against splitfunc if there is some tokens to emit it emits otherwise it ends with some data unflushed filelog input operator is able to keep track pointer to last data being flushed so in next iteration it takes data from that point meantime can be more data added to the end of the file and run against splitfunc if there is no new data added to the end of the file for forceperiod flusher force to flush it this is hardly for me to explain so i will add list just in case filelog input operator read file from previous pointer position for start at beginning eof for start at end pointer position if already in loop run against splitfunc and update pointer position a return tokens if any will update pointer position to the end of last token b force flush if needed no change since forceperiod will update pointer position at the eof c wait for more data will do not update pointer position go to i don t see a possibility to test full flow in one unit test that s why kind of integration test is needed
1
224,752
17,773,083,062
IssuesEvent
2021-08-30 15:44:50
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
ccl/serverccl: TestTenantGRPCServices failed
C-test-failure O-robot branch-master T-observability-inf
ccl/serverccl.TestTenantGRPCServices [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3296858&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3296858&tab=artifacts#/) on master @ [3928f1963833fbf51ae47bd2a42ae6a200ebbb14](https://github.com/cockroachdb/cockroach/commits/3928f1963833fbf51ae47bd2a42ae6a200ebbb14): ``` === RUN TestTenantGRPCServices test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestTenantGRPCServices582200065 test_log_scope.go:74: use -show-logs to present logs inline === CONT TestTenantGRPCServices tenant_grpc_test.go:152: -- test log scope end -- test logs left over in: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestTenantGRPCServices582200065 --- FAIL: TestTenantGRPCServices (20.23s) === RUN TestTenantGRPCServices/statements_endpoint_fans_out_request_to_multiple_pods tenant_grpc_test.go:93: Error Trace: tenant_grpc_test.go:93 Error: Received unexpected error: Get "http://127.0.0.1:38797/_status/statements": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Test: TestTenantGRPCServices/statements_endpoint_fans_out_request_to_multiple_pods --- FAIL: TestTenantGRPCServices/statements_endpoint_fans_out_request_to_multiple_pods (3.00s) ``` <details><summary>Reproduce</summary> <p> To reproduce, try: ```bash make stressrace TESTS=TestTenantGRPCServices PKG=./pkg/ccl/serverccl TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1 ``` Parameters in this failure: - GOFLAGS=-race -parallel=4 </p> </details> /cc @cockroachdb/server dhartunian <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestTenantGRPCServices.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
1.0
ccl/serverccl: TestTenantGRPCServices failed - ccl/serverccl.TestTenantGRPCServices [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3296858&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3296858&tab=artifacts#/) on master @ [3928f1963833fbf51ae47bd2a42ae6a200ebbb14](https://github.com/cockroachdb/cockroach/commits/3928f1963833fbf51ae47bd2a42ae6a200ebbb14): ``` === RUN TestTenantGRPCServices test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestTenantGRPCServices582200065 test_log_scope.go:74: use -show-logs to present logs inline === CONT TestTenantGRPCServices tenant_grpc_test.go:152: -- test log scope end -- test logs left over in: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestTenantGRPCServices582200065 --- FAIL: TestTenantGRPCServices (20.23s) === RUN TestTenantGRPCServices/statements_endpoint_fans_out_request_to_multiple_pods tenant_grpc_test.go:93: Error Trace: tenant_grpc_test.go:93 Error: Received unexpected error: Get "http://127.0.0.1:38797/_status/statements": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Test: TestTenantGRPCServices/statements_endpoint_fans_out_request_to_multiple_pods --- FAIL: TestTenantGRPCServices/statements_endpoint_fans_out_request_to_multiple_pods (3.00s) ``` <details><summary>Reproduce</summary> <p> To reproduce, try: ```bash make stressrace TESTS=TestTenantGRPCServices PKG=./pkg/ccl/serverccl TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1 ``` Parameters in this failure: - GOFLAGS=-race -parallel=4 </p> </details> /cc @cockroachdb/server dhartunian <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestTenantGRPCServices.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
ccl serverccl testtenantgrpcservices failed ccl serverccl testtenantgrpcservices with on master run testtenantgrpcservices test log scope go test logs captured to go src github com cockroachdb cockroach artifacts test log scope go use show logs to present logs inline cont testtenantgrpcservices tenant grpc test go test log scope end test logs left over in go src github com cockroachdb cockroach artifacts fail testtenantgrpcservices run testtenantgrpcservices statements endpoint fans out request to multiple pods tenant grpc test go error trace tenant grpc test go error received unexpected error get context deadline exceeded client timeout exceeded while awaiting headers test testtenantgrpcservices statements endpoint fans out request to multiple pods fail testtenantgrpcservices statements endpoint fans out request to multiple pods reproduce to reproduce try bash make stressrace tests testtenantgrpcservices pkg pkg ccl serverccl testtimeout stressflags timeout parameters in this failure goflags race parallel cc cockroachdb server dhartunian
1
121,943
12,137,649,460
IssuesEvent
2020-04-23 16:01:09
oasis-tcs/sarif-spec
https://api.github.com/repos/oasis-tcs/sarif-spec
opened
Clarify that offset/length and line/column region properties need to stand alone
future impact-documentation-only
The spec says that there are two ways to specify a text region: - With line/column properties. - With character offset/length properties. It does not _explicitly_ state that there is no third option (a mix of line/column and character/offset properties), although that constraint is implicit because it says "A text region **MAY** be specified in two ways", not "... in three ways". Add the explicit statement as a NOTE.
1.0
Clarify that offset/length and line/column region properties need to stand alone - The spec says that there are two ways to specify a text region: - With line/column properties. - With character offset/length properties. It does not _explicitly_ state that there is no third option (a mix of line/column and character/offset properties), although that constraint is implicit because it says "A text region **MAY** be specified in two ways", not "... in three ways". Add the explicit statement as a NOTE.
non_test
clarify that offset length and line column region properties need to stand alone the spec says that there are two ways to specify a text region with line column properties with character offset length properties it does not explicitly state that there is no third option a mix of line column and character offset properties although that constraint is implicit because it says a text region may be specified in two ways not in three ways add the explicit statement as a note
0
310,610
26,727,106,707
IssuesEvent
2023-01-29 21:02:36
python/cpython
https://api.github.com/repos/python/cpython
closed
`ast.NodeTransformer` is not tested
type-bug tests
There are no tests for `ast.NodeTransformer`: ``` ยป ag NodeTransformer Misc/NEWS.d/3.9.0a3.rst 817:In the :mod:`ast` module documentation, fix a misleading ``NodeTransformer`` Lib/inspect.py 2230: class RewriteSymbolics(ast.NodeTransformer): Lib/ast.py 411: traversing. For this a special visitor exists (`NodeTransformer`) that 453:class NodeTransformer(NodeVisitor): 458: The `NodeTransformer` will walk the AST and use the return value of the 467: class RewriteName(NodeTransformer): Doc/library/ast.rst 2122: (:class:`NodeTransformer`) that allows modifications. 2132:.. class:: NodeTransformer() 2137: The :class:`NodeTransformer` will walk the AST and use the return value of 2146: class RewriteName(NodeTransformer): 2163: If :class:`NodeTransformer` introduces new nodes (that weren't part of Doc/whatsnew/2.6.rst 2740::class:`NodeTransformer` classes for traversing and modifying an AST, ``` I think we should add at least some high-level tests that will check the desired behaviour. I will send a PR today. <!-- gh-linked-prs --> ### Linked PRs * gh-100521 <!-- /gh-linked-prs -->
1.0
`ast.NodeTransformer` is not tested - There are no tests for `ast.NodeTransformer`: ``` ยป ag NodeTransformer Misc/NEWS.d/3.9.0a3.rst 817:In the :mod:`ast` module documentation, fix a misleading ``NodeTransformer`` Lib/inspect.py 2230: class RewriteSymbolics(ast.NodeTransformer): Lib/ast.py 411: traversing. For this a special visitor exists (`NodeTransformer`) that 453:class NodeTransformer(NodeVisitor): 458: The `NodeTransformer` will walk the AST and use the return value of the 467: class RewriteName(NodeTransformer): Doc/library/ast.rst 2122: (:class:`NodeTransformer`) that allows modifications. 2132:.. class:: NodeTransformer() 2137: The :class:`NodeTransformer` will walk the AST and use the return value of 2146: class RewriteName(NodeTransformer): 2163: If :class:`NodeTransformer` introduces new nodes (that weren't part of Doc/whatsnew/2.6.rst 2740::class:`NodeTransformer` classes for traversing and modifying an AST, ``` I think we should add at least some high-level tests that will check the desired behaviour. I will send a PR today. <!-- gh-linked-prs --> ### Linked PRs * gh-100521 <!-- /gh-linked-prs -->
test
ast nodetransformer is not tested there are no tests for ast nodetransformer ยป ag nodetransformer misc news d rst in the mod ast module documentation fix a misleading nodetransformer lib inspect py class rewritesymbolics ast nodetransformer lib ast py traversing for this a special visitor exists nodetransformer that class nodetransformer nodevisitor the nodetransformer will walk the ast and use the return value of the class rewritename nodetransformer doc library ast rst class nodetransformer that allows modifications class nodetransformer the class nodetransformer will walk the ast and use the return value of class rewritename nodetransformer if class nodetransformer introduces new nodes that weren t part of doc whatsnew rst class nodetransformer classes for traversing and modifying an ast i think we should add at least some high level tests that will check the desired behaviour i will send a pr today linked prs gh
1
4,750
7,610,710,972
IssuesEvent
2018-05-01 09:56:26
TeamPotry/tutorial_text
https://api.github.com/repos/TeamPotry/tutorial_text
opened
ํŠœํ† ๋ฆฌ์–ผ ํ…์ŠคํŠธ์˜ ๊ตฌ์กฐ ๋ณ€๊ฒฝ
enhancement in_process
- [ ] **๊ธฐ์กด์˜ ํ•˜๋‚˜์˜ ์ปจํ”ฝ๋งŒ ์‚ฌ์šฉํ•˜๋˜ ๋กœ์ง์„ ๋ณ€๊ฒฝ** `configs/tutorial_text` ํด๋” ๋‚ด์˜ ๋ชจ๋“  ํ…์ŠคํŠธ ํŒŒ์ผ์„ ์ฝ๊ณ  ํŒŒ์ผ ์ด๋ฆ„์„ ๊ด‘์—ญ `stringmap`์— ์ €์žฅ ๊ธฐ์กด์— ํ…Œ์ŠคํŠธ ๋ฉ”๋‰ด์— ์‚ฌ์šฉ๋ฌ๋˜ ์ปจํ”ฝ์€ ๊ทธ๋ƒฅ ํ…Œ์ŠคํŠธ์šฉ์œผ๋กœ ์กด์žฌํ•˜๋„๋ก ๋ณ€๊ฒฝ - [ ] **`TTextKeyValue (KeyValues ์ƒ์†)` ์„ค๊ณ„ ๋ฐ ๊ธฐ์กด `KeyValues` ๊ด€๋ จ ์Šคํ†ก ์ด์ „** - [ ] **์ฟ ํ‚ค ๊ตฌ์กฐ ๋ณ€๊ฒฝ** ํ˜„์žฌ: `messageId` ๊ธฐ์ค€ ๊ตฌ์กฐ ๋ณ€๊ฒฝ: `filename_messageId` ๊ธฐ์ค€
1.0
ํŠœํ† ๋ฆฌ์–ผ ํ…์ŠคํŠธ์˜ ๊ตฌ์กฐ ๋ณ€๊ฒฝ - - [ ] **๊ธฐ์กด์˜ ํ•˜๋‚˜์˜ ์ปจํ”ฝ๋งŒ ์‚ฌ์šฉํ•˜๋˜ ๋กœ์ง์„ ๋ณ€๊ฒฝ** `configs/tutorial_text` ํด๋” ๋‚ด์˜ ๋ชจ๋“  ํ…์ŠคํŠธ ํŒŒ์ผ์„ ์ฝ๊ณ  ํŒŒ์ผ ์ด๋ฆ„์„ ๊ด‘์—ญ `stringmap`์— ์ €์žฅ ๊ธฐ์กด์— ํ…Œ์ŠคํŠธ ๋ฉ”๋‰ด์— ์‚ฌ์šฉ๋ฌ๋˜ ์ปจํ”ฝ์€ ๊ทธ๋ƒฅ ํ…Œ์ŠคํŠธ์šฉ์œผ๋กœ ์กด์žฌํ•˜๋„๋ก ๋ณ€๊ฒฝ - [ ] **`TTextKeyValue (KeyValues ์ƒ์†)` ์„ค๊ณ„ ๋ฐ ๊ธฐ์กด `KeyValues` ๊ด€๋ จ ์Šคํ†ก ์ด์ „** - [ ] **์ฟ ํ‚ค ๊ตฌ์กฐ ๋ณ€๊ฒฝ** ํ˜„์žฌ: `messageId` ๊ธฐ์ค€ ๊ตฌ์กฐ ๋ณ€๊ฒฝ: `filename_messageId` ๊ธฐ์ค€
non_test
ํŠœํ† ๋ฆฌ์–ผ ํ…์ŠคํŠธ์˜ ๊ตฌ์กฐ ๋ณ€๊ฒฝ ๊ธฐ์กด์˜ ํ•˜๋‚˜์˜ ์ปจํ”ฝ๋งŒ ์‚ฌ์šฉํ•˜๋˜ ๋กœ์ง์„ ๋ณ€๊ฒฝ configs tutorial text ํด๋” ๋‚ด์˜ ๋ชจ๋“  ํ…์ŠคํŠธ ํŒŒ์ผ์„ ์ฝ๊ณ  ํŒŒ์ผ ์ด๋ฆ„์„ ๊ด‘์—ญ stringmap ์— ์ €์žฅ ๊ธฐ์กด์— ํ…Œ์ŠคํŠธ ๋ฉ”๋‰ด์— ์‚ฌ์šฉ๋ฌ๋˜ ์ปจํ”ฝ์€ ๊ทธ๋ƒฅ ํ…Œ์ŠคํŠธ์šฉ์œผ๋กœ ์กด์žฌํ•˜๋„๋ก ๋ณ€๊ฒฝ ttextkeyvalue keyvalues ์ƒ์† ์„ค๊ณ„ ๋ฐ ๊ธฐ์กด keyvalues ๊ด€๋ จ ์Šคํ†ก ์ด์ „ ์ฟ ํ‚ค ๊ตฌ์กฐ ๋ณ€๊ฒฝ ํ˜„์žฌ messageid ๊ธฐ์ค€ ๊ตฌ์กฐ ๋ณ€๊ฒฝ filename messageid ๊ธฐ์ค€
0
205,865
15,695,870,425
IssuesEvent
2021-03-26 00:39:10
backend-br/vagas
https://api.github.com/repos/backend-br/vagas
closed
[Remoto] Back-end Developer @ SupplyLabs
CI Docker Full Time Git GoLang GraphQL Kubernetes Linux PJ Pleno Remoto Stale Testes automatizados
## Descriรงรฃo da vaga A SupplyLabs foi criada com o propรณsito de entregar uma experiรชncia 100% digital para gestรฃo de cadeia de suprimentos e ESG. Estamos ร  procura de uma pessoa que irรก integrar o time de desenvolvimento de produtos e serรก responsรกvel pela evoluรงรฃo e manutenรงรฃo de nossa API e Serviรงos. Vocรช serรก responsรกvel por: Codificar APIs e Serviรงos modulares, desacoplados e escalรกveis; Participar de discussรตes para levantamento de hipรณteses de soluรงรฃo, nos ajudando a pensar novos produtos e novos mรณdulos. Testar e publicar o software desenvolvido em ambientes de staging e produรงรฃo; Garantir a performance e o monitoramento dos serviรงos; Participar ativamente de decisรตes de tecnologias, infraestrutura, etc. ## Local Remoto #### Diferenciais Ambiente descontraรญdo de trabalho e em plena expansรฃo; Full time home-office; ## Requisitos **Obrigatรณrios:** Ter trabalhado em ao menos um ano em um projeto utilizando Golang (imprescindรญvel)! Saiba pensar APIs para os diversos tipos de frontend. Saiba versionar cรณdigo e interagir com GIT. Que tenha noรงรตes de testes automatizados em APIs. Possuir conhecimentos intermediรกrios em servidores Linux. Gostar de trabalhar em grupo, ter empatia e humildade para liderar discussรตes. Ser curioso e estudioso. Ter lado analรญtico, preocupado com a ciรชncia por trรกs das decisรตes. **Diferenciais:** Ter atuado em outras startups รฉ um diferencial. Ter trabalhado com GraphQL/ gRPC/ Docker ou Kubernetes serรก um diferencial. ## Contrataรงรฃo PJ a combinar ## Nossa empresa A SupplyLabs foi criada com o propรณsito de entregar uma experiรชncia 100% para gestรฃo de cadeia de suprimentos. Somos uma empresa jovem, 100% focada em tecnologia, utilizando uma stack de ponta, que cresceu 100% no รบltimo ano. Nosso time รฉ composto por gente bacana, qualificada e amigรกvel. Nossos desafios como plataforma de software sรฃo imensos, desde qualidade de software, escalabilidade e reproveitamento de cรณdigo, anรกlise de dados, entre outros. Como time, esperamos que os membros sejam disciplinados, colaborativos, humildes e honestos. Esta รฉ uma posiรงรฃo que exige dedicaรงรฃo exclusiva, e disponibilidade integral durante nosso horรกrio de trabalho. ## Como se candidatar Envie um e-mail para rh@supplylabs.com.br, com assunto: Vaga Desenvolvedor Golang; Inclua portfolio, repositรณrio Github, currรญculo, Linkedin e todo material que demonstre seu nรญvel tรฉcnico e histรณrico profissional e acadรชmico; Conte na mensagem o quรช lhe interessou na vaga; Informe sua pretensรฃo salarial; ## Processo Seletivo 100% online; Anรกlise do material enviado ou referenciado no e-mail de candidatura; Conversa para conhecer a(o) candidata(o), e para apresentar a empresa; Conversa sobre engenharia de software; Teste de codificaรงรฃo; Conversa com os gestores e o time dev; ## Tempo mรฉdio de feedbacks Costumamos enviar feedbacks em atรฉ 15 dias apรณs cada processo. E-mail para contato em caso de nรฃo haver resposta: rh@supplylabs.com.br ## Labels - PJ - Pleno - Remoto
1.0
[Remoto] Back-end Developer @ SupplyLabs - ## Descriรงรฃo da vaga A SupplyLabs foi criada com o propรณsito de entregar uma experiรชncia 100% digital para gestรฃo de cadeia de suprimentos e ESG. Estamos ร  procura de uma pessoa que irรก integrar o time de desenvolvimento de produtos e serรก responsรกvel pela evoluรงรฃo e manutenรงรฃo de nossa API e Serviรงos. Vocรช serรก responsรกvel por: Codificar APIs e Serviรงos modulares, desacoplados e escalรกveis; Participar de discussรตes para levantamento de hipรณteses de soluรงรฃo, nos ajudando a pensar novos produtos e novos mรณdulos. Testar e publicar o software desenvolvido em ambientes de staging e produรงรฃo; Garantir a performance e o monitoramento dos serviรงos; Participar ativamente de decisรตes de tecnologias, infraestrutura, etc. ## Local Remoto #### Diferenciais Ambiente descontraรญdo de trabalho e em plena expansรฃo; Full time home-office; ## Requisitos **Obrigatรณrios:** Ter trabalhado em ao menos um ano em um projeto utilizando Golang (imprescindรญvel)! Saiba pensar APIs para os diversos tipos de frontend. Saiba versionar cรณdigo e interagir com GIT. Que tenha noรงรตes de testes automatizados em APIs. Possuir conhecimentos intermediรกrios em servidores Linux. Gostar de trabalhar em grupo, ter empatia e humildade para liderar discussรตes. Ser curioso e estudioso. Ter lado analรญtico, preocupado com a ciรชncia por trรกs das decisรตes. **Diferenciais:** Ter atuado em outras startups รฉ um diferencial. Ter trabalhado com GraphQL/ gRPC/ Docker ou Kubernetes serรก um diferencial. ## Contrataรงรฃo PJ a combinar ## Nossa empresa A SupplyLabs foi criada com o propรณsito de entregar uma experiรชncia 100% para gestรฃo de cadeia de suprimentos. Somos uma empresa jovem, 100% focada em tecnologia, utilizando uma stack de ponta, que cresceu 100% no รบltimo ano. Nosso time รฉ composto por gente bacana, qualificada e amigรกvel. Nossos desafios como plataforma de software sรฃo imensos, desde qualidade de software, escalabilidade e reproveitamento de cรณdigo, anรกlise de dados, entre outros. Como time, esperamos que os membros sejam disciplinados, colaborativos, humildes e honestos. Esta รฉ uma posiรงรฃo que exige dedicaรงรฃo exclusiva, e disponibilidade integral durante nosso horรกrio de trabalho. ## Como se candidatar Envie um e-mail para rh@supplylabs.com.br, com assunto: Vaga Desenvolvedor Golang; Inclua portfolio, repositรณrio Github, currรญculo, Linkedin e todo material que demonstre seu nรญvel tรฉcnico e histรณrico profissional e acadรชmico; Conte na mensagem o quรช lhe interessou na vaga; Informe sua pretensรฃo salarial; ## Processo Seletivo 100% online; Anรกlise do material enviado ou referenciado no e-mail de candidatura; Conversa para conhecer a(o) candidata(o), e para apresentar a empresa; Conversa sobre engenharia de software; Teste de codificaรงรฃo; Conversa com os gestores e o time dev; ## Tempo mรฉdio de feedbacks Costumamos enviar feedbacks em atรฉ 15 dias apรณs cada processo. E-mail para contato em caso de nรฃo haver resposta: rh@supplylabs.com.br ## Labels - PJ - Pleno - Remoto
test
back end developer supplylabs descriรงรฃo da vaga a supplylabs foi criada com o propรณsito de entregar uma experiรชncia digital para gestรฃo de cadeia de suprimentos e esg estamos ร  procura de uma pessoa que irรก integrar o time de desenvolvimento de produtos e serรก responsรกvel pela evoluรงรฃo e manutenรงรฃo de nossa api e serviรงos vocรช serรก responsรกvel por codificar apis e serviรงos modulares desacoplados e escalรกveis participar de discussรตes para levantamento de hipรณteses de soluรงรฃo nos ajudando a pensar novos produtos e novos mรณdulos testar e publicar o software desenvolvido em ambientes de staging e produรงรฃo garantir a performance e o monitoramento dos serviรงos participar ativamente de decisรตes de tecnologias infraestrutura etc local remoto diferenciais ambiente descontraรญdo de trabalho e em plena expansรฃo full time home office requisitos obrigatรณrios ter trabalhado em ao menos um ano em um projeto utilizando golang imprescindรญvel saiba pensar apis para os diversos tipos de frontend saiba versionar cรณdigo e interagir com git que tenha noรงรตes de testes automatizados em apis possuir conhecimentos intermediรกrios em servidores linux gostar de trabalhar em grupo ter empatia e humildade para liderar discussรตes ser curioso e estudioso ter lado analรญtico preocupado com a ciรชncia por trรกs das decisรตes diferenciais ter atuado em outras startups รฉ um diferencial ter trabalhado com graphql grpc docker ou kubernetes serรก um diferencial contrataรงรฃo pj a combinar nossa empresa a supplylabs foi criada com o propรณsito de entregar uma experiรชncia para gestรฃo de cadeia de suprimentos somos uma empresa jovem focada em tecnologia utilizando uma stack de ponta que cresceu no รบltimo ano nosso time รฉ composto por gente bacana qualificada e amigรกvel nossos desafios como plataforma de software sรฃo imensos desde qualidade de software escalabilidade e reproveitamento de cรณdigo anรกlise de dados entre outros como time esperamos que os membros sejam disciplinados colaborativos humildes e honestos esta รฉ uma posiรงรฃo que exige dedicaรงรฃo exclusiva e disponibilidade integral durante nosso horรกrio de trabalho como se candidatar envie um e mail para rh supplylabs com br com assunto vaga desenvolvedor golang inclua portfolio repositรณrio github currรญculo linkedin e todo material que demonstre seu nรญvel tรฉcnico e histรณrico profissional e acadรชmico conte na mensagem o quรช lhe interessou na vaga informe sua pretensรฃo salarial processo seletivo online anรกlise do material enviado ou referenciado no e mail de candidatura conversa para conhecer a o candidata o e para apresentar a empresa conversa sobre engenharia de software teste de codificaรงรฃo conversa com os gestores e o time dev tempo mรฉdio de feedbacks costumamos enviar feedbacks em atรฉ dias apรณs cada processo e mail para contato em caso de nรฃo haver resposta rh supplylabs com br labels pj pleno remoto
1
168,475
13,090,982,999
IssuesEvent
2020-08-03 05:19:35
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
[CI] EnableSecurityOnBasicLicenseIT.testSecuritySetup failure
:Security/Security >test-failure Team:Security
**Build scan**: https://gradle-enterprise.elastic.co/s/ususuu432yrsu **Repro line**: ``` ./gradlew ':x-pack:plugin:security:qa:basic-enable-security:integTestRunner' --tests "org.elasticsearch.xpack.security.EnableSecurityOnBasicLicenseIT.testSecuritySetup" -Dtests.seed=EB6BC4DAFFD9A916 -Dtests.security.manager=true -Dtests.locale=be-BY -Dtests.timezone=America/Araguaina -Druntime.java=8 ``` **Reproduces locally?**: No **Failure history**: It's been failing once in a while in the past 6 months. https://build-stats.elastic.co/app/kibana#/discover?_g=(refreshInterval:(pause:!t,value:0),time:(from:now-6M,mode:quick,to:now))&_a=(columns:!(_source),index:b646ed00-7efc-11e8-bf69-63c8ef516157,interval:auto,query:(language:lucene,query:testSecuritySetup),sort:!(process.time-start,desc)) **Failure excerpt**: ``` org.elasticsearch.xpack.security.EnableSecurityOnBasicLicenseIT > testSecuritySetup FAILED org.elasticsearch.client.ResponseException: method [GET], host [http://[::1]:43207], URI [/_license], status line [HTTP/1.1 404 Not Found] { } at __randomizedtesting.SeedInfo.seed([EB6BC4DAFFD9A916:960DB491A93081CF]:0) at org.elasticsearch.client.RestClient.convertResponse(RestClient.java:302) at org.elasticsearch.client.RestClient.performRequest(RestClient.java:272) at org.elasticsearch.client.RestClient.performRequest(RestClient.java:246) at org.elasticsearch.test.rest.ESRestTestCase.getAsMap(ESRestTestCase.java:1222) at org.elasticsearch.xpack.security.EnableSecurityOnBasicLicenseIT.checkBasicLicenseType(EnableSecurityOnBasicLicenseIT.java:101) at org.elasticsearch.xpack.security.EnableSecurityOnBasicLicenseIT.testSecuritySetup(EnableSecurityOnBasicLicenseIT.java:68) ```
1.0
[CI] EnableSecurityOnBasicLicenseIT.testSecuritySetup failure - **Build scan**: https://gradle-enterprise.elastic.co/s/ususuu432yrsu **Repro line**: ``` ./gradlew ':x-pack:plugin:security:qa:basic-enable-security:integTestRunner' --tests "org.elasticsearch.xpack.security.EnableSecurityOnBasicLicenseIT.testSecuritySetup" -Dtests.seed=EB6BC4DAFFD9A916 -Dtests.security.manager=true -Dtests.locale=be-BY -Dtests.timezone=America/Araguaina -Druntime.java=8 ``` **Reproduces locally?**: No **Failure history**: It's been failing once in a while in the past 6 months. https://build-stats.elastic.co/app/kibana#/discover?_g=(refreshInterval:(pause:!t,value:0),time:(from:now-6M,mode:quick,to:now))&_a=(columns:!(_source),index:b646ed00-7efc-11e8-bf69-63c8ef516157,interval:auto,query:(language:lucene,query:testSecuritySetup),sort:!(process.time-start,desc)) **Failure excerpt**: ``` org.elasticsearch.xpack.security.EnableSecurityOnBasicLicenseIT > testSecuritySetup FAILED org.elasticsearch.client.ResponseException: method [GET], host [http://[::1]:43207], URI [/_license], status line [HTTP/1.1 404 Not Found] { } at __randomizedtesting.SeedInfo.seed([EB6BC4DAFFD9A916:960DB491A93081CF]:0) at org.elasticsearch.client.RestClient.convertResponse(RestClient.java:302) at org.elasticsearch.client.RestClient.performRequest(RestClient.java:272) at org.elasticsearch.client.RestClient.performRequest(RestClient.java:246) at org.elasticsearch.test.rest.ESRestTestCase.getAsMap(ESRestTestCase.java:1222) at org.elasticsearch.xpack.security.EnableSecurityOnBasicLicenseIT.checkBasicLicenseType(EnableSecurityOnBasicLicenseIT.java:101) at org.elasticsearch.xpack.security.EnableSecurityOnBasicLicenseIT.testSecuritySetup(EnableSecurityOnBasicLicenseIT.java:68) ```
test
enablesecurityonbasiclicenseit testsecuritysetup failure build scan repro line gradlew x pack plugin security qa basic enable security integtestrunner tests org elasticsearch xpack security enablesecurityonbasiclicenseit testsecuritysetup dtests seed dtests security manager true dtests locale be by dtests timezone america araguaina druntime java reproduces locally no failure history it s been failing once in a while in the past months failure excerpt org elasticsearch xpack security enablesecurityonbasiclicenseit testsecuritysetup failed org elasticsearch client responseexception method host uri status line at randomizedtesting seedinfo seed at org elasticsearch client restclient convertresponse restclient java at org elasticsearch client restclient performrequest restclient java at org elasticsearch client restclient performrequest restclient java at org elasticsearch test rest esresttestcase getasmap esresttestcase java at org elasticsearch xpack security enablesecurityonbasiclicenseit checkbasiclicensetype enablesecurityonbasiclicenseit java at org elasticsearch xpack security enablesecurityonbasiclicenseit testsecuritysetup enablesecurityonbasiclicenseit java
1
5,347
7,874,754,040
IssuesEvent
2018-06-25 18:06:30
ualbertalib/discovery
https://api.github.com/repos/ualbertalib/discovery
opened
BPSC Form
need-requirements
As a user of the Discovery service, I want to be able to send a request to the Bruce Peel Special Collections to view non-requestable/non-loanable materials held there. It needs to allow me to: - Select where I want to view the material (RCRF or BPSC) - Notify BPSC staff with my contact information so that they can contact me and book an appointment with me for me to come in and view the material - Autopopulate bib-record information into the form. Currently, when user sees a non-loanable/non-requestable item, they can't take any action on the item to proceed with a request. ![image](https://user-images.githubusercontent.com/2047614/41866353-1bcf595a-786d-11e8-90aa-84f7fd44c5a8.png) Assuming that the development work is to add: ![screenshot from 2018-06-25 11-52-14](https://user-images.githubusercontent.com/2047614/41867013-1eced8f4-786f-11e8-971b-a11ded7d192d.png) @redlibrarian could you please verify? BPSC currently has forms for both Retrieval Request Form or Archives Retrieval Request Form: http://bpsclibrarynews.blogspot.com/2017/11/retrieval-request-form.html Requirements need to be verified with BPSC. Questions: 1. Do we need to differentiate different types of materials (Book or Archives) on the item record page, and present them with different forms based on the type of materials. 2. Could we consolidate the forms into one, with auto-populated information?
1.0
BPSC Form - As a user of the Discovery service, I want to be able to send a request to the Bruce Peel Special Collections to view non-requestable/non-loanable materials held there. It needs to allow me to: - Select where I want to view the material (RCRF or BPSC) - Notify BPSC staff with my contact information so that they can contact me and book an appointment with me for me to come in and view the material - Autopopulate bib-record information into the form. Currently, when user sees a non-loanable/non-requestable item, they can't take any action on the item to proceed with a request. ![image](https://user-images.githubusercontent.com/2047614/41866353-1bcf595a-786d-11e8-90aa-84f7fd44c5a8.png) Assuming that the development work is to add: ![screenshot from 2018-06-25 11-52-14](https://user-images.githubusercontent.com/2047614/41867013-1eced8f4-786f-11e8-971b-a11ded7d192d.png) @redlibrarian could you please verify? BPSC currently has forms for both Retrieval Request Form or Archives Retrieval Request Form: http://bpsclibrarynews.blogspot.com/2017/11/retrieval-request-form.html Requirements need to be verified with BPSC. Questions: 1. Do we need to differentiate different types of materials (Book or Archives) on the item record page, and present them with different forms based on the type of materials. 2. Could we consolidate the forms into one, with auto-populated information?
non_test
bpsc form as a user of the discovery service i want to be able to send a request to the bruce peel special collections to view non requestable non loanable materials held there it needs to allow me to select where i want to view the material rcrf or bpsc notify bpsc staff with my contact information so that they can contact me and book an appointment with me for me to come in and view the material autopopulate bib record information into the form currently when user sees a non loanable non requestable item they can t take any action on the item to proceed with a request assuming that the development work is to add redlibrarian could you please verify bpsc currently has forms for both retrieval request form or archives retrieval request form requirements need to be verified with bpsc questions do we need to differentiate different types of materials book or archives on the item record page and present them with different forms based on the type of materials could we consolidate the forms into one with auto populated information
0
320,953
27,494,207,498
IssuesEvent
2023-03-05 00:41:09
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[PL/SQL] [ORACLE] [VISUAL BASIC] [C#] [TAMBร‰M PCD] Analista III de Desenvolvimento de Software na [TOTVS]
SALVADOR DESENVOLVIMENTO DE SOFTWARE C# PL/SQL ORACLE DB TESTES DE SOFTWARE Lร“GICA DE PROGRAMAร‡รƒO VISUAL BASIC HELP WANTED VAGA PARA PCD TAMBร‰M Stale
<!-- ================================================== POR FAVOR, Sร“ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS! Use: "Desenvolvedor Front-end" ao invรฉs de "Front-End Developer" \o/ Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]` ================================================== --> ## Descriรงรฃo da vaga - Nossa tecnologia dita o ritmo do mercado. Afinal, 25% do PIB brasileiro passa pelos softwares presentes em mais de 30 mil empresas clientes da TOTVS. - Para que isso aconteรงa precisamos ser รกgeis, pois as coisas aqui acontecem em uma velocidade acima da mรฉdia. No nosso dia-dia, aprendemos muito e muito rรกpido, em um ambiente leve, divertido e desafiador. - Nosso time de tecnologia tem a ousadia para colocar todo esse aprendizado e conhecimento em prรกtica. Assim, transformando todos os setores da economia por meio da tecnologia. - Topa esse desafio? **Responsabilidades e atribuiรงรตes:** - Desenvolver soluรงรตes tรฉcnicas de mรฉdia e alta complexidade, com baixo nรญvel de supervisรฃo e pouco apoio, atravรฉs do desenvolvimento de requisitos de softwares utilizando-se das tecnologias envolvidas no projeto; - Conhecer a fundo a arquitetura do software e ser capaz de atuar nela atravรฉs da anรกlise de seus componentes, organizaรงรฃo e funcionalidades; - Criar testes de alta complexidade, sejam estes manuais ou automatizados, que cubram o comportamento desejado dos componentes tรฉcnicos desenvolvidos utilizando os requisitos e ferramentas de testes; - Documentar as etapas de desenvolvimento realizando o registro do desenvolvimento nas ferramentas de versionamento e documentaรงรฃo; - Orientar e trabalhar com o time de desenvolvimento para encontrar soluรงรตes de negรณcio, arquitetura e desenho atravรฉs da comunicaรงรฃo e trocas de conhecimento; - Manter a qualidade do software verificando os processos, ferramentas e boas prรกticas disponรญveis para o produto e tecnologia utilizados; - Orientar desenvolvedores de menor senioridade apoiando suas dรบvidas, disseminando seu conhecimento, desenhando e arquitetando soluรงรตes ou indicando documentaรงรฃo de apoio; - Utiliza frameworks jรก definidos para o desenvolvimento de soluรงรตes. ## Local - Salvador ## Benefรญcios - Seguro de Vida - Assistรชncia Mรฉdico-Hospitalar Bradesco - Assistรชncia Odontolรณgica Bradesco - Previdรชncia Privada - Vale Refeiรงรฃo - Vale Transporte - Auxรญlio Creche - PLR - Gympass - Universidade em Rede TOTVS, uma universidade Corporativa com conteรบdos e certificaรงรตes gratuitos para os nossos TOTVERS; - Programa +Saudรกveis, que cuidamos dos nossos TOTVERS com assessoria e aรงรตes voltadas para o bem estar em corpo, mente e finanรงas pessoais; - Programa +Vantagens, a maior rede de descontos da Amรฉrica Latina, exclusivos para nossos TOTVERS; ## Requisitos **Obrigatรณrios:** - Lรณgica de Programaรงรฃo - PL/SQL e banco de dados Oracle โ€“ nรญvel avanรงado; - Linguagem de programaรงรฃo Visual Basic 6.0 e C# - Fundamento em testes de software - nรญvel avanรงado; - Mรฉtodos รกgeis de desenvolvimento de software โ€“ desejรกvel nรญvel intermediรกrio; **Desejรกveis:** - Superior completo em tecnologia e afins (desejรกvel). **Diferenciais:** - Arquitetura web - Conhecimento do framework Angular; - Vivรชncia ou conhecimento em regras de negรณcio da รกrea de Varejo. ## Contrataรงรฃo - a combinar ## Nossa empresa - Aqui na TOTVS a ousadia e a transformaรงรฃo fazem parte da nossa rotina. - Somos a maior empresa de tecnologia do Brasil e nosso poder de transformaรงรฃo faz com que nossos softwares estejam presentes em mais de 40mil empresas no Brasil e no mundo. - O nosso grande segredo para estar sempre um passo a frente do mercado รฉ a ousadia e a transformaรงรฃo que nossos TOTVERS promovem todos os dias. - Ousadia para movimentar 25% do PIB brasileiro por meio dos clientes que usam os nossos softwares. - Ousadia para ser a primeira empresa de tecnologia do Brasil a ser listada na Ibovespa. - Ousadia para fazer da mudanรงa a nossa rotina, com mais de 40 aquisiรงรตes ao longo de nossa histรณria. - Nossa curiosidade nos impulsiona e, o melhor, todo mundo colabora com todo mundo, nรฃo importa quem deu a ideia: todos estรฃo dispostos a ajudar e se envolvem legitimamente para fazer dar certo e entregar resultados. - Tudo isso acontece em um ambiente leve e divertido, mas jรก deixamos claro: a intensidade e a rapidez com que as coisas acontecem aqui รฉ acima da mรฉdia. - Vem fazer parte de tudo isso! ## Como se candidatar - [Clique aqui para se candidatar](https://totvs.gupy.io/jobs/720677?jobBoardSource=gupy_public_page)
1.0
[PL/SQL] [ORACLE] [VISUAL BASIC] [C#] [TAMBร‰M PCD] Analista III de Desenvolvimento de Software na [TOTVS] - <!-- ================================================== POR FAVOR, Sร“ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS! Use: "Desenvolvedor Front-end" ao invรฉs de "Front-End Developer" \o/ Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]` ================================================== --> ## Descriรงรฃo da vaga - Nossa tecnologia dita o ritmo do mercado. Afinal, 25% do PIB brasileiro passa pelos softwares presentes em mais de 30 mil empresas clientes da TOTVS. - Para que isso aconteรงa precisamos ser รกgeis, pois as coisas aqui acontecem em uma velocidade acima da mรฉdia. No nosso dia-dia, aprendemos muito e muito rรกpido, em um ambiente leve, divertido e desafiador. - Nosso time de tecnologia tem a ousadia para colocar todo esse aprendizado e conhecimento em prรกtica. Assim, transformando todos os setores da economia por meio da tecnologia. - Topa esse desafio? **Responsabilidades e atribuiรงรตes:** - Desenvolver soluรงรตes tรฉcnicas de mรฉdia e alta complexidade, com baixo nรญvel de supervisรฃo e pouco apoio, atravรฉs do desenvolvimento de requisitos de softwares utilizando-se das tecnologias envolvidas no projeto; - Conhecer a fundo a arquitetura do software e ser capaz de atuar nela atravรฉs da anรกlise de seus componentes, organizaรงรฃo e funcionalidades; - Criar testes de alta complexidade, sejam estes manuais ou automatizados, que cubram o comportamento desejado dos componentes tรฉcnicos desenvolvidos utilizando os requisitos e ferramentas de testes; - Documentar as etapas de desenvolvimento realizando o registro do desenvolvimento nas ferramentas de versionamento e documentaรงรฃo; - Orientar e trabalhar com o time de desenvolvimento para encontrar soluรงรตes de negรณcio, arquitetura e desenho atravรฉs da comunicaรงรฃo e trocas de conhecimento; - Manter a qualidade do software verificando os processos, ferramentas e boas prรกticas disponรญveis para o produto e tecnologia utilizados; - Orientar desenvolvedores de menor senioridade apoiando suas dรบvidas, disseminando seu conhecimento, desenhando e arquitetando soluรงรตes ou indicando documentaรงรฃo de apoio; - Utiliza frameworks jรก definidos para o desenvolvimento de soluรงรตes. ## Local - Salvador ## Benefรญcios - Seguro de Vida - Assistรชncia Mรฉdico-Hospitalar Bradesco - Assistรชncia Odontolรณgica Bradesco - Previdรชncia Privada - Vale Refeiรงรฃo - Vale Transporte - Auxรญlio Creche - PLR - Gympass - Universidade em Rede TOTVS, uma universidade Corporativa com conteรบdos e certificaรงรตes gratuitos para os nossos TOTVERS; - Programa +Saudรกveis, que cuidamos dos nossos TOTVERS com assessoria e aรงรตes voltadas para o bem estar em corpo, mente e finanรงas pessoais; - Programa +Vantagens, a maior rede de descontos da Amรฉrica Latina, exclusivos para nossos TOTVERS; ## Requisitos **Obrigatรณrios:** - Lรณgica de Programaรงรฃo - PL/SQL e banco de dados Oracle โ€“ nรญvel avanรงado; - Linguagem de programaรงรฃo Visual Basic 6.0 e C# - Fundamento em testes de software - nรญvel avanรงado; - Mรฉtodos รกgeis de desenvolvimento de software โ€“ desejรกvel nรญvel intermediรกrio; **Desejรกveis:** - Superior completo em tecnologia e afins (desejรกvel). **Diferenciais:** - Arquitetura web - Conhecimento do framework Angular; - Vivรชncia ou conhecimento em regras de negรณcio da รกrea de Varejo. ## Contrataรงรฃo - a combinar ## Nossa empresa - Aqui na TOTVS a ousadia e a transformaรงรฃo fazem parte da nossa rotina. - Somos a maior empresa de tecnologia do Brasil e nosso poder de transformaรงรฃo faz com que nossos softwares estejam presentes em mais de 40mil empresas no Brasil e no mundo. - O nosso grande segredo para estar sempre um passo a frente do mercado รฉ a ousadia e a transformaรงรฃo que nossos TOTVERS promovem todos os dias. - Ousadia para movimentar 25% do PIB brasileiro por meio dos clientes que usam os nossos softwares. - Ousadia para ser a primeira empresa de tecnologia do Brasil a ser listada na Ibovespa. - Ousadia para fazer da mudanรงa a nossa rotina, com mais de 40 aquisiรงรตes ao longo de nossa histรณria. - Nossa curiosidade nos impulsiona e, o melhor, todo mundo colabora com todo mundo, nรฃo importa quem deu a ideia: todos estรฃo dispostos a ajudar e se envolvem legitimamente para fazer dar certo e entregar resultados. - Tudo isso acontece em um ambiente leve e divertido, mas jรก deixamos claro: a intensidade e a rapidez com que as coisas acontecem aqui รฉ acima da mรฉdia. - Vem fazer parte de tudo isso! ## Como se candidatar - [Clique aqui para se candidatar](https://totvs.gupy.io/jobs/720677?jobBoardSource=gupy_public_page)
test
analista iii de desenvolvimento de software na por favor sรณ poste se a vaga for para salvador e cidades vizinhas use desenvolvedor front end ao invรฉs de front end developer o exemplo desenvolvedor front end na descriรงรฃo da vaga nossa tecnologia dita o ritmo do mercado afinal do pib brasileiro passa pelos softwares presentes em mais de mil empresas clientes da totvs para que isso aconteรงa precisamos ser รกgeis pois as coisas aqui acontecem em uma velocidade acima da mรฉdia no nosso dia dia aprendemos muito e muito rรกpido em um ambiente leve divertido e desafiador nosso time de tecnologia tem a ousadia para colocar todo esse aprendizado e conhecimento em prรกtica assim transformando todos os setores da economia por meio da tecnologia topa esse desafio responsabilidades e atribuiรงรตes desenvolver soluรงรตes tรฉcnicas de mรฉdia e alta complexidade com baixo nรญvel de supervisรฃo e pouco apoio atravรฉs do desenvolvimento de requisitos de softwares utilizando se das tecnologias envolvidas no projeto conhecer a fundo a arquitetura do software e ser capaz de atuar nela atravรฉs da anรกlise de seus componentes organizaรงรฃo e funcionalidades criar testes de alta complexidade sejam estes manuais ou automatizados que cubram o comportamento desejado dos componentes tรฉcnicos desenvolvidos utilizando os requisitos e ferramentas de testes documentar as etapas de desenvolvimento realizando o registro do desenvolvimento nas ferramentas de versionamento e documentaรงรฃo orientar e trabalhar com o time de desenvolvimento para encontrar soluรงรตes de negรณcio arquitetura e desenho atravรฉs da comunicaรงรฃo e trocas de conhecimento manter a qualidade do software verificando os processos ferramentas e boas prรกticas disponรญveis para o produto e tecnologia utilizados orientar desenvolvedores de menor senioridade apoiando suas dรบvidas disseminando seu conhecimento desenhando e arquitetando soluรงรตes ou indicando documentaรงรฃo de apoio utiliza frameworks jรก definidos para o desenvolvimento de soluรงรตes local salvador benefรญcios seguro de vida assistรชncia mรฉdico hospitalar bradesco assistรชncia odontolรณgica bradesco previdรชncia privada vale refeiรงรฃo vale transporte auxรญlio creche plr gympass universidade em rede totvs uma universidade corporativa com conteรบdos e certificaรงรตes gratuitos para os nossos totvers programa saudรกveis que cuidamos dos nossos totvers com assessoria e aรงรตes voltadas para o bem estar em corpo mente e finanรงas pessoais programa vantagens a maior rede de descontos da amรฉrica latina exclusivos para nossos totvers requisitos obrigatรณrios lรณgica de programaรงรฃo pl sql e banco de dados oracle โ€“ nรญvel avanรงado linguagem de programaรงรฃo visual basic e c fundamento em testes de software nรญvel avanรงado mรฉtodos รกgeis de desenvolvimento de software โ€“ desejรกvel nรญvel intermediรกrio desejรกveis superior completo em tecnologia e afins desejรกvel diferenciais arquitetura web conhecimento do framework angular vivรชncia ou conhecimento em regras de negรณcio da รกrea de varejo contrataรงรฃo a combinar nossa empresa aqui na totvs a ousadia e a transformaรงรฃo fazem parte da nossa rotina somos a maior empresa de tecnologia do brasil e nosso poder de transformaรงรฃo faz com que nossos softwares estejam presentes em mais de empresas no brasil e no mundo o nosso grande segredo para estar sempre um passo a frente do mercado รฉ a ousadia e a transformaรงรฃo que nossos totvers promovem todos os dias ousadia para movimentar do pib brasileiro por meio dos clientes que usam os nossos softwares ousadia para ser a primeira empresa de tecnologia do brasil a ser listada na ibovespa ousadia para fazer da mudanรงa a nossa rotina com mais de aquisiรงรตes ao longo de nossa histรณria nossa curiosidade nos impulsiona e o melhor todo mundo colabora com todo mundo nรฃo importa quem deu a ideia todos estรฃo dispostos a ajudar e se envolvem legitimamente para fazer dar certo e entregar resultados tudo isso acontece em um ambiente leve e divertido mas jรก deixamos claro a intensidade e a rapidez com que as coisas acontecem aqui รฉ acima da mรฉdia vem fazer parte de tudo isso como se candidatar
1
290,615
25,080,429,700
IssuesEvent
2022-11-07 18:49:05
ismail0234/Subnautica-Below-Zero-Multiplayer
https://api.github.com/repos/ismail0234/Subnautica-Below-Zero-Multiplayer
closed
The furniture looks wrong for the player outside.
Bug In Progress Pending Test
When a player builds furniture inside the building, the furniture looks wrong for the player outside.
1.0
The furniture looks wrong for the player outside. - When a player builds furniture inside the building, the furniture looks wrong for the player outside.
test
the furniture looks wrong for the player outside when a player builds furniture inside the building the furniture looks wrong for the player outside
1
324,080
27,776,203,960
IssuesEvent
2023-03-16 17:23:30
BoBAdministration/QA-Bug-Reports
https://api.github.com/repos/BoBAdministration/QA-Bug-Reports
closed
Untextured rocks surrounding Shrines on Forest Island.
graphics bug live branch map bug Fixed-PendingTesting
**Describe the Bug** Rocks surrounding shrines on Forest Island have no texture. **To Reproduce** 1. Log onto any server with Forest Island 2. Enter Spectate 3. Go to shrine 4. Observe rocks. **Expected behavior** Rocks have rocky texture. **Actual behavior** Rocks have no texture. **Screenshots & Video** ![image](https://user-images.githubusercontent.com/77513445/135940211-ed85e97c-520e-47e7-85f3-80a453fcb108.png) **Branch Version** Tester and Live. **Character Information** Any pawn. **Additional Information** Forest Island only.
1.0
Untextured rocks surrounding Shrines on Forest Island. - **Describe the Bug** Rocks surrounding shrines on Forest Island have no texture. **To Reproduce** 1. Log onto any server with Forest Island 2. Enter Spectate 3. Go to shrine 4. Observe rocks. **Expected behavior** Rocks have rocky texture. **Actual behavior** Rocks have no texture. **Screenshots & Video** ![image](https://user-images.githubusercontent.com/77513445/135940211-ed85e97c-520e-47e7-85f3-80a453fcb108.png) **Branch Version** Tester and Live. **Character Information** Any pawn. **Additional Information** Forest Island only.
test
untextured rocks surrounding shrines on forest island describe the bug rocks surrounding shrines on forest island have no texture to reproduce log onto any server with forest island enter spectate go to shrine observe rocks expected behavior rocks have rocky texture actual behavior rocks have no texture screenshots video branch version tester and live character information any pawn additional information forest island only
1
5,459
2,610,188,033
IssuesEvent
2015-02-26 18:59:35
chrsmith/quchuseban
https://api.github.com/repos/chrsmith/quchuseban
opened
่ต„่ฎฏ่‰ฒๆ–‘ๅฆ‚ไฝ•ๆ‰่ƒฝๅŽป้™ค
auto-migrated Priority-Medium Type-Defect
``` ใ€Šๆ‘˜่ฆใ€‹ ๅ“ญ็š„ๆ—ถๅ€™๏ผŒๆˆ‘ไผš้—ญไธŠ็œผ็›ไธ่ฎฉๅฎƒๆตๆณช๏ผ›ๅญค็‹ฌๅฏ‚ๅฏž็š„ๆ—ถๅ€™๏ผŒๆˆ‘๏ฟฝ๏ฟฝ ๏ฟฝ้™้™็š„ๆƒณ็€ๆŸไบบ๏ผ›ไผคๅฟƒ็š„ๆ—ถๅ€™๏ผŒๆˆ‘ไผšๆ‰พไธชๅœฐๆ–น้™้™็š„ๅ‘ๅ‘†๏ผŒ๏ฟฝ ๏ฟฝ๏ฟฝๅŽๅ‘Š่ฏ‰่‡ชๅทฑ๏ผŒ่ฟ˜ๆ˜ฏ่ฆ้ขๅฏนๅšๆŒไธ‹ๅŽป๏ผ›้šพ่ฟ‡็š„ๆ—ถๅ€™๏ผŒๆˆ‘ไผšไผช่ฃ… ่‡ชๅทฑ๏ผŒๅฏนๅˆซไบบ่ฏด๏ผšๆˆ‘ๅพˆๅฅฝใ€ๆˆ‘ๅพˆๅผ€ๅฟƒ๏ผ›ๅคฑ่ฝ็š„ๆ—ถๅ€™๏ผŒๆˆ‘ไผš็ฌ‘๏ฟฝ๏ฟฝ ๏ฟฝๅฏน่‡ชๅทฑ่ฏด๏ผŒๆฒกไบ‹็š„๏ผŒไธ€ๅˆ‡ๆ€ปไผš่ฟ‡ๅŽปใ€‚้ป„่คๆ–‘ๆœ€ไธป่ฆ็š„ๅฐฑๆ˜ฏไธ๏ฟฝ ๏ฟฝ๏ฟฝ็ญ‰้•ฟไบ†้ป„่คๆ–‘ๆ‰ๆƒณ่ฆ็ฅ›ๆ–‘๏ผŒ้‚ฃๆ˜ฏๅพ—ไธๅฟๅคฑ็š„๏ผŒๅŠณ็ฅž่ดนๆ‰็š„ไบ‹ ๆƒ…ใ€‚ๅŒๆ—ถๅœจๅŽป้™ค้ป„่คๆ–‘็š„ๆ—ถๅ€™ไฝ ่ฆ็Ÿฅ้“้ข้ƒจ้ป„่คๆ–‘ๆ˜ฏๆ€Žไนˆๆฅ๏ฟฝ๏ฟฝ ๏ฟฝ๏ผŒ่ฟ™ๆ‰่ƒฝๆœ‰ๆ•ˆๅŽป้™ค้ป„่คๆ–‘ใ€‚่‰ฒๆ–‘ๅฆ‚ไฝ•ๆ‰่ƒฝๅŽป้™ค๏ผŒ ใ€Šๅฎขๆˆทๆกˆไพ‹ใ€‹ ใ€€ใ€€ไธบไบ†ๆˆ‘่„ธไธŠ็š„้ป„่คๆ–‘๏ผŒๅœจ็ฝ‘ไธŠๅฅ”ๆณขไบ†ๅฅฝ้•ฟๆ—ถ้—ดใ€‚ๅŽๆฅๅ‘๏ฟฝ๏ฟฝ ๏ฟฝใ€Œ้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒใ€ๅๅ“ไธ้”™๏ผŒ่ฟ˜ๆ–นไพฟ๏ผŒไธ็”จๆ•ดๅคฉๆถ‚ๆถ‚ๆŠนๆŠน๏ฟฝ ๏ฟฝ๏ฟฝไบŽๆ˜ฏๅฐฑๆŸฅไบ†ๅพˆๅคš็›ธๅ…ณ็š„่ต„ๆ–™ใ€‚็œ‹ๆฅๅ‘็Žฐๅฅฝไธœ่ฅฟ๏ผŒ่ฟ˜ๅพ—ๆ‹ฅๆœ‰ไธ€ ๅŒโ€œๆ…ง็œผโ€ๅ•Š๏ผŒ่ฏดใ€Œ้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒใ€ๅฅฝ๏ผŒๆœ€ไธป่ฆ็š„่ฟ˜ๆ˜ฏๅฎƒ๏ฟฝ๏ฟฝ ๏ฟฝๆˆ‘ๅŽปๆމๆˆ‘้‚ฃไบ›ๆผไบบ็š„้ป„่คๆ–‘ใ€‚<br> ใ€€ใ€€็Ÿฅ้“ไบ†ใ€Œ้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒใ€ๅŽ๏ผŒๅ‘ไธ“ๅฎถ่ฏ‰่ฏดๅฟƒไธญ็š„็ƒฆๆผ๏ฟฝ๏ฟฝ ๏ฟฝ็œ‹ๆ˜ฏๅฆๆœ‰ไธ“้—จ้’ˆๅฏน็”ทๆ€ง็š„็ฅ›ๆ–‘ไบงๅ“ใ€‚ไธ“ๅฎถๅˆ†ๆž่ฏด๏ผŒๆฎๆˆ‘็š„ๆ๏ฟฝ ๏ฟฝ๏ฟฝ๏ผŒไปŽๅค–ๅœจๅ› ็ด ็œ‹๏ผŒ้•ฟๆ–‘ไธป่ฆๆ˜ฏ่ทŸ้•ฟๆ—ถ้—ด็š„็”ต่„‘่พๅฐ„ๆœ‰ๅ…ณใ€‚ไธ“ ๅฎถ่ฏดใ€Œ้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒใ€ๅฐฑๆ˜ฏไปŽDNA็ฒพๅŽๆˆ‘ไนŸ่ฎฐไธๆธ…ไบ†๏ผŒๅๆญฃ๏ฟฝ๏ฟฝ ๏ฟฝๆŒบๅฎ‰ๅ…จ็š„๏ผŒ่€Œไธ”ไฝฟ็”จไบ†็บณ็ฑณๆŠ€ๆœฏๅฎนๆ˜“ๅธๆ”ถ๏ผŒ่ฏด็”ทๅญฉๅญๅฎŒๅ…จๅฏ๏ฟฝ ๏ฟฝ๏ฟฝ็”จ็š„๏ผŒๅฏไธ”่ฟ˜ๅฏไปฅ้€ๆˆ‘ๅฏ่ดๅฐ”้ข่ดด่†œ้…ๅˆไฝฟ็”จๆ•ˆๆžœไผšๆ›ดๅฟซใ€‚ ไธ็ฎกไบ†๏ผŒๅช่ฆ่ƒฝ็ป™ๆˆ‘็š„้ป„่คๆ–‘ๅŽปๆމๅฐฑ่กŒ๏ผŒไบŽๆ˜ฏๆˆ‘ไธ€็‹ ๅฟƒ่ฎข่ดญ๏ฟฝ๏ฟฝ ๏ฟฝไธ‰ไธชๅ‘จๆœŸ๏ผŒๆฏๅคฉๅšๆŒ๏ผŒไปŽๆฅๆฒกๆœ‰่€ฝ่ฏฏ่ฟ‡ใ€‚ๆœ‹ๅ‹ไปฌ่ฟ˜็ฌ‘่ฟ‡่ทŸไธช๏ฟฝ ๏ฟฝ๏ฟฝไบบไธ€ๆ ท่ฟ˜ๆ˜ฏๆŽ’ๆฏ’ๅ…ป้ขœ็ฒพๅŽๆถฒๅ•Šใ€‚ๅ“Ž๏ผŒ่ฟ˜็œŸๅˆซ่ฏด๏ผŒไธ€ไธชๅ‘จๆœŸๅฎŒ ไบ†ไน‹ๅŽๆ•ˆๆžœ่ฟ˜็œŸไธ้”™๏ผŒไบŽๆ˜ฏๆ›ดๅŠ ๆœ‰ไฟกๅฟƒไบ†๏ผŒไนŸไธ็ฎกๅˆซไบบๆ€Žไนˆ๏ฟฝ๏ฟฝ ๏ฟฝ๏ผŒ่ƒฝๅŽปๆމ้ป„่คๆ–‘ๅฐฑๆ˜ฏๆญฃไบ‹ใ€‚็Žฐๅœจไธ‰ไธชๅ‘จๆœŸไฝฟ็”จๅฎŒไบ†่„ธไธŠ็š„้ป„๏ฟฝ ๏ฟฝ๏ฟฝๆ–‘่ฟ˜็œŸ็š„ๆฒกไบ†๏ผŒ่€Œไธ”ๆฒกๆœ‰็•™ไธ‹ไปปไฝ•็—•่ฟน๏ผŒ็šฎ่‚คไนŸๅ˜ๅพ—ๆฏ”ไปฅๅ‰ ๅฅฝไบ†ๅ‘ขใ€‚ๅ˜ฟๅ˜ฟ<br> ใ€€ใ€€่ฟ™ๅฐฑๆ˜ฏๆˆ‘็š„็ฅ›ๆ–‘็ปๅކ๏ผŒไธบไบ†ๅธ…ๆฐ”็š„่„ธ๏ผŒไธบไบ†ๆˆ‘้‚ฃ็ฝ—ๆ›ผ่’‚๏ฟฝ๏ฟฝ ๏ฟฝ็š„็ˆฑๆƒ…๏ผŒๆˆ‘ๅฏๆฒกๅฐ‘่ดนๅฟƒๆ€ๅ•Š๏ผŒ่ฟ˜ๅฅฝๆœ€็ปˆ่ตขๅพ—ไบ†่ƒœๅˆฉใ€‚ๆœ€ๆƒณๆ„Ÿ๏ฟฝ ๏ฟฝ๏ฟฝ็š„ๅฐฑๆ˜ฏใ€Œ้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒใ€ไบ†๏ผŒๅคšไบๆœ‰่ฟ™ไนˆๅฅฝ็š„ไบงๅ“!ๆœ€ๆƒณ๏ฟฝ๏ฟฝ ๏ฟฝ่ฏ„็š„ๅฐฑๆ˜ฏ่‡ชๅทฑ๏ผŒไปฅๅŽไธ่ƒฝๅ†ๅคฉๅคฉๆฒ‰่ฟทไบŽๆธธๆˆไบ†ใ€‚ ้˜…่ฏปไบ†่‰ฒๆ–‘ๅฆ‚ไฝ•ๆ‰่ƒฝๅŽป้™ค๏ผŒๅ†็œ‹่„ธไธŠๅฎนๆ˜“้•ฟๆ–‘็š„ๅŽŸๅ› ๏ผš ใ€Š่‰ฒๆ–‘ๅฝขๆˆๅŽŸๅ› ใ€‹ ใ€€ใ€€ๅ†…้ƒจๅ› ็ด  ใ€€ใ€€ไธ€ใ€ๅŽ‹ๅŠ› ใ€€ใ€€ๅฝ“ไบบๅ—ๅˆฐๅŽ‹ๅŠ›ๆ—ถ๏ผŒๅฐฑไผšๅˆ†ๆณŒ่‚พไธŠ่…บ็ด ๏ผŒไธบๅฏนไป˜ๅŽ‹ๅŠ›่€Œๅš๏ฟฝ๏ฟฝ ๏ฟฝๅค‡ใ€‚ๅฆ‚ๆžœ้•ฟๆœŸๅ—ๅˆฐๅŽ‹ๅŠ›๏ผŒไบบไฝ“ๆ–ฐ้™ˆไปฃ่ฐข็š„ๅนณ่กกๅฐฑไผš้ญๅˆฐ็ ดๅ๏ฟฝ ๏ฟฝ๏ฟฝ็šฎ่‚คๆ‰€้œ€็š„่ฅๅ…ปไพ›ๅบ”่ถ‹ไบŽ็ผ“ๆ…ข๏ผŒ่‰ฒ็ด ๆฏ็ป†่ƒžๅฐฑไผšๅ˜ๅพ—ๅพˆๆดป่ทƒ ใ€‚ ใ€€ใ€€ไบŒใ€่ทๅฐ”่’™ๅˆ†ๆณŒๅคฑ่ฐƒ ใ€€ใ€€้ฟๅญ•่ฏ้‡Œๆ‰€ๅซ็š„ๅฅณๆ€ง่ทๅฐ”่’™้›Œๆฟ€็ด ๏ผŒไผšๅˆบๆฟ€้บฆๆ‹‰ๅฎ็ป†่ƒž๏ฟฝ๏ฟฝ ๏ฟฝๅˆ†ๆณŒ่€Œๅฝขๆˆไธๅ‡ๅŒ€็š„ๆ–‘็‚น๏ผŒๅ› ้ฟๅญ•่ฏ่€Œๅฝขๆˆ็š„ๆ–‘็‚น๏ผŒ่™ฝ็„ถๅœจ๏ฟฝ ๏ฟฝ๏ฟฝ่ฏไธญๆ–ญๅŽไผšๅœๆญข๏ผŒไฝ†ไปไผšๅœจ็šฎ่‚คไธŠๅœ็•™ๅพˆ้•ฟไธ€ๆฎตๆ—ถ้—ดใ€‚ๆ€€ๅญ• ไธญๅ› ๅฅณๆ€ง่ทๅฐ”่’™้›Œๆฟ€็ด ็š„ๅขžๅŠ ๏ผŒไปŽๆ€€ๅญ•4โ€”5ไธชๆœˆๅผ€ๅง‹ไผšๅฎนๆ˜“ๅ‡บ ็Žฐๆ–‘๏ผŒ่ฟ™ๆ—ถๅ€™ๅ‡บ็Žฐ็š„ๆ–‘็‚นๅœจไบงๅŽๅคง้ƒจๅˆ†ไผšๆถˆๅคฑใ€‚ๅฏๆ˜ฏ๏ผŒๆ–ฐ้™ˆ๏ฟฝ๏ฟฝ ๏ฟฝ่ฐขไธๆญฃๅธธใ€่‚Œ่‚ค่ฃธ้œฒๅœจๅผบ็ƒˆ็š„็ดซๅค–็บฟไธ‹ใ€็ฒพ็ฅžไธŠๅ—ๅˆฐๅŽ‹ๅŠ›็ญ‰๏ฟฝ ๏ฟฝ๏ฟฝๅ› ๏ผŒ้ƒฝไผšไฝฟๆ–‘ๅŠ ๆทฑใ€‚ๆœ‰ๆ—ถๆ–ฐ้•ฟๅ‡บ็š„ๆ–‘๏ผŒไบงๅŽไนŸไธไผšๆถˆๅคฑ๏ผŒๆ‰€ ไปฅ้œ€่ฆๆ›ดๅŠ ๆณจๆ„ใ€‚ ใ€€ใ€€ไธ‰ใ€ๆ–ฐ้™ˆไปฃ่ฐข็ผ“ๆ…ข ใ€€ใ€€่‚็š„ๆ–ฐ้™ˆไปฃ่ฐขๅŠŸ่ƒฝไธๆญฃๅธธๆˆ–ๅตๅทขๅŠŸ่ƒฝๅ‡้€€ๆ—ถไนŸไผšๅ‡บ็Žฐๆ–‘๏ฟฝ๏ฟฝ ๏ฟฝๅ› ไธบๆ–ฐ้™ˆไปฃ่ฐขไธ้กบ็•…ใ€ๆˆ–ๅ†…ๅˆ†ๆณŒๅคฑ่ฐƒ๏ผŒไฝฟ่บซไฝ“ๅค„ไบŽๆ•ๆ„Ÿ็Šถๆ€๏ฟฝ ๏ฟฝ๏ฟฝ๏ผŒไปŽ่€ŒๅŠ ๅ‰ง่‰ฒ็ด ้—ฎ้ข˜ใ€‚ๆˆ‘ไปฌๅธธ่ฏด็š„ไพฟ็ง˜ไผšๅฝขๆˆๆ–‘๏ผŒๅ…ถๅฎžๅฐฑๆ˜ฏ ๅ†…ๅˆ†ๆณŒๅคฑ่ฐƒๅฏผ่‡ด่ฟ‡ๆ•ไฝ“่ดจ่€Œๅฝขๆˆ็š„ใ€‚ๅฆๅค–๏ผŒ่บซไฝ“็Šถๆ€ไธๆญฃๅธธ๏ฟฝ๏ฟฝ ๏ฟฝๆ—ถๅ€™๏ผŒ็ดซๅค–็บฟ็š„็…งๅฐ„ไนŸไผšๅŠ ้€Ÿๆ–‘็š„ๅฝขๆˆใ€‚ ใ€€ใ€€ๅ››ใ€้”™่ฏฏ็š„ไฝฟ็”จๅŒ–ๅฆ†ๅ“ ใ€€ใ€€ไฝฟ็”จไบ†ไธ้€‚ๅˆ่‡ชๅทฑ็šฎ่‚ค็š„ๅŒ–ๅฆ†ๅ“๏ผŒไผšๅฏผ่‡ด็šฎ่‚ค่ฟ‡ๆ•ใ€‚ๅœจ๏ฟฝ๏ฟฝ ๏ฟฝ็–—็š„่ฟ‡็จ‹ไธญๅฆ‚่ฟ‡้‡็…งๅฐ„ๅˆฐ็ดซๅค–็บฟ๏ผŒ็šฎ่‚คไผšไธบไบ†ๆŠตๅพกๅค–็•Œ็š„ไพต๏ฟฝ ๏ฟฝ๏ฟฝ๏ผŒๅœจๆœ‰็‚Ž็—‡็š„้ƒจไฝ่š้›†้บฆๆ‹‰ๅฎ่‰ฒ็ด ๏ผŒ่ฟ™ๆ ทไผšๅ‡บ็Žฐ่‰ฒ็ด ๆฒ‰็€็š„ ้—ฎ้ข˜ใ€‚ ใ€€ใ€€ๅค–้ƒจๅ› ็ด  ใ€€ใ€€ไธ€ใ€็ดซๅค–็บฟ ใ€€ใ€€็…งๅฐ„็ดซๅค–็บฟ็š„ๆ—ถๅ€™๏ผŒไบบไฝ“ไธบไบ†ไฟๆŠค็šฎ่‚ค๏ผŒไผšๅœจๅŸบๅบ•ๅฑ‚ไบง๏ฟฝ๏ฟฝ ๏ฟฝๅพˆๅคš้บฆๆ‹‰ๅฎ่‰ฒ็ด ใ€‚ๆ‰€ไปฅไธบไบ†ไฟๆŠค็šฎ่‚ค๏ผŒไผšๅœจๆ•ๆ„Ÿ้ƒจไฝ่š้›†ๆ›ด๏ฟฝ ๏ฟฝ๏ฟฝ็š„่‰ฒ็ด ใ€‚็ปๅธธ่ฃธ้œฒๅœจๅผบ็ƒˆ็š„้˜ณๅ…‰ๅบ•ไธ‹ไธไป…ไฟƒ่ฟ›็šฎ่‚ค็š„่€ๅŒ–๏ผŒ ่ฟ˜ไผšๅผ•่ตท้ป‘ๆ–‘ใ€้›€ๆ–‘็ญ‰่‰ฒ็ด ๆฒ‰็€็š„็šฎ่‚ค็–พๆ‚ฃใ€‚ ใ€€ใ€€ไบŒใ€ไธ่‰ฏ็š„ๆธ…ๆดไน ๆƒฏ ใ€€ใ€€ๅ› ๅผบ็ƒˆ็š„ๆธ…ๆดไน ๆƒฏไฝฟ็šฎ่‚คๅ˜ๅพ—ๆ•ๆ„Ÿ๏ผŒ่ฟ™ๆ ทไผšๅˆบๆฟ€็šฎ่‚คใ€‚๏ฟฝ๏ฟฝ ๏ฟฝ็šฎ่‚คๆ•ๆ„Ÿๆ—ถ๏ผŒไบบไฝ“ไธบไบ†ไฟๆŠค็šฎ่‚ค๏ผŒ้ป‘่‰ฒ็ด ็ป†่ƒžไผšๅˆ†ๆณŒๅพˆๅคš้บฆ๏ฟฝ ๏ฟฝ๏ฟฝๅฎ่‰ฒ็ด ๏ผŒๅฝ“่‰ฒ็ด ่ฟ‡ๅ‰ฉๆ—ถๅฐฑๅ‡บ็Žฐไบ†ๆ–‘ใ€็‘•็–ต็ญ‰็šฎ่‚ค่‰ฒ็ด ๆฒ‰็€็š„ ้—ฎ้ข˜ใ€‚ ใ€€ใ€€ไธ‰ใ€้—ไผ ๅŸบๅ›  ใ€€ใ€€็ˆถๆฏไธญๆœ‰้•ฟๆ–‘็š„๏ผŒๅˆ™ๆœฌไบบ้•ฟๆ–‘็š„ๆฆ‚็އๅฐฑๅพˆ้ซ˜๏ผŒ่ฟ™็งๆƒ…ๅ†ต๏ฟฝ๏ฟฝ ๏ฟฝไธ€ๅฎš็จ‹ๅบฆไธŠๅฐฑๅฏๅˆคๅฎšๆ˜ฏ้—ไผ ๅŸบๅ› ็š„ไฝœ็”จใ€‚ๆ‰€ไปฅๅฎถ้‡Œ็‰นๅˆซๆ˜ฏ้•ฟ๏ฟฝ ๏ฟฝ๏ฟฝๆœ‰้•ฟๆ–‘็š„ไบบ๏ผŒ่ฆๆณจๆ„้ฟๅ…ๅผ•ๅ‘้•ฟๆ–‘็š„้‡่ฆๅ› ็ด ไน‹ไธ€โ€”โ€”็ดซๅค– ็บฟ็…งๅฐ„๏ผŒ่ฟ™ๆ˜ฏ้ข„้˜ฒๆ–‘ๅฟ…้กปๆณจๆ„็š„ใ€‚ ใ€Šๆœ‰็–‘้—ฎๅธฎไฝ ่งฃๅ†ณใ€‹ ใ€€ใ€€1,้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒ็œŸ็š„ๆœ‰ๆ•ˆๆžœๅ—?็œŸ็š„ๅฏไปฅๆŠŠ่„ธไธŠ็š„้ป„่ค๏ฟฝ๏ฟฝ ๏ฟฝๅŽปๆމๅ—? ใ€€ใ€€็ญ”๏ผš้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒDNA็ฒพๅŽ่ƒฝๅคŸๆœ‰ๆ•ˆ็š„ไฟฎๅคๅ‘จๅ›ด้šพไปฅ่งฆ๏ฟฝ๏ฟฝ ๏ฟฝ็š„่‰ฒๆ–‘๏ผŒๅ…ถ็‹ฌๆœ‰็š„็บณ่ฑ†ๆˆๅˆ†ไธบ็šฎ่‚ค็š„็พŽ็™ฝไธŽ้“ไธฝ๏ผŒๆไพ›ไบ†ๅฟ…๏ฟฝ ๏ฟฝ๏ฟฝๅฏๅฐ‘็š„่ฅๅ…ป็‰ฉ่ดจ๏ผŒๅฏไปฅๆœ‰ๆ•ˆ็š„ๅŽป้™ค้ป„่คๆ–‘๏ผŒ้ป„่คๆ–‘๏ผŒ้ป„่คๆ–‘ ๏ผŒ่ด่ถๆ–‘๏ผŒๆ™’ๆ–‘ใ€ๅฆŠๅจ ๆ–‘็ญ‰ใ€‚ๅฎƒๅฎƒๅฎŒๅ…จ็ช็ ดไบ†ไผ ็ปŸ็š„็พŽ่‚คๆ—ถ๏ฟฝ๏ฟฝ ๏ฟฝ๏ผŒๅฎ›ๅฆ‚ๅœจ็šฎ่‚คไธญๆณจๅ…ฅไบ†ไธ€ๆฏๅ…ผๅ…ทๆดปๅŒ–ใ€ๅ†็”Ÿใ€ๆป‹ๅ…ป็ญ‰ๅŠŸๆ•ˆ็š„๏ฟฝ ๏ฟฝ๏ฟฝๅฐพ้…’๏ผŒๅŒๆ—ถไธบ่„ธ้ƒจๆไพ›ๅคง้‡ๆœ‰ๆœบ็ปด็”Ÿ็ด ็ฒพๅŽ๏ผŒ่„ธ้ƒจ็š„ๆ”นๅ˜ๆ˜พ ่€Œๆ˜“่งใ€‚่‡ชไบงๅ“ไธŠๅธ‚ไปฅๆฅ๏ผŒ่€้กพๅฎข็บท็บทไป‹็ปๆ–ฐ้กพๅฎข๏ผŒ71%็š„ๆ–ฐ๏ฟฝ๏ฟฝ ๏ฟฝๅฎข้ƒฝๆ˜ฏ้€š่ฟ‡่€้กพๅฎขไป‹็ป่€Œๆฅ๏ผŒๅฃ็ข‘็”ฑๆญค่€Œๆฅ! ใ€€ใ€€2๏ผŒๆœ็”จ้ป›่Š™่–‡ๅฐ”็พŽ็™ฝ๏ผŒไผšไผค่บซไฝ“ๅ—?ๆœ‰ๅ‰ฏไฝœ็”จๅ—? ใ€€ใ€€็ญ”๏ผš้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒๅบ”็”จไบ†็ฒพ็บฏๅคๅˆ้…ๆ–นๅ’Œ้ข†ๅ…ˆ็š„ๅˆ†็ฑป๏ฟฝ๏ฟฝ ๏ฟฝๆ–‘็ง‘ๆŠ€๏ผŒๅนถๅฐ†โ€œDNA็พŽ่‚ค็ณป็ปŸโ€็–—ๆณ•ๅบ”็”จๅˆฐไบ†่ฏฅไบงๅ“ไธญ๏ผŒ่ƒฝๅฝป๏ฟฝ ๏ฟฝ๏ฟฝ็ฅ›้™ค้ป„่คๆ–‘๏ผŒ่ด่ถๆ–‘๏ผŒๅฆŠๅจ ๆ–‘๏ผŒๆ™’ๆ–‘๏ผŒ้ป„่คๆ–‘๏ผŒ่€ๅนดๆ–‘๏ผŒๆœ‰ ๆ•ˆๆทกๅŒ–้ป„่คๆ–‘่‡ณๆŽฅ่ฟ‘่‚ค่‰ฒใ€‚้ป›่Š™่–‡ๅฐ”้€š่ฟ‡ๆณ•ๅ›ฝใ€็พŽๅ›ฝใ€ๅฐๆนพ๏ฟฝ๏ฟฝ ๏ฟฝๅœฐ็š„ไธ“ๅฎถ้€šๅŠ›ๅไฝœ๏ผŒ่ถ…่ฟ‡10ๅนด็š„็ ”็ฉถไปฅๅ…จๆ–ฐ็š„DNA่‚Œ่‚คไฟฎๅคๆŠ€๏ฟฝ๏ฟฝ ๏ฟฝ๏ผŒๆŒ‘ๆˆ˜ไผ ็ปŸๅŒ–ๅญฆๆŠค่‚ค็†ๅฟต๏ผŒไธๆ‡ˆ่ฟฝๅฏปๅ‘็Žฐ็ ด่ฏ‘ๅคง่‡ช็„ถ็š„็พŽไธฝ๏ฟฝ ๏ฟฝ๏ฟฝ่ฟน๏ผŒไปคๆฏไธ€ไฝ็ˆฑ็พŽ็š„ๅฅณๆ€ง้ƒฝ่ƒฝไบซๅ—ๅˆฐ็ง‘ๆŠ€ๅˆ›ๆ–ฐๆ‰€ๅธฆๆฅ็š„่‡ช็„ถ ไน‹็พŽใ€‚ ไธ“ไธบไบšๆดฒๅฅณๆ€ง่‚ค่ดจ็ ”ๅˆถ๏ผŒ็ฒพๅฟƒๅ‘ตๆŠคๅฅณๆ€ง็พŽไธฝ๏ผŒๅคšๅนดๆฅ๏ผŒไธบๆ•ฐ๏ฟฝ๏ฟฝ ๏ฟฝ็™พไธ‡่ฎก็š„ๅฅณๆ€ง่งฃ้™คไบ†้ป„่คๆ–‘ๅ›ฐๆ‰ฐใ€‚ๆทฑๅพ—ๅนฟๅคงๅฅณๆ€งๆœ‹ๅ‹็š„ไฟก่ต–! ใ€€ใ€€3๏ผŒๅŽป้™ค้ป„่คๆ–‘ไน‹ๅŽ๏ผŒไผšๅๅผนๅ—? ใ€€ใ€€็ญ”๏ผšๅพˆๅคšๆ›พ็ป้•ฟไบ†้ป„่คๆ–‘็š„ไบบๅฃซ๏ผŒ่‡ชไปŽ้€‰ๆ‹ฉไบ†้ป›่Š™่–‡ๅฐ”๏ฟฝ๏ฟฝ ๏ฟฝ็™ฝ๏ผŒๅฐฑไธ€ๅŠณๆฐธ้€ธใ€‚่ฟ™ๆฌพ็ฅ›ๆ–‘ไบงๅ“ๆ˜ฏ็ป่ฟ‡ๆ•ฐๅไฝๆƒๅจ็ฅ›ๆ–‘ไธ“ๅฎถ๏ฟฝ ๏ฟฝ๏ฟฝๆฎๆ–‘็š„ๅฝขๆˆๅŽŸๅ› ็ฒพๅฟƒ็ ”ๅˆถ่€Œๆˆ็”จไบ‹ๅฎž่ฏด่ฏ๏ผŒ่ฎฉๆถˆ่ดน่€…ๆ‰“ๅˆ†ใ€‚ ๆ ‘็ซ‹ๆƒๅจๅ“็‰Œ!ๆˆ‘ไปฌ็š„ๅพˆๅคšๆ–ฐๅฎขๆˆท้ƒฝๆ˜ฏ่€ๅฎขๆˆทไป‹็ป่€Œๆฅ๏ผŒ่ฏท้—ฎ๏ฟฝ ๏ฟฝ๏ฟฝๅฆ‚ๆžœๆ•ˆๆžœไธๅฅฝ๏ผŒไผšๆœ‰ๅฎขๆˆท่ฝฌไป‹็ปๅ—? ใ€€ใ€€4๏ผŒไฝ ไปฌ็š„ไปทๆ ผๆœ‰็‚น่ดต๏ผŒ่ƒฝไธ่ƒฝไพฟๅฎœไธ€็‚น? ใ€€ใ€€็ญ”๏ผšๅฆ‚ๆžœๆ‚จไฝฟ็”จ่ฅฟ่ฏๆœ€ๅฐ‘้œ€่ฆ2000ๅ…ƒ๏ผŒ็…Žๆœ็š„่ฏๆœ€ๅฐ‘้œ€่ฆ3 000ๅ…ƒ๏ผŒๅšๆ‰‹ๆœฏๆœ€ๅฐ‘ๆ˜ฏ5000ๅ…ƒ๏ผŒ่€Œ่ฟ™ไบ›ๆฏซๆ— ็–‘้—ฎ๏ผŒไธไผšๅฏนๅฝปๅบ•ๅŽป๏ฟฝ ๏ฟฝ๏ฟฝไฝ ็š„ๆ–‘็‚นๆœ‰ไปปไฝ•ๅธฎๅŠฉ!ไธ€ๅˆ†ไปท้’ฑ๏ผŒไธ€ไปฝไปทๅ€ผ๏ผŒๆˆ‘ไปฌ็Žฐๅœจๅš็š„๏ฟฝ๏ฟฝ ๏ฟฝๆ˜ฏไธ€ไธชๅฃ็ข‘๏ผŒไธ€ไธชๅ“็‰Œ๏ผŒไปท้’ฑๅนถไธ้ซ˜ใ€‚ๅฆ‚ๆžœ่Šฑ่ฟ™็‚น้’ฑๆŠŠไฝ ็š„๏ฟฝ ๏ฟฝ๏ฟฝ่คๆ–‘ๅฝปๅบ•ๅŽป้™ค๏ผŒไฝ ่ฟ˜ไผš่ง‰ๅพ—่ดตๅ—?ไฝ ่ฟ˜ไผšๅ†ๅŽป่Šฑ้‚ฃไนˆๅคšๅ†คๆž‰๏ฟฝ๏ฟฝ ๏ฟฝ๏ผŒไธไฝ†ๆ–‘ๆฒกๅŽปๆމ๏ผŒ่ฟ˜ๆŠŠ่‡ชๅทฑ็š„็šฎ่‚คๅผ„็š„่ถŠๆฅ่ถŠ็ณŸๅ— ใ€€ใ€€5๏ผŒๆˆ‘้€‚ๅˆ็”จ้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒๅ—? ใ€€ใ€€็ญ”๏ผš้ป›่Š™่–‡ๅฐ”้€‚็”จไบบ็พค๏ผš ใ€€ใ€€1ใ€็”Ÿ็†็ดŠไนฑๅผ•่ตท็š„้ป„่คๆ–‘ไบบ็พค ใ€€ใ€€2ใ€็”Ÿ่‚ฒๅผ•่ตท็š„ๅฆŠๅจ ๆ–‘ไบบ็พค ใ€€ใ€€3ใ€ๅนด็บชๅขž้•ฟๅผ•่ตท็š„่€ๅนดๆ–‘ไบบ็พค ใ€€ใ€€4ใ€ๅŒ–ๅฆ†ๅ“่‰ฒ็ด ๆฒ‰็งฏใ€่พๅฐ„ๆ–‘ไบบ็พค ใ€€ใ€€5ใ€้•ฟๆœŸๆ—ฅ็…งๅผ•่ตท็š„ๆ—ฅๆ™’ๆ–‘ไบบ็พค ใ€€ใ€€6ใ€่‚Œ่‚คๆš—ๆทกๆ€ฅ้œ€็พŽ็™ฝ็š„ไบบ็พค ใ€Š็ฅ›ๆ–‘ๅฐๆ–นๆณ•ใ€‹ ่‰ฒๆ–‘ๅฆ‚ไฝ•ๆ‰่ƒฝๅŽป้™ค๏ผŒๅŒๆ—ถไธบๆ‚จๅˆ†ไบซ็ฅ›ๆ–‘ๅฐๆ–นๆณ• ่ŒถๆฐดๅŽปๆ–‘็พŽ็™ฝ ๆ–นๆณ•ไธ€๏ผšๆด—่„ธๅŽ๏ผŒๅฐ†่Œถๆฐดๆถ‚ๅˆฐ่„ธไธŠ๏ผŒๅนถ็”จๆ‰‹่ฝป่ฝปๆ‹่„ธใ€‚ ๆ–นๆณ•ไบŒ๏ผšๅฐ†่˜ธไบ†่Œถๆฐด็š„่„ฑ่„‚ๆฃ‰้™„ๅœจ่„ธไธŠ2-3ๅˆ†้’Ÿ๏ผŒ็„ถๅŽๆธ…ๆฐดๆด—๏ฟฝ๏ฟฝ ๏ฟฝ๏ผŒๆœ‰้™ค่‰ฒๆ–‘ใ€็พŽ็™ฝ็š„ๆ•ˆๆžœใ€‚ ``` ----- Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 4:22
1.0
่ต„่ฎฏ่‰ฒๆ–‘ๅฆ‚ไฝ•ๆ‰่ƒฝๅŽป้™ค - ``` ใ€Šๆ‘˜่ฆใ€‹ ๅ“ญ็š„ๆ—ถๅ€™๏ผŒๆˆ‘ไผš้—ญไธŠ็œผ็›ไธ่ฎฉๅฎƒๆตๆณช๏ผ›ๅญค็‹ฌๅฏ‚ๅฏž็š„ๆ—ถๅ€™๏ผŒๆˆ‘๏ฟฝ๏ฟฝ ๏ฟฝ้™้™็š„ๆƒณ็€ๆŸไบบ๏ผ›ไผคๅฟƒ็š„ๆ—ถๅ€™๏ผŒๆˆ‘ไผšๆ‰พไธชๅœฐๆ–น้™้™็š„ๅ‘ๅ‘†๏ผŒ๏ฟฝ ๏ฟฝ๏ฟฝๅŽๅ‘Š่ฏ‰่‡ชๅทฑ๏ผŒ่ฟ˜ๆ˜ฏ่ฆ้ขๅฏนๅšๆŒไธ‹ๅŽป๏ผ›้šพ่ฟ‡็š„ๆ—ถๅ€™๏ผŒๆˆ‘ไผšไผช่ฃ… ่‡ชๅทฑ๏ผŒๅฏนๅˆซไบบ่ฏด๏ผšๆˆ‘ๅพˆๅฅฝใ€ๆˆ‘ๅพˆๅผ€ๅฟƒ๏ผ›ๅคฑ่ฝ็š„ๆ—ถๅ€™๏ผŒๆˆ‘ไผš็ฌ‘๏ฟฝ๏ฟฝ ๏ฟฝๅฏน่‡ชๅทฑ่ฏด๏ผŒๆฒกไบ‹็š„๏ผŒไธ€ๅˆ‡ๆ€ปไผš่ฟ‡ๅŽปใ€‚้ป„่คๆ–‘ๆœ€ไธป่ฆ็š„ๅฐฑๆ˜ฏไธ๏ฟฝ ๏ฟฝ๏ฟฝ็ญ‰้•ฟไบ†้ป„่คๆ–‘ๆ‰ๆƒณ่ฆ็ฅ›ๆ–‘๏ผŒ้‚ฃๆ˜ฏๅพ—ไธๅฟๅคฑ็š„๏ผŒๅŠณ็ฅž่ดนๆ‰็š„ไบ‹ ๆƒ…ใ€‚ๅŒๆ—ถๅœจๅŽป้™ค้ป„่คๆ–‘็š„ๆ—ถๅ€™ไฝ ่ฆ็Ÿฅ้“้ข้ƒจ้ป„่คๆ–‘ๆ˜ฏๆ€Žไนˆๆฅ๏ฟฝ๏ฟฝ ๏ฟฝ๏ผŒ่ฟ™ๆ‰่ƒฝๆœ‰ๆ•ˆๅŽป้™ค้ป„่คๆ–‘ใ€‚่‰ฒๆ–‘ๅฆ‚ไฝ•ๆ‰่ƒฝๅŽป้™ค๏ผŒ ใ€Šๅฎขๆˆทๆกˆไพ‹ใ€‹ ใ€€ใ€€ไธบไบ†ๆˆ‘่„ธไธŠ็š„้ป„่คๆ–‘๏ผŒๅœจ็ฝ‘ไธŠๅฅ”ๆณขไบ†ๅฅฝ้•ฟๆ—ถ้—ดใ€‚ๅŽๆฅๅ‘๏ฟฝ๏ฟฝ ๏ฟฝใ€Œ้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒใ€ๅๅ“ไธ้”™๏ผŒ่ฟ˜ๆ–นไพฟ๏ผŒไธ็”จๆ•ดๅคฉๆถ‚ๆถ‚ๆŠนๆŠน๏ฟฝ ๏ฟฝ๏ฟฝไบŽๆ˜ฏๅฐฑๆŸฅไบ†ๅพˆๅคš็›ธๅ…ณ็š„่ต„ๆ–™ใ€‚็œ‹ๆฅๅ‘็Žฐๅฅฝไธœ่ฅฟ๏ผŒ่ฟ˜ๅพ—ๆ‹ฅๆœ‰ไธ€ ๅŒโ€œๆ…ง็œผโ€ๅ•Š๏ผŒ่ฏดใ€Œ้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒใ€ๅฅฝ๏ผŒๆœ€ไธป่ฆ็š„่ฟ˜ๆ˜ฏๅฎƒ๏ฟฝ๏ฟฝ ๏ฟฝๆˆ‘ๅŽปๆމๆˆ‘้‚ฃไบ›ๆผไบบ็š„้ป„่คๆ–‘ใ€‚<br> ใ€€ใ€€็Ÿฅ้“ไบ†ใ€Œ้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒใ€ๅŽ๏ผŒๅ‘ไธ“ๅฎถ่ฏ‰่ฏดๅฟƒไธญ็š„็ƒฆๆผ๏ฟฝ๏ฟฝ ๏ฟฝ็œ‹ๆ˜ฏๅฆๆœ‰ไธ“้—จ้’ˆๅฏน็”ทๆ€ง็š„็ฅ›ๆ–‘ไบงๅ“ใ€‚ไธ“ๅฎถๅˆ†ๆž่ฏด๏ผŒๆฎๆˆ‘็š„ๆ๏ฟฝ ๏ฟฝ๏ฟฝ๏ผŒไปŽๅค–ๅœจๅ› ็ด ็œ‹๏ผŒ้•ฟๆ–‘ไธป่ฆๆ˜ฏ่ทŸ้•ฟๆ—ถ้—ด็š„็”ต่„‘่พๅฐ„ๆœ‰ๅ…ณใ€‚ไธ“ ๅฎถ่ฏดใ€Œ้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒใ€ๅฐฑๆ˜ฏไปŽDNA็ฒพๅŽๆˆ‘ไนŸ่ฎฐไธๆธ…ไบ†๏ผŒๅๆญฃ๏ฟฝ๏ฟฝ ๏ฟฝๆŒบๅฎ‰ๅ…จ็š„๏ผŒ่€Œไธ”ไฝฟ็”จไบ†็บณ็ฑณๆŠ€ๆœฏๅฎนๆ˜“ๅธๆ”ถ๏ผŒ่ฏด็”ทๅญฉๅญๅฎŒๅ…จๅฏ๏ฟฝ ๏ฟฝ๏ฟฝ็”จ็š„๏ผŒๅฏไธ”่ฟ˜ๅฏไปฅ้€ๆˆ‘ๅฏ่ดๅฐ”้ข่ดด่†œ้…ๅˆไฝฟ็”จๆ•ˆๆžœไผšๆ›ดๅฟซใ€‚ ไธ็ฎกไบ†๏ผŒๅช่ฆ่ƒฝ็ป™ๆˆ‘็š„้ป„่คๆ–‘ๅŽปๆމๅฐฑ่กŒ๏ผŒไบŽๆ˜ฏๆˆ‘ไธ€็‹ ๅฟƒ่ฎข่ดญ๏ฟฝ๏ฟฝ ๏ฟฝไธ‰ไธชๅ‘จๆœŸ๏ผŒๆฏๅคฉๅšๆŒ๏ผŒไปŽๆฅๆฒกๆœ‰่€ฝ่ฏฏ่ฟ‡ใ€‚ๆœ‹ๅ‹ไปฌ่ฟ˜็ฌ‘่ฟ‡่ทŸไธช๏ฟฝ ๏ฟฝ๏ฟฝไบบไธ€ๆ ท่ฟ˜ๆ˜ฏๆŽ’ๆฏ’ๅ…ป้ขœ็ฒพๅŽๆถฒๅ•Šใ€‚ๅ“Ž๏ผŒ่ฟ˜็œŸๅˆซ่ฏด๏ผŒไธ€ไธชๅ‘จๆœŸๅฎŒ ไบ†ไน‹ๅŽๆ•ˆๆžœ่ฟ˜็œŸไธ้”™๏ผŒไบŽๆ˜ฏๆ›ดๅŠ ๆœ‰ไฟกๅฟƒไบ†๏ผŒไนŸไธ็ฎกๅˆซไบบๆ€Žไนˆ๏ฟฝ๏ฟฝ ๏ฟฝ๏ผŒ่ƒฝๅŽปๆމ้ป„่คๆ–‘ๅฐฑๆ˜ฏๆญฃไบ‹ใ€‚็Žฐๅœจไธ‰ไธชๅ‘จๆœŸไฝฟ็”จๅฎŒไบ†่„ธไธŠ็š„้ป„๏ฟฝ ๏ฟฝ๏ฟฝๆ–‘่ฟ˜็œŸ็š„ๆฒกไบ†๏ผŒ่€Œไธ”ๆฒกๆœ‰็•™ไธ‹ไปปไฝ•็—•่ฟน๏ผŒ็šฎ่‚คไนŸๅ˜ๅพ—ๆฏ”ไปฅๅ‰ ๅฅฝไบ†ๅ‘ขใ€‚ๅ˜ฟๅ˜ฟ<br> ใ€€ใ€€่ฟ™ๅฐฑๆ˜ฏๆˆ‘็š„็ฅ›ๆ–‘็ปๅކ๏ผŒไธบไบ†ๅธ…ๆฐ”็š„่„ธ๏ผŒไธบไบ†ๆˆ‘้‚ฃ็ฝ—ๆ›ผ่’‚๏ฟฝ๏ฟฝ ๏ฟฝ็š„็ˆฑๆƒ…๏ผŒๆˆ‘ๅฏๆฒกๅฐ‘่ดนๅฟƒๆ€ๅ•Š๏ผŒ่ฟ˜ๅฅฝๆœ€็ปˆ่ตขๅพ—ไบ†่ƒœๅˆฉใ€‚ๆœ€ๆƒณๆ„Ÿ๏ฟฝ ๏ฟฝ๏ฟฝ็š„ๅฐฑๆ˜ฏใ€Œ้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒใ€ไบ†๏ผŒๅคšไบๆœ‰่ฟ™ไนˆๅฅฝ็š„ไบงๅ“!ๆœ€ๆƒณ๏ฟฝ๏ฟฝ ๏ฟฝ่ฏ„็š„ๅฐฑๆ˜ฏ่‡ชๅทฑ๏ผŒไปฅๅŽไธ่ƒฝๅ†ๅคฉๅคฉๆฒ‰่ฟทไบŽๆธธๆˆไบ†ใ€‚ ้˜…่ฏปไบ†่‰ฒๆ–‘ๅฆ‚ไฝ•ๆ‰่ƒฝๅŽป้™ค๏ผŒๅ†็œ‹่„ธไธŠๅฎนๆ˜“้•ฟๆ–‘็š„ๅŽŸๅ› ๏ผš ใ€Š่‰ฒๆ–‘ๅฝขๆˆๅŽŸๅ› ใ€‹ ใ€€ใ€€ๅ†…้ƒจๅ› ็ด  ใ€€ใ€€ไธ€ใ€ๅŽ‹ๅŠ› ใ€€ใ€€ๅฝ“ไบบๅ—ๅˆฐๅŽ‹ๅŠ›ๆ—ถ๏ผŒๅฐฑไผšๅˆ†ๆณŒ่‚พไธŠ่…บ็ด ๏ผŒไธบๅฏนไป˜ๅŽ‹ๅŠ›่€Œๅš๏ฟฝ๏ฟฝ ๏ฟฝๅค‡ใ€‚ๅฆ‚ๆžœ้•ฟๆœŸๅ—ๅˆฐๅŽ‹ๅŠ›๏ผŒไบบไฝ“ๆ–ฐ้™ˆไปฃ่ฐข็š„ๅนณ่กกๅฐฑไผš้ญๅˆฐ็ ดๅ๏ฟฝ ๏ฟฝ๏ฟฝ็šฎ่‚คๆ‰€้œ€็š„่ฅๅ…ปไพ›ๅบ”่ถ‹ไบŽ็ผ“ๆ…ข๏ผŒ่‰ฒ็ด ๆฏ็ป†่ƒžๅฐฑไผšๅ˜ๅพ—ๅพˆๆดป่ทƒ ใ€‚ ใ€€ใ€€ไบŒใ€่ทๅฐ”่’™ๅˆ†ๆณŒๅคฑ่ฐƒ ใ€€ใ€€้ฟๅญ•่ฏ้‡Œๆ‰€ๅซ็š„ๅฅณๆ€ง่ทๅฐ”่’™้›Œๆฟ€็ด ๏ผŒไผšๅˆบๆฟ€้บฆๆ‹‰ๅฎ็ป†่ƒž๏ฟฝ๏ฟฝ ๏ฟฝๅˆ†ๆณŒ่€Œๅฝขๆˆไธๅ‡ๅŒ€็š„ๆ–‘็‚น๏ผŒๅ› ้ฟๅญ•่ฏ่€Œๅฝขๆˆ็š„ๆ–‘็‚น๏ผŒ่™ฝ็„ถๅœจ๏ฟฝ ๏ฟฝ๏ฟฝ่ฏไธญๆ–ญๅŽไผšๅœๆญข๏ผŒไฝ†ไปไผšๅœจ็šฎ่‚คไธŠๅœ็•™ๅพˆ้•ฟไธ€ๆฎตๆ—ถ้—ดใ€‚ๆ€€ๅญ• ไธญๅ› ๅฅณๆ€ง่ทๅฐ”่’™้›Œๆฟ€็ด ็š„ๅขžๅŠ ๏ผŒไปŽๆ€€ๅญ•4โ€”5ไธชๆœˆๅผ€ๅง‹ไผšๅฎนๆ˜“ๅ‡บ ็Žฐๆ–‘๏ผŒ่ฟ™ๆ—ถๅ€™ๅ‡บ็Žฐ็š„ๆ–‘็‚นๅœจไบงๅŽๅคง้ƒจๅˆ†ไผšๆถˆๅคฑใ€‚ๅฏๆ˜ฏ๏ผŒๆ–ฐ้™ˆ๏ฟฝ๏ฟฝ ๏ฟฝ่ฐขไธๆญฃๅธธใ€่‚Œ่‚ค่ฃธ้œฒๅœจๅผบ็ƒˆ็š„็ดซๅค–็บฟไธ‹ใ€็ฒพ็ฅžไธŠๅ—ๅˆฐๅŽ‹ๅŠ›็ญ‰๏ฟฝ ๏ฟฝ๏ฟฝๅ› ๏ผŒ้ƒฝไผšไฝฟๆ–‘ๅŠ ๆทฑใ€‚ๆœ‰ๆ—ถๆ–ฐ้•ฟๅ‡บ็š„ๆ–‘๏ผŒไบงๅŽไนŸไธไผšๆถˆๅคฑ๏ผŒๆ‰€ ไปฅ้œ€่ฆๆ›ดๅŠ ๆณจๆ„ใ€‚ ใ€€ใ€€ไธ‰ใ€ๆ–ฐ้™ˆไปฃ่ฐข็ผ“ๆ…ข ใ€€ใ€€่‚็š„ๆ–ฐ้™ˆไปฃ่ฐขๅŠŸ่ƒฝไธๆญฃๅธธๆˆ–ๅตๅทขๅŠŸ่ƒฝๅ‡้€€ๆ—ถไนŸไผšๅ‡บ็Žฐๆ–‘๏ฟฝ๏ฟฝ ๏ฟฝๅ› ไธบๆ–ฐ้™ˆไปฃ่ฐขไธ้กบ็•…ใ€ๆˆ–ๅ†…ๅˆ†ๆณŒๅคฑ่ฐƒ๏ผŒไฝฟ่บซไฝ“ๅค„ไบŽๆ•ๆ„Ÿ็Šถๆ€๏ฟฝ ๏ฟฝ๏ฟฝ๏ผŒไปŽ่€ŒๅŠ ๅ‰ง่‰ฒ็ด ้—ฎ้ข˜ใ€‚ๆˆ‘ไปฌๅธธ่ฏด็š„ไพฟ็ง˜ไผšๅฝขๆˆๆ–‘๏ผŒๅ…ถๅฎžๅฐฑๆ˜ฏ ๅ†…ๅˆ†ๆณŒๅคฑ่ฐƒๅฏผ่‡ด่ฟ‡ๆ•ไฝ“่ดจ่€Œๅฝขๆˆ็š„ใ€‚ๅฆๅค–๏ผŒ่บซไฝ“็Šถๆ€ไธๆญฃๅธธ๏ฟฝ๏ฟฝ ๏ฟฝๆ—ถๅ€™๏ผŒ็ดซๅค–็บฟ็š„็…งๅฐ„ไนŸไผšๅŠ ้€Ÿๆ–‘็š„ๅฝขๆˆใ€‚ ใ€€ใ€€ๅ››ใ€้”™่ฏฏ็š„ไฝฟ็”จๅŒ–ๅฆ†ๅ“ ใ€€ใ€€ไฝฟ็”จไบ†ไธ้€‚ๅˆ่‡ชๅทฑ็šฎ่‚ค็š„ๅŒ–ๅฆ†ๅ“๏ผŒไผšๅฏผ่‡ด็šฎ่‚ค่ฟ‡ๆ•ใ€‚ๅœจ๏ฟฝ๏ฟฝ ๏ฟฝ็–—็š„่ฟ‡็จ‹ไธญๅฆ‚่ฟ‡้‡็…งๅฐ„ๅˆฐ็ดซๅค–็บฟ๏ผŒ็šฎ่‚คไผšไธบไบ†ๆŠตๅพกๅค–็•Œ็š„ไพต๏ฟฝ ๏ฟฝ๏ฟฝ๏ผŒๅœจๆœ‰็‚Ž็—‡็š„้ƒจไฝ่š้›†้บฆๆ‹‰ๅฎ่‰ฒ็ด ๏ผŒ่ฟ™ๆ ทไผšๅ‡บ็Žฐ่‰ฒ็ด ๆฒ‰็€็š„ ้—ฎ้ข˜ใ€‚ ใ€€ใ€€ๅค–้ƒจๅ› ็ด  ใ€€ใ€€ไธ€ใ€็ดซๅค–็บฟ ใ€€ใ€€็…งๅฐ„็ดซๅค–็บฟ็š„ๆ—ถๅ€™๏ผŒไบบไฝ“ไธบไบ†ไฟๆŠค็šฎ่‚ค๏ผŒไผšๅœจๅŸบๅบ•ๅฑ‚ไบง๏ฟฝ๏ฟฝ ๏ฟฝๅพˆๅคš้บฆๆ‹‰ๅฎ่‰ฒ็ด ใ€‚ๆ‰€ไปฅไธบไบ†ไฟๆŠค็šฎ่‚ค๏ผŒไผšๅœจๆ•ๆ„Ÿ้ƒจไฝ่š้›†ๆ›ด๏ฟฝ ๏ฟฝ๏ฟฝ็š„่‰ฒ็ด ใ€‚็ปๅธธ่ฃธ้œฒๅœจๅผบ็ƒˆ็š„้˜ณๅ…‰ๅบ•ไธ‹ไธไป…ไฟƒ่ฟ›็šฎ่‚ค็š„่€ๅŒ–๏ผŒ ่ฟ˜ไผšๅผ•่ตท้ป‘ๆ–‘ใ€้›€ๆ–‘็ญ‰่‰ฒ็ด ๆฒ‰็€็š„็šฎ่‚ค็–พๆ‚ฃใ€‚ ใ€€ใ€€ไบŒใ€ไธ่‰ฏ็š„ๆธ…ๆดไน ๆƒฏ ใ€€ใ€€ๅ› ๅผบ็ƒˆ็š„ๆธ…ๆดไน ๆƒฏไฝฟ็šฎ่‚คๅ˜ๅพ—ๆ•ๆ„Ÿ๏ผŒ่ฟ™ๆ ทไผšๅˆบๆฟ€็šฎ่‚คใ€‚๏ฟฝ๏ฟฝ ๏ฟฝ็šฎ่‚คๆ•ๆ„Ÿๆ—ถ๏ผŒไบบไฝ“ไธบไบ†ไฟๆŠค็šฎ่‚ค๏ผŒ้ป‘่‰ฒ็ด ็ป†่ƒžไผšๅˆ†ๆณŒๅพˆๅคš้บฆ๏ฟฝ ๏ฟฝ๏ฟฝๅฎ่‰ฒ็ด ๏ผŒๅฝ“่‰ฒ็ด ่ฟ‡ๅ‰ฉๆ—ถๅฐฑๅ‡บ็Žฐไบ†ๆ–‘ใ€็‘•็–ต็ญ‰็šฎ่‚ค่‰ฒ็ด ๆฒ‰็€็š„ ้—ฎ้ข˜ใ€‚ ใ€€ใ€€ไธ‰ใ€้—ไผ ๅŸบๅ›  ใ€€ใ€€็ˆถๆฏไธญๆœ‰้•ฟๆ–‘็š„๏ผŒๅˆ™ๆœฌไบบ้•ฟๆ–‘็š„ๆฆ‚็އๅฐฑๅพˆ้ซ˜๏ผŒ่ฟ™็งๆƒ…ๅ†ต๏ฟฝ๏ฟฝ ๏ฟฝไธ€ๅฎš็จ‹ๅบฆไธŠๅฐฑๅฏๅˆคๅฎšๆ˜ฏ้—ไผ ๅŸบๅ› ็š„ไฝœ็”จใ€‚ๆ‰€ไปฅๅฎถ้‡Œ็‰นๅˆซๆ˜ฏ้•ฟ๏ฟฝ ๏ฟฝ๏ฟฝๆœ‰้•ฟๆ–‘็š„ไบบ๏ผŒ่ฆๆณจๆ„้ฟๅ…ๅผ•ๅ‘้•ฟๆ–‘็š„้‡่ฆๅ› ็ด ไน‹ไธ€โ€”โ€”็ดซๅค– ็บฟ็…งๅฐ„๏ผŒ่ฟ™ๆ˜ฏ้ข„้˜ฒๆ–‘ๅฟ…้กปๆณจๆ„็š„ใ€‚ ใ€Šๆœ‰็–‘้—ฎๅธฎไฝ ่งฃๅ†ณใ€‹ ใ€€ใ€€1,้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒ็œŸ็š„ๆœ‰ๆ•ˆๆžœๅ—?็œŸ็š„ๅฏไปฅๆŠŠ่„ธไธŠ็š„้ป„่ค๏ฟฝ๏ฟฝ ๏ฟฝๅŽปๆމๅ—? ใ€€ใ€€็ญ”๏ผš้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒDNA็ฒพๅŽ่ƒฝๅคŸๆœ‰ๆ•ˆ็š„ไฟฎๅคๅ‘จๅ›ด้šพไปฅ่งฆ๏ฟฝ๏ฟฝ ๏ฟฝ็š„่‰ฒๆ–‘๏ผŒๅ…ถ็‹ฌๆœ‰็š„็บณ่ฑ†ๆˆๅˆ†ไธบ็šฎ่‚ค็š„็พŽ็™ฝไธŽ้“ไธฝ๏ผŒๆไพ›ไบ†ๅฟ…๏ฟฝ ๏ฟฝ๏ฟฝๅฏๅฐ‘็š„่ฅๅ…ป็‰ฉ่ดจ๏ผŒๅฏไปฅๆœ‰ๆ•ˆ็š„ๅŽป้™ค้ป„่คๆ–‘๏ผŒ้ป„่คๆ–‘๏ผŒ้ป„่คๆ–‘ ๏ผŒ่ด่ถๆ–‘๏ผŒๆ™’ๆ–‘ใ€ๅฆŠๅจ ๆ–‘็ญ‰ใ€‚ๅฎƒๅฎƒๅฎŒๅ…จ็ช็ ดไบ†ไผ ็ปŸ็š„็พŽ่‚คๆ—ถ๏ฟฝ๏ฟฝ ๏ฟฝ๏ผŒๅฎ›ๅฆ‚ๅœจ็šฎ่‚คไธญๆณจๅ…ฅไบ†ไธ€ๆฏๅ…ผๅ…ทๆดปๅŒ–ใ€ๅ†็”Ÿใ€ๆป‹ๅ…ป็ญ‰ๅŠŸๆ•ˆ็š„๏ฟฝ ๏ฟฝ๏ฟฝๅฐพ้…’๏ผŒๅŒๆ—ถไธบ่„ธ้ƒจๆไพ›ๅคง้‡ๆœ‰ๆœบ็ปด็”Ÿ็ด ็ฒพๅŽ๏ผŒ่„ธ้ƒจ็š„ๆ”นๅ˜ๆ˜พ ่€Œๆ˜“่งใ€‚่‡ชไบงๅ“ไธŠๅธ‚ไปฅๆฅ๏ผŒ่€้กพๅฎข็บท็บทไป‹็ปๆ–ฐ้กพๅฎข๏ผŒ71%็š„ๆ–ฐ๏ฟฝ๏ฟฝ ๏ฟฝๅฎข้ƒฝๆ˜ฏ้€š่ฟ‡่€้กพๅฎขไป‹็ป่€Œๆฅ๏ผŒๅฃ็ข‘็”ฑๆญค่€Œๆฅ! ใ€€ใ€€2๏ผŒๆœ็”จ้ป›่Š™่–‡ๅฐ”็พŽ็™ฝ๏ผŒไผšไผค่บซไฝ“ๅ—?ๆœ‰ๅ‰ฏไฝœ็”จๅ—? ใ€€ใ€€็ญ”๏ผš้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒๅบ”็”จไบ†็ฒพ็บฏๅคๅˆ้…ๆ–นๅ’Œ้ข†ๅ…ˆ็š„ๅˆ†็ฑป๏ฟฝ๏ฟฝ ๏ฟฝๆ–‘็ง‘ๆŠ€๏ผŒๅนถๅฐ†โ€œDNA็พŽ่‚ค็ณป็ปŸโ€็–—ๆณ•ๅบ”็”จๅˆฐไบ†่ฏฅไบงๅ“ไธญ๏ผŒ่ƒฝๅฝป๏ฟฝ ๏ฟฝ๏ฟฝ็ฅ›้™ค้ป„่คๆ–‘๏ผŒ่ด่ถๆ–‘๏ผŒๅฆŠๅจ ๆ–‘๏ผŒๆ™’ๆ–‘๏ผŒ้ป„่คๆ–‘๏ผŒ่€ๅนดๆ–‘๏ผŒๆœ‰ ๆ•ˆๆทกๅŒ–้ป„่คๆ–‘่‡ณๆŽฅ่ฟ‘่‚ค่‰ฒใ€‚้ป›่Š™่–‡ๅฐ”้€š่ฟ‡ๆณ•ๅ›ฝใ€็พŽๅ›ฝใ€ๅฐๆนพ๏ฟฝ๏ฟฝ ๏ฟฝๅœฐ็š„ไธ“ๅฎถ้€šๅŠ›ๅไฝœ๏ผŒ่ถ…่ฟ‡10ๅนด็š„็ ”็ฉถไปฅๅ…จๆ–ฐ็š„DNA่‚Œ่‚คไฟฎๅคๆŠ€๏ฟฝ๏ฟฝ ๏ฟฝ๏ผŒๆŒ‘ๆˆ˜ไผ ็ปŸๅŒ–ๅญฆๆŠค่‚ค็†ๅฟต๏ผŒไธๆ‡ˆ่ฟฝๅฏปๅ‘็Žฐ็ ด่ฏ‘ๅคง่‡ช็„ถ็š„็พŽไธฝ๏ฟฝ ๏ฟฝ๏ฟฝ่ฟน๏ผŒไปคๆฏไธ€ไฝ็ˆฑ็พŽ็š„ๅฅณๆ€ง้ƒฝ่ƒฝไบซๅ—ๅˆฐ็ง‘ๆŠ€ๅˆ›ๆ–ฐๆ‰€ๅธฆๆฅ็š„่‡ช็„ถ ไน‹็พŽใ€‚ ไธ“ไธบไบšๆดฒๅฅณๆ€ง่‚ค่ดจ็ ”ๅˆถ๏ผŒ็ฒพๅฟƒๅ‘ตๆŠคๅฅณๆ€ง็พŽไธฝ๏ผŒๅคšๅนดๆฅ๏ผŒไธบๆ•ฐ๏ฟฝ๏ฟฝ ๏ฟฝ็™พไธ‡่ฎก็š„ๅฅณๆ€ง่งฃ้™คไบ†้ป„่คๆ–‘ๅ›ฐๆ‰ฐใ€‚ๆทฑๅพ—ๅนฟๅคงๅฅณๆ€งๆœ‹ๅ‹็š„ไฟก่ต–! ใ€€ใ€€3๏ผŒๅŽป้™ค้ป„่คๆ–‘ไน‹ๅŽ๏ผŒไผšๅๅผนๅ—? ใ€€ใ€€็ญ”๏ผšๅพˆๅคšๆ›พ็ป้•ฟไบ†้ป„่คๆ–‘็š„ไบบๅฃซ๏ผŒ่‡ชไปŽ้€‰ๆ‹ฉไบ†้ป›่Š™่–‡ๅฐ”๏ฟฝ๏ฟฝ ๏ฟฝ็™ฝ๏ผŒๅฐฑไธ€ๅŠณๆฐธ้€ธใ€‚่ฟ™ๆฌพ็ฅ›ๆ–‘ไบงๅ“ๆ˜ฏ็ป่ฟ‡ๆ•ฐๅไฝๆƒๅจ็ฅ›ๆ–‘ไธ“ๅฎถ๏ฟฝ ๏ฟฝ๏ฟฝๆฎๆ–‘็š„ๅฝขๆˆๅŽŸๅ› ็ฒพๅฟƒ็ ”ๅˆถ่€Œๆˆ็”จไบ‹ๅฎž่ฏด่ฏ๏ผŒ่ฎฉๆถˆ่ดน่€…ๆ‰“ๅˆ†ใ€‚ ๆ ‘็ซ‹ๆƒๅจๅ“็‰Œ!ๆˆ‘ไปฌ็š„ๅพˆๅคšๆ–ฐๅฎขๆˆท้ƒฝๆ˜ฏ่€ๅฎขๆˆทไป‹็ป่€Œๆฅ๏ผŒ่ฏท้—ฎ๏ฟฝ ๏ฟฝ๏ฟฝๅฆ‚ๆžœๆ•ˆๆžœไธๅฅฝ๏ผŒไผšๆœ‰ๅฎขๆˆท่ฝฌไป‹็ปๅ—? ใ€€ใ€€4๏ผŒไฝ ไปฌ็š„ไปทๆ ผๆœ‰็‚น่ดต๏ผŒ่ƒฝไธ่ƒฝไพฟๅฎœไธ€็‚น? ใ€€ใ€€็ญ”๏ผšๅฆ‚ๆžœๆ‚จไฝฟ็”จ่ฅฟ่ฏๆœ€ๅฐ‘้œ€่ฆ2000ๅ…ƒ๏ผŒ็…Žๆœ็š„่ฏๆœ€ๅฐ‘้œ€่ฆ3 000ๅ…ƒ๏ผŒๅšๆ‰‹ๆœฏๆœ€ๅฐ‘ๆ˜ฏ5000ๅ…ƒ๏ผŒ่€Œ่ฟ™ไบ›ๆฏซๆ— ็–‘้—ฎ๏ผŒไธไผšๅฏนๅฝปๅบ•ๅŽป๏ฟฝ ๏ฟฝ๏ฟฝไฝ ็š„ๆ–‘็‚นๆœ‰ไปปไฝ•ๅธฎๅŠฉ!ไธ€ๅˆ†ไปท้’ฑ๏ผŒไธ€ไปฝไปทๅ€ผ๏ผŒๆˆ‘ไปฌ็Žฐๅœจๅš็š„๏ฟฝ๏ฟฝ ๏ฟฝๆ˜ฏไธ€ไธชๅฃ็ข‘๏ผŒไธ€ไธชๅ“็‰Œ๏ผŒไปท้’ฑๅนถไธ้ซ˜ใ€‚ๅฆ‚ๆžœ่Šฑ่ฟ™็‚น้’ฑๆŠŠไฝ ็š„๏ฟฝ ๏ฟฝ๏ฟฝ่คๆ–‘ๅฝปๅบ•ๅŽป้™ค๏ผŒไฝ ่ฟ˜ไผš่ง‰ๅพ—่ดตๅ—?ไฝ ่ฟ˜ไผšๅ†ๅŽป่Šฑ้‚ฃไนˆๅคšๅ†คๆž‰๏ฟฝ๏ฟฝ ๏ฟฝ๏ผŒไธไฝ†ๆ–‘ๆฒกๅŽปๆމ๏ผŒ่ฟ˜ๆŠŠ่‡ชๅทฑ็š„็šฎ่‚คๅผ„็š„่ถŠๆฅ่ถŠ็ณŸๅ— ใ€€ใ€€5๏ผŒๆˆ‘้€‚ๅˆ็”จ้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒๅ—? ใ€€ใ€€็ญ”๏ผš้ป›่Š™่–‡ๅฐ”้€‚็”จไบบ็พค๏ผš ใ€€ใ€€1ใ€็”Ÿ็†็ดŠไนฑๅผ•่ตท็š„้ป„่คๆ–‘ไบบ็พค ใ€€ใ€€2ใ€็”Ÿ่‚ฒๅผ•่ตท็š„ๅฆŠๅจ ๆ–‘ไบบ็พค ใ€€ใ€€3ใ€ๅนด็บชๅขž้•ฟๅผ•่ตท็š„่€ๅนดๆ–‘ไบบ็พค ใ€€ใ€€4ใ€ๅŒ–ๅฆ†ๅ“่‰ฒ็ด ๆฒ‰็งฏใ€่พๅฐ„ๆ–‘ไบบ็พค ใ€€ใ€€5ใ€้•ฟๆœŸๆ—ฅ็…งๅผ•่ตท็š„ๆ—ฅๆ™’ๆ–‘ไบบ็พค ใ€€ใ€€6ใ€่‚Œ่‚คๆš—ๆทกๆ€ฅ้œ€็พŽ็™ฝ็š„ไบบ็พค ใ€Š็ฅ›ๆ–‘ๅฐๆ–นๆณ•ใ€‹ ่‰ฒๆ–‘ๅฆ‚ไฝ•ๆ‰่ƒฝๅŽป้™ค๏ผŒๅŒๆ—ถไธบๆ‚จๅˆ†ไบซ็ฅ›ๆ–‘ๅฐๆ–นๆณ• ่ŒถๆฐดๅŽปๆ–‘็พŽ็™ฝ ๆ–นๆณ•ไธ€๏ผšๆด—่„ธๅŽ๏ผŒๅฐ†่Œถๆฐดๆถ‚ๅˆฐ่„ธไธŠ๏ผŒๅนถ็”จๆ‰‹่ฝป่ฝปๆ‹่„ธใ€‚ ๆ–นๆณ•ไบŒ๏ผšๅฐ†่˜ธไบ†่Œถๆฐด็š„่„ฑ่„‚ๆฃ‰้™„ๅœจ่„ธไธŠ2-3ๅˆ†้’Ÿ๏ผŒ็„ถๅŽๆธ…ๆฐดๆด—๏ฟฝ๏ฟฝ ๏ฟฝ๏ผŒๆœ‰้™ค่‰ฒๆ–‘ใ€็พŽ็™ฝ็š„ๆ•ˆๆžœใ€‚ ``` ----- Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 4:22
non_test
่ต„่ฎฏ่‰ฒๆ–‘ๅฆ‚ไฝ•ๆ‰่ƒฝๅŽป้™ค ใ€Šๆ‘˜่ฆใ€‹ ๅ“ญ็š„ๆ—ถๅ€™๏ผŒๆˆ‘ไผš้—ญไธŠ็œผ็›ไธ่ฎฉๅฎƒๆตๆณช๏ผ›ๅญค็‹ฌๅฏ‚ๅฏž็š„ๆ—ถๅ€™๏ผŒๆˆ‘๏ฟฝ๏ฟฝ ๏ฟฝ้™้™็š„ๆƒณ็€ๆŸไบบ๏ผ›ไผคๅฟƒ็š„ๆ—ถๅ€™๏ผŒๆˆ‘ไผšๆ‰พไธชๅœฐๆ–น้™้™็š„ๅ‘ๅ‘†๏ผŒ๏ฟฝ ๏ฟฝ๏ฟฝๅŽๅ‘Š่ฏ‰่‡ชๅทฑ๏ผŒ่ฟ˜ๆ˜ฏ่ฆ้ขๅฏนๅšๆŒไธ‹ๅŽป๏ผ›้šพ่ฟ‡็š„ๆ—ถๅ€™๏ผŒๆˆ‘ไผšไผช่ฃ… ่‡ชๅทฑ๏ผŒๅฏนๅˆซไบบ่ฏด๏ผšๆˆ‘ๅพˆๅฅฝใ€ๆˆ‘ๅพˆๅผ€ๅฟƒ๏ผ›ๅคฑ่ฝ็š„ๆ—ถๅ€™๏ผŒๆˆ‘ไผš็ฌ‘๏ฟฝ๏ฟฝ ๏ฟฝๅฏน่‡ชๅทฑ่ฏด๏ผŒๆฒกไบ‹็š„๏ผŒไธ€ๅˆ‡ๆ€ปไผš่ฟ‡ๅŽปใ€‚้ป„่คๆ–‘ๆœ€ไธป่ฆ็š„ๅฐฑๆ˜ฏไธ๏ฟฝ ๏ฟฝ๏ฟฝ็ญ‰้•ฟไบ†้ป„่คๆ–‘ๆ‰ๆƒณ่ฆ็ฅ›ๆ–‘๏ผŒ้‚ฃๆ˜ฏๅพ—ไธๅฟๅคฑ็š„๏ผŒๅŠณ็ฅž่ดนๆ‰็š„ไบ‹ ๆƒ…ใ€‚ๅŒๆ—ถๅœจๅŽป้™ค้ป„่คๆ–‘็š„ๆ—ถๅ€™ไฝ ่ฆ็Ÿฅ้“้ข้ƒจ้ป„่คๆ–‘ๆ˜ฏๆ€Žไนˆๆฅ๏ฟฝ๏ฟฝ ๏ฟฝ๏ผŒ่ฟ™ๆ‰่ƒฝๆœ‰ๆ•ˆๅŽป้™ค้ป„่คๆ–‘ใ€‚่‰ฒๆ–‘ๅฆ‚ไฝ•ๆ‰่ƒฝๅŽป้™ค๏ผŒ ใ€Šๅฎขๆˆทๆกˆไพ‹ใ€‹ ใ€€ใ€€ไธบไบ†ๆˆ‘่„ธไธŠ็š„้ป„่คๆ–‘๏ผŒๅœจ็ฝ‘ไธŠๅฅ”ๆณขไบ†ๅฅฝ้•ฟๆ—ถ้—ดใ€‚ๅŽๆฅๅ‘๏ฟฝ๏ฟฝ ๏ฟฝใ€Œ้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒใ€ๅๅ“ไธ้”™๏ผŒ่ฟ˜ๆ–นไพฟ๏ผŒไธ็”จๆ•ดๅคฉๆถ‚ๆถ‚ๆŠนๆŠน๏ฟฝ ๏ฟฝ๏ฟฝไบŽๆ˜ฏๅฐฑๆŸฅไบ†ๅพˆๅคš็›ธๅ…ณ็š„่ต„ๆ–™ใ€‚็œ‹ๆฅๅ‘็Žฐๅฅฝไธœ่ฅฟ๏ผŒ่ฟ˜ๅพ—ๆ‹ฅๆœ‰ไธ€ ๅŒโ€œๆ…ง็œผโ€ๅ•Š๏ผŒ่ฏดใ€Œ้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒใ€ๅฅฝ๏ผŒๆœ€ไธป่ฆ็š„่ฟ˜ๆ˜ฏๅฎƒ๏ฟฝ๏ฟฝ ๏ฟฝๆˆ‘ๅŽปๆމๆˆ‘้‚ฃไบ›ๆผไบบ็š„้ป„่คๆ–‘ใ€‚ ใ€€ใ€€็Ÿฅ้“ไบ†ใ€Œ้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒใ€ๅŽ๏ผŒๅ‘ไธ“ๅฎถ่ฏ‰่ฏดๅฟƒไธญ็š„็ƒฆๆผ๏ฟฝ๏ฟฝ ๏ฟฝ็œ‹ๆ˜ฏๅฆๆœ‰ไธ“้—จ้’ˆๅฏน็”ทๆ€ง็š„็ฅ›ๆ–‘ไบงๅ“ใ€‚ไธ“ๅฎถๅˆ†ๆž่ฏด๏ผŒๆฎๆˆ‘็š„ๆ๏ฟฝ ๏ฟฝ๏ฟฝ๏ผŒไปŽๅค–ๅœจๅ› ็ด ็œ‹๏ผŒ้•ฟๆ–‘ไธป่ฆๆ˜ฏ่ทŸ้•ฟๆ—ถ้—ด็š„็”ต่„‘่พๅฐ„ๆœ‰ๅ…ณใ€‚ไธ“ ๅฎถ่ฏดใ€Œ้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒใ€ๅฐฑๆ˜ฏไปŽdna็ฒพๅŽๆˆ‘ไนŸ่ฎฐไธๆธ…ไบ†๏ผŒๅๆญฃ๏ฟฝ๏ฟฝ ๏ฟฝๆŒบๅฎ‰ๅ…จ็š„๏ผŒ่€Œไธ”ไฝฟ็”จไบ†็บณ็ฑณๆŠ€ๆœฏๅฎนๆ˜“ๅธๆ”ถ๏ผŒ่ฏด็”ทๅญฉๅญๅฎŒๅ…จๅฏ๏ฟฝ ๏ฟฝ๏ฟฝ็”จ็š„๏ผŒๅฏไธ”่ฟ˜ๅฏไปฅ้€ๆˆ‘ๅฏ่ดๅฐ”้ข่ดด่†œ้…ๅˆไฝฟ็”จๆ•ˆๆžœไผšๆ›ดๅฟซใ€‚ ไธ็ฎกไบ†๏ผŒๅช่ฆ่ƒฝ็ป™ๆˆ‘็š„้ป„่คๆ–‘ๅŽปๆމๅฐฑ่กŒ๏ผŒไบŽๆ˜ฏๆˆ‘ไธ€็‹ ๅฟƒ่ฎข่ดญ๏ฟฝ๏ฟฝ ๏ฟฝไธ‰ไธชๅ‘จๆœŸ๏ผŒๆฏๅคฉๅšๆŒ๏ผŒไปŽๆฅๆฒกๆœ‰่€ฝ่ฏฏ่ฟ‡ใ€‚ๆœ‹ๅ‹ไปฌ่ฟ˜็ฌ‘่ฟ‡่ทŸไธช๏ฟฝ ๏ฟฝ๏ฟฝไบบไธ€ๆ ท่ฟ˜ๆ˜ฏๆŽ’ๆฏ’ๅ…ป้ขœ็ฒพๅŽๆถฒๅ•Šใ€‚ๅ“Ž๏ผŒ่ฟ˜็œŸๅˆซ่ฏด๏ผŒไธ€ไธชๅ‘จๆœŸๅฎŒ ไบ†ไน‹ๅŽๆ•ˆๆžœ่ฟ˜็œŸไธ้”™๏ผŒไบŽๆ˜ฏๆ›ดๅŠ ๆœ‰ไฟกๅฟƒไบ†๏ผŒไนŸไธ็ฎกๅˆซไบบๆ€Žไนˆ๏ฟฝ๏ฟฝ ๏ฟฝ๏ผŒ่ƒฝๅŽปๆމ้ป„่คๆ–‘ๅฐฑๆ˜ฏๆญฃไบ‹ใ€‚็Žฐๅœจไธ‰ไธชๅ‘จๆœŸไฝฟ็”จๅฎŒไบ†่„ธไธŠ็š„้ป„๏ฟฝ ๏ฟฝ๏ฟฝๆ–‘่ฟ˜็œŸ็š„ๆฒกไบ†๏ผŒ่€Œไธ”ๆฒกๆœ‰็•™ไธ‹ไปปไฝ•็—•่ฟน๏ผŒ็šฎ่‚คไนŸๅ˜ๅพ—ๆฏ”ไปฅๅ‰ ๅฅฝไบ†ๅ‘ขใ€‚ๅ˜ฟๅ˜ฟ ใ€€ใ€€่ฟ™ๅฐฑๆ˜ฏๆˆ‘็š„็ฅ›ๆ–‘็ปๅކ๏ผŒไธบไบ†ๅธ…ๆฐ”็š„่„ธ๏ผŒไธบไบ†ๆˆ‘้‚ฃ็ฝ—ๆ›ผ่’‚๏ฟฝ๏ฟฝ ๏ฟฝ็š„็ˆฑๆƒ…๏ผŒๆˆ‘ๅฏๆฒกๅฐ‘่ดนๅฟƒๆ€ๅ•Š๏ผŒ่ฟ˜ๅฅฝๆœ€็ปˆ่ตขๅพ—ไบ†่ƒœๅˆฉใ€‚ๆœ€ๆƒณๆ„Ÿ๏ฟฝ ๏ฟฝ๏ฟฝ็š„ๅฐฑๆ˜ฏใ€Œ้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒใ€ไบ†๏ผŒๅคšไบๆœ‰่ฟ™ไนˆๅฅฝ็š„ไบงๅ“ ๆœ€ๆƒณ๏ฟฝ๏ฟฝ ๏ฟฝ่ฏ„็š„ๅฐฑๆ˜ฏ่‡ชๅทฑ๏ผŒไปฅๅŽไธ่ƒฝๅ†ๅคฉๅคฉๆฒ‰่ฟทไบŽๆธธๆˆไบ†ใ€‚ ้˜…่ฏปไบ†่‰ฒๆ–‘ๅฆ‚ไฝ•ๆ‰่ƒฝๅŽป้™ค๏ผŒๅ†็œ‹่„ธไธŠๅฎนๆ˜“้•ฟๆ–‘็š„ๅŽŸๅ› ๏ผš ใ€Š่‰ฒๆ–‘ๅฝขๆˆๅŽŸๅ› ใ€‹ ใ€€ใ€€ๅ†…้ƒจๅ› ็ด  ใ€€ใ€€ไธ€ใ€ๅŽ‹ๅŠ› ใ€€ใ€€ๅฝ“ไบบๅ—ๅˆฐๅŽ‹ๅŠ›ๆ—ถ๏ผŒๅฐฑไผšๅˆ†ๆณŒ่‚พไธŠ่…บ็ด ๏ผŒไธบๅฏนไป˜ๅŽ‹ๅŠ›่€Œๅš๏ฟฝ๏ฟฝ ๏ฟฝๅค‡ใ€‚ๅฆ‚ๆžœ้•ฟๆœŸๅ—ๅˆฐๅŽ‹ๅŠ›๏ผŒไบบไฝ“ๆ–ฐ้™ˆไปฃ่ฐข็š„ๅนณ่กกๅฐฑไผš้ญๅˆฐ็ ดๅ๏ฟฝ ๏ฟฝ๏ฟฝ็šฎ่‚คๆ‰€้œ€็š„่ฅๅ…ปไพ›ๅบ”่ถ‹ไบŽ็ผ“ๆ…ข๏ผŒ่‰ฒ็ด ๆฏ็ป†่ƒžๅฐฑไผšๅ˜ๅพ—ๅพˆๆดป่ทƒ ใ€‚ ใ€€ใ€€ไบŒใ€่ทๅฐ”่’™ๅˆ†ๆณŒๅคฑ่ฐƒ ใ€€ใ€€้ฟๅญ•่ฏ้‡Œๆ‰€ๅซ็š„ๅฅณๆ€ง่ทๅฐ”่’™้›Œๆฟ€็ด ๏ผŒไผšๅˆบๆฟ€้บฆๆ‹‰ๅฎ็ป†่ƒž๏ฟฝ๏ฟฝ ๏ฟฝๅˆ†ๆณŒ่€Œๅฝขๆˆไธๅ‡ๅŒ€็š„ๆ–‘็‚น๏ผŒๅ› ้ฟๅญ•่ฏ่€Œๅฝขๆˆ็š„ๆ–‘็‚น๏ผŒ่™ฝ็„ถๅœจ๏ฟฝ ๏ฟฝ๏ฟฝ่ฏไธญๆ–ญๅŽไผšๅœๆญข๏ผŒไฝ†ไปไผšๅœจ็šฎ่‚คไธŠๅœ็•™ๅพˆ้•ฟไธ€ๆฎตๆ—ถ้—ดใ€‚ๆ€€ๅญ• ไธญๅ› ๅฅณๆ€ง่ทๅฐ”่’™้›Œๆฟ€็ด ็š„ๅขžๅŠ ๏ผŒ โ€” ็Žฐๆ–‘๏ผŒ่ฟ™ๆ—ถๅ€™ๅ‡บ็Žฐ็š„ๆ–‘็‚นๅœจไบงๅŽๅคง้ƒจๅˆ†ไผšๆถˆๅคฑใ€‚ๅฏๆ˜ฏ๏ผŒๆ–ฐ้™ˆ๏ฟฝ๏ฟฝ ๏ฟฝ่ฐขไธๆญฃๅธธใ€่‚Œ่‚ค่ฃธ้œฒๅœจๅผบ็ƒˆ็š„็ดซๅค–็บฟไธ‹ใ€็ฒพ็ฅžไธŠๅ—ๅˆฐๅŽ‹ๅŠ›็ญ‰๏ฟฝ ๏ฟฝ๏ฟฝๅ› ๏ผŒ้ƒฝไผšไฝฟๆ–‘ๅŠ ๆทฑใ€‚ๆœ‰ๆ—ถๆ–ฐ้•ฟๅ‡บ็š„ๆ–‘๏ผŒไบงๅŽไนŸไธไผšๆถˆๅคฑ๏ผŒๆ‰€ ไปฅ้œ€่ฆๆ›ดๅŠ ๆณจๆ„ใ€‚ ใ€€ใ€€ไธ‰ใ€ๆ–ฐ้™ˆไปฃ่ฐข็ผ“ๆ…ข ใ€€ใ€€่‚็š„ๆ–ฐ้™ˆไปฃ่ฐขๅŠŸ่ƒฝไธๆญฃๅธธๆˆ–ๅตๅทขๅŠŸ่ƒฝๅ‡้€€ๆ—ถไนŸไผšๅ‡บ็Žฐๆ–‘๏ฟฝ๏ฟฝ ๏ฟฝๅ› ไธบๆ–ฐ้™ˆไปฃ่ฐขไธ้กบ็•…ใ€ๆˆ–ๅ†…ๅˆ†ๆณŒๅคฑ่ฐƒ๏ผŒไฝฟ่บซไฝ“ๅค„ไบŽๆ•ๆ„Ÿ็Šถๆ€๏ฟฝ ๏ฟฝ๏ฟฝ๏ผŒไปŽ่€ŒๅŠ ๅ‰ง่‰ฒ็ด ้—ฎ้ข˜ใ€‚ๆˆ‘ไปฌๅธธ่ฏด็š„ไพฟ็ง˜ไผšๅฝขๆˆๆ–‘๏ผŒๅ…ถๅฎžๅฐฑๆ˜ฏ ๅ†…ๅˆ†ๆณŒๅคฑ่ฐƒๅฏผ่‡ด่ฟ‡ๆ•ไฝ“่ดจ่€Œๅฝขๆˆ็š„ใ€‚ๅฆๅค–๏ผŒ่บซไฝ“็Šถๆ€ไธๆญฃๅธธ๏ฟฝ๏ฟฝ ๏ฟฝๆ—ถๅ€™๏ผŒ็ดซๅค–็บฟ็š„็…งๅฐ„ไนŸไผšๅŠ ้€Ÿๆ–‘็š„ๅฝขๆˆใ€‚ ใ€€ใ€€ๅ››ใ€้”™่ฏฏ็š„ไฝฟ็”จๅŒ–ๅฆ†ๅ“ ใ€€ใ€€ไฝฟ็”จไบ†ไธ้€‚ๅˆ่‡ชๅทฑ็šฎ่‚ค็š„ๅŒ–ๅฆ†ๅ“๏ผŒไผšๅฏผ่‡ด็šฎ่‚ค่ฟ‡ๆ•ใ€‚ๅœจ๏ฟฝ๏ฟฝ ๏ฟฝ็–—็š„่ฟ‡็จ‹ไธญๅฆ‚่ฟ‡้‡็…งๅฐ„ๅˆฐ็ดซๅค–็บฟ๏ผŒ็šฎ่‚คไผšไธบไบ†ๆŠตๅพกๅค–็•Œ็š„ไพต๏ฟฝ ๏ฟฝ๏ฟฝ๏ผŒๅœจๆœ‰็‚Ž็—‡็š„้ƒจไฝ่š้›†้บฆๆ‹‰ๅฎ่‰ฒ็ด ๏ผŒ่ฟ™ๆ ทไผšๅ‡บ็Žฐ่‰ฒ็ด ๆฒ‰็€็š„ ้—ฎ้ข˜ใ€‚ ใ€€ใ€€ๅค–้ƒจๅ› ็ด  ใ€€ใ€€ไธ€ใ€็ดซๅค–็บฟ ใ€€ใ€€็…งๅฐ„็ดซๅค–็บฟ็š„ๆ—ถๅ€™๏ผŒไบบไฝ“ไธบไบ†ไฟๆŠค็šฎ่‚ค๏ผŒไผšๅœจๅŸบๅบ•ๅฑ‚ไบง๏ฟฝ๏ฟฝ ๏ฟฝๅพˆๅคš้บฆๆ‹‰ๅฎ่‰ฒ็ด ใ€‚ๆ‰€ไปฅไธบไบ†ไฟๆŠค็šฎ่‚ค๏ผŒไผšๅœจๆ•ๆ„Ÿ้ƒจไฝ่š้›†ๆ›ด๏ฟฝ ๏ฟฝ๏ฟฝ็š„่‰ฒ็ด ใ€‚็ปๅธธ่ฃธ้œฒๅœจๅผบ็ƒˆ็š„้˜ณๅ…‰ๅบ•ไธ‹ไธไป…ไฟƒ่ฟ›็šฎ่‚ค็š„่€ๅŒ–๏ผŒ ่ฟ˜ไผšๅผ•่ตท้ป‘ๆ–‘ใ€้›€ๆ–‘็ญ‰่‰ฒ็ด ๆฒ‰็€็š„็šฎ่‚ค็–พๆ‚ฃใ€‚ ใ€€ใ€€ไบŒใ€ไธ่‰ฏ็š„ๆธ…ๆดไน ๆƒฏ ใ€€ใ€€ๅ› ๅผบ็ƒˆ็š„ๆธ…ๆดไน ๆƒฏไฝฟ็šฎ่‚คๅ˜ๅพ—ๆ•ๆ„Ÿ๏ผŒ่ฟ™ๆ ทไผšๅˆบๆฟ€็šฎ่‚คใ€‚๏ฟฝ๏ฟฝ ๏ฟฝ็šฎ่‚คๆ•ๆ„Ÿๆ—ถ๏ผŒไบบไฝ“ไธบไบ†ไฟๆŠค็šฎ่‚ค๏ผŒ้ป‘่‰ฒ็ด ็ป†่ƒžไผšๅˆ†ๆณŒๅพˆๅคš้บฆ๏ฟฝ ๏ฟฝ๏ฟฝๅฎ่‰ฒ็ด ๏ผŒๅฝ“่‰ฒ็ด ่ฟ‡ๅ‰ฉๆ—ถๅฐฑๅ‡บ็Žฐไบ†ๆ–‘ใ€็‘•็–ต็ญ‰็šฎ่‚ค่‰ฒ็ด ๆฒ‰็€็š„ ้—ฎ้ข˜ใ€‚ ใ€€ใ€€ไธ‰ใ€้—ไผ ๅŸบๅ›  ใ€€ใ€€็ˆถๆฏไธญๆœ‰้•ฟๆ–‘็š„๏ผŒๅˆ™ๆœฌไบบ้•ฟๆ–‘็š„ๆฆ‚็އๅฐฑๅพˆ้ซ˜๏ผŒ่ฟ™็งๆƒ…ๅ†ต๏ฟฝ๏ฟฝ ๏ฟฝไธ€ๅฎš็จ‹ๅบฆไธŠๅฐฑๅฏๅˆคๅฎšๆ˜ฏ้—ไผ ๅŸบๅ› ็š„ไฝœ็”จใ€‚ๆ‰€ไปฅๅฎถ้‡Œ็‰นๅˆซๆ˜ฏ้•ฟ๏ฟฝ ๏ฟฝ๏ฟฝๆœ‰้•ฟๆ–‘็š„ไบบ๏ผŒ่ฆๆณจๆ„้ฟๅ…ๅผ•ๅ‘้•ฟๆ–‘็š„้‡่ฆๅ› ็ด ไน‹ไธ€โ€”โ€”็ดซๅค– ็บฟ็…งๅฐ„๏ผŒ่ฟ™ๆ˜ฏ้ข„้˜ฒๆ–‘ๅฟ…้กปๆณจๆ„็š„ใ€‚ ใ€Šๆœ‰็–‘้—ฎๅธฎไฝ ่งฃๅ†ณใ€‹ ใ€€ใ€€ ้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒ็œŸ็š„ๆœ‰ๆ•ˆๆžœๅ— ็œŸ็š„ๅฏไปฅๆŠŠ่„ธไธŠ็š„้ป„่ค๏ฟฝ๏ฟฝ ๏ฟฝๅŽปๆމๅ— ใ€€ใ€€็ญ”๏ผš้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒdna็ฒพๅŽ่ƒฝๅคŸๆœ‰ๆ•ˆ็š„ไฟฎๅคๅ‘จๅ›ด้šพไปฅ่งฆ๏ฟฝ๏ฟฝ ๏ฟฝ็š„่‰ฒๆ–‘๏ผŒๅ…ถ็‹ฌๆœ‰็š„็บณ่ฑ†ๆˆๅˆ†ไธบ็šฎ่‚ค็š„็พŽ็™ฝไธŽ้“ไธฝ๏ผŒๆไพ›ไบ†ๅฟ…๏ฟฝ ๏ฟฝ๏ฟฝๅฏๅฐ‘็š„่ฅๅ…ป็‰ฉ่ดจ๏ผŒๅฏไปฅๆœ‰ๆ•ˆ็š„ๅŽป้™ค้ป„่คๆ–‘๏ผŒ้ป„่คๆ–‘๏ผŒ้ป„่คๆ–‘ ๏ผŒ่ด่ถๆ–‘๏ผŒๆ™’ๆ–‘ใ€ๅฆŠๅจ ๆ–‘็ญ‰ใ€‚ๅฎƒๅฎƒๅฎŒๅ…จ็ช็ ดไบ†ไผ ็ปŸ็š„็พŽ่‚คๆ—ถ๏ฟฝ๏ฟฝ ๏ฟฝ๏ผŒๅฎ›ๅฆ‚ๅœจ็šฎ่‚คไธญๆณจๅ…ฅไบ†ไธ€ๆฏๅ…ผๅ…ทๆดปๅŒ–ใ€ๅ†็”Ÿใ€ๆป‹ๅ…ป็ญ‰ๅŠŸๆ•ˆ็š„๏ฟฝ ๏ฟฝ๏ฟฝๅฐพ้…’๏ผŒๅŒๆ—ถไธบ่„ธ้ƒจๆไพ›ๅคง้‡ๆœ‰ๆœบ็ปด็”Ÿ็ด ็ฒพๅŽ๏ผŒ่„ธ้ƒจ็š„ๆ”นๅ˜ๆ˜พ ่€Œๆ˜“่งใ€‚่‡ชไบงๅ“ไธŠๅธ‚ไปฅๆฅ๏ผŒ่€้กพๅฎข็บท็บทไป‹็ปๆ–ฐ้กพๅฎข๏ผŒ ็š„ๆ–ฐ๏ฟฝ๏ฟฝ ๏ฟฝๅฎข้ƒฝๆ˜ฏ้€š่ฟ‡่€้กพๅฎขไป‹็ป่€Œๆฅ๏ผŒๅฃ็ข‘็”ฑๆญค่€Œๆฅ ใ€€ใ€€ ๏ผŒๆœ็”จ้ป›่Š™่–‡ๅฐ”็พŽ็™ฝ๏ผŒไผšไผค่บซไฝ“ๅ— ๆœ‰ๅ‰ฏไฝœ็”จๅ— ใ€€ใ€€็ญ”๏ผš้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒๅบ”็”จไบ†็ฒพ็บฏๅคๅˆ้…ๆ–นๅ’Œ้ข†ๅ…ˆ็š„ๅˆ†็ฑป๏ฟฝ๏ฟฝ ๏ฟฝๆ–‘็ง‘ๆŠ€๏ผŒๅนถๅฐ†โ€œdna็พŽ่‚ค็ณป็ปŸโ€็–—ๆณ•ๅบ”็”จๅˆฐไบ†่ฏฅไบงๅ“ไธญ๏ผŒ่ƒฝๅฝป๏ฟฝ ๏ฟฝ๏ฟฝ็ฅ›้™ค้ป„่คๆ–‘๏ผŒ่ด่ถๆ–‘๏ผŒๅฆŠๅจ ๆ–‘๏ผŒๆ™’ๆ–‘๏ผŒ้ป„่คๆ–‘๏ผŒ่€ๅนดๆ–‘๏ผŒๆœ‰ ๆ•ˆๆทกๅŒ–้ป„่คๆ–‘่‡ณๆŽฅ่ฟ‘่‚ค่‰ฒใ€‚้ป›่Š™่–‡ๅฐ”้€š่ฟ‡ๆณ•ๅ›ฝใ€็พŽๅ›ฝใ€ๅฐๆนพ๏ฟฝ๏ฟฝ ๏ฟฝๅœฐ็š„ไธ“ๅฎถ้€šๅŠ›ๅไฝœ๏ผŒ ๏ฟฝ๏ฟฝ ๏ฟฝ๏ผŒๆŒ‘ๆˆ˜ไผ ็ปŸๅŒ–ๅญฆๆŠค่‚ค็†ๅฟต๏ผŒไธๆ‡ˆ่ฟฝๅฏปๅ‘็Žฐ็ ด่ฏ‘ๅคง่‡ช็„ถ็š„็พŽไธฝ๏ฟฝ ๏ฟฝ๏ฟฝ่ฟน๏ผŒไปคๆฏไธ€ไฝ็ˆฑ็พŽ็š„ๅฅณๆ€ง้ƒฝ่ƒฝไบซๅ—ๅˆฐ็ง‘ๆŠ€ๅˆ›ๆ–ฐๆ‰€ๅธฆๆฅ็š„่‡ช็„ถ ไน‹็พŽใ€‚ ไธ“ไธบไบšๆดฒๅฅณๆ€ง่‚ค่ดจ็ ”ๅˆถ๏ผŒ็ฒพๅฟƒๅ‘ตๆŠคๅฅณๆ€ง็พŽไธฝ๏ผŒๅคšๅนดๆฅ๏ผŒไธบๆ•ฐ๏ฟฝ๏ฟฝ ๏ฟฝ็™พไธ‡่ฎก็š„ๅฅณๆ€ง่งฃ้™คไบ†้ป„่คๆ–‘ๅ›ฐๆ‰ฐใ€‚ๆทฑๅพ—ๅนฟๅคงๅฅณๆ€งๆœ‹ๅ‹็š„ไฟก่ต– ใ€€ใ€€ ๏ผŒๅŽป้™ค้ป„่คๆ–‘ไน‹ๅŽ๏ผŒไผšๅๅผนๅ— ใ€€ใ€€็ญ”๏ผšๅพˆๅคšๆ›พ็ป้•ฟไบ†้ป„่คๆ–‘็š„ไบบๅฃซ๏ผŒ่‡ชไปŽ้€‰ๆ‹ฉไบ†้ป›่Š™่–‡ๅฐ”๏ฟฝ๏ฟฝ ๏ฟฝ็™ฝ๏ผŒๅฐฑไธ€ๅŠณๆฐธ้€ธใ€‚่ฟ™ๆฌพ็ฅ›ๆ–‘ไบงๅ“ๆ˜ฏ็ป่ฟ‡ๆ•ฐๅไฝๆƒๅจ็ฅ›ๆ–‘ไธ“ๅฎถ๏ฟฝ ๏ฟฝ๏ฟฝๆฎๆ–‘็š„ๅฝขๆˆๅŽŸๅ› ็ฒพๅฟƒ็ ”ๅˆถ่€Œๆˆ็”จไบ‹ๅฎž่ฏด่ฏ๏ผŒ่ฎฉๆถˆ่ดน่€…ๆ‰“ๅˆ†ใ€‚ ๆ ‘็ซ‹ๆƒๅจๅ“็‰Œ ๆˆ‘ไปฌ็š„ๅพˆๅคšๆ–ฐๅฎขๆˆท้ƒฝๆ˜ฏ่€ๅฎขๆˆทไป‹็ป่€Œๆฅ๏ผŒ่ฏท้—ฎ๏ฟฝ ๏ฟฝ๏ฟฝๅฆ‚ๆžœๆ•ˆๆžœไธๅฅฝ๏ผŒไผšๆœ‰ๅฎขๆˆท่ฝฌไป‹็ปๅ— ใ€€ใ€€ ๏ผŒไฝ ไปฌ็š„ไปทๆ ผๆœ‰็‚น่ดต๏ผŒ่ƒฝไธ่ƒฝไพฟๅฎœไธ€็‚น ใ€€ใ€€็ญ”๏ผš ๏ผŒ ๏ผŒ ๏ผŒ่€Œ่ฟ™ไบ›ๆฏซๆ— ็–‘้—ฎ๏ผŒไธไผšๅฏนๅฝปๅบ•ๅŽป๏ฟฝ ๏ฟฝ๏ฟฝไฝ ็š„ๆ–‘็‚นๆœ‰ไปปไฝ•ๅธฎๅŠฉ ไธ€ๅˆ†ไปท้’ฑ๏ผŒไธ€ไปฝไปทๅ€ผ๏ผŒๆˆ‘ไปฌ็Žฐๅœจๅš็š„๏ฟฝ๏ฟฝ ๏ฟฝๆ˜ฏไธ€ไธชๅฃ็ข‘๏ผŒไธ€ไธชๅ“็‰Œ๏ผŒไปท้’ฑๅนถไธ้ซ˜ใ€‚ๅฆ‚ๆžœ่Šฑ่ฟ™็‚น้’ฑๆŠŠไฝ ็š„๏ฟฝ ๏ฟฝ๏ฟฝ่คๆ–‘ๅฝปๅบ•ๅŽป้™ค๏ผŒไฝ ่ฟ˜ไผš่ง‰ๅพ—่ดตๅ— ไฝ ่ฟ˜ไผšๅ†ๅŽป่Šฑ้‚ฃไนˆๅคšๅ†คๆž‰๏ฟฝ๏ฟฝ ๏ฟฝ๏ผŒไธไฝ†ๆ–‘ๆฒกๅŽปๆމ๏ผŒ่ฟ˜ๆŠŠ่‡ชๅทฑ็š„็šฎ่‚คๅผ„็š„่ถŠๆฅ่ถŠ็ณŸๅ— ใ€€ใ€€ ๏ผŒๆˆ‘้€‚ๅˆ็”จ้ป›่Š™่–‡ๅฐ”็ฒพๅŽๆถฒๅ— ใ€€ใ€€็ญ”๏ผš้ป›่Š™่–‡ๅฐ”้€‚็”จไบบ็พค๏ผš ใ€€ใ€€ ใ€็”Ÿ็†็ดŠไนฑๅผ•่ตท็š„้ป„่คๆ–‘ไบบ็พค ใ€€ใ€€ ใ€็”Ÿ่‚ฒๅผ•่ตท็š„ๅฆŠๅจ ๆ–‘ไบบ็พค ใ€€ใ€€ ใ€ๅนด็บชๅขž้•ฟๅผ•่ตท็š„่€ๅนดๆ–‘ไบบ็พค ใ€€ใ€€ ใ€ๅŒ–ๅฆ†ๅ“่‰ฒ็ด ๆฒ‰็งฏใ€่พๅฐ„ๆ–‘ไบบ็พค ใ€€ใ€€ ใ€้•ฟๆœŸๆ—ฅ็…งๅผ•่ตท็š„ๆ—ฅๆ™’ๆ–‘ไบบ็พค ใ€€ใ€€ ใ€่‚Œ่‚คๆš—ๆทกๆ€ฅ้œ€็พŽ็™ฝ็š„ไบบ็พค ใ€Š็ฅ›ๆ–‘ๅฐๆ–นๆณ•ใ€‹ ่‰ฒๆ–‘ๅฆ‚ไฝ•ๆ‰่ƒฝๅŽป้™ค๏ผŒๅŒๆ—ถไธบๆ‚จๅˆ†ไบซ็ฅ›ๆ–‘ๅฐๆ–นๆณ• ่ŒถๆฐดๅŽปๆ–‘็พŽ็™ฝ ๆ–นๆณ•ไธ€๏ผšๆด—่„ธๅŽ๏ผŒๅฐ†่Œถๆฐดๆถ‚ๅˆฐ่„ธไธŠ๏ผŒๅนถ็”จๆ‰‹่ฝป่ฝปๆ‹่„ธใ€‚ ๆ–นๆณ•ไบŒ๏ผš ๏ผŒ็„ถๅŽๆธ…ๆฐดๆด—๏ฟฝ๏ฟฝ ๏ฟฝ๏ผŒๆœ‰้™ค่‰ฒๆ–‘ใ€็พŽ็™ฝ็š„ๆ•ˆๆžœใ€‚ original issue reported on code google com by additive gmail com on jul at
0
161,253
6,111,647,712
IssuesEvent
2017-06-21 17:31:24
vmware/vic
https://api.github.com/repos/vmware/vic
closed
Diagnostics: logging from bootstrap scripts not present in debug log
component/container-base kind/diagnostics priority/low
The messages in isos/bootstrap/bootstrap script are not present in the containerVM debug logs. We have the ISOLINUX message from the bootloader and the messages from the tether, but nothing in between.
1.0
Diagnostics: logging from bootstrap scripts not present in debug log - The messages in isos/bootstrap/bootstrap script are not present in the containerVM debug logs. We have the ISOLINUX message from the bootloader and the messages from the tether, but nothing in between.
non_test
diagnostics logging from bootstrap scripts not present in debug log the messages in isos bootstrap bootstrap script are not present in the containervm debug logs we have the isolinux message from the bootloader and the messages from the tether but nothing in between
0
68,514
9,198,021,112
IssuesEvent
2019-03-07 11:27:46
Pageworks/fuel-pjax
https://api.github.com/repos/Pageworks/fuel-pjax
opened
Better Debugging
documentation enhancement
**What:** Cleanup and modify the debug console logs to match the format that all the other projects are currently following. We should also modify the initial debug status dump on `init()` so that it provides some potentially useful information such as the active version of Pjax.
1.0
Better Debugging - **What:** Cleanup and modify the debug console logs to match the format that all the other projects are currently following. We should also modify the initial debug status dump on `init()` so that it provides some potentially useful information such as the active version of Pjax.
non_test
better debugging what cleanup and modify the debug console logs to match the format that all the other projects are currently following we should also modify the initial debug status dump on init so that it provides some potentially useful information such as the active version of pjax
0
399,042
11,742,643,715
IssuesEvent
2020-03-12 01:33:18
thaliawww/concrexit
https://api.github.com/repos/thaliawww/concrexit
closed
Banner Dimensions Not in Styleguide
bug easy and fun priority: low style
In GitLab by thijsdejong on Apr 10, 2019, 13:12 ### One-sentence description ### Current behaviour The banner dimensions are not shown at the current page of the styleguide. ### Expected behaviour Implement the banner dimensions on the styleguide page.
1.0
Banner Dimensions Not in Styleguide - In GitLab by thijsdejong on Apr 10, 2019, 13:12 ### One-sentence description ### Current behaviour The banner dimensions are not shown at the current page of the styleguide. ### Expected behaviour Implement the banner dimensions on the styleguide page.
non_test
banner dimensions not in styleguide in gitlab by thijsdejong on apr one sentence description current behaviour the banner dimensions are not shown at the current page of the styleguide expected behaviour implement the banner dimensions on the styleguide page
0
275,718
23,932,980,583
IssuesEvent
2022-09-10 20:41:06
bitcoin/bitcoin
https://api.github.com/repos/bitcoin/bitcoin
closed
feature_config_args.py failure
Tests Questions and Help
Run test: `python3 feature_config_args.py` Results in error: > > 2022-08-01T21:05:29.574000Z TestFramework (INFO): Initializing test directory /tmp/bitcoin_func_test_die2tixp > 2022-08-01T21:05:30.306000Z TestFramework (INFO): Test config args logging > 2022-08-01T21:05:30.564000Z TestFramework (INFO): Test seed peers > 2022-08-01T21:05:40.697000Z TestFramework (ERROR): Assertion failed > Traceback (most recent call last): > File "/home/user/bitcoin/test/functional/test_framework/test_framework.py", line 133, in main > self.run_test() > File "/home/user/bitcoin/test/functional/feature_config_args.py", line 229, in run_test > self.test_seed_peers() > File "/home/user/bitcoin/test/functional/feature_config_args.py", line 175, in test_seed_peers > self.start_node(0, extra_args=['-dnsseed=1', '-fixedseeds=1', f'-mocktime={start}']) > File "/usr/lib/python3.9/contextlib.py", line 124, in __exit__ > next(self.gen) > File "/home/user/bitcoin/test/functional/test_framework/test_node.py", line 423, in assert_debug_log > self._raise_assertion_error('Expected messages "{}" does not partially match log:\n\n{}\n\n'.format(str(expected_msgs), print_log)) > File "/home/user/bitcoin/test/functional/test_framework/test_node.py", line 167, in _raise_assertion_error > raise AssertionError(self._node_msg(msg)) > AssertionError: [node 0] Expected messages "['Loaded 0 addresses from peers.dat', '0 addresses found from DNS seeds', 'opencon thread start']" does not partially match log: > > > > > > > 2022-08-01T21:05:30.776965Z [init] [init/common.cpp:127] [LogPackageVersion] Bitcoin Core version v23.99.0-ce3b75690d10 (release build) > 2022-08-01T21:05:30.777017Z [init] [init.cpp:657] [InitParameterInteraction] InitParameterInteraction: parameter interaction: -bind set -> setting -listen=1 > 2022-08-01T21:05:30.777088Z [init] [init.cpp:922] [AppInitParameterInteraction] Validating signatures for all blocks. > 2022-08-01T21:05:30.777094Z [init] [init.cpp:933] [AppInitParameterInteraction] Setting nMinimumChainWork=0000000000000000000000000000000000000000000000000000000000000000 > 2022-08-01T21:05:30.777164Z (mocktime: 2022-08-01T21:05:30Z) [init] [kernel/context.cpp:21] [Context] Using the 'sse4(1way),sse41(4way),avx2(8way)' SHA256 implementation > 2022-08-01T21:05:30.777175Z (mocktime: 2022-08-01T21:05:30Z) [init] [random.cpp:100] [ReportHardwareRand] Using RdSeed as an additional entropy source > 2022-08-01T21:05:30.777180Z (mocktime: 2022-08-01T21:05:30Z) [init] [random.cpp:103] [ReportHardwareRand] Using RdRand as an additional entropy source > 2022-08-01T21:05:30.778975Z (mocktime: 2022-08-01T21:05:30Z) [init] [init/common.cpp:98] [StartLogging] Default data directory /home/user/.bitcoin > 2022-08-01T21:05:30.778985Z (mocktime: 2022-08-01T21:05:30Z) [init] [init/common.cpp:99] [StartLogging] Using data directory /tmp/bitcoin_func_test_die2tixp/node0/regtest > 2022-08-01T21:05:30.778998Z (mocktime: 2022-08-01T21:05:30Z) [init] [init/common.cpp:104] [StartLogging] Config file: /tmp/bitcoin_func_test_die2tixp/node0/bitcoin.conf > 2022-08-01T21:05:30.779015Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: regtest="1" > 2022-08-01T21:05:30.779025Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] bind="127.0.0.1" > 2022-08-01T21:05:30.779032Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] discover="0" > 2022-08-01T21:05:30.779039Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] dnsseed="0" > 2022-08-01T21:05:30.779047Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] fallbackfee="0.0002" > 2022-08-01T21:05:30.779053Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] fixedseeds="0" > 2022-08-01T21:05:30.779060Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] keypool="1" > 2022-08-01T21:05:30.779067Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] listenonion="0" > 2022-08-01T21:05:30.779074Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] natpmp="0" > 2022-08-01T21:05:30.779082Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] peertimeout="999999999" > 2022-08-01T21:05:30.779089Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] port="11093" > 2022-08-01T21:05:30.779096Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] printtoconsole="0" > 2022-08-01T21:05:30.779104Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] rpcdoccheck="1" > 2022-08-01T21:05:30.779112Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] rpcport="16093" > 2022-08-01T21:05:30.779119Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] server="1" > 2022-08-01T21:05:30.779127Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] shrinkdebugfile="0" > 2022-08-01T21:05:30.779134Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] unsafesqlitesync="1" > 2022-08-01T21:05:30.779141Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] upnp="0" > 2022-08-01T21:05:30.779149Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: datadir="/tmp/bitcoin_func_test_die2tixp/node0" > 2022-08-01T21:05:30.779156Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: debug="" > 2022-08-01T21:05:30.779164Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: debugexclude="libevent" > 2022-08-01T21:05:30.779171Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: debugexclude="leveldb" > 2022-08-01T21:05:30.779178Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: dnsseed="1" > 2022-08-01T21:05:30.779184Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: fixedseeds="1" > 2022-08-01T21:05:30.779192Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: logsourcelocations="" > 2022-08-01T21:05:30.779199Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: logthreadnames="" > 2022-08-01T21:05:30.779206Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: logtimemicros="" > 2022-08-01T21:05:30.779213Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: mocktime="1659387930" > 2022-08-01T21:05:30.779224Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: uacomment="testnode0" > 2022-08-01T21:05:30.779232Z (mocktime: 2022-08-01T21:05:30Z) [init] [init.cpp:1148] [AppInitMain] Using at most 125 automatic connections (1024 file descriptors available) > 2022-08-01T21:05:30.788089Z (mocktime: 2022-08-01T21:05:30Z) [init] [script/sigcache.cpp:101] [InitSignatureCache] Using 16 MiB out of 32/2 requested for signature cache, able to store 524288 elements > 2022-08-01T21:05:30.796940Z (mocktime: 2022-08-01T21:05:30Z) [init] [validation.cpp:1669] [InitScriptExecutionCache] Using 16 MiB out of 32/2 requested for script execution cache, able to store 524288 elements > 2022-08-01T21:05:30.796987Z (mocktime: 2022-08-01T21:05:30Z) [init] [init.cpp:1175] [AppInitMain] Script verification uses 1 additional threads > 2022-08-01T21:05:30.797361Z (mocktime: 2022-08-01T21:05:30Z) [scheduler] [util/thread.cpp:18] [TraceThread] scheduler thread start > 2022-08-01T21:05:30.806770Z (mocktime: 2022-08-01T21:05:30Z) [init] [httpserver.cpp:184] [InitHTTPAllowList] [http] Allowing HTTP connections from: 127.0.0.0/8 ::1/128 > 2022-08-01T21:05:30.806890Z (mocktime: 2022-08-01T21:05:30Z) [init] [httpserver.cpp:321] [HTTPBindAddresses] [http] Binding RPC on address ::1 port 16093 > 2022-08-01T21:05:30.807109Z (mocktime: 2022-08-01T21:05:30Z) [init] [httpserver.cpp:321] [HTTPBindAddresses] [http] Binding RPC on address 127.0.0.1 port 16093 > 2022-08-01T21:05:30.807155Z (mocktime: 2022-08-01T21:05:30Z) [init] [httpserver.cpp:401] [InitHTTPServer] [http] Initialized HTTP server > 2022-08-01T21:05:30.807167Z (mocktime: 2022-08-01T21:05:30Z) [init] [httpserver.cpp:403] [InitHTTPServer] [http] creating work queue of depth 16 > 2022-08-01T21:05:30.807177Z (mocktime: 2022-08-01T21:05:30Z) [init] [rpc/server.cpp:291] [StartRPC] [rpc] Starting RPC > 2022-08-01T21:05:30.807193Z (mocktime: 2022-08-01T21:05:30Z) [init] [httprpc.cpp:296] [StartHTTPRPC] [rpc] Starting HTTP RPC server > 2022-08-01T21:05:30.807201Z (mocktime: 2022-08-01T21:05:30Z) [init] [httprpc.cpp:245] [InitRPCAuthentication] Using random cookie authentication. > 2022-08-01T21:05:30.807297Z (mocktime: 2022-08-01T21:05:30Z) [init] [rpc/request.cpp:106] [GenerateAuthCookie] Generated RPC authentication cookie /tmp/bitcoin_func_test_die2tixp/node0/regtest/.cookie > 2022-08-01T21:05:30.807312Z (mocktime: 2022-08-01T21:05:30Z) [init] [httpserver.cpp:676] [RegisterHTTPHandler] [http] Registering HTTP handler for / (exactmatch 1) > 2022-08-01T21:05:30.807321Z (mocktime: 2022-08-01T21:05:30Z) [init] [httpserver.cpp:676] [RegisterHTTPHandler] [http] Registering HTTP handler for /wallet/ (exactmatch 0) > 2022-08-01T21:05:30.807335Z (mocktime: 2022-08-01T21:05:30Z) [init] [httpserver.cpp:425] [StartHTTPServer] [http] Starting HTTP server > 2022-08-01T21:05:30.807343Z (mocktime: 2022-08-01T21:05:30Z) [init] [httpserver.cpp:427] [StartHTTPServer] [http] starting 4 worker threads > 2022-08-01T21:05:30.807552Z (mocktime: 2022-08-01T21:05:30Z) [http] [httpserver.cpp:287] [ThreadHTTP] [http] Entering http event loop > 2022-08-01T21:05:30.807784Z (mocktime: 2022-08-01T21:05:30Z) [init] [wallet/load.cpp:51] [VerifyWallets] Using wallet directory /tmp/bitcoin_func_test_die2tixp/node0/regtest/wallets > 2022-08-01T21:05:30.807801Z (mocktime: 2022-08-01T21:05:30Z) [init] [noui.cpp:56] [noui_InitMessage] init message: Verifying wallet(s)โ€ฆ > 2022-08-01T21:05:30.807840Z (mocktime: 2022-08-01T21:05:30Z) [init] [init.cpp:1261] [AppInitMain] Using /16 prefix for IP bucketing > 2022-08-01T21:05:30.807850Z (mocktime: 2022-08-01T21:05:30Z) [init] [noui.cpp:56] [noui_InitMessage] init message: Loading P2P addressesโ€ฆ > 2022-08-01T21:05:30.808116Z (mocktime: 2022-08-01T21:05:30Z) [init] [logging/timer.h:57] [Log] [addrman] CheckAddrman: new 0, tried 0, total 0 started > 2022-08-01T21:05:30.808228Z (mocktime: 2022-08-01T21:05:30Z) [init] [logging/timer.h:57] [Log] [addrman] CheckAddrman: completed (0.00ms) > 2022-08-01T21:05:30.808242Z (mocktime: 2022-08-01T21:05:30Z) [init] [addrdb.cpp:194] [LoadAddrman] Loaded 0 addresses from peers.dat 0ms > 2022-08-01T21:05:30.808586Z (mocktime: 2022-08-01T21:05:30Z) [init] [noui.cpp:56] [noui_InitMessage] init message: Loading banlistโ€ฆ > 2022-08-01T21:05:30.808626Z (mocktime: 2022-08-01T21:05:30Z) [init] [banman.cpp:38] [LoadBanlist] [net] Loaded 0 banned node addresses/subnets 0ms > 2022-08-01T21:05:30.808645Z (mocktime: 2022-08-01T21:05:30Z) [init] [net.cpp:1523] [SetTryNewOutboundPeer] [net] setting try another outbound peer=false > 2022-08-01T21:05:30.808654Z (mocktime: 2022-08-01T21:05:30Z) [init] [net.cpp:2170] [SetNetworkActive] SetNetworkActive: true > 2022-08-01T21:05:30.810409Z (mocktime: 2022-08-01T21:05:30Z) [init] [policy/fees.cpp:455] [Read] [estimatefee] Reading estimates: 190 buckets counting confirms up to 48 blocks > 2022-08-01T21:05:30.810608Z (mocktime: 2022-08-01T21:05:30Z) [init] [policy/fees.cpp:455] [Read] [estimatefee] Reading estimates: 190 buckets counting confirms up to 12 blocks > 2022-08-01T21:05:30.811134Z (mocktime: 2022-08-01T21:05:30Z) [init] [policy/fees.cpp:455] [Read] [estimatefee] Reading estimates: 190 buckets counting confirms up to 1008 blocks > 2022-08-01T21:05:30.811253Z (mocktime: 2022-08-01T21:05:30Z) [init] [init.cpp:1403] [AppInitMain] Cache configuration: > 2022-08-01T21:05:30.811267Z (mocktime: 2022-08-01T21:05:30Z) [init] [init.cpp:1404] [AppInitMain] * Using 2.0 MiB for block index database > 2022-08-01T21:05:30.811276Z (mocktime: 2022-08-01T21:05:30Z) [init] [init.cpp:1412] [AppInitMain] * Using 8.0 MiB for chain state database > 2022-08-01T21:05:30.811290Z (mocktime: 2022-08-01T21:05:30Z) [init] [init.cpp:1428] [AppInitMain] * Using 440.0 MiB for in-memory UTXO set (plus up to 286.1 MiB of unused mempool space) > 2022-08-01T21:05:30.811315Z (mocktime: 2022-08-01T21:05:30Z) [init] [noui.cpp:56] [noui_InitMessage] init message: Loading block indexโ€ฆ > 2022-08-01T21:05:30.811330Z (mocktime: 2022-08-01T21:05:30Z) [init] [validation.cpp:4710] [InitializeChainstate] Switching active chainstate to Chainstate [ibd] @ height -1 (null) > 2022-08-01T21:05:30.811367Z (mocktime: 2022-08-01T21:05:30Z) [init] [dbwrapper.cpp:150] [CDBWrapper] Opening LevelDB in /tmp/bitcoin_func_test_die2tixp/node0/regtest/blocks/index > 2022-08-01T21:05:30.826887Z (mocktime: 2022-08-01T21:05:30Z) [init] [dbwrapper.cpp:158] [CDBWrapper] Opened LevelDB successfully > 2022-08-01T21:05:30.826920Z (mocktime: 2022-08-01T21:05:30Z) [init] [dbwrapper.cpp:183] [CDBWrapper] Using obfuscation key for /tmp/bitcoin_func_test_die2tixp/node0/regtest/blocks/index: 0000000000000000 > 2022-08-01T21:05:30.827091Z (mocktime: 2022-08-01T21:05:30Z) [init] [node/blockstorage.cpp:331] [LoadBlockIndexDB] LoadBlockIndexDB: last block file = 0 > 2022-08-01T21:05:30.827110Z (mocktime: 2022-08-01T21:05:30Z) [init] [node/blockstorage.cpp:335] [LoadBlockIndexDB] LoadBlockIndexDB: last block file info: CBlockFileInfo(blocks=1, size=293, heights=0...0, time=2011-02-02...2011-02-02) > 2022-08-01T21:05:30.827119Z (mocktime: 2022-08-01T21:05:30Z) [init] [node/blockstorage.cpp:346] [LoadBlockIndexDB] Checking all blk files are present... > 2022-08-01T21:05:30.827164Z (mocktime: 2022-08-01T21:05:30Z) [init] [dbwrapper.cpp:150] [CDBWrapper] Opening LevelDB in /tmp/bitcoin_func_test_die2tixp/node0/regtest/chainstate > 2022-08-01T21:05:30.846337Z (mocktime: 2022-08-01T21:05:30Z) [init] [dbwrapper.cpp:158] [CDBWrapper] Opened LevelDB successfully > 2022-08-01T21:05:30.846444Z (mocktime: 2022-08-01T21:05:30Z) [init] [dbwrapper.cpp:183] [CDBWrapper] Using obfuscation key for /tmp/bitcoin_func_test_die2tixp/node0/regtest/chainstate: 223bb69586684f13 > 2022-08-01T21:05:30.846548Z (mocktime: 2022-08-01T21:05:30Z) [init] [validation.cpp:3894] [LoadChainTip] Loaded best chain: hashBestChain=0f9188f13cb7b2c71f2a335e3a4fc328bf5beb436012afca590b1a11466e2206 height=0 date=2011-02-02T23:16:42Z progress=1.000000 > 2022-08-01T21:05:30.846567Z (mocktime: 2022-08-01T21:05:30Z) [init] [noui.cpp:56] [noui_InitMessage] init message: Verifying blocksโ€ฆ > 2022-08-01T21:05:30.846583Z (mocktime: 2022-08-01T21:05:30Z) [init] [init.cpp:1474] [AppInitMain] block index 35ms > 2022-08-01T21:05:30.847557Z (mocktime: 2022-08-01T21:05:30Z) [init] [init.cpp:1631] [AppInitMain] block tree size = 1 > 2022-08-01T21:05:30.847581Z (mocktime: 2022-08-01T21:05:30Z) [init] [init.cpp:1643] [AppInitMain] nBestHeight = 0 > 2022-08-01T21:05:30.847637Z (mocktime: 2022-08-01T21:05:30Z) [loadblk] [util/thread.cpp:18] [TraceThread] loadblk thread start > 2022-08-01T21:05:30.847728Z (mocktime: 2022-08-01T21:05:30Z) [loadblk] [kernel/mempool_persist.cpp:121] [LoadMempool] Imported mempool transactions from disk: 0 succeeded, 0 failed, 0 expired, 0 already there, 0 waiting for initial broadcast > 2022-08-01T21:05:30.847748Z (mocktime: 2022-08-01T21:05:30Z) [loadblk] [util/thread.cpp:20] [TraceThread] loadblk thread exit > 2022-08-01T21:05:30.847790Z (mocktime: 2022-08-01T21:05:30Z) [init] [net.cpp:2103] [BindListenPort] Bound to 127.0.0.1:11093 > 2022-08-01T21:05:30.847830Z (mocktime: 2022-08-01T21:05:30Z) [init] [net.cpp:2103] [BindListenPort] Bound to 127.0.0.1:18445 > 2022-08-01T21:05:30.847866Z (mocktime: 2022-08-01T21:05:30Z) [init] [addrdb.cpp:228] [ReadAnchors] Loaded 0 addresses from "anchors.dat" > 2022-08-01T21:05:30.847936Z (mocktime: 2022-08-01T21:05:30Z) [init] [net.cpp:2277] [Start] 0 block-relay-only anchors will be tried for connections. > 2022-08-01T21:05:30.847950Z (mocktime: 2022-08-01T21:05:30Z) [init] [noui.cpp:56] [noui_InitMessage] init message: Starting network threadsโ€ฆ > 2022-08-01T21:05:30.848029Z (mocktime: 2022-08-01T21:05:30Z) [net] [util/thread.cpp:18] [TraceThread] net thread start > 2022-08-01T21:05:30.848098Z (mocktime: 2022-08-01T21:05:30Z) [addcon] [util/thread.cpp:18] [TraceThread] addcon thread start > 2022-08-01T21:05:30.848236Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [util/thread.cpp:18] [TraceThread] opencon thread start > 2022-08-01T21:05:30.848301Z (mocktime: 2022-08-01T21:05:30Z) [msghand] [util/thread.cpp:18] [TraceThread] msghand thread start > 2022-08-01T21:05:30.848322Z (mocktime: 2022-08-01T21:05:30Z) [init] [noui.cpp:56] [noui_InitMessage] init message: Done loading > 2022-08-01T21:05:30.848394Z (mocktime: 2022-08-01T21:05:30Z) [dnsseed] [util/thread.cpp:18] [TraceThread] dnsseed thread start > 2022-08-01T21:05:30.848416Z (mocktime: 2022-08-01T21:05:30Z) [dnsseed] [net.cpp:1456] [ThreadDNSAddressSeed] Loading addresses from DNS seed dummySeed.invalid. > 2022-08-01T21:05:30.857530Z (mocktime: 2022-08-01T21:05:30Z) [dnsseed] [addrman.cpp:612] [AddSingle] [addrman] Added 1.0.0.1:18444 mapped to AS0 to new[81][35] > 2022-08-01T21:05:30.857563Z (mocktime: 2022-08-01T21:05:30Z) [dnsseed] [addrman.cpp:684] [Add_] [addrman] Added 1 addresses (of 1) from jzva5xo5xttqfkfm.internal: 0 tried, 1 new > 2022-08-01T21:05:30.857575Z (mocktime: 2022-08-01T21:05:30Z) [dnsseed] [net.cpp:1485] [ThreadDNSAddressSeed] 1 addresses found from DNS seeds > 2022-08-01T21:05:30.857585Z (mocktime: 2022-08-01T21:05:30Z) [dnsseed] [util/thread.cpp:20] [TraceThread] dnsseed thread exit > 2022-08-01T21:05:30.921809Z (mocktime: 2022-08-01T21:05:30Z) [http] [httpserver.cpp:240] [http_request_cb] [http] Received a POST request for / from 127.0.0.1:43440 > 2022-08-01T21:05:30.921886Z (mocktime: 2022-08-01T21:05:30Z) [httpworker.1] [rpc/request.cpp:179] [parse] [rpc] ThreadRPCServer method=getblockcount user=__cookie__ > 2022-08-01T21:05:30.923628Z (mocktime: 2022-08-01T21:05:30Z) [http] [httpserver.cpp:240] [http_request_cb] [http] Received a POST request for / from 127.0.0.1:43440 > 2022-08-01T21:05:30.923856Z (mocktime: 2022-08-01T21:05:30Z) [httpworker.0] [rpc/request.cpp:179] [parse] [rpc] ThreadRPCServer method=getmempoolinfo user=__cookie__ > 2022-08-01T21:05:31.349127Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:31.349201Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [net.cpp:455] [ConnectNode] [net:debug] trying connection 1.0.0.1:18444 lastseen=146.9hrs > 2022-08-01T21:05:31.354156Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [net.cpp:2729] [CNode] [net] Added connection peer=0 > 2022-08-01T21:05:31.354233Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [net.cpp:2745] [PushMessage] [net] sending version (114 bytes) peer=0 > 2022-08-01T21:05:31.354306Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [net_processing.cpp:1303] [PushNodeVersion] [net] send version message: version 70016, blocks=0, txrelay=1, peer=0 > 2022-08-01T21:05:31.854950Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:32.356882Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:32.857649Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:33.359252Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:33.860731Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:34.361696Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:34.863290Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:35.364832Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:35.867734Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:36.369805Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:36.870817Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:37.372111Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:37.873116Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:38.374390Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:38.875975Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:39.377264Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:39.878563Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:40.380302Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > > > 2022-08-01T21:05:40.709000Z TestFramework (INFO): Stopping nodes > 2022-08-01T21:05:40.862000Z TestFramework (WARNING): Not cleaning up dir /tmp/bitcoin_func_test_die2tixp > 2022-08-01T21:05:40.862000Z TestFramework (ERROR): Test failed. Test logging available at /tmp/bitcoin_func_test_die2tixp/test_framework.log > 2022-08-01T21:05:40.862000Z TestFramework (ERROR): > 2022-08-01T21:05:40.862000Z TestFramework (ERROR): Hint: Call /home/user/bitcoin/test/functional/combine_logs.py '/tmp/bitcoin_func_test_die2tixp' to consolidate all logs > 2022-08-01T21:05:40.863000Z TestFramework (ERROR): > 2022-08-01T21:05:40.863000Z TestFramework (ERROR): If this failure happened unexpectedly or intermittently, please file a bug and provide a link or upload of the combined log. > 2022-08-01T21:05:40.863000Z TestFramework (ERROR): https://github.com/bitcoin/bitcoin/issues > 2022-08-01T21:05:40.863000Z TestFramework (ERROR): Logs: `python3 /home/user/bitcoin/test/functional/combine_logs.py '/tmp/bitcoin_func_test_die2tixp'` Pasted here: https://pastebin.com/SZtnGvEK
1.0
feature_config_args.py failure - Run test: `python3 feature_config_args.py` Results in error: > > 2022-08-01T21:05:29.574000Z TestFramework (INFO): Initializing test directory /tmp/bitcoin_func_test_die2tixp > 2022-08-01T21:05:30.306000Z TestFramework (INFO): Test config args logging > 2022-08-01T21:05:30.564000Z TestFramework (INFO): Test seed peers > 2022-08-01T21:05:40.697000Z TestFramework (ERROR): Assertion failed > Traceback (most recent call last): > File "/home/user/bitcoin/test/functional/test_framework/test_framework.py", line 133, in main > self.run_test() > File "/home/user/bitcoin/test/functional/feature_config_args.py", line 229, in run_test > self.test_seed_peers() > File "/home/user/bitcoin/test/functional/feature_config_args.py", line 175, in test_seed_peers > self.start_node(0, extra_args=['-dnsseed=1', '-fixedseeds=1', f'-mocktime={start}']) > File "/usr/lib/python3.9/contextlib.py", line 124, in __exit__ > next(self.gen) > File "/home/user/bitcoin/test/functional/test_framework/test_node.py", line 423, in assert_debug_log > self._raise_assertion_error('Expected messages "{}" does not partially match log:\n\n{}\n\n'.format(str(expected_msgs), print_log)) > File "/home/user/bitcoin/test/functional/test_framework/test_node.py", line 167, in _raise_assertion_error > raise AssertionError(self._node_msg(msg)) > AssertionError: [node 0] Expected messages "['Loaded 0 addresses from peers.dat', '0 addresses found from DNS seeds', 'opencon thread start']" does not partially match log: > > > > > > > 2022-08-01T21:05:30.776965Z [init] [init/common.cpp:127] [LogPackageVersion] Bitcoin Core version v23.99.0-ce3b75690d10 (release build) > 2022-08-01T21:05:30.777017Z [init] [init.cpp:657] [InitParameterInteraction] InitParameterInteraction: parameter interaction: -bind set -> setting -listen=1 > 2022-08-01T21:05:30.777088Z [init] [init.cpp:922] [AppInitParameterInteraction] Validating signatures for all blocks. > 2022-08-01T21:05:30.777094Z [init] [init.cpp:933] [AppInitParameterInteraction] Setting nMinimumChainWork=0000000000000000000000000000000000000000000000000000000000000000 > 2022-08-01T21:05:30.777164Z (mocktime: 2022-08-01T21:05:30Z) [init] [kernel/context.cpp:21] [Context] Using the 'sse4(1way),sse41(4way),avx2(8way)' SHA256 implementation > 2022-08-01T21:05:30.777175Z (mocktime: 2022-08-01T21:05:30Z) [init] [random.cpp:100] [ReportHardwareRand] Using RdSeed as an additional entropy source > 2022-08-01T21:05:30.777180Z (mocktime: 2022-08-01T21:05:30Z) [init] [random.cpp:103] [ReportHardwareRand] Using RdRand as an additional entropy source > 2022-08-01T21:05:30.778975Z (mocktime: 2022-08-01T21:05:30Z) [init] [init/common.cpp:98] [StartLogging] Default data directory /home/user/.bitcoin > 2022-08-01T21:05:30.778985Z (mocktime: 2022-08-01T21:05:30Z) [init] [init/common.cpp:99] [StartLogging] Using data directory /tmp/bitcoin_func_test_die2tixp/node0/regtest > 2022-08-01T21:05:30.778998Z (mocktime: 2022-08-01T21:05:30Z) [init] [init/common.cpp:104] [StartLogging] Config file: /tmp/bitcoin_func_test_die2tixp/node0/bitcoin.conf > 2022-08-01T21:05:30.779015Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: regtest="1" > 2022-08-01T21:05:30.779025Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] bind="127.0.0.1" > 2022-08-01T21:05:30.779032Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] discover="0" > 2022-08-01T21:05:30.779039Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] dnsseed="0" > 2022-08-01T21:05:30.779047Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] fallbackfee="0.0002" > 2022-08-01T21:05:30.779053Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] fixedseeds="0" > 2022-08-01T21:05:30.779060Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] keypool="1" > 2022-08-01T21:05:30.779067Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] listenonion="0" > 2022-08-01T21:05:30.779074Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] natpmp="0" > 2022-08-01T21:05:30.779082Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] peertimeout="999999999" > 2022-08-01T21:05:30.779089Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] port="11093" > 2022-08-01T21:05:30.779096Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] printtoconsole="0" > 2022-08-01T21:05:30.779104Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] rpcdoccheck="1" > 2022-08-01T21:05:30.779112Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] rpcport="16093" > 2022-08-01T21:05:30.779119Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] server="1" > 2022-08-01T21:05:30.779127Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] shrinkdebugfile="0" > 2022-08-01T21:05:30.779134Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] unsafesqlitesync="1" > 2022-08-01T21:05:30.779141Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Config file arg: [regtest] upnp="0" > 2022-08-01T21:05:30.779149Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: datadir="/tmp/bitcoin_func_test_die2tixp/node0" > 2022-08-01T21:05:30.779156Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: debug="" > 2022-08-01T21:05:30.779164Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: debugexclude="libevent" > 2022-08-01T21:05:30.779171Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: debugexclude="leveldb" > 2022-08-01T21:05:30.779178Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: dnsseed="1" > 2022-08-01T21:05:30.779184Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: fixedseeds="1" > 2022-08-01T21:05:30.779192Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: logsourcelocations="" > 2022-08-01T21:05:30.779199Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: logthreadnames="" > 2022-08-01T21:05:30.779206Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: logtimemicros="" > 2022-08-01T21:05:30.779213Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: mocktime="1659387930" > 2022-08-01T21:05:30.779224Z (mocktime: 2022-08-01T21:05:30Z) [init] [util/system.cpp:1118] [logArgsPrefix] Command-line arg: uacomment="testnode0" > 2022-08-01T21:05:30.779232Z (mocktime: 2022-08-01T21:05:30Z) [init] [init.cpp:1148] [AppInitMain] Using at most 125 automatic connections (1024 file descriptors available) > 2022-08-01T21:05:30.788089Z (mocktime: 2022-08-01T21:05:30Z) [init] [script/sigcache.cpp:101] [InitSignatureCache] Using 16 MiB out of 32/2 requested for signature cache, able to store 524288 elements > 2022-08-01T21:05:30.796940Z (mocktime: 2022-08-01T21:05:30Z) [init] [validation.cpp:1669] [InitScriptExecutionCache] Using 16 MiB out of 32/2 requested for script execution cache, able to store 524288 elements > 2022-08-01T21:05:30.796987Z (mocktime: 2022-08-01T21:05:30Z) [init] [init.cpp:1175] [AppInitMain] Script verification uses 1 additional threads > 2022-08-01T21:05:30.797361Z (mocktime: 2022-08-01T21:05:30Z) [scheduler] [util/thread.cpp:18] [TraceThread] scheduler thread start > 2022-08-01T21:05:30.806770Z (mocktime: 2022-08-01T21:05:30Z) [init] [httpserver.cpp:184] [InitHTTPAllowList] [http] Allowing HTTP connections from: 127.0.0.0/8 ::1/128 > 2022-08-01T21:05:30.806890Z (mocktime: 2022-08-01T21:05:30Z) [init] [httpserver.cpp:321] [HTTPBindAddresses] [http] Binding RPC on address ::1 port 16093 > 2022-08-01T21:05:30.807109Z (mocktime: 2022-08-01T21:05:30Z) [init] [httpserver.cpp:321] [HTTPBindAddresses] [http] Binding RPC on address 127.0.0.1 port 16093 > 2022-08-01T21:05:30.807155Z (mocktime: 2022-08-01T21:05:30Z) [init] [httpserver.cpp:401] [InitHTTPServer] [http] Initialized HTTP server > 2022-08-01T21:05:30.807167Z (mocktime: 2022-08-01T21:05:30Z) [init] [httpserver.cpp:403] [InitHTTPServer] [http] creating work queue of depth 16 > 2022-08-01T21:05:30.807177Z (mocktime: 2022-08-01T21:05:30Z) [init] [rpc/server.cpp:291] [StartRPC] [rpc] Starting RPC > 2022-08-01T21:05:30.807193Z (mocktime: 2022-08-01T21:05:30Z) [init] [httprpc.cpp:296] [StartHTTPRPC] [rpc] Starting HTTP RPC server > 2022-08-01T21:05:30.807201Z (mocktime: 2022-08-01T21:05:30Z) [init] [httprpc.cpp:245] [InitRPCAuthentication] Using random cookie authentication. > 2022-08-01T21:05:30.807297Z (mocktime: 2022-08-01T21:05:30Z) [init] [rpc/request.cpp:106] [GenerateAuthCookie] Generated RPC authentication cookie /tmp/bitcoin_func_test_die2tixp/node0/regtest/.cookie > 2022-08-01T21:05:30.807312Z (mocktime: 2022-08-01T21:05:30Z) [init] [httpserver.cpp:676] [RegisterHTTPHandler] [http] Registering HTTP handler for / (exactmatch 1) > 2022-08-01T21:05:30.807321Z (mocktime: 2022-08-01T21:05:30Z) [init] [httpserver.cpp:676] [RegisterHTTPHandler] [http] Registering HTTP handler for /wallet/ (exactmatch 0) > 2022-08-01T21:05:30.807335Z (mocktime: 2022-08-01T21:05:30Z) [init] [httpserver.cpp:425] [StartHTTPServer] [http] Starting HTTP server > 2022-08-01T21:05:30.807343Z (mocktime: 2022-08-01T21:05:30Z) [init] [httpserver.cpp:427] [StartHTTPServer] [http] starting 4 worker threads > 2022-08-01T21:05:30.807552Z (mocktime: 2022-08-01T21:05:30Z) [http] [httpserver.cpp:287] [ThreadHTTP] [http] Entering http event loop > 2022-08-01T21:05:30.807784Z (mocktime: 2022-08-01T21:05:30Z) [init] [wallet/load.cpp:51] [VerifyWallets] Using wallet directory /tmp/bitcoin_func_test_die2tixp/node0/regtest/wallets > 2022-08-01T21:05:30.807801Z (mocktime: 2022-08-01T21:05:30Z) [init] [noui.cpp:56] [noui_InitMessage] init message: Verifying wallet(s)โ€ฆ > 2022-08-01T21:05:30.807840Z (mocktime: 2022-08-01T21:05:30Z) [init] [init.cpp:1261] [AppInitMain] Using /16 prefix for IP bucketing > 2022-08-01T21:05:30.807850Z (mocktime: 2022-08-01T21:05:30Z) [init] [noui.cpp:56] [noui_InitMessage] init message: Loading P2P addressesโ€ฆ > 2022-08-01T21:05:30.808116Z (mocktime: 2022-08-01T21:05:30Z) [init] [logging/timer.h:57] [Log] [addrman] CheckAddrman: new 0, tried 0, total 0 started > 2022-08-01T21:05:30.808228Z (mocktime: 2022-08-01T21:05:30Z) [init] [logging/timer.h:57] [Log] [addrman] CheckAddrman: completed (0.00ms) > 2022-08-01T21:05:30.808242Z (mocktime: 2022-08-01T21:05:30Z) [init] [addrdb.cpp:194] [LoadAddrman] Loaded 0 addresses from peers.dat 0ms > 2022-08-01T21:05:30.808586Z (mocktime: 2022-08-01T21:05:30Z) [init] [noui.cpp:56] [noui_InitMessage] init message: Loading banlistโ€ฆ > 2022-08-01T21:05:30.808626Z (mocktime: 2022-08-01T21:05:30Z) [init] [banman.cpp:38] [LoadBanlist] [net] Loaded 0 banned node addresses/subnets 0ms > 2022-08-01T21:05:30.808645Z (mocktime: 2022-08-01T21:05:30Z) [init] [net.cpp:1523] [SetTryNewOutboundPeer] [net] setting try another outbound peer=false > 2022-08-01T21:05:30.808654Z (mocktime: 2022-08-01T21:05:30Z) [init] [net.cpp:2170] [SetNetworkActive] SetNetworkActive: true > 2022-08-01T21:05:30.810409Z (mocktime: 2022-08-01T21:05:30Z) [init] [policy/fees.cpp:455] [Read] [estimatefee] Reading estimates: 190 buckets counting confirms up to 48 blocks > 2022-08-01T21:05:30.810608Z (mocktime: 2022-08-01T21:05:30Z) [init] [policy/fees.cpp:455] [Read] [estimatefee] Reading estimates: 190 buckets counting confirms up to 12 blocks > 2022-08-01T21:05:30.811134Z (mocktime: 2022-08-01T21:05:30Z) [init] [policy/fees.cpp:455] [Read] [estimatefee] Reading estimates: 190 buckets counting confirms up to 1008 blocks > 2022-08-01T21:05:30.811253Z (mocktime: 2022-08-01T21:05:30Z) [init] [init.cpp:1403] [AppInitMain] Cache configuration: > 2022-08-01T21:05:30.811267Z (mocktime: 2022-08-01T21:05:30Z) [init] [init.cpp:1404] [AppInitMain] * Using 2.0 MiB for block index database > 2022-08-01T21:05:30.811276Z (mocktime: 2022-08-01T21:05:30Z) [init] [init.cpp:1412] [AppInitMain] * Using 8.0 MiB for chain state database > 2022-08-01T21:05:30.811290Z (mocktime: 2022-08-01T21:05:30Z) [init] [init.cpp:1428] [AppInitMain] * Using 440.0 MiB for in-memory UTXO set (plus up to 286.1 MiB of unused mempool space) > 2022-08-01T21:05:30.811315Z (mocktime: 2022-08-01T21:05:30Z) [init] [noui.cpp:56] [noui_InitMessage] init message: Loading block indexโ€ฆ > 2022-08-01T21:05:30.811330Z (mocktime: 2022-08-01T21:05:30Z) [init] [validation.cpp:4710] [InitializeChainstate] Switching active chainstate to Chainstate [ibd] @ height -1 (null) > 2022-08-01T21:05:30.811367Z (mocktime: 2022-08-01T21:05:30Z) [init] [dbwrapper.cpp:150] [CDBWrapper] Opening LevelDB in /tmp/bitcoin_func_test_die2tixp/node0/regtest/blocks/index > 2022-08-01T21:05:30.826887Z (mocktime: 2022-08-01T21:05:30Z) [init] [dbwrapper.cpp:158] [CDBWrapper] Opened LevelDB successfully > 2022-08-01T21:05:30.826920Z (mocktime: 2022-08-01T21:05:30Z) [init] [dbwrapper.cpp:183] [CDBWrapper] Using obfuscation key for /tmp/bitcoin_func_test_die2tixp/node0/regtest/blocks/index: 0000000000000000 > 2022-08-01T21:05:30.827091Z (mocktime: 2022-08-01T21:05:30Z) [init] [node/blockstorage.cpp:331] [LoadBlockIndexDB] LoadBlockIndexDB: last block file = 0 > 2022-08-01T21:05:30.827110Z (mocktime: 2022-08-01T21:05:30Z) [init] [node/blockstorage.cpp:335] [LoadBlockIndexDB] LoadBlockIndexDB: last block file info: CBlockFileInfo(blocks=1, size=293, heights=0...0, time=2011-02-02...2011-02-02) > 2022-08-01T21:05:30.827119Z (mocktime: 2022-08-01T21:05:30Z) [init] [node/blockstorage.cpp:346] [LoadBlockIndexDB] Checking all blk files are present... > 2022-08-01T21:05:30.827164Z (mocktime: 2022-08-01T21:05:30Z) [init] [dbwrapper.cpp:150] [CDBWrapper] Opening LevelDB in /tmp/bitcoin_func_test_die2tixp/node0/regtest/chainstate > 2022-08-01T21:05:30.846337Z (mocktime: 2022-08-01T21:05:30Z) [init] [dbwrapper.cpp:158] [CDBWrapper] Opened LevelDB successfully > 2022-08-01T21:05:30.846444Z (mocktime: 2022-08-01T21:05:30Z) [init] [dbwrapper.cpp:183] [CDBWrapper] Using obfuscation key for /tmp/bitcoin_func_test_die2tixp/node0/regtest/chainstate: 223bb69586684f13 > 2022-08-01T21:05:30.846548Z (mocktime: 2022-08-01T21:05:30Z) [init] [validation.cpp:3894] [LoadChainTip] Loaded best chain: hashBestChain=0f9188f13cb7b2c71f2a335e3a4fc328bf5beb436012afca590b1a11466e2206 height=0 date=2011-02-02T23:16:42Z progress=1.000000 > 2022-08-01T21:05:30.846567Z (mocktime: 2022-08-01T21:05:30Z) [init] [noui.cpp:56] [noui_InitMessage] init message: Verifying blocksโ€ฆ > 2022-08-01T21:05:30.846583Z (mocktime: 2022-08-01T21:05:30Z) [init] [init.cpp:1474] [AppInitMain] block index 35ms > 2022-08-01T21:05:30.847557Z (mocktime: 2022-08-01T21:05:30Z) [init] [init.cpp:1631] [AppInitMain] block tree size = 1 > 2022-08-01T21:05:30.847581Z (mocktime: 2022-08-01T21:05:30Z) [init] [init.cpp:1643] [AppInitMain] nBestHeight = 0 > 2022-08-01T21:05:30.847637Z (mocktime: 2022-08-01T21:05:30Z) [loadblk] [util/thread.cpp:18] [TraceThread] loadblk thread start > 2022-08-01T21:05:30.847728Z (mocktime: 2022-08-01T21:05:30Z) [loadblk] [kernel/mempool_persist.cpp:121] [LoadMempool] Imported mempool transactions from disk: 0 succeeded, 0 failed, 0 expired, 0 already there, 0 waiting for initial broadcast > 2022-08-01T21:05:30.847748Z (mocktime: 2022-08-01T21:05:30Z) [loadblk] [util/thread.cpp:20] [TraceThread] loadblk thread exit > 2022-08-01T21:05:30.847790Z (mocktime: 2022-08-01T21:05:30Z) [init] [net.cpp:2103] [BindListenPort] Bound to 127.0.0.1:11093 > 2022-08-01T21:05:30.847830Z (mocktime: 2022-08-01T21:05:30Z) [init] [net.cpp:2103] [BindListenPort] Bound to 127.0.0.1:18445 > 2022-08-01T21:05:30.847866Z (mocktime: 2022-08-01T21:05:30Z) [init] [addrdb.cpp:228] [ReadAnchors] Loaded 0 addresses from "anchors.dat" > 2022-08-01T21:05:30.847936Z (mocktime: 2022-08-01T21:05:30Z) [init] [net.cpp:2277] [Start] 0 block-relay-only anchors will be tried for connections. > 2022-08-01T21:05:30.847950Z (mocktime: 2022-08-01T21:05:30Z) [init] [noui.cpp:56] [noui_InitMessage] init message: Starting network threadsโ€ฆ > 2022-08-01T21:05:30.848029Z (mocktime: 2022-08-01T21:05:30Z) [net] [util/thread.cpp:18] [TraceThread] net thread start > 2022-08-01T21:05:30.848098Z (mocktime: 2022-08-01T21:05:30Z) [addcon] [util/thread.cpp:18] [TraceThread] addcon thread start > 2022-08-01T21:05:30.848236Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [util/thread.cpp:18] [TraceThread] opencon thread start > 2022-08-01T21:05:30.848301Z (mocktime: 2022-08-01T21:05:30Z) [msghand] [util/thread.cpp:18] [TraceThread] msghand thread start > 2022-08-01T21:05:30.848322Z (mocktime: 2022-08-01T21:05:30Z) [init] [noui.cpp:56] [noui_InitMessage] init message: Done loading > 2022-08-01T21:05:30.848394Z (mocktime: 2022-08-01T21:05:30Z) [dnsseed] [util/thread.cpp:18] [TraceThread] dnsseed thread start > 2022-08-01T21:05:30.848416Z (mocktime: 2022-08-01T21:05:30Z) [dnsseed] [net.cpp:1456] [ThreadDNSAddressSeed] Loading addresses from DNS seed dummySeed.invalid. > 2022-08-01T21:05:30.857530Z (mocktime: 2022-08-01T21:05:30Z) [dnsseed] [addrman.cpp:612] [AddSingle] [addrman] Added 1.0.0.1:18444 mapped to AS0 to new[81][35] > 2022-08-01T21:05:30.857563Z (mocktime: 2022-08-01T21:05:30Z) [dnsseed] [addrman.cpp:684] [Add_] [addrman] Added 1 addresses (of 1) from jzva5xo5xttqfkfm.internal: 0 tried, 1 new > 2022-08-01T21:05:30.857575Z (mocktime: 2022-08-01T21:05:30Z) [dnsseed] [net.cpp:1485] [ThreadDNSAddressSeed] 1 addresses found from DNS seeds > 2022-08-01T21:05:30.857585Z (mocktime: 2022-08-01T21:05:30Z) [dnsseed] [util/thread.cpp:20] [TraceThread] dnsseed thread exit > 2022-08-01T21:05:30.921809Z (mocktime: 2022-08-01T21:05:30Z) [http] [httpserver.cpp:240] [http_request_cb] [http] Received a POST request for / from 127.0.0.1:43440 > 2022-08-01T21:05:30.921886Z (mocktime: 2022-08-01T21:05:30Z) [httpworker.1] [rpc/request.cpp:179] [parse] [rpc] ThreadRPCServer method=getblockcount user=__cookie__ > 2022-08-01T21:05:30.923628Z (mocktime: 2022-08-01T21:05:30Z) [http] [httpserver.cpp:240] [http_request_cb] [http] Received a POST request for / from 127.0.0.1:43440 > 2022-08-01T21:05:30.923856Z (mocktime: 2022-08-01T21:05:30Z) [httpworker.0] [rpc/request.cpp:179] [parse] [rpc] ThreadRPCServer method=getmempoolinfo user=__cookie__ > 2022-08-01T21:05:31.349127Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:31.349201Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [net.cpp:455] [ConnectNode] [net:debug] trying connection 1.0.0.1:18444 lastseen=146.9hrs > 2022-08-01T21:05:31.354156Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [net.cpp:2729] [CNode] [net] Added connection peer=0 > 2022-08-01T21:05:31.354233Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [net.cpp:2745] [PushMessage] [net] sending version (114 bytes) peer=0 > 2022-08-01T21:05:31.354306Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [net_processing.cpp:1303] [PushNodeVersion] [net] send version message: version 70016, blocks=0, txrelay=1, peer=0 > 2022-08-01T21:05:31.854950Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:32.356882Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:32.857649Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:33.359252Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:33.860731Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:34.361696Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:34.863290Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:35.364832Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:35.867734Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:36.369805Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:36.870817Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:37.372111Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:37.873116Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:38.374390Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:38.875975Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:39.377264Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:39.878563Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > 2022-08-01T21:05:40.380302Z (mocktime: 2022-08-01T21:05:30Z) [opencon] [addrman.cpp:769] [Select_] [addrman] Selected 1.0.0.1:18444 from new > > > 2022-08-01T21:05:40.709000Z TestFramework (INFO): Stopping nodes > 2022-08-01T21:05:40.862000Z TestFramework (WARNING): Not cleaning up dir /tmp/bitcoin_func_test_die2tixp > 2022-08-01T21:05:40.862000Z TestFramework (ERROR): Test failed. Test logging available at /tmp/bitcoin_func_test_die2tixp/test_framework.log > 2022-08-01T21:05:40.862000Z TestFramework (ERROR): > 2022-08-01T21:05:40.862000Z TestFramework (ERROR): Hint: Call /home/user/bitcoin/test/functional/combine_logs.py '/tmp/bitcoin_func_test_die2tixp' to consolidate all logs > 2022-08-01T21:05:40.863000Z TestFramework (ERROR): > 2022-08-01T21:05:40.863000Z TestFramework (ERROR): If this failure happened unexpectedly or intermittently, please file a bug and provide a link or upload of the combined log. > 2022-08-01T21:05:40.863000Z TestFramework (ERROR): https://github.com/bitcoin/bitcoin/issues > 2022-08-01T21:05:40.863000Z TestFramework (ERROR): Logs: `python3 /home/user/bitcoin/test/functional/combine_logs.py '/tmp/bitcoin_func_test_die2tixp'` Pasted here: https://pastebin.com/SZtnGvEK
test
feature config args py failure run test feature config args py results in error testframework info initializing test directory tmp bitcoin func test testframework info test config args logging testframework info test seed peers testframework error assertion failed traceback most recent call last file home user bitcoin test functional test framework test framework py line in main self run test file home user bitcoin test functional feature config args py line in run test self test seed peers file home user bitcoin test functional feature config args py line in test seed peers self start node extra args file usr lib contextlib py line in exit next self gen file home user bitcoin test functional test framework test node py line in assert debug log self raise assertion error expected messages does not partially match log n n n n format str expected msgs print log file home user bitcoin test functional test framework test node py line in raise assertion error raise assertionerror self node msg msg assertionerror expected messages does not partially match log bitcoin core version release build initparameterinteraction parameter interaction bind set setting listen validating signatures for all blocks setting nminimumchainwork mocktime using the implementation mocktime using rdseed as an additional entropy source mocktime using rdrand as an additional entropy source mocktime default data directory home user bitcoin mocktime using data directory tmp bitcoin func test regtest mocktime config file tmp bitcoin func test bitcoin conf mocktime config file arg regtest mocktime config file arg bind mocktime config file arg discover mocktime config file arg dnsseed mocktime config file arg fallbackfee mocktime config file arg fixedseeds mocktime config file arg keypool mocktime config file arg listenonion mocktime config file arg natpmp mocktime config file arg peertimeout mocktime config file arg port mocktime config file arg printtoconsole mocktime config file arg rpcdoccheck mocktime config file arg rpcport mocktime config file arg server mocktime config file arg shrinkdebugfile mocktime config file arg unsafesqlitesync mocktime config file arg upnp mocktime command line arg datadir tmp bitcoin func test mocktime command line arg debug mocktime command line arg debugexclude libevent mocktime command line arg debugexclude leveldb mocktime command line arg dnsseed mocktime command line arg fixedseeds mocktime command line arg logsourcelocations mocktime command line arg logthreadnames mocktime command line arg logtimemicros mocktime command line arg mocktime mocktime command line arg uacomment mocktime using at most automatic connections file descriptors available mocktime using mib out of requested for signature cache able to store elements mocktime using mib out of requested for script execution cache able to store elements mocktime script verification uses additional threads mocktime scheduler thread start mocktime allowing http connections from mocktime binding rpc on address port mocktime binding rpc on address port mocktime initialized http server mocktime creating work queue of depth mocktime starting rpc mocktime starting http rpc server mocktime using random cookie authentication mocktime generated rpc authentication cookie tmp bitcoin func test regtest cookie mocktime registering http handler for exactmatch mocktime registering http handler for wallet exactmatch mocktime starting http server mocktime starting worker threads mocktime entering http event loop mocktime using wallet directory tmp bitcoin func test regtest wallets mocktime init message verifying wallet s โ€ฆ mocktime using prefix for ip bucketing mocktime init message loading addressesโ€ฆ mocktime checkaddrman new tried total started mocktime checkaddrman completed mocktime loaded addresses from peers dat mocktime init message loading banlistโ€ฆ mocktime loaded banned node addresses subnets mocktime setting try another outbound peer false mocktime setnetworkactive true mocktime reading estimates buckets counting confirms up to blocks mocktime reading estimates buckets counting confirms up to blocks mocktime reading estimates buckets counting confirms up to blocks mocktime cache configuration mocktime using mib for block index database mocktime using mib for chain state database mocktime using mib for in memory utxo set plus up to mib of unused mempool space mocktime init message loading block indexโ€ฆ mocktime switching active chainstate to chainstate height null mocktime opening leveldb in tmp bitcoin func test regtest blocks index mocktime opened leveldb successfully mocktime using obfuscation key for tmp bitcoin func test regtest blocks index mocktime loadblockindexdb last block file mocktime loadblockindexdb last block file info cblockfileinfo blocks size heights time mocktime checking all blk files are present mocktime opening leveldb in tmp bitcoin func test regtest chainstate mocktime opened leveldb successfully mocktime using obfuscation key for tmp bitcoin func test regtest chainstate mocktime loaded best chain hashbestchain height date progress mocktime init message verifying blocksโ€ฆ mocktime block index mocktime block tree size mocktime nbestheight mocktime loadblk thread start mocktime imported mempool transactions from disk succeeded failed expired already there waiting for initial broadcast mocktime loadblk thread exit mocktime bound to mocktime bound to mocktime loaded addresses from anchors dat mocktime block relay only anchors will be tried for connections mocktime init message starting network threadsโ€ฆ mocktime net thread start mocktime addcon thread start mocktime opencon thread start mocktime msghand thread start mocktime init message done loading mocktime dnsseed thread start mocktime loading addresses from dns seed dummyseed invalid mocktime added mapped to to new mocktime added addresses of from internal tried new mocktime addresses found from dns seeds mocktime dnsseed thread exit mocktime received a post request for from mocktime threadrpcserver method getblockcount user cookie mocktime received a post request for from mocktime threadrpcserver method getmempoolinfo user cookie mocktime selected from new mocktime trying connection lastseen mocktime added connection peer mocktime sending version bytes peer mocktime send version message version blocks txrelay peer mocktime selected from new mocktime selected from new mocktime selected from new mocktime selected from new mocktime selected from new mocktime selected from new mocktime selected from new mocktime selected from new mocktime selected from new mocktime selected from new mocktime selected from new mocktime selected from new mocktime selected from new mocktime selected from new mocktime selected from new mocktime selected from new mocktime selected from new mocktime selected from new testframework info stopping nodes testframework warning not cleaning up dir tmp bitcoin func test testframework error test failed test logging available at tmp bitcoin func test test framework log testframework error testframework error hint call home user bitcoin test functional combine logs py tmp bitcoin func test to consolidate all logs testframework error testframework error if this failure happened unexpectedly or intermittently please file a bug and provide a link or upload of the combined log testframework error testframework error logs home user bitcoin test functional combine logs py tmp bitcoin func test pasted here
1
113,859
24,500,716,529
IssuesEvent
2022-10-10 12:35:25
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
opened
executors: Add ability to whitelist certain network traffic
team/batchers user-code-execution
A customer is using a private gateway into their Sourcegraph instance. This gateway lives in the network IP space of `10.0.0.0/16`. We currently disallow ANY traffic on that network, so there's a bunch of things we should do here: - Add documentation on what traffic is allowed and which isn't, and why we do that - Add a flag to allow a certain additional CIDR (or multiple) to be permitted access to - Maybe: Investigation needed: Add a flag to allow a certain CIDR to be permitted from the executor, so it's a native feature and not dependent on the terraform module - If (3) is not feasible, add docs to the binary setup on how to white list a certain CIDR from firecracker VMs manually
1.0
executors: Add ability to whitelist certain network traffic - A customer is using a private gateway into their Sourcegraph instance. This gateway lives in the network IP space of `10.0.0.0/16`. We currently disallow ANY traffic on that network, so there's a bunch of things we should do here: - Add documentation on what traffic is allowed and which isn't, and why we do that - Add a flag to allow a certain additional CIDR (or multiple) to be permitted access to - Maybe: Investigation needed: Add a flag to allow a certain CIDR to be permitted from the executor, so it's a native feature and not dependent on the terraform module - If (3) is not feasible, add docs to the binary setup on how to white list a certain CIDR from firecracker VMs manually
non_test
executors add ability to whitelist certain network traffic a customer is using a private gateway into their sourcegraph instance this gateway lives in the network ip space of we currently disallow any traffic on that network so there s a bunch of things we should do here add documentation on what traffic is allowed and which isn t and why we do that add a flag to allow a certain additional cidr or multiple to be permitted access to maybe investigation needed add a flag to allow a certain cidr to be permitted from the executor so it s a native feature and not dependent on the terraform module if is not feasible add docs to the binary setup on how to white list a certain cidr from firecracker vms manually
0
37,023
6,562,395,619
IssuesEvent
2017-09-07 16:25:11
SAFETAG/SAFETAG
https://api.github.com/repos/SAFETAG/SAFETAG
closed
Create Debian package
documentation enhancement
The current make based installer is a dependency hell. As such, a debian package is the best path forward. There are concerns. - What is the best installation path for a package that we desire the user to edit. - Do we add basic git pull ability into the package so that the user can keep their SAFETAG content up to date before they build? - Should this be a seperate repository that pulls from the git content and places it in a specific directory? - How do we best incorporate GUI control into a seperate package like this?
1.0
Create Debian package - The current make based installer is a dependency hell. As such, a debian package is the best path forward. There are concerns. - What is the best installation path for a package that we desire the user to edit. - Do we add basic git pull ability into the package so that the user can keep their SAFETAG content up to date before they build? - Should this be a seperate repository that pulls from the git content and places it in a specific directory? - How do we best incorporate GUI control into a seperate package like this?
non_test
create debian package the current make based installer is a dependency hell as such a debian package is the best path forward there are concerns what is the best installation path for a package that we desire the user to edit do we add basic git pull ability into the package so that the user can keep their safetag content up to date before they build should this be a seperate repository that pulls from the git content and places it in a specific directory how do we best incorporate gui control into a seperate package like this
0
117,721
9,957,599,990
IssuesEvent
2019-07-05 17:28:50
easydigitaldownloads/easy-digital-downloads
https://api.github.com/repos/easydigitaldownloads/easy-digital-downloads
closed
3.0 - If paying with PayPal, you get redirected to wp-admin instead of PayPal
Bug Has PR Needs Testing
## Bug Report ### Expected behavior When I choose that I want to pay with PayPal, I expect to be redirected to PayPal upon clicking Complete, on the checkout page. ### Actual behavior I get redirected to wp-admin. ### Steps to reproduce the behavior 1) Enable PayPal as a gateway 2) Add an item to the cart and checkout using PayPal 3) You get redirected to wp-admin instead of PayPal. If logged out you hit the login screen. ### Information (if a specific version is affected): EDD Version (or branch): 3.0
1.0
3.0 - If paying with PayPal, you get redirected to wp-admin instead of PayPal - ## Bug Report ### Expected behavior When I choose that I want to pay with PayPal, I expect to be redirected to PayPal upon clicking Complete, on the checkout page. ### Actual behavior I get redirected to wp-admin. ### Steps to reproduce the behavior 1) Enable PayPal as a gateway 2) Add an item to the cart and checkout using PayPal 3) You get redirected to wp-admin instead of PayPal. If logged out you hit the login screen. ### Information (if a specific version is affected): EDD Version (or branch): 3.0
test
if paying with paypal you get redirected to wp admin instead of paypal bug report expected behavior when i choose that i want to pay with paypal i expect to be redirected to paypal upon clicking complete on the checkout page actual behavior i get redirected to wp admin steps to reproduce the behavior enable paypal as a gateway add an item to the cart and checkout using paypal you get redirected to wp admin instead of paypal if logged out you hit the login screen information if a specific version is affected edd version or branch
1
32,142
6,720,580,499
IssuesEvent
2017-10-16 08:24:45
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
Wrong Date Format with "M/d/yy" pattern
6.1.7 defect
Reported By PRO User; > I tried to make calendar input with format of "M/d/yy" but it gives me only one digit for month and one digit for day (I can't write manually 12/12/17, when I'm writing "1" it jumps to the next digit without giving an option to write 12). What can I do? (I know I can write MM/DD/YY but I need this format because I'm taking it from java method) The code used is as follows: <p:calendar id="startDate" required="true" mask="true" pattern="M/d/yy"></p:calendar>
1.0
Wrong Date Format with "M/d/yy" pattern - Reported By PRO User; > I tried to make calendar input with format of "M/d/yy" but it gives me only one digit for month and one digit for day (I can't write manually 12/12/17, when I'm writing "1" it jumps to the next digit without giving an option to write 12). What can I do? (I know I can write MM/DD/YY but I need this format because I'm taking it from java method) The code used is as follows: <p:calendar id="startDate" required="true" mask="true" pattern="M/d/yy"></p:calendar>
non_test
wrong date format with m d yy pattern reported by pro user i tried to make calendar input with format of m d yy but it gives me only one digit for month and one digit for day i can t write manually when i m writing it jumps to the next digit without giving an option to write what can i do i know i can write mm dd yy but i need this format because i m taking it from java method the code used is as follows p calendar id startdate required true mask true pattern m d yy
0
29,430
11,755,022,725
IssuesEvent
2020-03-13 08:38:01
f-peng/centreon
https://api.github.com/repos/f-peng/centreon
opened
CVE-2018-11694 (High) detected in opennms-opennms-source-24.1.2-1, node-sass-4.12.0.tgz
security vulnerability
## CVE-2018-11694 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.12.0.tgz</b></p></summary> <p> <details><summary><b>node-sass-4.12.0.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.12.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.12.0.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/centreon/package.json</p> <p>Path to vulnerable library: /centreon/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - :x: **node-sass-4.12.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/f-peng/centreon/commit/295f5b6a05f65c316bd0b01098e9d838c74ac2dc">295f5b6a05f65c316bd0b01098e9d838c74ac2dc</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in LibSass through 3.5.4. A NULL pointer dereference was found in the function Sass::Functions::selector_append which could be leveraged by an attacker to cause a denial of service (application crash) or possibly have unspecified other impact. <p>Publish Date: 2018-06-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11694>CVE-2018-11694</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11694">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11694</a></p> <p>Release Date: 2018-06-04</p> <p>Fix Resolution: 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-11694 (High) detected in opennms-opennms-source-24.1.2-1, node-sass-4.12.0.tgz - ## CVE-2018-11694 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.12.0.tgz</b></p></summary> <p> <details><summary><b>node-sass-4.12.0.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.12.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.12.0.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/centreon/package.json</p> <p>Path to vulnerable library: /centreon/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - :x: **node-sass-4.12.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/f-peng/centreon/commit/295f5b6a05f65c316bd0b01098e9d838c74ac2dc">295f5b6a05f65c316bd0b01098e9d838c74ac2dc</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in LibSass through 3.5.4. A NULL pointer dereference was found in the function Sass::Functions::selector_append which could be leveraged by an attacker to cause a denial of service (application crash) or possibly have unspecified other impact. <p>Publish Date: 2018-06-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11694>CVE-2018-11694</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11694">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11694</a></p> <p>Release Date: 2018-06-04</p> <p>Fix Resolution: 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in opennms opennms source node sass tgz cve high severity vulnerability vulnerable libraries node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file tmp ws scm centreon package json path to vulnerable library centreon node modules node sass package json dependency hierarchy x node sass tgz vulnerable library found in head commit a href vulnerability details an issue was discovered in libsass through a null pointer dereference was found in the function sass functions selector append which could be leveraged by an attacker to cause a denial of service application crash or possibly have unspecified other impact publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
119,643
10,059,574,021
IssuesEvent
2019-07-22 16:45:18
mantidproject/mantid
https://api.github.com/repos/mantidproject/mantid
closed
Python Smoke Tests
Quality: Smoke test
# Python smoke tests Please run these tests on the compiled package of Mantid; **not a locally compiled version**. When reporting back please include: - Your OS - Your facility/location - If you have other package versions of Mantid installed (yes/no) If you have any questions please don't hesitate to contact the creator of this issue. ## Checklist/directions * Test that the Python scripting window works, [directions here](http://docs.mantidproject.org/v3.7.2/interfaces/ScriptingWindow.html) - [ ] Editor options - [ ] Execution options - [ ] Script output - [ ] Perform some workspace algebra - [ ] Test numpy functionality - [ ] Use the scripting window to run some scripts - [ ] Run through some examples from [the documentation](http://www.mantidproject.org/Python_In_Mantid), 3 or 4 examples is enough
1.0
Python Smoke Tests - # Python smoke tests Please run these tests on the compiled package of Mantid; **not a locally compiled version**. When reporting back please include: - Your OS - Your facility/location - If you have other package versions of Mantid installed (yes/no) If you have any questions please don't hesitate to contact the creator of this issue. ## Checklist/directions * Test that the Python scripting window works, [directions here](http://docs.mantidproject.org/v3.7.2/interfaces/ScriptingWindow.html) - [ ] Editor options - [ ] Execution options - [ ] Script output - [ ] Perform some workspace algebra - [ ] Test numpy functionality - [ ] Use the scripting window to run some scripts - [ ] Run through some examples from [the documentation](http://www.mantidproject.org/Python_In_Mantid), 3 or 4 examples is enough
test
python smoke tests python smoke tests please run these tests on the compiled package of mantid not a locally compiled version when reporting back please include your os your facility location if you have other package versions of mantid installed yes no if you have any questions please don t hesitate to contact the creator of this issue checklist directions test that the python scripting window works editor options execution options script output perform some workspace algebra test numpy functionality use the scripting window to run some scripts run through some examples from or examples is enough
1
1,902
4,568,174,769
IssuesEvent
2016-09-15 13:45:12
AdguardTeam/AdguardForAndroid
https://api.github.com/repos/AdguardTeam/AdguardForAndroid
opened
Apple Music is incompatible with SSL filtering
Compatibility SSL
[Apple Music](https://play.google.com/store/apps/details?id=com.apple.android.music&hl=uk) ั ะฒะบะปัŽั‡ะตะฝะฝะพะน ะทะฐั‰ะธั‚ะพะน, ะฝะตะบะพั€ั€ะตะบั‚ะฝะพ ะธะฝะพะณะดะฐ ั€ะฐะฑะพั‚ะฐะตั‚! ะพะฑั€ั‹ะฒั‹ ะฒ ะฟั€ะพัะปัƒัˆะธะฒะฐะฝะธะธ ,ะฝะต ะฟะพะบะฐะทั‹ะฒะฐะตั‚ ั€ะตะบะพะผะตะฝะดะฐั†ะธะธ ะฒะพ ะฒะบะปะฐะดะบะต (ะดะปั ะฒะฐั) ะฟั€ะธั…ะพะดะธั‚ัั ะฒั‹ะบะปัŽั‡ะฐั‚ัŒ ะทะฐั‰ะธั‚ัƒ, ะฟะตั€ะตะทะฟัƒัะบะฐั‚ัŒ Apple Music, ะฟะพั‚ะพะผ ะฒะบะปัŽั‡ะฐั‚ัŒ ะทะฐั‰ะธั‚ัƒ ะพะฑั€ะฐั‚ะฝะพ. ะžั‚ะบะปัŽั‡ะตะฝะธะต HTTPS ั„ะธะปัŒั‚ั€ะฐั†ะธะธ ะฟะพะผะพะณะฐะตั‚.
True
Apple Music is incompatible with SSL filtering - [Apple Music](https://play.google.com/store/apps/details?id=com.apple.android.music&hl=uk) ั ะฒะบะปัŽั‡ะตะฝะฝะพะน ะทะฐั‰ะธั‚ะพะน, ะฝะตะบะพั€ั€ะตะบั‚ะฝะพ ะธะฝะพะณะดะฐ ั€ะฐะฑะพั‚ะฐะตั‚! ะพะฑั€ั‹ะฒั‹ ะฒ ะฟั€ะพัะปัƒัˆะธะฒะฐะฝะธะธ ,ะฝะต ะฟะพะบะฐะทั‹ะฒะฐะตั‚ ั€ะตะบะพะผะตะฝะดะฐั†ะธะธ ะฒะพ ะฒะบะปะฐะดะบะต (ะดะปั ะฒะฐั) ะฟั€ะธั…ะพะดะธั‚ัั ะฒั‹ะบะปัŽั‡ะฐั‚ัŒ ะทะฐั‰ะธั‚ัƒ, ะฟะตั€ะตะทะฟัƒัะบะฐั‚ัŒ Apple Music, ะฟะพั‚ะพะผ ะฒะบะปัŽั‡ะฐั‚ัŒ ะทะฐั‰ะธั‚ัƒ ะพะฑั€ะฐั‚ะฝะพ. ะžั‚ะบะปัŽั‡ะตะฝะธะต HTTPS ั„ะธะปัŒั‚ั€ะฐั†ะธะธ ะฟะพะผะพะณะฐะตั‚.
non_test
apple music is incompatible with ssl filtering ั ะฒะบะปัŽั‡ะตะฝะฝะพะน ะทะฐั‰ะธั‚ะพะน ะฝะตะบะพั€ั€ะตะบั‚ะฝะพ ะธะฝะพะณะดะฐ ั€ะฐะฑะพั‚ะฐะตั‚ ะพะฑั€ั‹ะฒั‹ ะฒ ะฟั€ะพัะปัƒัˆะธะฒะฐะฝะธะธ ะฝะต ะฟะพะบะฐะทั‹ะฒะฐะตั‚ ั€ะตะบะพะผะตะฝะดะฐั†ะธะธ ะฒะพ ะฒะบะปะฐะดะบะต ะดะปั ะฒะฐั ะฟั€ะธั…ะพะดะธั‚ัั ะฒั‹ะบะปัŽั‡ะฐั‚ัŒ ะทะฐั‰ะธั‚ัƒ ะฟะตั€ะตะทะฟัƒัะบะฐั‚ัŒ apple music ะฟะพั‚ะพะผ ะฒะบะปัŽั‡ะฐั‚ัŒ ะทะฐั‰ะธั‚ัƒ ะพะฑั€ะฐั‚ะฝะพ ะพั‚ะบะปัŽั‡ะตะฝะธะต https ั„ะธะปัŒั‚ั€ะฐั†ะธะธ ะฟะพะผะพะณะฐะตั‚
0
302,737
26,160,302,199
IssuesEvent
2022-12-31 12:10:53
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: schemachange/random-load failed
C-test-failure O-robot O-roachtest branch-release-22.1
roachtest.schemachange/random-load [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=8147270&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=8147270&tab=artifacts#/schemachange/random-load) on release-22.1 @ [000c9624b56b09d5fbd06557c559b2f910142a9c](https://github.com/cockroachdb/cockroach/commits/000c9624b56b09d5fbd06557c559b2f910142a9c): ``` | I221231 12:10:08.149551 224 workload/pgx_helpers.go:79 [-] 4 pgx logger [error]: Exec logParams=map[args:[] err:ERROR: database must have associated regions before a survival goal can be set (SQLSTATE 42602) pid:1380207 sql:ALTER DATABASE schemachange SURVIVE REGION FAILURE] | I221231 12:10:08.165902 213 workload/pgx_helpers.go:79 [-] 5 pgx logger [error]: Exec logParams=map[args:[] err:ERROR: relation "public.table2" does not exist (SQLSTATE 42P01) pid:1222218 sql:ALTER TABLE public.table2 ADD COLUMN IrrelevantColumnName string] | I221231 12:10:08.167336 221 workload/pgx_helpers.go:79 [-] 6 pgx logger [error]: Exec logParams=map[args:[] err:ERROR: Use of multi-region features requires an enterprise license. Your evaluation license expired on December 30, 2022. If you're interested in getting a new license, please contact subscriptions@cockroachlabs.com and we can help you out. (SQLSTATE XXC02) pid:1775429 sql:ALTER DATABASE schemachange PRIMARY REGION "europe-west2"] | Error: ***UNEXPECTED ERROR; Received an unexpected execution error. Dumping state before death: | Expected errors: ===========================Executed queries for generating errors: QUERY [SELECT region FROM [SHOW REGIONS FROM CLUSTER]] : us-east1,us-west1,europe-west2, | QUERY [SELECT region FROM [SHOW REGIONS FROM DATABASE]] : | QUERY [SHOW DATABASE] :schemachange | ===========================Previous statements [ALTER DATABASE schemachange PRIMARY REGION "europe-west2"]: ERROR: Use of multi-region features requires an enterprise license. Your evaluation license expired on December 30, 2022. If you're interested in getting a new license, please contact subscriptions@cockroachlabs.com and we can help you out. (SQLSTATE XXC02) | | stdout: | { | "workerId": 0, | "clientTimestamp": "12:10:08.046319", | "ops": [ | "BEGIN", | "ALTER DATABASE schemachange ADD REGION \"europe-west2\"" | ], | "expectedExecErrors": "42P12", | "expectedCommitErrors": "", | "message": "ROLLBACK; Successfully got expected execution error. Dumping state before death:\nExpected errors: 42P12===========================Executed queries for generating errors: QUERY [SELECT region FROM [SHOW REGIONS FROM CLUSTER]] : us-east1,us-west1,europe-west2,\nQUERY [SELECT region FROM [SHOW REGIONS FROM DATABASE]] : \nQUERY [SHOW DATABASE] :schemachange\n===========================Previous statements [ALTER DATABASE schemachange ADD REGION \"europe-west2\"]: ERROR: cannot add region \"europe-west2\" to database schemachange (SQLSTATE 42P12)" | } | { | "workerId": 0, | "clientTimestamp": "12:10:08.04797", | "ops": [ | "BEGIN", | "ALTER DATABASE schemachange SURVIVE REGION FAILURE" | ], | "expectedExecErrors": "22023,42602", | "expectedCommitErrors": "", | "message": "ROLLBACK; Successfully got expected execution error. Dumping state before death:\nExpected errors: 22023,42602===========================Executed queries for generating errors: QUERY [SELECT region FROM [SHOW REGIONS FROM DATABASE]] : \nQUERY [SHOW DATABASE] :schemachange\n===========================Previous statements [ALTER DATABASE schemachange SURVIVE REGION FAILURE]: ERROR: database must have associated regions before a survival goal can be set (SQLSTATE 42602)" | } | { | "workerId": 0, | "clientTimestamp": "12:10:08.045123", | "ops": [ | "BEGIN", | "ALTER DATABASE schemachange PRIMARY REGION \"europe-west2\"" | ], | "expectedExecErrors": "", | "expectedCommitErrors": "", | "message": "***UNEXPECTED ERROR; Received an unexpected execution error. Dumping state before death:\nExpected errors: ===========================Executed queries for generating errors: QUERY [SELECT region FROM [SHOW REGIONS FROM CLUSTER]] : us-east1,us-west1,europe-west2,\nQUERY [SELECT region FROM [SHOW REGIONS FROM DATABASE]] : \nQUERY [SHOW DATABASE] :schemachange\n===========================Previous statements [ALTER DATABASE schemachange PRIMARY REGION \"europe-west2\"]: ERROR: Use of multi-region features requires an enterprise license. Your evaluation license expired on December 30, 2022. If you're interested in getting a new license, please contact subscriptions@cockroachlabs.com and we can help you out. (SQLSTATE XXC02)" | } Wraps: (4) COMMAND_PROBLEM Wraps: (5) Node 1. Command with error: | `````` | ./workload run schemachange --verbose=1 --tolerate-errors=false --histograms=perf/stats.json --max-ops 5000 --concurrency 20 --txn-log /mnt/data1/cockroach/transactions.json | `````` Wraps: (6) exit status 1 Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *cluster.WithCommandDetails (4) errors.Cmd (5) *hintdetail.withDetail (6) *exec.ExitError ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #91353 roachtest: schemachange/random-load failed [C-test-failure O-roachtest O-robot T-sql-schema branch-release-22.2] </p> </details> /cc @cockroachdb/sql-schema <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*schemachange/random-load.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: schemachange/random-load failed - roachtest.schemachange/random-load [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=8147270&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=8147270&tab=artifacts#/schemachange/random-load) on release-22.1 @ [000c9624b56b09d5fbd06557c559b2f910142a9c](https://github.com/cockroachdb/cockroach/commits/000c9624b56b09d5fbd06557c559b2f910142a9c): ``` | I221231 12:10:08.149551 224 workload/pgx_helpers.go:79 [-] 4 pgx logger [error]: Exec logParams=map[args:[] err:ERROR: database must have associated regions before a survival goal can be set (SQLSTATE 42602) pid:1380207 sql:ALTER DATABASE schemachange SURVIVE REGION FAILURE] | I221231 12:10:08.165902 213 workload/pgx_helpers.go:79 [-] 5 pgx logger [error]: Exec logParams=map[args:[] err:ERROR: relation "public.table2" does not exist (SQLSTATE 42P01) pid:1222218 sql:ALTER TABLE public.table2 ADD COLUMN IrrelevantColumnName string] | I221231 12:10:08.167336 221 workload/pgx_helpers.go:79 [-] 6 pgx logger [error]: Exec logParams=map[args:[] err:ERROR: Use of multi-region features requires an enterprise license. Your evaluation license expired on December 30, 2022. If you're interested in getting a new license, please contact subscriptions@cockroachlabs.com and we can help you out. (SQLSTATE XXC02) pid:1775429 sql:ALTER DATABASE schemachange PRIMARY REGION "europe-west2"] | Error: ***UNEXPECTED ERROR; Received an unexpected execution error. Dumping state before death: | Expected errors: ===========================Executed queries for generating errors: QUERY [SELECT region FROM [SHOW REGIONS FROM CLUSTER]] : us-east1,us-west1,europe-west2, | QUERY [SELECT region FROM [SHOW REGIONS FROM DATABASE]] : | QUERY [SHOW DATABASE] :schemachange | ===========================Previous statements [ALTER DATABASE schemachange PRIMARY REGION "europe-west2"]: ERROR: Use of multi-region features requires an enterprise license. Your evaluation license expired on December 30, 2022. If you're interested in getting a new license, please contact subscriptions@cockroachlabs.com and we can help you out. (SQLSTATE XXC02) | | stdout: | { | "workerId": 0, | "clientTimestamp": "12:10:08.046319", | "ops": [ | "BEGIN", | "ALTER DATABASE schemachange ADD REGION \"europe-west2\"" | ], | "expectedExecErrors": "42P12", | "expectedCommitErrors": "", | "message": "ROLLBACK; Successfully got expected execution error. Dumping state before death:\nExpected errors: 42P12===========================Executed queries for generating errors: QUERY [SELECT region FROM [SHOW REGIONS FROM CLUSTER]] : us-east1,us-west1,europe-west2,\nQUERY [SELECT region FROM [SHOW REGIONS FROM DATABASE]] : \nQUERY [SHOW DATABASE] :schemachange\n===========================Previous statements [ALTER DATABASE schemachange ADD REGION \"europe-west2\"]: ERROR: cannot add region \"europe-west2\" to database schemachange (SQLSTATE 42P12)" | } | { | "workerId": 0, | "clientTimestamp": "12:10:08.04797", | "ops": [ | "BEGIN", | "ALTER DATABASE schemachange SURVIVE REGION FAILURE" | ], | "expectedExecErrors": "22023,42602", | "expectedCommitErrors": "", | "message": "ROLLBACK; Successfully got expected execution error. Dumping state before death:\nExpected errors: 22023,42602===========================Executed queries for generating errors: QUERY [SELECT region FROM [SHOW REGIONS FROM DATABASE]] : \nQUERY [SHOW DATABASE] :schemachange\n===========================Previous statements [ALTER DATABASE schemachange SURVIVE REGION FAILURE]: ERROR: database must have associated regions before a survival goal can be set (SQLSTATE 42602)" | } | { | "workerId": 0, | "clientTimestamp": "12:10:08.045123", | "ops": [ | "BEGIN", | "ALTER DATABASE schemachange PRIMARY REGION \"europe-west2\"" | ], | "expectedExecErrors": "", | "expectedCommitErrors": "", | "message": "***UNEXPECTED ERROR; Received an unexpected execution error. Dumping state before death:\nExpected errors: ===========================Executed queries for generating errors: QUERY [SELECT region FROM [SHOW REGIONS FROM CLUSTER]] : us-east1,us-west1,europe-west2,\nQUERY [SELECT region FROM [SHOW REGIONS FROM DATABASE]] : \nQUERY [SHOW DATABASE] :schemachange\n===========================Previous statements [ALTER DATABASE schemachange PRIMARY REGION \"europe-west2\"]: ERROR: Use of multi-region features requires an enterprise license. Your evaluation license expired on December 30, 2022. If you're interested in getting a new license, please contact subscriptions@cockroachlabs.com and we can help you out. (SQLSTATE XXC02)" | } Wraps: (4) COMMAND_PROBLEM Wraps: (5) Node 1. Command with error: | `````` | ./workload run schemachange --verbose=1 --tolerate-errors=false --histograms=perf/stats.json --max-ops 5000 --concurrency 20 --txn-log /mnt/data1/cockroach/transactions.json | `````` Wraps: (6) exit status 1 Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *cluster.WithCommandDetails (4) errors.Cmd (5) *hintdetail.withDetail (6) *exec.ExitError ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #91353 roachtest: schemachange/random-load failed [C-test-failure O-roachtest O-robot T-sql-schema branch-release-22.2] </p> </details> /cc @cockroachdb/sql-schema <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*schemachange/random-load.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
roachtest schemachange random load failed roachtest schemachange random load with on release workload pgx helpers go pgx logger exec logparams map err error database must have associated regions before a survival goal can be set sqlstate pid sql alter database schemachange survive region failure workload pgx helpers go pgx logger exec logparams map err error relation public does not exist sqlstate pid sql alter table public add column irrelevantcolumnname string workload pgx helpers go pgx logger exec logparams map err error use of multi region features requires an enterprise license your evaluation license expired on december if you re interested in getting a new license please contact subscriptions cockroachlabs com and we can help you out sqlstate pid sql alter database schemachange primary region europe error unexpected error received an unexpected execution error dumping state before death expected errors executed queries for generating errors query us us europe query query schemachange previous statements error use of multi region features requires an enterprise license your evaluation license expired on december if you re interested in getting a new license please contact subscriptions cockroachlabs com and we can help you out sqlstate stdout workerid clienttimestamp ops begin alter database schemachange add region europe expectedexecerrors expectedcommiterrors message rollback successfully got expected execution error dumping state before death nexpected errors executed queries for generating errors query us us europe nquery nquery schemachange n previous statements error cannot add region europe to database schemachange sqlstate workerid clienttimestamp ops begin alter database schemachange survive region failure expectedexecerrors expectedcommiterrors message rollback successfully got expected execution error dumping state before death nexpected errors executed queries for generating errors query nquery schemachange n previous statements error database must have associated regions before a survival goal can be set sqlstate workerid clienttimestamp ops begin alter database schemachange primary region europe expectedexecerrors expectedcommiterrors message unexpected error received an unexpected execution error dumping state before death nexpected errors executed queries for generating errors query us us europe nquery nquery schemachange n previous statements error use of multi region features requires an enterprise license your evaluation license expired on december if you re interested in getting a new license please contact subscriptions cockroachlabs com and we can help you out sqlstate wraps command problem wraps node command with error workload run schemachange verbose tolerate errors false histograms perf stats json max ops concurrency txn log mnt cockroach transactions json wraps exit status error types withstack withstack errutil withprefix cluster withcommanddetails errors cmd hintdetail withdetail exec exiterror help see see same failure on other branches roachtest schemachange random load failed cc cockroachdb sql schema
1
16,881
3,573,330,199
IssuesEvent
2016-01-27 05:31:15
Microsoft/vscode
https://api.github.com/repos/Microsoft/vscode
closed
Find Widget tab order
testplan-item
- [X] any platform @PJMeyer The find widget can be in 4 states, please check the tab order in all the 4 states: **no search string, replace hidden:** ![image](https://cloud.githubusercontent.com/assets/5047891/12195674/381b5156-b5f9-11e5-87fb-e2cc34ed643b.png) **no search string, replace revealed:** ![image](https://cloud.githubusercontent.com/assets/5047891/12195687/5b1207fe-b5f9-11e5-98f6-3263b394d12f.png) **with search string, replace hidden:** ![image](https://cloud.githubusercontent.com/assets/5047891/12195700/75f82684-b5f9-11e5-88c5-276bcb02c581.png) **with search string, replace revealed:** ![image](https://cloud.githubusercontent.com/assets/5047891/12195715/87bdb94c-b5f9-11e5-9861-bddcbc78b5ed.png) Moreover, the `Find in Selection` toggle can be enabled or disabled based on current editor selection. The nominal tab order (when everything is enabled) is the following: ![image](https://cloud.githubusercontent.com/assets/5047891/12195773/0b12ece0-b5fa-11e5-96f0-e7861ef3949f.png) Please check carefully with <kbd>Tab</kbd> and <kbd>Shift</kbd>+<kbd>Tab</kbd> as a lot of tricks are needed to implement this tab order (that is not the element order in the DOM)
1.0
Find Widget tab order - - [X] any platform @PJMeyer The find widget can be in 4 states, please check the tab order in all the 4 states: **no search string, replace hidden:** ![image](https://cloud.githubusercontent.com/assets/5047891/12195674/381b5156-b5f9-11e5-87fb-e2cc34ed643b.png) **no search string, replace revealed:** ![image](https://cloud.githubusercontent.com/assets/5047891/12195687/5b1207fe-b5f9-11e5-98f6-3263b394d12f.png) **with search string, replace hidden:** ![image](https://cloud.githubusercontent.com/assets/5047891/12195700/75f82684-b5f9-11e5-88c5-276bcb02c581.png) **with search string, replace revealed:** ![image](https://cloud.githubusercontent.com/assets/5047891/12195715/87bdb94c-b5f9-11e5-9861-bddcbc78b5ed.png) Moreover, the `Find in Selection` toggle can be enabled or disabled based on current editor selection. The nominal tab order (when everything is enabled) is the following: ![image](https://cloud.githubusercontent.com/assets/5047891/12195773/0b12ece0-b5fa-11e5-96f0-e7861ef3949f.png) Please check carefully with <kbd>Tab</kbd> and <kbd>Shift</kbd>+<kbd>Tab</kbd> as a lot of tricks are needed to implement this tab order (that is not the element order in the DOM)
test
find widget tab order any platform pjmeyer the find widget can be in states please check the tab order in all the states no search string replace hidden no search string replace revealed with search string replace hidden with search string replace revealed moreover the find in selection toggle can be enabled or disabled based on current editor selection the nominal tab order when everything is enabled is the following please check carefully with tab and shift tab as a lot of tricks are needed to implement this tab order that is not the element order in the dom
1
124,117
26,405,790,602
IssuesEvent
2023-01-13 07:51:26
yt-project/yt
https://api.github.com/repos/yt-project/yt
closed
NaNs generated in grid edges for 1D FLASH data
bug code frontends
### Bug report After loading 1D FLASH data index.grid_left_edge and index.grid_right edge contain NaN values for the z axis, which are not present in the HDF5 file. This appears to then lead to errors when trying to generate data arrays. Link to data: http://use.yt/upload/e175762f ```python ds = yt.load('pp_shock_hdf5_plt_cnt_0000') dens = ds.r['gas','density'] ``` ```python-traceback RuntimeError Traceback (most recent call last) Cell In[4], line 1 ----> 1 dens = ds.r['gas','density'] File /export/slavin/anaconda3/envs/py38/lib/python3.8/site-packages/yt/data_objects/region_expression.py:30, in RegionExpression.__getitem__(self, item) 25 def __getitem__(self, item): 26 # At first, we will only implement this as accepting a slice that is 27 # (optionally) unitful corresponding to a specific set of coordinates 28 # that result in a rectangular prism or a slice. 29 try: ---> 30 return self.all_data[item] 31 except (TypeError, YTFieldNotParseable): 32 pass File /export/slavin/anaconda3/envs/py38/lib/python3.8/site-packages/yt/data_objects/data_containers.py:264, in YTDataContainer.__getitem__(self, key) 262 return self.field_data[f] 263 else: --> 264 self.get_data(f) 265 # fi.units is the unit expression string. We depend on the registry 266 # hanging off the dataset to define this unit object. 267 # Note that this is less succinct so that we can account for the case 268 # when there are, for example, no elements in the object. 269 try: File /export/slavin/anaconda3/envs/py38/lib/python3.8/site-packages/yt/data_objects/selection_objects/data_selection_objects.py:131, in YTSelectionContainer.get_data(self, fields) 129 def get_data(self, fields=None): 130 if self._current_chunk is None: --> 131 self.index._identify_base_chunk(self) 132 if fields is None: 133 return File /export/slavin/anaconda3/envs/py38/lib/python3.8/site-packages/yt/geometry/grid_geometry_handler.py:336, in GridIndex._identify_base_chunk(self, dobj) 334 dobj._chunk_info[0] = weakref.proxy(dobj) 335 elif getattr(dobj, "_grids", None) is None: --> 336 gi = dobj.selector.select_grids( 337 self.grid_left_edge, self.grid_right_edge, self.grid_levels 338 ) 339 if any([g.filename is not None for g in self.grids[gi]]): 340 _gsort = _grid_sort_mixed File /export/slavin/anaconda3/envs/py38/lib/python3.8/site-packages/yt/data_objects/selection_objects/data_selection_objects.py:81, in YTSelectionContainer.selector(self) 77 self._selector = compose_selector( 78 self, self._data_source.selector, sclass(self) 79 ) 80 else: ---> 81 self._selector = sclass(self) 82 return self._selector File /export/slavin/anaconda3/envs/py38/lib/python3.8/site-packages/yt/geometry/_selection_routines/region_selector.pxi:39, in yt.geometry.selection_routines.RegionSelector.__init__() RuntimeError: Region right edge[2] < left edge: width = 0.0 ``` **Expected outcome** <!--A description of the expected outcome from the code snippet--> <!--If this used to work in an earlier version of yt, please note the version it used to work on--> dens should be a 1D unyt_array **Version Information** <!--Please specify your platform and versions of the relevant libraries you are using:--> * Operating System: Linux 4.18.0-425.3.1.el8.x86_64 * Python Version: Python 3.8.13 * yt version: 4.1.1 * Other Libraries (if applicable): <!--Please tell us how you installed yt and python e.g., from source, pip, conda. If you installed from conda, please specify which channel you used if not the default--> conda edited by @neutrinoceros (formatting)
1.0
NaNs generated in grid edges for 1D FLASH data - ### Bug report After loading 1D FLASH data index.grid_left_edge and index.grid_right edge contain NaN values for the z axis, which are not present in the HDF5 file. This appears to then lead to errors when trying to generate data arrays. Link to data: http://use.yt/upload/e175762f ```python ds = yt.load('pp_shock_hdf5_plt_cnt_0000') dens = ds.r['gas','density'] ``` ```python-traceback RuntimeError Traceback (most recent call last) Cell In[4], line 1 ----> 1 dens = ds.r['gas','density'] File /export/slavin/anaconda3/envs/py38/lib/python3.8/site-packages/yt/data_objects/region_expression.py:30, in RegionExpression.__getitem__(self, item) 25 def __getitem__(self, item): 26 # At first, we will only implement this as accepting a slice that is 27 # (optionally) unitful corresponding to a specific set of coordinates 28 # that result in a rectangular prism or a slice. 29 try: ---> 30 return self.all_data[item] 31 except (TypeError, YTFieldNotParseable): 32 pass File /export/slavin/anaconda3/envs/py38/lib/python3.8/site-packages/yt/data_objects/data_containers.py:264, in YTDataContainer.__getitem__(self, key) 262 return self.field_data[f] 263 else: --> 264 self.get_data(f) 265 # fi.units is the unit expression string. We depend on the registry 266 # hanging off the dataset to define this unit object. 267 # Note that this is less succinct so that we can account for the case 268 # when there are, for example, no elements in the object. 269 try: File /export/slavin/anaconda3/envs/py38/lib/python3.8/site-packages/yt/data_objects/selection_objects/data_selection_objects.py:131, in YTSelectionContainer.get_data(self, fields) 129 def get_data(self, fields=None): 130 if self._current_chunk is None: --> 131 self.index._identify_base_chunk(self) 132 if fields is None: 133 return File /export/slavin/anaconda3/envs/py38/lib/python3.8/site-packages/yt/geometry/grid_geometry_handler.py:336, in GridIndex._identify_base_chunk(self, dobj) 334 dobj._chunk_info[0] = weakref.proxy(dobj) 335 elif getattr(dobj, "_grids", None) is None: --> 336 gi = dobj.selector.select_grids( 337 self.grid_left_edge, self.grid_right_edge, self.grid_levels 338 ) 339 if any([g.filename is not None for g in self.grids[gi]]): 340 _gsort = _grid_sort_mixed File /export/slavin/anaconda3/envs/py38/lib/python3.8/site-packages/yt/data_objects/selection_objects/data_selection_objects.py:81, in YTSelectionContainer.selector(self) 77 self._selector = compose_selector( 78 self, self._data_source.selector, sclass(self) 79 ) 80 else: ---> 81 self._selector = sclass(self) 82 return self._selector File /export/slavin/anaconda3/envs/py38/lib/python3.8/site-packages/yt/geometry/_selection_routines/region_selector.pxi:39, in yt.geometry.selection_routines.RegionSelector.__init__() RuntimeError: Region right edge[2] < left edge: width = 0.0 ``` **Expected outcome** <!--A description of the expected outcome from the code snippet--> <!--If this used to work in an earlier version of yt, please note the version it used to work on--> dens should be a 1D unyt_array **Version Information** <!--Please specify your platform and versions of the relevant libraries you are using:--> * Operating System: Linux 4.18.0-425.3.1.el8.x86_64 * Python Version: Python 3.8.13 * yt version: 4.1.1 * Other Libraries (if applicable): <!--Please tell us how you installed yt and python e.g., from source, pip, conda. If you installed from conda, please specify which channel you used if not the default--> conda edited by @neutrinoceros (formatting)
non_test
nans generated in grid edges for flash data bug report after loading flash data index grid left edge and index grid right edge contain nan values for the z axis which are not present in the file this appears to then lead to errors when trying to generate data arrays link to data python ds yt load pp shock plt cnt dens ds r python traceback runtimeerror traceback most recent call last cell in line dens ds r file export slavin envs lib site packages yt data objects region expression py in regionexpression getitem self item def getitem self item at first we will only implement this as accepting a slice that is optionally unitful corresponding to a specific set of coordinates that result in a rectangular prism or a slice try return self all data except typeerror ytfieldnotparseable pass file export slavin envs lib site packages yt data objects data containers py in ytdatacontainer getitem self key return self field data else self get data f fi units is the unit expression string we depend on the registry hanging off the dataset to define this unit object note that this is less succinct so that we can account for the case when there are for example no elements in the object try file export slavin envs lib site packages yt data objects selection objects data selection objects py in ytselectioncontainer get data self fields def get data self fields none if self current chunk is none self index identify base chunk self if fields is none return file export slavin envs lib site packages yt geometry grid geometry handler py in gridindex identify base chunk self dobj dobj chunk info weakref proxy dobj elif getattr dobj grids none is none gi dobj selector select grids self grid left edge self grid right edge self grid levels if any gsort grid sort mixed file export slavin envs lib site packages yt data objects selection objects data selection objects py in ytselectioncontainer selector self self selector compose selector self self data source selector sclass self else self selector sclass self return self selector file export slavin envs lib site packages yt geometry selection routines region selector pxi in yt geometry selection routines regionselector init runtimeerror region right edge left edge width expected outcome if this used to work in an earlier version of yt please note the version it used to work on dens should be a unyt array version information please specify your platform and versions of the relevant libraries you are using operating system linux python version python yt version other libraries if applicable please tell us how you installed yt and python e g from source pip conda if you installed from conda please specify which channel you used if not the default conda edited by neutrinoceros formatting
0
222,560
7,433,785,730
IssuesEvent
2018-03-26 08:50:54
quintel/etmodel
https://api.github.com/repos/quintel/etmodel
closed
CO2 sheet total cut off
Priority
In some cases, the total ghg emissions in the co2 sheet are not displayed correctly. @antw Is it possible to increase the margin a little bit (and deploy the changes to production)? ![monosnap 2018-03-26 09-18-19 1](https://user-images.githubusercontent.com/32056448/37892025-b9e489c8-30d6-11e8-8eef-b9b41498de6f.png)
1.0
CO2 sheet total cut off - In some cases, the total ghg emissions in the co2 sheet are not displayed correctly. @antw Is it possible to increase the margin a little bit (and deploy the changes to production)? ![monosnap 2018-03-26 09-18-19 1](https://user-images.githubusercontent.com/32056448/37892025-b9e489c8-30d6-11e8-8eef-b9b41498de6f.png)
non_test
sheet total cut off in some cases the total ghg emissions in the sheet are not displayed correctly antw is it possible to increase the margin a little bit and deploy the changes to production
0
636,158
20,593,758,784
IssuesEvent
2022-03-05 06:40:27
apache/dolphinscheduler
https://api.github.com/repos/apache/dolphinscheduler
closed
[Bug][UI-Next][V1.0.0-Alpha] click timing error in process definition page
bug UI priority:high
### Search before asking - [X] I had searched in the [issues](https://github.com/apache/dolphinscheduler/issues?q=is%3Aissue) and found no similar issues. ### What happened ![](https://vip2.loli.io/2022/03/01/7cNiWD5Uyw18EFB.png) ### What you expected to happen click successfully. ### How to reproduce above. ### Anything else _No response_ ### Version dev ### Are you willing to submit PR? - [ ] Yes I am willing to submit a PR! ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
1.0
[Bug][UI-Next][V1.0.0-Alpha] click timing error in process definition page - ### Search before asking - [X] I had searched in the [issues](https://github.com/apache/dolphinscheduler/issues?q=is%3Aissue) and found no similar issues. ### What happened ![](https://vip2.loli.io/2022/03/01/7cNiWD5Uyw18EFB.png) ### What you expected to happen click successfully. ### How to reproduce above. ### Anything else _No response_ ### Version dev ### Are you willing to submit PR? - [ ] Yes I am willing to submit a PR! ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
non_test
click timing error in process definition page search before asking i had searched in the and found no similar issues what happened what you expected to happen click successfully how to reproduce above anything else no response version dev are you willing to submit pr yes i am willing to submit a pr code of conduct i agree to follow this project s
0
210,053
16,083,926,495
IssuesEvent
2021-04-26 08:57:31
ARMmbed/connectedhomeip
https://api.github.com/repos/ARMmbed/connectedhomeip
opened
'Mutex 0x0 error' when executting some of unit-tests
testing unit-tests
#### Problem Running some of unit-tests (group) causes to 'Mutex 0x0 error' occur. #### Proposed Solution 1. Isolate tests which causes that problem. 2. Debug it by investigating one of problematic tests 3. Fix it!
2.0
'Mutex 0x0 error' when executting some of unit-tests - #### Problem Running some of unit-tests (group) causes to 'Mutex 0x0 error' occur. #### Proposed Solution 1. Isolate tests which causes that problem. 2. Debug it by investigating one of problematic tests 3. Fix it!
test
mutex error when executting some of unit tests problem running some of unit tests group causes to mutex error occur proposed solution isolate tests which causes that problem debug it by investigating one of problematic tests fix it
1
140,294
11,307,977,073
IssuesEvent
2020-01-19 01:24:01
OpenAPITools/openapi-generator
https://api.github.com/repos/OpenAPITools/openapi-generator
closed
Upgrade CircleCI JDK8 to a newer version to catch JavaDoc errors
Enhancement: CI/Test
##### Description Current JDK8 in CircleCI build won't detect the following Javadoc errors: ``` [ERROR] /Users/williamcheng/Code/openapi-generator/samples/client/petstore/java/rest-assured/src/main/java/org/openapitools/client/api/UserApi.java:271: error: unknown tag: User [ERROR] * @param user (List<User>) List of user object (required) [ERROR] ^ [ERROR] [ERROR] Command line was: /Library/Java/JavaVirtualMachines/jdk1.8.0_172.jdk/Contents/Home/jre/../bin/javadoc @options @packages [ERROR] [ERROR] Refer to the generated Javadoc files in '/Users/williamcheng/Code/openapi-generator/samples/client/petstore/java/rest-assured/target/apidocs' dir. [ERROR] [ERROR] -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException ``` Current JDK version in CircleCI: ``` openjdk version "1.8.0_141" ``` ##### openapi-generator version Latest master ##### Related issues/PRs https://github.com/OpenAPITools/openapi-generator/pull/697#issuecomment-409573524 ##### Suggest a fix/enhancement Upgrade JDK to a new version (8.x)
1.0
Upgrade CircleCI JDK8 to a newer version to catch JavaDoc errors - ##### Description Current JDK8 in CircleCI build won't detect the following Javadoc errors: ``` [ERROR] /Users/williamcheng/Code/openapi-generator/samples/client/petstore/java/rest-assured/src/main/java/org/openapitools/client/api/UserApi.java:271: error: unknown tag: User [ERROR] * @param user (List<User>) List of user object (required) [ERROR] ^ [ERROR] [ERROR] Command line was: /Library/Java/JavaVirtualMachines/jdk1.8.0_172.jdk/Contents/Home/jre/../bin/javadoc @options @packages [ERROR] [ERROR] Refer to the generated Javadoc files in '/Users/williamcheng/Code/openapi-generator/samples/client/petstore/java/rest-assured/target/apidocs' dir. [ERROR] [ERROR] -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException ``` Current JDK version in CircleCI: ``` openjdk version "1.8.0_141" ``` ##### openapi-generator version Latest master ##### Related issues/PRs https://github.com/OpenAPITools/openapi-generator/pull/697#issuecomment-409573524 ##### Suggest a fix/enhancement Upgrade JDK to a new version (8.x)
test
upgrade circleci to a newer version to catch javadoc errors description current in circleci build won t detect the following javadoc errors users williamcheng code openapi generator samples client petstore java rest assured src main java org openapitools client api userapi java error unknown tag user param user list list of user object required command line was library java javavirtualmachines jdk contents home jre bin javadoc options packages refer to the generated javadoc files in users williamcheng code openapi generator samples client petstore java rest assured target apidocs dir to see the full stack trace of the errors re run maven with the e switch re run maven using the x switch to enable full debug logging for more information about the errors and possible solutions please read the following articles current jdk version in circleci openjdk version openapi generator version latest master related issues prs suggest a fix enhancement upgrade jdk to a new version x
1
163,287
12,715,869,904
IssuesEvent
2020-06-24 00:03:34
Thy-Vipe/BeastsOfBermuda-issues
https://api.github.com/repos/Thy-Vipe/BeastsOfBermuda-issues
closed
[Bug] jungle surface mesh hole
Map bug tester-team
_Originally written by **Cleafspear | 76561198077984700**_ Game Version: 1.1.946 *===== System Specs ===== CPU Brand: Intel(R) Core(TM) i7-8700K CPU @ 3.70GHz Vendor: GenuineIntel GPU Brand: NVIDIA GeForce GTX 1080 GPU Driver Info: Unknown Num CPU Cores: 6 ===================* Map: Rival_Shores location where there are 3 triangels in the mesh that have no collision, you can fall through on to a rock below Location: X=110341.047 Y=88280.383 Z=7504.488
1.0
[Bug] jungle surface mesh hole - _Originally written by **Cleafspear | 76561198077984700**_ Game Version: 1.1.946 *===== System Specs ===== CPU Brand: Intel(R) Core(TM) i7-8700K CPU @ 3.70GHz Vendor: GenuineIntel GPU Brand: NVIDIA GeForce GTX 1080 GPU Driver Info: Unknown Num CPU Cores: 6 ===================* Map: Rival_Shores location where there are 3 triangels in the mesh that have no collision, you can fall through on to a rock below Location: X=110341.047 Y=88280.383 Z=7504.488
test
jungle surface mesh hole originally written by cleafspear game version system specs cpu brand intel r core tm cpu vendor genuineintel gpu brand nvidia geforce gtx gpu driver info unknown num cpu cores map rival shores location where there are triangels in the mesh that have no collision you can fall through on to a rock below location x y z
1
300,516
25,973,625,091
IssuesEvent
2022-12-19 13:15:35
ubtue/DatenProbleme
https://api.github.com/repos/ubtue/DatenProbleme
closed
ISSN 1988-7949 | International E-Journal of Criminal Sciences (OJS) | Bandangabe
ready for testing Zotero_SEMI-AUTO
#### URL https://ojs.ehu.eus/index.php/inecs/issue/view/1906 #### Import-Translator Einzel- und Mehrfachimport: ubtue_Open Journal Systems Standard.js ### Problembeschreibung Diese ZS hat eine durchgรคngige Heftzรคhlung ohne zusรคtzliche Bandangabe, weswegen die Zรคhlung in die Bandangabe laut Handbuch muss. Kรถnnte das beim Import oder Export entsprechend fรผr diese ZS angepasst werden?
1.0
ISSN 1988-7949 | International E-Journal of Criminal Sciences (OJS) | Bandangabe - #### URL https://ojs.ehu.eus/index.php/inecs/issue/view/1906 #### Import-Translator Einzel- und Mehrfachimport: ubtue_Open Journal Systems Standard.js ### Problembeschreibung Diese ZS hat eine durchgรคngige Heftzรคhlung ohne zusรคtzliche Bandangabe, weswegen die Zรคhlung in die Bandangabe laut Handbuch muss. Kรถnnte das beim Import oder Export entsprechend fรผr diese ZS angepasst werden?
test
issn international e journal of criminal sciences ojs bandangabe url import translator einzel und mehrfachimport ubtue open journal systems standard js problembeschreibung diese zs hat eine durchgรคngige heftzรคhlung ohne zusรคtzliche bandangabe weswegen die zรคhlung in die bandangabe laut handbuch muss kรถnnte das beim import oder export entsprechend fรผr diese zs angepasst werden
1
744,470
25,944,513,202
IssuesEvent
2022-12-16 22:21:30
LoMaTiInformatik/Smarthome_Ventilator
https://api.github.com/repos/LoMaTiInformatik/Smarthome_Ventilator
opened
Implementation
enhancement highpriority
Put the arduinos in the fan. Add some debugging features such as a external reset switch and usb connectors for both arduinos. Powersupplies needed. Wiring?
1.0
Implementation - Put the arduinos in the fan. Add some debugging features such as a external reset switch and usb connectors for both arduinos. Powersupplies needed. Wiring?
non_test
implementation put the arduinos in the fan add some debugging features such as a external reset switch and usb connectors for both arduinos powersupplies needed wiring
0
4,678
2,610,140,007
IssuesEvent
2015-02-26 18:44:08
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
Invincibility isn't invincible
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Use the "Invincibility" drop item 2. Use "Kamikaze" 3. Profit What is the expected output? What do you see instead? I would expect the hedgehog to be invincible and therefore unable to be killed, yet the Kamikaze move still results in death, hence non-invincible. What version of the product are you using? On what operating system? The latest version, downloaded today. OS X 10.6.4. Please provide any additional information below. It resulted in me losing a round. :( ``` ----- Original issue reported on code.google.com by `MathewT...@gmail.com` on 12 Nov 2010 at 5:02
1.0
Invincibility isn't invincible - ``` What steps will reproduce the problem? 1. Use the "Invincibility" drop item 2. Use "Kamikaze" 3. Profit What is the expected output? What do you see instead? I would expect the hedgehog to be invincible and therefore unable to be killed, yet the Kamikaze move still results in death, hence non-invincible. What version of the product are you using? On what operating system? The latest version, downloaded today. OS X 10.6.4. Please provide any additional information below. It resulted in me losing a round. :( ``` ----- Original issue reported on code.google.com by `MathewT...@gmail.com` on 12 Nov 2010 at 5:02
non_test
invincibility isn t invincible what steps will reproduce the problem use the invincibility drop item use kamikaze profit what is the expected output what do you see instead i would expect the hedgehog to be invincible and therefore unable to be killed yet the kamikaze move still results in death hence non invincible what version of the product are you using on what operating system the latest version downloaded today os x please provide any additional information below it resulted in me losing a round original issue reported on code google com by mathewt gmail com on nov at
0
785,569
27,618,471,036
IssuesEvent
2023-03-09 21:26:35
arfc/saltproc
https://api.github.com/repos/arfc/saltproc
opened
[Bug]: SaltProc does not store more than one array of `materials` results in `saltproc_results.h5` for OpenMC-coupled simulations
Type:Bug Status:4-In Progress Priority:1-Critical
## Expected behavior <!-- A clear and concise description of what you expected to happen. --> `saltproc_results.h5` should contain material compositions for before and after material reprocessing at every depletion step. ## Describe the bug <!-- A clear and concise description of what the bug is. --> For OpenMC runs, there appears to be an issue where there is only one array of material compositions stored. The other attributes (mass, burnup, etc) and neutronic information (beta-eff, lambda) all seem to be stored correctly. ## To Reproduce Steps to reproduce the behavior: 1. Decrese the simulation parameters of `msbr_openmc.json` and `settings.xml` in `examples/msbr` to something that will run quickly (I used 20 active, 8 inactice, 100 particles, 3 depletion steps) 2. Run the SaltProc simulation 3. Load the `saltproc_results.h5` file and observe. ## Supporting Informatioon <!-- If applicable, add program output with triple backticks, a gist of your output, or a screenshot to help explain your problem. --> ## System Specifications: <!-- At a minimum, include the SaltProc version and versions of software directly related to the bug--> - Software version: - Platform/OS: - Subsystem: ## How can this issue be closed? I can think of couple potential culprits: - Errors in the implementation of `_fix_nuclide_discrepancy` - Errors in the implementation of `read_depleted_materials` I'll need to do some interactive debugging to find the root cause. This issue can be closed when a PR exists that fixes the issue.
1.0
[Bug]: SaltProc does not store more than one array of `materials` results in `saltproc_results.h5` for OpenMC-coupled simulations - ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> `saltproc_results.h5` should contain material compositions for before and after material reprocessing at every depletion step. ## Describe the bug <!-- A clear and concise description of what the bug is. --> For OpenMC runs, there appears to be an issue where there is only one array of material compositions stored. The other attributes (mass, burnup, etc) and neutronic information (beta-eff, lambda) all seem to be stored correctly. ## To Reproduce Steps to reproduce the behavior: 1. Decrese the simulation parameters of `msbr_openmc.json` and `settings.xml` in `examples/msbr` to something that will run quickly (I used 20 active, 8 inactice, 100 particles, 3 depletion steps) 2. Run the SaltProc simulation 3. Load the `saltproc_results.h5` file and observe. ## Supporting Informatioon <!-- If applicable, add program output with triple backticks, a gist of your output, or a screenshot to help explain your problem. --> ## System Specifications: <!-- At a minimum, include the SaltProc version and versions of software directly related to the bug--> - Software version: - Platform/OS: - Subsystem: ## How can this issue be closed? I can think of couple potential culprits: - Errors in the implementation of `_fix_nuclide_discrepancy` - Errors in the implementation of `read_depleted_materials` I'll need to do some interactive debugging to find the root cause. This issue can be closed when a PR exists that fixes the issue.
non_test
saltproc does not store more than one array of materials results in saltproc results for openmc coupled simulations expected behavior saltproc results should contain material compositions for before and after material reprocessing at every depletion step describe the bug for openmc runs there appears to be an issue where there is only one array of material compositions stored the other attributes mass burnup etc and neutronic information beta eff lambda all seem to be stored correctly to reproduce steps to reproduce the behavior decrese the simulation parameters of msbr openmc json and settings xml in examples msbr to something that will run quickly i used active inactice particles depletion steps run the saltproc simulation load the saltproc results file and observe supporting informatioon system specifications software version platform os subsystem how can this issue be closed i can think of couple potential culprits errors in the implementation of fix nuclide discrepancy errors in the implementation of read depleted materials i ll need to do some interactive debugging to find the root cause this issue can be closed when a pr exists that fixes the issue
0
97,440
8,655,863,956
IssuesEvent
2018-11-27 16:54:46
FCP-INDI/C-PAC
https://api.github.com/repos/FCP-INDI/C-PAC
closed
Skull Stripping not customizable
testing
I noticed that we currently can't make custom adjustments to how skull-stripping is performed. This might be ok for some young adult datasets, but it really is a sticking point for working with child or older adult data, where for example, FSL's skull stripping parameters really need to be customized to work well/at all. Would it be possible to include a dropdown for choosing your own parameters?
1.0
Skull Stripping not customizable - I noticed that we currently can't make custom adjustments to how skull-stripping is performed. This might be ok for some young adult datasets, but it really is a sticking point for working with child or older adult data, where for example, FSL's skull stripping parameters really need to be customized to work well/at all. Would it be possible to include a dropdown for choosing your own parameters?
test
skull stripping not customizable i noticed that we currently can t make custom adjustments to how skull stripping is performed this might be ok for some young adult datasets but it really is a sticking point for working with child or older adult data where for example fsl s skull stripping parameters really need to be customized to work well at all would it be possible to include a dropdown for choosing your own parameters
1
152,161
12,093,499,356
IssuesEvent
2020-04-19 19:53:39
JoshKarpel/chtc-bot
https://api.github.com/repos/JoshKarpel/chtc-bot
closed
Add functional test for /knobs and /jobads
slash-commands testing
The unit tests are fine, but frequently miss problems. Add another test to each one's test suite which combines all of the unit test cases by calling get_*_descrption(). More end-to-end testing will have to wait for mocks, probably.
1.0
Add functional test for /knobs and /jobads - The unit tests are fine, but frequently miss problems. Add another test to each one's test suite which combines all of the unit test cases by calling get_*_descrption(). More end-to-end testing will have to wait for mocks, probably.
test
add functional test for knobs and jobads the unit tests are fine but frequently miss problems add another test to each one s test suite which combines all of the unit test cases by calling get descrption more end to end testing will have to wait for mocks probably
1
327,422
28,061,941,294
IssuesEvent
2023-03-29 13:10:42
pulp/pulp_rpm
https://api.github.com/repos/pulp/pulp_rpm
closed
Refactor `test_advisory_conflict` test module to use pytest fixtures
Task Tests
The `test_advisory_conflict` module needs to be refactored to use pytest fixtures.
1.0
Refactor `test_advisory_conflict` test module to use pytest fixtures - The `test_advisory_conflict` module needs to be refactored to use pytest fixtures.
test
refactor test advisory conflict test module to use pytest fixtures the test advisory conflict module needs to be refactored to use pytest fixtures
1
553,149
16,358,664,065
IssuesEvent
2021-05-14 05:25:19
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
opened
Support union types for Configurable
Priority/High Team/jBallerina Type/Improvement
**Description:** Need to support config enum values via toml and cli.
1.0
Support union types for Configurable - **Description:** Need to support config enum values via toml and cli.
non_test
support union types for configurable description need to support config enum values via toml and cli
0
133,727
10,861,008,291
IssuesEvent
2019-11-14 10:13:00
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
opened
`Send Tip` alignment issue in beta build 0.72.x
QA/Test-Plan-Specified QA/Yes bug feature/rewards
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> `Send Tip` alignment issue in beta build 0.72.x ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Clean profile staging 2. Enable rewards and claim grants 3. Open DDG 4. Open BR panel and click on `SEND A TIP` 5. `Send Tip` button isn't aligned with other Tip amount buttons ## Actual result: <!--Please add screenshots if needed--> `Send Tip` button isn't aligned with other Tip amount buttons ![image](https://user-images.githubusercontent.com/38657976/68846857-b7482680-06f3-11ea-8ed6-462393d705ba.png) ## Expected result: `Send Tip` button should be aligned with other Tip amount buttons ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Easy ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 0.72.124 Chromium: 78.0.3904.97ย (Official Build)ย betaย (64-bit) -- | -- Revision | 021b9028c246d820be17a10e5b393ee90f41375e-refs/branch-heads/3904@{#859} OS | Windowsย 10 OS Version 1803 (Build 17134.1006) ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? No - Can you reproduce this issue with the beta channel? Yes - Can you reproduce this issue with the dev channel? Yes - Can you reproduce this issue with the nightly channel? Yes ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? NA - Does the issue resolve itself when disabling Brave Rewards? NA - Is the issue reproducible on the latest version of Chrome? NA ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> cc: @NejcZdovc @brave/legacy_qa
1.0
`Send Tip` alignment issue in beta build 0.72.x - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> `Send Tip` alignment issue in beta build 0.72.x ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Clean profile staging 2. Enable rewards and claim grants 3. Open DDG 4. Open BR panel and click on `SEND A TIP` 5. `Send Tip` button isn't aligned with other Tip amount buttons ## Actual result: <!--Please add screenshots if needed--> `Send Tip` button isn't aligned with other Tip amount buttons ![image](https://user-images.githubusercontent.com/38657976/68846857-b7482680-06f3-11ea-8ed6-462393d705ba.png) ## Expected result: `Send Tip` button should be aligned with other Tip amount buttons ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Easy ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 0.72.124 Chromium: 78.0.3904.97ย (Official Build)ย betaย (64-bit) -- | -- Revision | 021b9028c246d820be17a10e5b393ee90f41375e-refs/branch-heads/3904@{#859} OS | Windowsย 10 OS Version 1803 (Build 17134.1006) ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? No - Can you reproduce this issue with the beta channel? Yes - Can you reproduce this issue with the dev channel? Yes - Can you reproduce this issue with the nightly channel? Yes ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? NA - Does the issue resolve itself when disabling Brave Rewards? NA - Is the issue reproducible on the latest version of Chrome? NA ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> cc: @NejcZdovc @brave/legacy_qa
test
send tip alignment issue in beta build x have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description send tip alignment issue in beta build x steps to reproduce clean profile staging enable rewards and claim grants open ddg open br panel and click on send a tip send tip button isn t aligned with other tip amount buttons actual result send tip button isn t aligned with other tip amount buttons expected result send tip button should be aligned with other tip amount buttons reproduces how often easy brave version brave version info brave chromium ย  official build ย betaย  bit revision refs branch heads os windowsย  os version build version channel information can you reproduce this issue with the current release no can you reproduce this issue with the beta channel yes can you reproduce this issue with the dev channel yes can you reproduce this issue with the nightly channel yes other additional information does the issue resolve itself when disabling brave shields na does the issue resolve itself when disabling brave rewards na is the issue reproducible on the latest version of chrome na miscellaneous information cc nejczdovc brave legacy qa
1
206,164
15,712,907,636
IssuesEvent
2021-03-27 14:06:16
SymbiFlow/sv-tests
https://api.github.com/repos/SymbiFlow/sv-tests
closed
Incorrect tests: forward typedefs
bug tests
Forwarding typedefs are required to resolve to a valid type declaration (see 6.18). Tests that I noticed have this problem: enum_test_0 struct_test_0 union_test_0 You might just want to make these expect to fail; other tools incorrectly allow this.
1.0
Incorrect tests: forward typedefs - Forwarding typedefs are required to resolve to a valid type declaration (see 6.18). Tests that I noticed have this problem: enum_test_0 struct_test_0 union_test_0 You might just want to make these expect to fail; other tools incorrectly allow this.
test
incorrect tests forward typedefs forwarding typedefs are required to resolve to a valid type declaration see tests that i noticed have this problem enum test struct test union test you might just want to make these expect to fail other tools incorrectly allow this
1
339,932
30,486,216,373
IssuesEvent
2023-07-18 02:34:35
ClickHouse/ClickHouse
https://api.github.com/repos/ClickHouse/ClickHouse
closed
Logical error: 'Different order of columns in UNION subquery: globalIn(CounterID, _subquery8), CounterID and CounterID'.
testing fuzz
**Describe the bug** https://s3.amazonaws.com/clickhouse-test-reports/42143/82850eb56046c857be8df56a8c425304b2670425/fuzzer_astfuzzerubsan//report.html **How to reproduce** ```sql CREATE TABLE globalin (CounterID UInt32, StartDate Date ) ENGINE = Memory; SELECT * FROM (SELECT -1, CounterID FROM remote('127.0.0.2', currentDatabase(), 'globalin') WHERE CounterID GLOBAL IN (SELECT toUInt32(34)) GROUP BY CounterID UNION ALL SELECT CounterID GLOBAL IN (SELECT NULL, toUInt32('65535')), CounterID FROM remote('127.0.0.2', currentDatabase(), 'globalin') WHERE CounterID GLOBAL IN (SELECT toUInt32(-9223372036854775808)) GROUP BY CounterID); ``` ``` [pc] 2022.10.18 14:10:22.642809 [ 38124 ] {980a9e79-fb2e-4229-9992-5cee2fc4bde9} <Fatal> : Logical error: 'Different order of columns in UNION subquery: globalIn(CounterID, _subquery8), CounterID and CounterID'. [pc] 2022.10.18 14:10:22.643610 [ 38390 ] <Fatal> BaseDaemon: ######################################## [pc] 2022.10.18 14:10:22.643894 [ 38390 ] <Fatal> BaseDaemon: (version 22.10.1.1, build id: F07FC74C69356811396CCE060D389408D3821E41) (from thread 38124) (query_id: 980a9e79-fb2e-4229-9992-5cee2fc4bde9) (query: SELECT * FROM (SELECT -1, CounterID FROM remote('127.0.0.2', currentDatabase(), 'globalin') WHERE CounterID GLOBAL IN (SELECT toUInt32(34)) GROUP BY CounterID UNION ALL SELECT CounterID GLOBAL IN (SELECT NULL, toUInt32('65535')), CounterID FROM remote('127.0.0.2', currentDatabase(), 'globalin') WHERE CounterID GLOBAL IN (SELECT toUInt32(-9223372036854775808)) GROUP BY CounterID)) Received signal Aborted (6) [pc] 2022.10.18 14:10:22.644130 [ 38390 ] <Fatal> BaseDaemon: [pc] 2022.10.18 14:10:22.644292 [ 38390 ] <Fatal> BaseDaemon: Stack trace: 0x7f49695fb64c 0x7f49695ab958 0x7f496959553d 0x213b4757 0x213b4815 0x213b48d2 0x29a9df87 0x29a9ada3 0x29a99463 0x29a20015 0x29a0eb75 0x29a06bc9 0x29a05bbe 0x29a9c658 0x29a9d23c 0x29f8b7ea 0x29f8779b 0x2b2c9c78 0x2b2dab52 0x305793b9 0x30579c28 0x307f5081 0x307f169a 0x307f0335 0x7f49695f98fd 0x7f496967ba60 [pc] 2022.10.18 14:10:22.644424 [ 38390 ] <Fatal> BaseDaemon: 4. ? @ 0x7f49695fb64c in ? [pc] 2022.10.18 14:10:22.644496 [ 38390 ] <Fatal> BaseDaemon: 5. gsignal @ 0x7f49695ab958 in ? [pc] 2022.10.18 14:10:22.644603 [ 38390 ] <Fatal> BaseDaemon: 6. abort @ 0x7f496959553d in ? [pc] 2022.10.18 14:10:22.721989 [ 38390 ] <Fatal> BaseDaemon: 7. /home/jakalletti/ClickHouse/ClickHouse/src/Common/Exception.cpp:40: DB::abortOnFailedAssertion(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&) @ 0x213b4757 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:22.797320 [ 38390 ] <Fatal> BaseDaemon: 8. /home/jakalletti/ClickHouse/ClickHouse/src/Common/Exception.cpp:63: DB::handle_error_code(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, int, bool, std::__1::vector<void*, std::__1::allocator<void*> > const&) @ 0x213b4815 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:22.862013 [ 38390 ] <Fatal> BaseDaemon: 9. /home/jakalletti/ClickHouse/ClickHouse/src/Common/Exception.cpp:70: DB::Exception::Exception(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, int, bool) @ 0x213b48d2 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:23.032600 [ 38390 ] <Fatal> BaseDaemon: 10. /home/jakalletti/ClickHouse/ClickHouse/src/Common/Exception.h:37: DB::Exception::Exception<fmt::v8::join_view<std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, char>, fmt::v8::join_view<std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, char> >(int, fmt::v8::basic_format_string<char, fmt::v8::type_identity<fmt::v8::join_view<std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, char> >::type, fmt::v8::type_identity<fmt::v8::join_view<std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, char> >::type>, fmt::v8::join_view<std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, char>&&, fmt::v8::join_view<std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, char>&&) @ 0x29a9df87 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:23.199942 [ 38390 ] <Fatal> BaseDaemon: 11. /home/jakalletti/ClickHouse/ClickHouse/src/Interpreters/InterpreterSelectWithUnionQuery.cpp:170: DB::InterpreterSelectWithUnionQuery::InterpreterSelectWithUnionQuery(std::__1::shared_ptr<DB::IAST> const&, std::__1::shared_ptr<DB::Context>, DB::SelectQueryOptions const&, std::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > > > const&) @ 0x29a9ada3 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:23.366898 [ 38390 ] <Fatal> BaseDaemon: 12. /home/jakalletti/ClickHouse/ClickHouse/src/Interpreters/InterpreterSelectWithUnionQuery.cpp:36: DB::InterpreterSelectWithUnionQuery::InterpreterSelectWithUnionQuery(std::__1::shared_ptr<DB::IAST> const&, std::__1::shared_ptr<DB::Context const>, DB::SelectQueryOptions const&, std::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > > > const&) @ 0x29a99463 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:23.878475 [ 38390 ] <Fatal> BaseDaemon: 13. /home/jakalletti/ClickHouse/ClickHouse/contrib/libcxx/include/__memory/unique_ptr.h:725: std::__1::__unique_if<DB::InterpreterSelectWithUnionQuery>::__unique_single std::__1::make_unique<DB::InterpreterSelectWithUnionQuery, std::__1::shared_ptr<DB::IAST>&, std::__1::shared_ptr<DB::Context const>, DB::SelectQueryOptions&, std::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > > >&>(std::__1::shared_ptr<DB::IAST>&, std::__1::shared_ptr<DB::Context const>&&, DB::SelectQueryOptions&, std::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > > >&) @ 0x29a20015 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:24.253812 [ 38390 ] <Fatal> BaseDaemon: 14. /home/jakalletti/ClickHouse/ClickHouse/src/Interpreters/InterpreterSelectQuery.cpp:2189: DB::InterpreterSelectQuery::executeFetchColumns(DB::QueryProcessingStage::Enum, DB::QueryPlan&) @ 0x29a0eb75 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:24.639299 [ 38390 ] <Fatal> BaseDaemon: 15. /home/jakalletti/ClickHouse/ClickHouse/src/Interpreters/InterpreterSelectQuery.cpp:0: DB::InterpreterSelectQuery::executeImpl(DB::QueryPlan&, std::__1::optional<DB::Pipe>) @ 0x29a06bc9 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:25.025454 [ 38390 ] <Fatal> BaseDaemon: 16. /home/jakalletti/ClickHouse/ClickHouse/src/Interpreters/InterpreterSelectQuery.cpp:723: DB::InterpreterSelectQuery::buildQueryPlan(DB::QueryPlan&) @ 0x29a05bbe in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:25.194901 [ 38390 ] <Fatal> BaseDaemon: 17. /home/jakalletti/ClickHouse/ClickHouse/src/Interpreters/InterpreterSelectWithUnionQuery.cpp:294: DB::InterpreterSelectWithUnionQuery::buildQueryPlan(DB::QueryPlan&) @ 0x29a9c658 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:25.364098 [ 38390 ] <Fatal> BaseDaemon: 18. /home/jakalletti/ClickHouse/ClickHouse/src/Interpreters/InterpreterSelectWithUnionQuery.cpp:368: DB::InterpreterSelectWithUnionQuery::execute() @ 0x29a9d23c in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:25.578080 [ 38390 ] <Fatal> BaseDaemon: 19. /home/jakalletti/ClickHouse/ClickHouse/src/Interpreters/executeQuery.cpp:709: DB::executeQueryImpl(char const*, char const*, std::__1::shared_ptr<DB::Context>, bool, DB::QueryProcessingStage::Enum, DB::ReadBuffer*) @ 0x29f8b7ea in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:25.807903 [ 38390 ] <Fatal> BaseDaemon: 20. /home/jakalletti/ClickHouse/ClickHouse/src/Interpreters/executeQuery.cpp:1105: DB::executeQuery(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, std::__1::shared_ptr<DB::Context>, bool, DB::QueryProcessingStage::Enum) @ 0x29f8779b in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:26.029811 [ 38390 ] <Fatal> BaseDaemon: 21. /home/jakalletti/ClickHouse/ClickHouse/src/Server/TCPHandler.cpp:374: DB::TCPHandler::runImpl() @ 0x2b2c9c78 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:26.272469 [ 38390 ] <Fatal> BaseDaemon: 22. /home/jakalletti/ClickHouse/ClickHouse/src/Server/TCPHandler.cpp:1899: DB::TCPHandler::run() @ 0x2b2dab52 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:26.328831 [ 38390 ] <Fatal> BaseDaemon: 23. /home/jakalletti/ClickHouse/ClickHouse/contrib/poco/Net/src/TCPServerConnection.cpp:43: Poco::Net::TCPServerConnection::start() @ 0x305793b9 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:26.395751 [ 38390 ] <Fatal> BaseDaemon: 24. /home/jakalletti/ClickHouse/ClickHouse/contrib/poco/Net/src/TCPServerDispatcher.cpp:115: Poco::Net::TCPServerDispatcher::run() @ 0x30579c28 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:26.465802 [ 38390 ] <Fatal> BaseDaemon: 25. /home/jakalletti/ClickHouse/ClickHouse/contrib/poco/Foundation/src/ThreadPool.cpp:199: Poco::PooledThread::run() @ 0x307f5081 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:26.535301 [ 38390 ] <Fatal> BaseDaemon: 26. /home/jakalletti/ClickHouse/ClickHouse/contrib/poco/Foundation/src/Thread.cpp:56: Poco::(anonymous namespace)::RunnableHolder::run() @ 0x307f169a in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:26.603992 [ 38390 ] <Fatal> BaseDaemon: 27. /home/jakalletti/ClickHouse/ClickHouse/contrib/poco/Foundation/src/Thread_POSIX.cpp:345: Poco::ThreadImpl::runnableEntry(void*) @ 0x307f0335 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:26.604118 [ 38390 ] <Fatal> BaseDaemon: 28. ? @ 0x7f49695f98fd in ? [pc] 2022.10.18 14:10:26.604221 [ 38390 ] <Fatal> BaseDaemon: 29. ? @ 0x7f496967ba60 in ? [pc] 2022.10.18 14:10:27.203586 [ 38390 ] <Fatal> BaseDaemon: Integrity check of the executable skipped because the reference checksum could not be read. (calculated checksum: 55D5CF69AA15DAFADACD284D5B99C435) Exception on client: Code: 32. DB::Exception: Attempt to read after eof: while receiving packet from localhost:9000. (ATTEMPT_TO_READ_AFTER_EOF) ```
1.0
Logical error: 'Different order of columns in UNION subquery: globalIn(CounterID, _subquery8), CounterID and CounterID'. - **Describe the bug** https://s3.amazonaws.com/clickhouse-test-reports/42143/82850eb56046c857be8df56a8c425304b2670425/fuzzer_astfuzzerubsan//report.html **How to reproduce** ```sql CREATE TABLE globalin (CounterID UInt32, StartDate Date ) ENGINE = Memory; SELECT * FROM (SELECT -1, CounterID FROM remote('127.0.0.2', currentDatabase(), 'globalin') WHERE CounterID GLOBAL IN (SELECT toUInt32(34)) GROUP BY CounterID UNION ALL SELECT CounterID GLOBAL IN (SELECT NULL, toUInt32('65535')), CounterID FROM remote('127.0.0.2', currentDatabase(), 'globalin') WHERE CounterID GLOBAL IN (SELECT toUInt32(-9223372036854775808)) GROUP BY CounterID); ``` ``` [pc] 2022.10.18 14:10:22.642809 [ 38124 ] {980a9e79-fb2e-4229-9992-5cee2fc4bde9} <Fatal> : Logical error: 'Different order of columns in UNION subquery: globalIn(CounterID, _subquery8), CounterID and CounterID'. [pc] 2022.10.18 14:10:22.643610 [ 38390 ] <Fatal> BaseDaemon: ######################################## [pc] 2022.10.18 14:10:22.643894 [ 38390 ] <Fatal> BaseDaemon: (version 22.10.1.1, build id: F07FC74C69356811396CCE060D389408D3821E41) (from thread 38124) (query_id: 980a9e79-fb2e-4229-9992-5cee2fc4bde9) (query: SELECT * FROM (SELECT -1, CounterID FROM remote('127.0.0.2', currentDatabase(), 'globalin') WHERE CounterID GLOBAL IN (SELECT toUInt32(34)) GROUP BY CounterID UNION ALL SELECT CounterID GLOBAL IN (SELECT NULL, toUInt32('65535')), CounterID FROM remote('127.0.0.2', currentDatabase(), 'globalin') WHERE CounterID GLOBAL IN (SELECT toUInt32(-9223372036854775808)) GROUP BY CounterID)) Received signal Aborted (6) [pc] 2022.10.18 14:10:22.644130 [ 38390 ] <Fatal> BaseDaemon: [pc] 2022.10.18 14:10:22.644292 [ 38390 ] <Fatal> BaseDaemon: Stack trace: 0x7f49695fb64c 0x7f49695ab958 0x7f496959553d 0x213b4757 0x213b4815 0x213b48d2 0x29a9df87 0x29a9ada3 0x29a99463 0x29a20015 0x29a0eb75 0x29a06bc9 0x29a05bbe 0x29a9c658 0x29a9d23c 0x29f8b7ea 0x29f8779b 0x2b2c9c78 0x2b2dab52 0x305793b9 0x30579c28 0x307f5081 0x307f169a 0x307f0335 0x7f49695f98fd 0x7f496967ba60 [pc] 2022.10.18 14:10:22.644424 [ 38390 ] <Fatal> BaseDaemon: 4. ? @ 0x7f49695fb64c in ? [pc] 2022.10.18 14:10:22.644496 [ 38390 ] <Fatal> BaseDaemon: 5. gsignal @ 0x7f49695ab958 in ? [pc] 2022.10.18 14:10:22.644603 [ 38390 ] <Fatal> BaseDaemon: 6. abort @ 0x7f496959553d in ? [pc] 2022.10.18 14:10:22.721989 [ 38390 ] <Fatal> BaseDaemon: 7. /home/jakalletti/ClickHouse/ClickHouse/src/Common/Exception.cpp:40: DB::abortOnFailedAssertion(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&) @ 0x213b4757 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:22.797320 [ 38390 ] <Fatal> BaseDaemon: 8. /home/jakalletti/ClickHouse/ClickHouse/src/Common/Exception.cpp:63: DB::handle_error_code(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, int, bool, std::__1::vector<void*, std::__1::allocator<void*> > const&) @ 0x213b4815 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:22.862013 [ 38390 ] <Fatal> BaseDaemon: 9. /home/jakalletti/ClickHouse/ClickHouse/src/Common/Exception.cpp:70: DB::Exception::Exception(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, int, bool) @ 0x213b48d2 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:23.032600 [ 38390 ] <Fatal> BaseDaemon: 10. /home/jakalletti/ClickHouse/ClickHouse/src/Common/Exception.h:37: DB::Exception::Exception<fmt::v8::join_view<std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, char>, fmt::v8::join_view<std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, char> >(int, fmt::v8::basic_format_string<char, fmt::v8::type_identity<fmt::v8::join_view<std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, char> >::type, fmt::v8::type_identity<fmt::v8::join_view<std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, char> >::type>, fmt::v8::join_view<std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, char>&&, fmt::v8::join_view<std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, std::__1::__wrap_iter<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const*>, char>&&) @ 0x29a9df87 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:23.199942 [ 38390 ] <Fatal> BaseDaemon: 11. /home/jakalletti/ClickHouse/ClickHouse/src/Interpreters/InterpreterSelectWithUnionQuery.cpp:170: DB::InterpreterSelectWithUnionQuery::InterpreterSelectWithUnionQuery(std::__1::shared_ptr<DB::IAST> const&, std::__1::shared_ptr<DB::Context>, DB::SelectQueryOptions const&, std::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > > > const&) @ 0x29a9ada3 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:23.366898 [ 38390 ] <Fatal> BaseDaemon: 12. /home/jakalletti/ClickHouse/ClickHouse/src/Interpreters/InterpreterSelectWithUnionQuery.cpp:36: DB::InterpreterSelectWithUnionQuery::InterpreterSelectWithUnionQuery(std::__1::shared_ptr<DB::IAST> const&, std::__1::shared_ptr<DB::Context const>, DB::SelectQueryOptions const&, std::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > > > const&) @ 0x29a99463 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:23.878475 [ 38390 ] <Fatal> BaseDaemon: 13. /home/jakalletti/ClickHouse/ClickHouse/contrib/libcxx/include/__memory/unique_ptr.h:725: std::__1::__unique_if<DB::InterpreterSelectWithUnionQuery>::__unique_single std::__1::make_unique<DB::InterpreterSelectWithUnionQuery, std::__1::shared_ptr<DB::IAST>&, std::__1::shared_ptr<DB::Context const>, DB::SelectQueryOptions&, std::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > > >&>(std::__1::shared_ptr<DB::IAST>&, std::__1::shared_ptr<DB::Context const>&&, DB::SelectQueryOptions&, std::__1::vector<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, std::__1::allocator<std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > > >&) @ 0x29a20015 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:24.253812 [ 38390 ] <Fatal> BaseDaemon: 14. /home/jakalletti/ClickHouse/ClickHouse/src/Interpreters/InterpreterSelectQuery.cpp:2189: DB::InterpreterSelectQuery::executeFetchColumns(DB::QueryProcessingStage::Enum, DB::QueryPlan&) @ 0x29a0eb75 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:24.639299 [ 38390 ] <Fatal> BaseDaemon: 15. /home/jakalletti/ClickHouse/ClickHouse/src/Interpreters/InterpreterSelectQuery.cpp:0: DB::InterpreterSelectQuery::executeImpl(DB::QueryPlan&, std::__1::optional<DB::Pipe>) @ 0x29a06bc9 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:25.025454 [ 38390 ] <Fatal> BaseDaemon: 16. /home/jakalletti/ClickHouse/ClickHouse/src/Interpreters/InterpreterSelectQuery.cpp:723: DB::InterpreterSelectQuery::buildQueryPlan(DB::QueryPlan&) @ 0x29a05bbe in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:25.194901 [ 38390 ] <Fatal> BaseDaemon: 17. /home/jakalletti/ClickHouse/ClickHouse/src/Interpreters/InterpreterSelectWithUnionQuery.cpp:294: DB::InterpreterSelectWithUnionQuery::buildQueryPlan(DB::QueryPlan&) @ 0x29a9c658 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:25.364098 [ 38390 ] <Fatal> BaseDaemon: 18. /home/jakalletti/ClickHouse/ClickHouse/src/Interpreters/InterpreterSelectWithUnionQuery.cpp:368: DB::InterpreterSelectWithUnionQuery::execute() @ 0x29a9d23c in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:25.578080 [ 38390 ] <Fatal> BaseDaemon: 19. /home/jakalletti/ClickHouse/ClickHouse/src/Interpreters/executeQuery.cpp:709: DB::executeQueryImpl(char const*, char const*, std::__1::shared_ptr<DB::Context>, bool, DB::QueryProcessingStage::Enum, DB::ReadBuffer*) @ 0x29f8b7ea in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:25.807903 [ 38390 ] <Fatal> BaseDaemon: 20. /home/jakalletti/ClickHouse/ClickHouse/src/Interpreters/executeQuery.cpp:1105: DB::executeQuery(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, std::__1::shared_ptr<DB::Context>, bool, DB::QueryProcessingStage::Enum) @ 0x29f8779b in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:26.029811 [ 38390 ] <Fatal> BaseDaemon: 21. /home/jakalletti/ClickHouse/ClickHouse/src/Server/TCPHandler.cpp:374: DB::TCPHandler::runImpl() @ 0x2b2c9c78 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:26.272469 [ 38390 ] <Fatal> BaseDaemon: 22. /home/jakalletti/ClickHouse/ClickHouse/src/Server/TCPHandler.cpp:1899: DB::TCPHandler::run() @ 0x2b2dab52 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:26.328831 [ 38390 ] <Fatal> BaseDaemon: 23. /home/jakalletti/ClickHouse/ClickHouse/contrib/poco/Net/src/TCPServerConnection.cpp:43: Poco::Net::TCPServerConnection::start() @ 0x305793b9 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:26.395751 [ 38390 ] <Fatal> BaseDaemon: 24. /home/jakalletti/ClickHouse/ClickHouse/contrib/poco/Net/src/TCPServerDispatcher.cpp:115: Poco::Net::TCPServerDispatcher::run() @ 0x30579c28 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:26.465802 [ 38390 ] <Fatal> BaseDaemon: 25. /home/jakalletti/ClickHouse/ClickHouse/contrib/poco/Foundation/src/ThreadPool.cpp:199: Poco::PooledThread::run() @ 0x307f5081 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:26.535301 [ 38390 ] <Fatal> BaseDaemon: 26. /home/jakalletti/ClickHouse/ClickHouse/contrib/poco/Foundation/src/Thread.cpp:56: Poco::(anonymous namespace)::RunnableHolder::run() @ 0x307f169a in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:26.603992 [ 38390 ] <Fatal> BaseDaemon: 27. /home/jakalletti/ClickHouse/ClickHouse/contrib/poco/Foundation/src/Thread_POSIX.cpp:345: Poco::ThreadImpl::runnableEntry(void*) @ 0x307f0335 in /home/jakalletti/ClickHouse/build/programs/clickhouse [pc] 2022.10.18 14:10:26.604118 [ 38390 ] <Fatal> BaseDaemon: 28. ? @ 0x7f49695f98fd in ? [pc] 2022.10.18 14:10:26.604221 [ 38390 ] <Fatal> BaseDaemon: 29. ? @ 0x7f496967ba60 in ? [pc] 2022.10.18 14:10:27.203586 [ 38390 ] <Fatal> BaseDaemon: Integrity check of the executable skipped because the reference checksum could not be read. (calculated checksum: 55D5CF69AA15DAFADACD284D5B99C435) Exception on client: Code: 32. DB::Exception: Attempt to read after eof: while receiving packet from localhost:9000. (ATTEMPT_TO_READ_AFTER_EOF) ```
test
logical error different order of columns in union subquery globalin counterid counterid and counterid describe the bug how to reproduce sql create table globalin counterid startdate date engine memory select from select counterid from remote currentdatabase globalin where counterid global in select group by counterid union all select counterid global in select null counterid from remote currentdatabase globalin where counterid global in select group by counterid logical error different order of columns in union subquery globalin counterid counterid and counterid basedaemon basedaemon version build id from thread query id query select from select counterid from remote currentdatabase globalin where counterid global in select group by counterid union all select counterid global in select null counterid from remote currentdatabase globalin where counterid global in select group by counterid received signal aborted basedaemon basedaemon stack trace basedaemon in basedaemon gsignal in basedaemon abort in basedaemon home jakalletti clickhouse clickhouse src common exception cpp db abortonfailedassertion std basic string std allocator const in home jakalletti clickhouse build programs clickhouse basedaemon home jakalletti clickhouse clickhouse src common exception cpp db handle error code std basic string std allocator const int bool std vector const in home jakalletti clickhouse build programs clickhouse basedaemon home jakalletti clickhouse clickhouse src common exception cpp db exception exception std basic string std allocator const int bool in home jakalletti clickhouse build programs clickhouse basedaemon home jakalletti clickhouse clickhouse src common exception h db exception exception std allocator const std wrap iter std allocator const char fmt join view std allocator const std wrap iter std allocator const char int fmt basic format string std allocator const std wrap iter std allocator const char type fmt type identity std allocator const std wrap iter std allocator const char type fmt join view std allocator const std wrap iter std allocator const char fmt join view std allocator const std wrap iter std allocator const char in home jakalletti clickhouse build programs clickhouse basedaemon home jakalletti clickhouse clickhouse src interpreters interpreterselectwithunionquery cpp db interpreterselectwithunionquery interpreterselectwithunionquery std shared ptr const std shared ptr db selectqueryoptions const std vector std allocator std allocator std allocator const in home jakalletti clickhouse build programs clickhouse basedaemon home jakalletti clickhouse clickhouse src interpreters interpreterselectwithunionquery cpp db interpreterselectwithunionquery interpreterselectwithunionquery std shared ptr const std shared ptr db selectqueryoptions const std vector std allocator std allocator std allocator const in home jakalletti clickhouse build programs clickhouse basedaemon home jakalletti clickhouse clickhouse contrib libcxx include memory unique ptr h std unique if unique single std make unique std shared ptr db selectqueryoptions std vector std allocator std allocator std allocator std shared ptr std shared ptr db selectqueryoptions std vector std allocator std allocator std allocator in home jakalletti clickhouse build programs clickhouse basedaemon home jakalletti clickhouse clickhouse src interpreters interpreterselectquery cpp db interpreterselectquery executefetchcolumns db queryprocessingstage enum db queryplan in home jakalletti clickhouse build programs clickhouse basedaemon home jakalletti clickhouse clickhouse src interpreters interpreterselectquery cpp db interpreterselectquery executeimpl db queryplan std optional in home jakalletti clickhouse build programs clickhouse basedaemon home jakalletti clickhouse clickhouse src interpreters interpreterselectquery cpp db interpreterselectquery buildqueryplan db queryplan in home jakalletti clickhouse build programs clickhouse basedaemon home jakalletti clickhouse clickhouse src interpreters interpreterselectwithunionquery cpp db interpreterselectwithunionquery buildqueryplan db queryplan in home jakalletti clickhouse build programs clickhouse basedaemon home jakalletti clickhouse clickhouse src interpreters interpreterselectwithunionquery cpp db interpreterselectwithunionquery execute in home jakalletti clickhouse build programs clickhouse basedaemon home jakalletti clickhouse clickhouse src interpreters executequery cpp db executequeryimpl char const char const std shared ptr bool db queryprocessingstage enum db readbuffer in home jakalletti clickhouse build programs clickhouse basedaemon home jakalletti clickhouse clickhouse src interpreters executequery cpp db executequery std basic string std allocator const std shared ptr bool db queryprocessingstage enum in home jakalletti clickhouse build programs clickhouse basedaemon home jakalletti clickhouse clickhouse src server tcphandler cpp db tcphandler runimpl in home jakalletti clickhouse build programs clickhouse basedaemon home jakalletti clickhouse clickhouse src server tcphandler cpp db tcphandler run in home jakalletti clickhouse build programs clickhouse basedaemon home jakalletti clickhouse clickhouse contrib poco net src tcpserverconnection cpp poco net tcpserverconnection start in home jakalletti clickhouse build programs clickhouse basedaemon home jakalletti clickhouse clickhouse contrib poco net src tcpserverdispatcher cpp poco net tcpserverdispatcher run in home jakalletti clickhouse build programs clickhouse basedaemon home jakalletti clickhouse clickhouse contrib poco foundation src threadpool cpp poco pooledthread run in home jakalletti clickhouse build programs clickhouse basedaemon home jakalletti clickhouse clickhouse contrib poco foundation src thread cpp poco anonymous namespace runnableholder run in home jakalletti clickhouse build programs clickhouse basedaemon home jakalletti clickhouse clickhouse contrib poco foundation src thread posix cpp poco threadimpl runnableentry void in home jakalletti clickhouse build programs clickhouse basedaemon in basedaemon in basedaemon integrity check of the executable skipped because the reference checksum could not be read calculated checksum exception on client code db exception attempt to read after eof while receiving packet from localhost attempt to read after eof
1
179,818
13,905,342,655
IssuesEvent
2020-10-20 09:45:15
imixs/imixs-saga
https://api.github.com/repos/imixs/imixs-saga
closed
Drop BatchEvents replaced by AsyncEvents
Testing feature
Drop BatchEvents as they are replaced by AsyncEvents in Imixs-Workflow v5.2
1.0
Drop BatchEvents replaced by AsyncEvents - Drop BatchEvents as they are replaced by AsyncEvents in Imixs-Workflow v5.2
test
drop batchevents replaced by asyncevents drop batchevents as they are replaced by asyncevents in imixs workflow
1
10,110
7,090,836,397
IssuesEvent
2018-01-12 10:31:28
sonarwhal/sonarwhal
https://api.github.com/repos/sonarwhal/sonarwhal
closed
Improve sniffing detection for the `http-compression` rule
area:rule enhacement rule-category:performance
When using the `jsdom` connector `google.com` is sniffing us quite well in all requests resulting in the `Should be served compressed with Brotli over HTTPS.` error, when that shouldn't be the case. Using the `Chrome` connector `sonarwhal` actually catches the sniffing, giving us the: `Should be served compressed with Brotli over HTTPS regardless of the user agent. error.
True
Improve sniffing detection for the `http-compression` rule - When using the `jsdom` connector `google.com` is sniffing us quite well in all requests resulting in the `Should be served compressed with Brotli over HTTPS.` error, when that shouldn't be the case. Using the `Chrome` connector `sonarwhal` actually catches the sniffing, giving us the: `Should be served compressed with Brotli over HTTPS regardless of the user agent. error.
non_test
improve sniffing detection for the http compression rule when using the jsdom connector google com is sniffing us quite well in all requests resulting in the should be served compressed with brotli over https error when that shouldn t be the case using the chrome connector sonarwhal actually catches the sniffing giving us the should be served compressed with brotli over https regardless of the user agent error
0
11,564
3,506,306,904
IssuesEvent
2016-01-08 05:33:38
RBMHTechnology/eventuate
https://api.github.com/repos/RBMHTechnology/eventuate
opened
Provide documentation for each release.
documentation
At the moment, there is only [documentation for the current development snapshot](http://rbmhtechnology.github.io/eventuate/). Additionally, there should be documentation for each released version separately.
1.0
Provide documentation for each release. - At the moment, there is only [documentation for the current development snapshot](http://rbmhtechnology.github.io/eventuate/). Additionally, there should be documentation for each released version separately.
non_test
provide documentation for each release at the moment there is only additionally there should be documentation for each released version separately
0
389,957
11,519,886,761
IssuesEvent
2020-02-14 13:46:37
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
listen.tidal.com - Firefox is unsupported browser
browser-firefox engine-gecko form-v2-experiment priority-normal severity-critical type-unsupported
<!-- @browser: Firefox 72.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:72.0) Gecko/20100101 Firefox/72.0 --> <!-- @reported_with: --> <!-- @extra_labels: form-v2-experiment --> **URL**: https://listen.tidal.com/ **Browser / Version**: Firefox 72.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Something else **Description**: Widevine not recognized **Steps to Reproduce**: Firefox has widevine as netflix plays proeprly but Tidal says there's no widevine so can't play HiFi. Works in Chrome though. [Screenshot](https://webcompat.com/uploads/2020/1/f740f5cc-63d2-46dd-8ca2-4e7b93281971.jpg) <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with โค๏ธ_
1.0
listen.tidal.com - Firefox is unsupported browser - <!-- @browser: Firefox 72.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:72.0) Gecko/20100101 Firefox/72.0 --> <!-- @reported_with: --> <!-- @extra_labels: form-v2-experiment --> **URL**: https://listen.tidal.com/ **Browser / Version**: Firefox 72.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Something else **Description**: Widevine not recognized **Steps to Reproduce**: Firefox has widevine as netflix plays proeprly but Tidal says there's no widevine so can't play HiFi. Works in Chrome though. [Screenshot](https://webcompat.com/uploads/2020/1/f740f5cc-63d2-46dd-8ca2-4e7b93281971.jpg) <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with โค๏ธ_
non_test
listen tidal com firefox is unsupported browser url browser version firefox operating system windows tested another browser yes chrome problem type something else description widevine not recognized steps to reproduce firefox has widevine as netflix plays proeprly but tidal says there s no widevine so can t play hifi works in chrome though browser configuration none from with โค๏ธ
0
9,222
2,615,140,046
IssuesEvent
2015-03-01 06:14:18
chrsmith/reaver-wps
https://api.github.com/repos/chrsmith/reaver-wps
closed
Cisco E3200 AP timeouts, out of order packets, failed attack.
auto-migrated Priority-Low Type-Defect
``` What steps will reproduce the problem? Run reaver against a Cisco/Linksys E3200 AP What is the expected output? What do you see instead? Expect a successful attack resulting in pin disclosure. Seeing receive timeouts, message processing errors, out of order packets, no pin change, retransmits. Product and OS version info: Backtrack 5 Linux 2.6.39.4 #1 SMP Thu Aug 18 13:38:02 NZST 2011 i686 GNU/Linux wlan0: 03:00.0 Network controller: Intel Corporation Centrino Advanced-N 6200 (rev 35) Cisco E3200 AP Firmware 1.0.02 root@root:~/reaver-wps-read-only/src# svn info Path: . URL: http://reaver-wps.googlecode.com/svn/trunk/src Repository Root: http://reaver-wps.googlecode.com/svn Repository UUID: 027a3e96-2d37-f1c0-85d6-5ce5a08386c2 Revision: 37 Node Kind: directory Schedule: normal Last Changed Author: cheffner@tacnetsol.com Last Changed Rev: 37 Last Changed Date: 2012-01-02 10:30:32 -0500 (Mon, 02 Jan 2012) iwconfig wlan0 mode monitor root@root:~/reaver-wps-read-only/src# ./reaver -i wlan0 -b 58:6d:8f:07:62:0d -c 11 -vv > /root/cisco-e3200-reaverlog.txt [+] Waiting for beacon from 58:6D:8F:07:62:0D [+] Switching wlan0 to channel 11 [!] WARNING: Failed to associate with 58:6D:8F:07:62:0D (ESSID: Cisco47272) [!] WARNING: Failed to associate with 58:6D:8F:07:62:0D (ESSID: Cisco47272) [+] Associated with 58:6D:8F:07:62:0D (ESSID: Cisco47272) [+] Trying pin 59631507 [!] WARNING: Receive timeout occurred [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [+] Trying pin 59631507 [!] WARNING: Receive timeout occurred [!] WARNING: Receive timeout occurred [!] WARNING: Receive timeout occurred [!] WARNING: Receive timeout occurred ``` Original issue reported on code.google.com by `pah...@gmail.com` on 2 Jan 2012 at 5:06 Attachments: * [reaver-wps-r37-cisco-e3200-eap.cap](https://storage.googleapis.com/google-code-attachments/reaver-wps/issue-53/comment-0/reaver-wps-r37-cisco-e3200-eap.cap)
1.0
Cisco E3200 AP timeouts, out of order packets, failed attack. - ``` What steps will reproduce the problem? Run reaver against a Cisco/Linksys E3200 AP What is the expected output? What do you see instead? Expect a successful attack resulting in pin disclosure. Seeing receive timeouts, message processing errors, out of order packets, no pin change, retransmits. Product and OS version info: Backtrack 5 Linux 2.6.39.4 #1 SMP Thu Aug 18 13:38:02 NZST 2011 i686 GNU/Linux wlan0: 03:00.0 Network controller: Intel Corporation Centrino Advanced-N 6200 (rev 35) Cisco E3200 AP Firmware 1.0.02 root@root:~/reaver-wps-read-only/src# svn info Path: . URL: http://reaver-wps.googlecode.com/svn/trunk/src Repository Root: http://reaver-wps.googlecode.com/svn Repository UUID: 027a3e96-2d37-f1c0-85d6-5ce5a08386c2 Revision: 37 Node Kind: directory Schedule: normal Last Changed Author: cheffner@tacnetsol.com Last Changed Rev: 37 Last Changed Date: 2012-01-02 10:30:32 -0500 (Mon, 02 Jan 2012) iwconfig wlan0 mode monitor root@root:~/reaver-wps-read-only/src# ./reaver -i wlan0 -b 58:6d:8f:07:62:0d -c 11 -vv > /root/cisco-e3200-reaverlog.txt [+] Waiting for beacon from 58:6D:8F:07:62:0D [+] Switching wlan0 to channel 11 [!] WARNING: Failed to associate with 58:6D:8F:07:62:0D (ESSID: Cisco47272) [!] WARNING: Failed to associate with 58:6D:8F:07:62:0D (ESSID: Cisco47272) [+] Associated with 58:6D:8F:07:62:0D (ESSID: Cisco47272) [+] Trying pin 59631507 [!] WARNING: Receive timeout occurred [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [!] WARNING: Last message not processed properly, reverting state to previous message [!] WARNING: Out of order packet received, re-trasmitting last message [+] Trying pin 59631507 [!] WARNING: Receive timeout occurred [!] WARNING: Receive timeout occurred [!] WARNING: Receive timeout occurred [!] WARNING: Receive timeout occurred ``` Original issue reported on code.google.com by `pah...@gmail.com` on 2 Jan 2012 at 5:06 Attachments: * [reaver-wps-r37-cisco-e3200-eap.cap](https://storage.googleapis.com/google-code-attachments/reaver-wps/issue-53/comment-0/reaver-wps-r37-cisco-e3200-eap.cap)
non_test
cisco ap timeouts out of order packets failed attack what steps will reproduce the problem run reaver against a cisco linksys ap what is the expected output what do you see instead expect a successful attack resulting in pin disclosure seeing receive timeouts message processing errors out of order packets no pin change retransmits product and os version info backtrack linux smp thu aug nzst gnu linux network controller intel corporation centrino advanced n rev cisco ap firmware root root reaver wps read only src svn info path url repository root repository uuid revision node kind directory schedule normal last changed author cheffner tacnetsol com last changed rev last changed date mon jan iwconfig mode monitor root root reaver wps read only src reaver i b c vv root cisco reaverlog txt waiting for beacon from switching to channel warning failed to associate with essid warning failed to associate with essid associated with essid trying pin warning receive timeout occurred warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message warning last message not processed properly reverting state to previous message warning out of order packet received re trasmitting last message trying pin warning receive timeout occurred warning receive timeout occurred warning receive timeout occurred warning receive timeout occurred original issue reported on code google com by pah gmail com on jan at attachments
0
150,406
11,959,248,925
IssuesEvent
2020-04-04 21:12:16
Oldes/Rebol-issues
https://api.github.com/repos/Oldes/Rebol-issues
closed
MOLD of unsigned vector doesn't show 'unsigned in result
Datatype: vector! Test.written Type.bug
_Submitted by:_ **BrianH** When you MOLD a vector of unsigned values, the word 'unsigned should appear in the result so we can better understand its behavior. ``` rebol ; With unsigned vector >> v: make vector! [unsigned 32 20] v/1: -1 v == make vector! [integer! 32 20 [ ; should be unsigned integer! 4294967295 0 0 0 0 0 0 0 ; ...and this is why 0 0 0 0 0 0 0 0 0 0 0 0 ]] ; With signed integer >> v: make vector! [32 20] v/1: -1 v == make vector! [integer! 32 20 [ -1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ]] ``` --- <sup>**Imported from:** **[CureCode](https://www.curecode.org/rebol3/ticket.rsp?id=756)** [ Version: alpha 51 Type: Bug Platform: All Category: n/a Reproduce: Always Fixed-in:alpha 55 ]</sup> <sup>**Imported from**: https://github.com/rebol/rebol-issues/issues/756</sup> Comments: --- --- > **Rebolbot** added the **Type.bug** on Jan 12, 2016 --- > **Oldes** commented on Jan 22, 2020: I cannot reproduce it: ``` >> v: make vector! [unsigned 32 20] v/1: -1 v == make vector! [unsigned integer! 32 20 [ 4294967295 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ]] >> v: make vector! [32 20] v/1: -1 v == make vector! [integer! 32 20 [ -1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ]] ``` ---
1.0
MOLD of unsigned vector doesn't show 'unsigned in result - _Submitted by:_ **BrianH** When you MOLD a vector of unsigned values, the word 'unsigned should appear in the result so we can better understand its behavior. ``` rebol ; With unsigned vector >> v: make vector! [unsigned 32 20] v/1: -1 v == make vector! [integer! 32 20 [ ; should be unsigned integer! 4294967295 0 0 0 0 0 0 0 ; ...and this is why 0 0 0 0 0 0 0 0 0 0 0 0 ]] ; With signed integer >> v: make vector! [32 20] v/1: -1 v == make vector! [integer! 32 20 [ -1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ]] ``` --- <sup>**Imported from:** **[CureCode](https://www.curecode.org/rebol3/ticket.rsp?id=756)** [ Version: alpha 51 Type: Bug Platform: All Category: n/a Reproduce: Always Fixed-in:alpha 55 ]</sup> <sup>**Imported from**: https://github.com/rebol/rebol-issues/issues/756</sup> Comments: --- --- > **Rebolbot** added the **Type.bug** on Jan 12, 2016 --- > **Oldes** commented on Jan 22, 2020: I cannot reproduce it: ``` >> v: make vector! [unsigned 32 20] v/1: -1 v == make vector! [unsigned integer! 32 20 [ 4294967295 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ]] >> v: make vector! [32 20] v/1: -1 v == make vector! [integer! 32 20 [ -1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ]] ``` ---
test
mold of unsigned vector doesn t show unsigned in result submitted by brianh when you mold a vector of unsigned values the word unsigned should appear in the result so we can better understand its behavior rebol with unsigned vector v make vector v v make vector integer should be unsigned integer and this is why with signed integer v make vector v v make vector integer imported from imported from comments rebolbot added the type bug on jan oldes commented on jan i cannot reproduce it v make vector v v make vector unsigned integer v make vector v v make vector integer
1
746,715
26,042,712,329
IssuesEvent
2022-12-22 11:51:19
aiidateam/aiida-core
https://api.github.com/repos/aiidateam/aiida-core
closed
`verdi code delete` falsely claims it actually deleted a node
type/bug priority/critical-blocking topic/storage
It claims that the code is deleted, but it is not actually the case. Seems that the problem is with the `PsqlDosBackend.delete_nodes_and_connections` that does perform the deletion but changes are not persisted. This is most likely related to recent refactorings of the `PsqlDosBackend.transaction` method.
1.0
`verdi code delete` falsely claims it actually deleted a node - It claims that the code is deleted, but it is not actually the case. Seems that the problem is with the `PsqlDosBackend.delete_nodes_and_connections` that does perform the deletion but changes are not persisted. This is most likely related to recent refactorings of the `PsqlDosBackend.transaction` method.
non_test
verdi code delete falsely claims it actually deleted a node it claims that the code is deleted but it is not actually the case seems that the problem is with the psqldosbackend delete nodes and connections that does perform the deletion but changes are not persisted this is most likely related to recent refactorings of the psqldosbackend transaction method
0
286,492
24,755,651,198
IssuesEvent
2022-10-21 17:27:20
Thy-Vipe/BeastsOfBermuda-issues
https://api.github.com/repos/Thy-Vipe/BeastsOfBermuda-issues
opened
[Bug] Para Alarm Call Bug
bug Gameplay public_testing
_Originally written by **Priuloch | 76561198144474637**_ Game Version: 1.1.1615 *===== System Specs ===== CPU Brand: Intel(R) Core(TM) i7-6700HQ CPU @ 2.60GHz Vendor: GenuineIntel GPU Brand: NVIDIA GeForce GTX 970M GPU Driver Info: Unknown Num CPU Cores: 4 ===================* Context: **Para** Map: Islatitania Hold Offspring Alarm call on enemy (trope was called, if this matters) Alarm call works with Offspring in mouth.
1.0
[Bug] Para Alarm Call Bug - _Originally written by **Priuloch | 76561198144474637**_ Game Version: 1.1.1615 *===== System Specs ===== CPU Brand: Intel(R) Core(TM) i7-6700HQ CPU @ 2.60GHz Vendor: GenuineIntel GPU Brand: NVIDIA GeForce GTX 970M GPU Driver Info: Unknown Num CPU Cores: 4 ===================* Context: **Para** Map: Islatitania Hold Offspring Alarm call on enemy (trope was called, if this matters) Alarm call works with Offspring in mouth.
test
para alarm call bug originally written by priuloch game version system specs cpu brand intel r core tm cpu vendor genuineintel gpu brand nvidia geforce gtx gpu driver info unknown num cpu cores context para map islatitania hold offspring alarm call on enemy trope was called if this matters alarm call works with offspring in mouth
1
184,343
14,288,760,619
IssuesEvent
2020-11-23 18:11:04
freqtrade/freqtrade
https://api.github.com/repos/freqtrade/freqtrade
closed
Very slow start for backtesting / hyperopt
Backtest Question
## Describe your environment * Operating system: WSL2 on Windows 10 20H2 ( using freqtrade docker image ) * Python Version: image's python * CCXT version: image's CCXT * Freqtrade Version: 2020.10 ## Describe the problem: Starting hyperopt or backtesting takes around 10 minutes ! It was okay until some time ago , it may have something to do with me migrating to WSL2 from my windows10 classic env Note : look at the timestamps of the last 10 lines of the logs, we see 10+ m on one step Also my CPU is sitting at around 30% during those 10m , so it's not a compute power bottleneck ^^" And the data is stored on a SSD , and i got way enough RAM ### Steps to reproduce: idk if you can reproduce it , all i have to do is start backtesting. ### Observed Results: * What happened? * What did you expect to happen? ### Relevant code exceptions or logs Note: Please copy/paste text of the messages, no screenshots of logs please. ``` 2020-11-22 17:36:17,513 - freqtrade.configuration.configuration - INFO - Using config: user_data/config.json ... 2020-11-22 17:36:17,514 - freqtrade.loggers - INFO - Verbosity set to 0 2020-11-22 17:36:17,514 - freqtrade.configuration.configuration - INFO - Using max_open_trades: 10 ... 2020-11-22 17:36:17,514 - freqtrade.configuration.configuration - INFO - Using user-data directory: /freqtrade/user_data ... 2020-11-22 17:36:17,515 - freqtrade.configuration.configuration - INFO - Using data directory: /freqtrade/user_data/data/binance ... 2020-11-22 17:36:17,515 - freqtrade.configuration.configuration - INFO - Using Hyperopt class name: A_STAR_HYPEROPT 2020-11-22 17:36:17,515 - freqtrade.configuration.configuration - INFO - Parameter --epochs detected ... Will run Hyperopt with for 1000 epochs ... 2020-11-22 17:36:17,515 - freqtrade.configuration.configuration - INFO - Parameter -s/--spaces detected: ['sell'] 2020-11-22 17:36:17,515 - freqtrade.configuration.configuration - INFO - Parameter -j/--job-workers detected: -1 2020-11-22 17:36:17,515 - freqtrade.configuration.configuration - INFO - Parameter --min-trades detected: 200 2020-11-22 17:36:17,515 - freqtrade.configuration.configuration - INFO - Using Hyperopt loss class name: ShortTradeDurHyperOptLoss 2020-11-22 17:36:17,515 - freqtrade.configuration.configuration - INFO - Using "json" to store OHLCV data. 2020-11-22 17:36:17,515 - freqtrade.configuration.check_exchange - INFO - Checking exchange... 2020-11-22 17:36:17,515 - freqtrade.configuration.check_exchange - INFO - Exchange "binance" is officially supported by the Freqtrade development team. 2020-11-22 17:36:17,515 - freqtrade.configuration.configuration - INFO - Using pairlist from configuration. 2020-11-22 17:36:17,515 - freqtrade.configuration.config_validation - INFO - Validating configuration ... 2020-11-22 17:36:17,517 - freqtrade.commands.optimize_commands - INFO - Starting freqtrade in Hyperopt mode 2020-11-22 17:36:17,517 - filelock - INFO - Lock 140685194654288 acquired on /freqtrade/user_data/hyperopt.lock 2020-11-22 17:36:17,517 - freqtrade.exchange.exchange - INFO - Instance is running with dry_run enabled 2020-11-22 17:36:17,517 - freqtrade.exchange.exchange - INFO - Using CCXT 1.36.85 2020-11-22 17:36:17,517 - freqtrade.exchange.exchange - INFO - Applying additional ccxt config: {'enableRateLimit': True, 'adjustForTimeDifference': True} 2020-11-22 17:36:17,521 - freqtrade.exchange.exchange - INFO - Applying additional ccxt config: {'enableRateLimit': True, 'adjustForTimeDifference': True, 'rateLimit': 200} 2020-11-22 17:36:17,526 - freqtrade.exchange.exchange - INFO - Using Exchange "Binance" 2020-11-22 17:36:18,682 - freqtrade.resolvers.exchange_resolver - INFO - Using resolved exchange 'Binance'... 2020-11-22 17:36:18,703 - freqtrade.resolvers.iresolver - INFO - Using resolved strategy A_STAR_STRAT from '/freqtrade/user_data/strategies/A_STAR_STRAT.py'... 2020-11-22 17:36:18,703 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'timeframe' with value in config file: 1m. 2020-11-22 17:36:18,703 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'order_types' with value in config file: {'buy': 'market', 'sell': 'market', 'stoploss_on_exchange': True, 'stoploss': 'market'}. 2020-11-22 17:36:18,704 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'stake_currency' with value in config file: USDT. 2020-11-22 17:36:18,704 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'stake_amount' with value in config file: 20. 2020-11-22 17:36:18,704 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'unfilledtimeout' with value in config file: {'buy': 5, 'sell': 5}. 2020-11-22 17:36:18,704 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'use_sell_signal' with value in config file: True. 2020-11-22 17:36:18,704 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using minimal_roi: {'0': 0.2} 2020-11-22 17:36:18,704 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using timeframe: 1m 2020-11-22 17:36:18,704 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using stoploss: -0.04244 2020-11-22 17:36:18,704 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_stop: False 2020-11-22 17:36:18,705 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_stop_positive: 0.16361 2020-11-22 17:36:18,705 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_stop_positive_offset: 0.16679 2020-11-22 17:36:18,705 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_only_offset_is_reached: True 2020-11-22 17:36:18,705 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using process_only_new_candles: False 2020-11-22 17:36:18,705 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using order_types: {'buy': 'market', 'sell': 'market', 'stoploss_on_exchange': True, 'stoploss': 'market'} 2020-11-22 17:36:18,705 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using order_time_in_force: {'buy': 'gtc', 'sell': 'gtc'} 2020-11-22 17:36:18,705 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using stake_currency: USDT 2020-11-22 17:36:18,705 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using stake_amount: 20 2020-11-22 17:36:18,705 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using startup_candle_count: 0 2020-11-22 17:36:18,706 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using unfilledtimeout: {'buy': 5, 'sell': 5} 2020-11-22 17:36:18,706 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using use_sell_signal: True 2020-11-22 17:36:18,706 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using sell_profit_only: True 2020-11-22 17:36:18,706 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using ignore_roi_if_buy_signal: True 2020-11-22 17:36:18,706 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using disable_dataframe_checks: False 2020-11-22 17:36:18,706 - freqtrade.configuration.config_validation - INFO - Validating configuration ... 2020-11-22 17:36:18,709 - freqtrade.resolvers.iresolver - INFO - Using resolved pairlist StaticPairList from '/freqtrade/freqtrade/pairlist/StaticPairList.py'... 2020-11-22 17:36:18,712 - freqtrade.resolvers.iresolver - INFO - Using resolved pairlist AgeFilter from '/freqtrade/freqtrade/pairlist/AgeFilter.py'... 2020-11-22 17:36:19,563 - freqtrade.exchange.exchange - INFO - Downloaded data for BTC/USDT with length 31. 2020-11-22 17:36:19,836 - freqtrade.exchange.exchange - INFO - Downloaded data for ETH/USDT with length 31. 2020-11-22 17:36:20,094 - freqtrade.exchange.exchange - INFO - Downloaded data for XRP/USDT with length 31. 2020-11-22 17:36:20,345 - freqtrade.exchange.exchange - INFO - Downloaded data for TRX/USDT with length 31. 2020-11-22 17:36:20,600 - freqtrade.exchange.exchange - INFO - Downloaded data for NEO/USDT with length 31. 2020-11-22 17:36:20,851 - freqtrade.exchange.exchange - INFO - Downloaded data for LTC/USDT with length 31. 2020-11-22 17:36:21,105 - freqtrade.exchange.exchange - INFO - Downloaded data for XMR/USDT with length 31. 2020-11-22 17:36:21,417 - freqtrade.exchange.exchange - INFO - Downloaded data for ADA/USDT with length 31. 2020-11-22 17:36:21,419 - freqtrade.resolvers.iresolver - INFO - Using resolved hyperopt A_STAR_HYPEROPT from '/freqtrade/user_data/hyperopts/A_STAR_HYPEROPT.py'... 2020-11-22 17:36:21,419 - freqtrade.resolvers.hyperopt_resolver - INFO - Hyperopt class does not provide populate_indicators() method. Using populate_indicators from the strategy. 2020-11-22 17:36:21,419 - freqtrade.resolvers.hyperopt_resolver - INFO - Hyperopt class does not provide populate_buy_trend() method. Using populate_buy_trend from the strategy. 2020-11-22 17:36:21,419 - freqtrade.resolvers.hyperopt_resolver - INFO - Hyperopt class does not provide populate_sell_trend() method. Using populate_sell_trend from the strategy. 2020-11-22 17:36:21,421 - freqtrade.resolvers.iresolver - INFO - Using resolved hyperoptloss ShortTradeDurHyperOptLoss from '/freqtrade/freqtrade/optimize/default_hyperopt_loss.py'... 2020-11-22 17:36:21,421 - freqtrade.optimize.hyperopt - INFO - Removing `/freqtrade/user_data/hyperopt_results/hyperopt_tickerdata.pkl`. 2020-11-22 17:36:21,488 - freqtrade.optimize.hyperopt - INFO - Using optimizer random state: 55436 2020-11-22 17:36:22,203 - numexpr.utils - INFO - NumExpr defaulting to 6 threads. 2020-11-22 17:36:22,360 - freqtrade.data.converter - INFO - Missing data fillup for BTC/USDT: before: 369182 - after: 369542 2020-11-22 17:36:23,149 - freqtrade.data.converter - INFO - Missing data fillup for ETH/USDT: before: 369183 - after: 369543 2020-11-22 17:36:23,978 - freqtrade.data.converter - INFO - Missing data fillup for XRP/USDT: before: 369183 - after: 369543 2020-11-22 17:36:24,751 - freqtrade.data.converter - INFO - Missing data fillup for TRX/USDT: before: 369183 - after: 369543 2020-11-22 17:36:25,524 - freqtrade.data.converter - INFO - Missing data fillup for NEO/USDT: before: 369183 - after: 369543 2020-11-22 17:36:26,387 - freqtrade.data.converter - INFO - Missing data fillup for LTC/USDT: before: 369183 - after: 369543 2020-11-22 17:36:27,161 - freqtrade.data.converter - INFO - Missing data fillup for XMR/USDT: before: 369183 - after: 369543 2020-11-22 17:36:27,891 - freqtrade.data.converter - INFO - Missing data fillup for ADA/USDT: before: 346146 - after: 346506 2020-11-22 17:36:27,910 - freqtrade.optimize.backtesting - INFO - Loading data from 2020-03-11 00:00:00 up to 2020-11-22 15:05:00 (256 days).. 2020-11-22 17:36:27,910 - freqtrade.configuration.timerange - WARNING - Moving start-date by 0 candles to account for startup time. 2020-11-22 17:49:07,451 - freqtrade.optimize.hyperopt - INFO - Hyperopting with data from 2020-03-11 00:00:00 up to 2020-11-22 15:05:00 (256 days).. 2020-11-22 17:49:08,163 - freqtrade.optimize.hyperopt - INFO - Found 6 CPU cores. Let's make them scream! 2020-11-22 17:49:08,163 - freqtrade.optimize.hyperopt - INFO - Number of parallel jobs set as: -1 2020-11-22 17:49:08,175 - freqtrade.optimize.hyperopt - INFO - Effective number of parallel workers used: 6 ```
1.0
Very slow start for backtesting / hyperopt - ## Describe your environment * Operating system: WSL2 on Windows 10 20H2 ( using freqtrade docker image ) * Python Version: image's python * CCXT version: image's CCXT * Freqtrade Version: 2020.10 ## Describe the problem: Starting hyperopt or backtesting takes around 10 minutes ! It was okay until some time ago , it may have something to do with me migrating to WSL2 from my windows10 classic env Note : look at the timestamps of the last 10 lines of the logs, we see 10+ m on one step Also my CPU is sitting at around 30% during those 10m , so it's not a compute power bottleneck ^^" And the data is stored on a SSD , and i got way enough RAM ### Steps to reproduce: idk if you can reproduce it , all i have to do is start backtesting. ### Observed Results: * What happened? * What did you expect to happen? ### Relevant code exceptions or logs Note: Please copy/paste text of the messages, no screenshots of logs please. ``` 2020-11-22 17:36:17,513 - freqtrade.configuration.configuration - INFO - Using config: user_data/config.json ... 2020-11-22 17:36:17,514 - freqtrade.loggers - INFO - Verbosity set to 0 2020-11-22 17:36:17,514 - freqtrade.configuration.configuration - INFO - Using max_open_trades: 10 ... 2020-11-22 17:36:17,514 - freqtrade.configuration.configuration - INFO - Using user-data directory: /freqtrade/user_data ... 2020-11-22 17:36:17,515 - freqtrade.configuration.configuration - INFO - Using data directory: /freqtrade/user_data/data/binance ... 2020-11-22 17:36:17,515 - freqtrade.configuration.configuration - INFO - Using Hyperopt class name: A_STAR_HYPEROPT 2020-11-22 17:36:17,515 - freqtrade.configuration.configuration - INFO - Parameter --epochs detected ... Will run Hyperopt with for 1000 epochs ... 2020-11-22 17:36:17,515 - freqtrade.configuration.configuration - INFO - Parameter -s/--spaces detected: ['sell'] 2020-11-22 17:36:17,515 - freqtrade.configuration.configuration - INFO - Parameter -j/--job-workers detected: -1 2020-11-22 17:36:17,515 - freqtrade.configuration.configuration - INFO - Parameter --min-trades detected: 200 2020-11-22 17:36:17,515 - freqtrade.configuration.configuration - INFO - Using Hyperopt loss class name: ShortTradeDurHyperOptLoss 2020-11-22 17:36:17,515 - freqtrade.configuration.configuration - INFO - Using "json" to store OHLCV data. 2020-11-22 17:36:17,515 - freqtrade.configuration.check_exchange - INFO - Checking exchange... 2020-11-22 17:36:17,515 - freqtrade.configuration.check_exchange - INFO - Exchange "binance" is officially supported by the Freqtrade development team. 2020-11-22 17:36:17,515 - freqtrade.configuration.configuration - INFO - Using pairlist from configuration. 2020-11-22 17:36:17,515 - freqtrade.configuration.config_validation - INFO - Validating configuration ... 2020-11-22 17:36:17,517 - freqtrade.commands.optimize_commands - INFO - Starting freqtrade in Hyperopt mode 2020-11-22 17:36:17,517 - filelock - INFO - Lock 140685194654288 acquired on /freqtrade/user_data/hyperopt.lock 2020-11-22 17:36:17,517 - freqtrade.exchange.exchange - INFO - Instance is running with dry_run enabled 2020-11-22 17:36:17,517 - freqtrade.exchange.exchange - INFO - Using CCXT 1.36.85 2020-11-22 17:36:17,517 - freqtrade.exchange.exchange - INFO - Applying additional ccxt config: {'enableRateLimit': True, 'adjustForTimeDifference': True} 2020-11-22 17:36:17,521 - freqtrade.exchange.exchange - INFO - Applying additional ccxt config: {'enableRateLimit': True, 'adjustForTimeDifference': True, 'rateLimit': 200} 2020-11-22 17:36:17,526 - freqtrade.exchange.exchange - INFO - Using Exchange "Binance" 2020-11-22 17:36:18,682 - freqtrade.resolvers.exchange_resolver - INFO - Using resolved exchange 'Binance'... 2020-11-22 17:36:18,703 - freqtrade.resolvers.iresolver - INFO - Using resolved strategy A_STAR_STRAT from '/freqtrade/user_data/strategies/A_STAR_STRAT.py'... 2020-11-22 17:36:18,703 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'timeframe' with value in config file: 1m. 2020-11-22 17:36:18,703 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'order_types' with value in config file: {'buy': 'market', 'sell': 'market', 'stoploss_on_exchange': True, 'stoploss': 'market'}. 2020-11-22 17:36:18,704 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'stake_currency' with value in config file: USDT. 2020-11-22 17:36:18,704 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'stake_amount' with value in config file: 20. 2020-11-22 17:36:18,704 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'unfilledtimeout' with value in config file: {'buy': 5, 'sell': 5}. 2020-11-22 17:36:18,704 - freqtrade.resolvers.strategy_resolver - INFO - Override strategy 'use_sell_signal' with value in config file: True. 2020-11-22 17:36:18,704 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using minimal_roi: {'0': 0.2} 2020-11-22 17:36:18,704 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using timeframe: 1m 2020-11-22 17:36:18,704 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using stoploss: -0.04244 2020-11-22 17:36:18,704 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_stop: False 2020-11-22 17:36:18,705 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_stop_positive: 0.16361 2020-11-22 17:36:18,705 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_stop_positive_offset: 0.16679 2020-11-22 17:36:18,705 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using trailing_only_offset_is_reached: True 2020-11-22 17:36:18,705 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using process_only_new_candles: False 2020-11-22 17:36:18,705 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using order_types: {'buy': 'market', 'sell': 'market', 'stoploss_on_exchange': True, 'stoploss': 'market'} 2020-11-22 17:36:18,705 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using order_time_in_force: {'buy': 'gtc', 'sell': 'gtc'} 2020-11-22 17:36:18,705 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using stake_currency: USDT 2020-11-22 17:36:18,705 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using stake_amount: 20 2020-11-22 17:36:18,705 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using startup_candle_count: 0 2020-11-22 17:36:18,706 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using unfilledtimeout: {'buy': 5, 'sell': 5} 2020-11-22 17:36:18,706 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using use_sell_signal: True 2020-11-22 17:36:18,706 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using sell_profit_only: True 2020-11-22 17:36:18,706 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using ignore_roi_if_buy_signal: True 2020-11-22 17:36:18,706 - freqtrade.resolvers.strategy_resolver - INFO - Strategy using disable_dataframe_checks: False 2020-11-22 17:36:18,706 - freqtrade.configuration.config_validation - INFO - Validating configuration ... 2020-11-22 17:36:18,709 - freqtrade.resolvers.iresolver - INFO - Using resolved pairlist StaticPairList from '/freqtrade/freqtrade/pairlist/StaticPairList.py'... 2020-11-22 17:36:18,712 - freqtrade.resolvers.iresolver - INFO - Using resolved pairlist AgeFilter from '/freqtrade/freqtrade/pairlist/AgeFilter.py'... 2020-11-22 17:36:19,563 - freqtrade.exchange.exchange - INFO - Downloaded data for BTC/USDT with length 31. 2020-11-22 17:36:19,836 - freqtrade.exchange.exchange - INFO - Downloaded data for ETH/USDT with length 31. 2020-11-22 17:36:20,094 - freqtrade.exchange.exchange - INFO - Downloaded data for XRP/USDT with length 31. 2020-11-22 17:36:20,345 - freqtrade.exchange.exchange - INFO - Downloaded data for TRX/USDT with length 31. 2020-11-22 17:36:20,600 - freqtrade.exchange.exchange - INFO - Downloaded data for NEO/USDT with length 31. 2020-11-22 17:36:20,851 - freqtrade.exchange.exchange - INFO - Downloaded data for LTC/USDT with length 31. 2020-11-22 17:36:21,105 - freqtrade.exchange.exchange - INFO - Downloaded data for XMR/USDT with length 31. 2020-11-22 17:36:21,417 - freqtrade.exchange.exchange - INFO - Downloaded data for ADA/USDT with length 31. 2020-11-22 17:36:21,419 - freqtrade.resolvers.iresolver - INFO - Using resolved hyperopt A_STAR_HYPEROPT from '/freqtrade/user_data/hyperopts/A_STAR_HYPEROPT.py'... 2020-11-22 17:36:21,419 - freqtrade.resolvers.hyperopt_resolver - INFO - Hyperopt class does not provide populate_indicators() method. Using populate_indicators from the strategy. 2020-11-22 17:36:21,419 - freqtrade.resolvers.hyperopt_resolver - INFO - Hyperopt class does not provide populate_buy_trend() method. Using populate_buy_trend from the strategy. 2020-11-22 17:36:21,419 - freqtrade.resolvers.hyperopt_resolver - INFO - Hyperopt class does not provide populate_sell_trend() method. Using populate_sell_trend from the strategy. 2020-11-22 17:36:21,421 - freqtrade.resolvers.iresolver - INFO - Using resolved hyperoptloss ShortTradeDurHyperOptLoss from '/freqtrade/freqtrade/optimize/default_hyperopt_loss.py'... 2020-11-22 17:36:21,421 - freqtrade.optimize.hyperopt - INFO - Removing `/freqtrade/user_data/hyperopt_results/hyperopt_tickerdata.pkl`. 2020-11-22 17:36:21,488 - freqtrade.optimize.hyperopt - INFO - Using optimizer random state: 55436 2020-11-22 17:36:22,203 - numexpr.utils - INFO - NumExpr defaulting to 6 threads. 2020-11-22 17:36:22,360 - freqtrade.data.converter - INFO - Missing data fillup for BTC/USDT: before: 369182 - after: 369542 2020-11-22 17:36:23,149 - freqtrade.data.converter - INFO - Missing data fillup for ETH/USDT: before: 369183 - after: 369543 2020-11-22 17:36:23,978 - freqtrade.data.converter - INFO - Missing data fillup for XRP/USDT: before: 369183 - after: 369543 2020-11-22 17:36:24,751 - freqtrade.data.converter - INFO - Missing data fillup for TRX/USDT: before: 369183 - after: 369543 2020-11-22 17:36:25,524 - freqtrade.data.converter - INFO - Missing data fillup for NEO/USDT: before: 369183 - after: 369543 2020-11-22 17:36:26,387 - freqtrade.data.converter - INFO - Missing data fillup for LTC/USDT: before: 369183 - after: 369543 2020-11-22 17:36:27,161 - freqtrade.data.converter - INFO - Missing data fillup for XMR/USDT: before: 369183 - after: 369543 2020-11-22 17:36:27,891 - freqtrade.data.converter - INFO - Missing data fillup for ADA/USDT: before: 346146 - after: 346506 2020-11-22 17:36:27,910 - freqtrade.optimize.backtesting - INFO - Loading data from 2020-03-11 00:00:00 up to 2020-11-22 15:05:00 (256 days).. 2020-11-22 17:36:27,910 - freqtrade.configuration.timerange - WARNING - Moving start-date by 0 candles to account for startup time. 2020-11-22 17:49:07,451 - freqtrade.optimize.hyperopt - INFO - Hyperopting with data from 2020-03-11 00:00:00 up to 2020-11-22 15:05:00 (256 days).. 2020-11-22 17:49:08,163 - freqtrade.optimize.hyperopt - INFO - Found 6 CPU cores. Let's make them scream! 2020-11-22 17:49:08,163 - freqtrade.optimize.hyperopt - INFO - Number of parallel jobs set as: -1 2020-11-22 17:49:08,175 - freqtrade.optimize.hyperopt - INFO - Effective number of parallel workers used: 6 ```
test
very slow start for backtesting hyperopt describe your environment operating system on windows using freqtrade docker image python version image s python ccxt version image s ccxt freqtrade version describe the problem starting hyperopt or backtesting takes around minutes it was okay until some time ago it may have something to do with me migrating to from my classic env note look at the timestamps of the last lines of the logs we see m on one step also my cpu is sitting at around during those so it s not a compute power bottleneck and the data is stored on a ssd and i got way enough ram steps to reproduce idk if you can reproduce it all i have to do is start backtesting observed results what happened what did you expect to happen relevant code exceptions or logs note please copy paste text of the messages no screenshots of logs please freqtrade configuration configuration info using config user data config json freqtrade loggers info verbosity set to freqtrade configuration configuration info using max open trades freqtrade configuration configuration info using user data directory freqtrade user data freqtrade configuration configuration info using data directory freqtrade user data data binance freqtrade configuration configuration info using hyperopt class name a star hyperopt freqtrade configuration configuration info parameter epochs detected will run hyperopt with for epochs freqtrade configuration configuration info parameter s spaces detected freqtrade configuration configuration info parameter j job workers detected freqtrade configuration configuration info parameter min trades detected freqtrade configuration configuration info using hyperopt loss class name shorttradedurhyperoptloss freqtrade configuration configuration info using json to store ohlcv data freqtrade configuration check exchange info checking exchange freqtrade configuration check exchange info exchange binance is officially supported by the freqtrade development team freqtrade configuration configuration info using pairlist from configuration freqtrade configuration config validation info validating configuration freqtrade commands optimize commands info starting freqtrade in hyperopt mode filelock info lock acquired on freqtrade user data hyperopt lock freqtrade exchange exchange info instance is running with dry run enabled freqtrade exchange exchange info using ccxt freqtrade exchange exchange info applying additional ccxt config enableratelimit true adjustfortimedifference true freqtrade exchange exchange info applying additional ccxt config enableratelimit true adjustfortimedifference true ratelimit freqtrade exchange exchange info using exchange binance freqtrade resolvers exchange resolver info using resolved exchange binance freqtrade resolvers iresolver info using resolved strategy a star strat from freqtrade user data strategies a star strat py freqtrade resolvers strategy resolver info override strategy timeframe with value in config file freqtrade resolvers strategy resolver info override strategy order types with value in config file buy market sell market stoploss on exchange true stoploss market freqtrade resolvers strategy resolver info override strategy stake currency with value in config file usdt freqtrade resolvers strategy resolver info override strategy stake amount with value in config file freqtrade resolvers strategy resolver info override strategy unfilledtimeout with value in config file buy sell freqtrade resolvers strategy resolver info override strategy use sell signal with value in config file true freqtrade resolvers strategy resolver info strategy using minimal roi freqtrade resolvers strategy resolver info strategy using timeframe freqtrade resolvers strategy resolver info strategy using stoploss freqtrade resolvers strategy resolver info strategy using trailing stop false freqtrade resolvers strategy resolver info strategy using trailing stop positive freqtrade resolvers strategy resolver info strategy using trailing stop positive offset freqtrade resolvers strategy resolver info strategy using trailing only offset is reached true freqtrade resolvers strategy resolver info strategy using process only new candles false freqtrade resolvers strategy resolver info strategy using order types buy market sell market stoploss on exchange true stoploss market freqtrade resolvers strategy resolver info strategy using order time in force buy gtc sell gtc freqtrade resolvers strategy resolver info strategy using stake currency usdt freqtrade resolvers strategy resolver info strategy using stake amount freqtrade resolvers strategy resolver info strategy using startup candle count freqtrade resolvers strategy resolver info strategy using unfilledtimeout buy sell freqtrade resolvers strategy resolver info strategy using use sell signal true freqtrade resolvers strategy resolver info strategy using sell profit only true freqtrade resolvers strategy resolver info strategy using ignore roi if buy signal true freqtrade resolvers strategy resolver info strategy using disable dataframe checks false freqtrade configuration config validation info validating configuration freqtrade resolvers iresolver info using resolved pairlist staticpairlist from freqtrade freqtrade pairlist staticpairlist py freqtrade resolvers iresolver info using resolved pairlist agefilter from freqtrade freqtrade pairlist agefilter py freqtrade exchange exchange info downloaded data for btc usdt with length freqtrade exchange exchange info downloaded data for eth usdt with length freqtrade exchange exchange info downloaded data for xrp usdt with length freqtrade exchange exchange info downloaded data for trx usdt with length freqtrade exchange exchange info downloaded data for neo usdt with length freqtrade exchange exchange info downloaded data for ltc usdt with length freqtrade exchange exchange info downloaded data for xmr usdt with length freqtrade exchange exchange info downloaded data for ada usdt with length freqtrade resolvers iresolver info using resolved hyperopt a star hyperopt from freqtrade user data hyperopts a star hyperopt py freqtrade resolvers hyperopt resolver info hyperopt class does not provide populate indicators method using populate indicators from the strategy freqtrade resolvers hyperopt resolver info hyperopt class does not provide populate buy trend method using populate buy trend from the strategy freqtrade resolvers hyperopt resolver info hyperopt class does not provide populate sell trend method using populate sell trend from the strategy freqtrade resolvers iresolver info using resolved hyperoptloss shorttradedurhyperoptloss from freqtrade freqtrade optimize default hyperopt loss py freqtrade optimize hyperopt info removing freqtrade user data hyperopt results hyperopt tickerdata pkl freqtrade optimize hyperopt info using optimizer random state numexpr utils info numexpr defaulting to threads freqtrade data converter info missing data fillup for btc usdt before after freqtrade data converter info missing data fillup for eth usdt before after freqtrade data converter info missing data fillup for xrp usdt before after freqtrade data converter info missing data fillup for trx usdt before after freqtrade data converter info missing data fillup for neo usdt before after freqtrade data converter info missing data fillup for ltc usdt before after freqtrade data converter info missing data fillup for xmr usdt before after freqtrade data converter info missing data fillup for ada usdt before after freqtrade optimize backtesting info loading data from up to days freqtrade configuration timerange warning moving start date by candles to account for startup time freqtrade optimize hyperopt info hyperopting with data from up to days freqtrade optimize hyperopt info found cpu cores let s make them scream freqtrade optimize hyperopt info number of parallel jobs set as freqtrade optimize hyperopt info effective number of parallel workers used
1
79,654
15,586,193,550
IssuesEvent
2021-03-18 01:23:10
ziednov007/JavaSpring
https://api.github.com/repos/ziednov007/JavaSpring
opened
CVE-2020-24750 (High) detected in jackson-databind-2.9.6.jar
security vulnerability
## CVE-2020-24750 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: JavaSpring/app/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar,/root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.0.5.RELEASE.jar (Root Library) - spring-boot-starter-json-2.0.5.RELEASE.jar - :x: **jackson-databind-2.9.6.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to com.pastdev.httpcomponents.configuration.JndiConfiguration. <p>Publish Date: 2020-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24750>CVE-2020-24750</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p> <p>Release Date: 2020-08-28</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-24750 (High) detected in jackson-databind-2.9.6.jar - ## CVE-2020-24750 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: JavaSpring/app/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar,/root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.0.5.RELEASE.jar (Root Library) - spring-boot-starter-json-2.0.5.RELEASE.jar - :x: **jackson-databind-2.9.6.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to com.pastdev.httpcomponents.configuration.JndiConfiguration. <p>Publish Date: 2020-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24750>CVE-2020-24750</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p> <p>Release Date: 2020-08-28</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file javaspring app build gradle path to vulnerable library root gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar root gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library spring boot starter json release jar x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com pastdev httpcomponents configuration jndiconfiguration publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
0
174,147
13,459,064,010
IssuesEvent
2020-09-09 11:40:32
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
FrozenPartitionTableTest.partitionTable_shouldBeFixed_whenMemberRestarts_usingNewUuid failure
Module: Partitioning Source: Internal Team: Core Type: Test-Failure
**Error Message** ``` com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - bd5d6faf-9735-4300-8038-c4ac23167ce8 this, expected-target: null, partitionId: 0, replicaIndex: 0, operation: com.hazelcast.internal.partition.impl.FrozenPartitionTableTest$NonRetryablePartitionOperation, service: null ``` **Stacktrace** ``` java.util.concurrent.CompletionException: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - bd5d6faf-9735-4300-8038-c4ac23167ce8 this, expected-target: null, partitionId: 0, replicaIndex: 0, operation: com.hazelcast.internal.partition.impl.FrozenPartitionTableTest$NonRetryablePartitionOperation, service: null at com.hazelcast.spi.impl.AbstractInvocationFuture.returnOrThrowWithJoinConventions(AbstractInvocationFuture.java:761) at com.hazelcast.spi.impl.AbstractInvocationFuture.resolveAndThrowWithJoinConvention(AbstractInvocationFuture.java:777) at com.hazelcast.spi.impl.AbstractInvocationFuture.join(AbstractInvocationFuture.java:537) at com.hazelcast.internal.partition.impl.FrozenPartitionTableTest.partitionTable_shouldBeFixed_whenMemberRestarts_usingNewUuid(FrozenPartitionTableTest.java:231) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:114) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:106) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.lang.Thread.run(Thread.java:748) Caused by: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - bd5d6faf-9735-4300-8038-c4ac23167ce8 this, expected-target: null, partitionId: 0, replicaIndex: 0, operation: com.hazelcast.internal.partition.impl.FrozenPartitionTableTest$NonRetryablePartitionOperation, service: null at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.ensureNoPartitionProblems(OperationRunnerImpl.java:375) at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:231) at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:452) at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(OperationThread.java:166) at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(OperationThread.java:136) at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.executeRun(OperationThread.java:123) at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:102) ``` http://jenkins.hazelcast.com/job/Hazelcast-pr-builder/2262/
1.0
FrozenPartitionTableTest.partitionTable_shouldBeFixed_whenMemberRestarts_usingNewUuid failure - **Error Message** ``` com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - bd5d6faf-9735-4300-8038-c4ac23167ce8 this, expected-target: null, partitionId: 0, replicaIndex: 0, operation: com.hazelcast.internal.partition.impl.FrozenPartitionTableTest$NonRetryablePartitionOperation, service: null ``` **Stacktrace** ``` java.util.concurrent.CompletionException: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - bd5d6faf-9735-4300-8038-c4ac23167ce8 this, expected-target: null, partitionId: 0, replicaIndex: 0, operation: com.hazelcast.internal.partition.impl.FrozenPartitionTableTest$NonRetryablePartitionOperation, service: null at com.hazelcast.spi.impl.AbstractInvocationFuture.returnOrThrowWithJoinConventions(AbstractInvocationFuture.java:761) at com.hazelcast.spi.impl.AbstractInvocationFuture.resolveAndThrowWithJoinConvention(AbstractInvocationFuture.java:777) at com.hazelcast.spi.impl.AbstractInvocationFuture.join(AbstractInvocationFuture.java:537) at com.hazelcast.internal.partition.impl.FrozenPartitionTableTest.partitionTable_shouldBeFixed_whenMemberRestarts_usingNewUuid(FrozenPartitionTableTest.java:231) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:114) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:106) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.lang.Thread.run(Thread.java:748) Caused by: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - bd5d6faf-9735-4300-8038-c4ac23167ce8 this, expected-target: null, partitionId: 0, replicaIndex: 0, operation: com.hazelcast.internal.partition.impl.FrozenPartitionTableTest$NonRetryablePartitionOperation, service: null at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.ensureNoPartitionProblems(OperationRunnerImpl.java:375) at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:231) at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:452) at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(OperationThread.java:166) at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(OperationThread.java:136) at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.executeRun(OperationThread.java:123) at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:102) ``` http://jenkins.hazelcast.com/job/Hazelcast-pr-builder/2262/
test
frozenpartitiontabletest partitiontable shouldbefixed whenmemberrestarts usingnewuuid failure error message com hazelcast spi exception wrongtargetexception wrongtarget local member this expected target null partitionid replicaindex operation com hazelcast internal partition impl frozenpartitiontabletest nonretryablepartitionoperation service null stacktrace java util concurrent completionexception com hazelcast spi exception wrongtargetexception wrongtarget local member this expected target null partitionid replicaindex operation com hazelcast internal partition impl frozenpartitiontabletest nonretryablepartitionoperation service null at com hazelcast spi impl abstractinvocationfuture returnorthrowwithjoinconventions abstractinvocationfuture java at com hazelcast spi impl abstractinvocationfuture resolveandthrowwithjoinconvention abstractinvocationfuture java at com hazelcast spi impl abstractinvocationfuture join abstractinvocationfuture java at com hazelcast internal partition impl frozenpartitiontabletest partitiontable shouldbefixed whenmemberrestarts usingnewuuid frozenpartitiontabletest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at java util concurrent futuretask run futuretask java at java lang thread run thread java caused by com hazelcast spi exception wrongtargetexception wrongtarget local member this expected target null partitionid replicaindex operation com hazelcast internal partition impl frozenpartitiontabletest nonretryablepartitionoperation service null at com hazelcast spi impl operationservice impl operationrunnerimpl ensurenopartitionproblems operationrunnerimpl java at com hazelcast spi impl operationservice impl operationrunnerimpl run operationrunnerimpl java at com hazelcast spi impl operationservice impl operationrunnerimpl run operationrunnerimpl java at com hazelcast spi impl operationexecutor impl operationthread process operationthread java at com hazelcast spi impl operationexecutor impl operationthread process operationthread java at com hazelcast spi impl operationexecutor impl operationthread executerun operationthread java at com hazelcast internal util executor hazelcastmanagedthread run hazelcastmanagedthread java
1
151,870
23,885,991,052
IssuesEvent
2022-09-08 07:42:58
NethmiRodrigo/poolin
https://api.github.com/repos/NethmiRodrigo/poolin
closed
Upload driver's license UI Design
feature design
**Is your feature request related to a problem? Please describe the user story.** <!---A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> **Describe the solution you'd like. Preferably the workflow of the feature** <!---A clear and concise description of what you want to happen. --> - [ ] Overly redirecting to upload license when offering ride - [ ] Upload button and verification status on profile screen - [ ] Upload screen - [ ] Upload successful overlay **Describe alternatives you've considered** <!--A clear and concise description of any alternative solutions or features you've considered. --> **Additional context (Screenshots of the UI, design)** <!--Add any other context or screenshots about the feature request here.-->
1.0
Upload driver's license UI Design - **Is your feature request related to a problem? Please describe the user story.** <!---A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> **Describe the solution you'd like. Preferably the workflow of the feature** <!---A clear and concise description of what you want to happen. --> - [ ] Overly redirecting to upload license when offering ride - [ ] Upload button and verification status on profile screen - [ ] Upload screen - [ ] Upload successful overlay **Describe alternatives you've considered** <!--A clear and concise description of any alternative solutions or features you've considered. --> **Additional context (Screenshots of the UI, design)** <!--Add any other context or screenshots about the feature request here.-->
non_test
upload driver s license ui design is your feature request related to a problem please describe the user story describe the solution you d like preferably the workflow of the feature overly redirecting to upload license when offering ride upload button and verification status on profile screen upload screen upload successful overlay describe alternatives you ve considered additional context screenshots of the ui design
0
40,898
5,321,325,638
IssuesEvent
2017-02-14 13:14:00
ukwa/w3act
https://api.github.com/repos/ukwa/w3act
closed
Document Title is not required for Journal Issues
1 - Ready to test on stage ddhapt Enhancement
Issues don't often have titles, so this field should not be filled out for Journal Issues. Titles are not allowed to be NULL, but can be the empty string instead. I am also assuming this means the title should not appear in the SIP export for Journal Issues.
1.0
Document Title is not required for Journal Issues - Issues don't often have titles, so this field should not be filled out for Journal Issues. Titles are not allowed to be NULL, but can be the empty string instead. I am also assuming this means the title should not appear in the SIP export for Journal Issues.
test
document title is not required for journal issues issues don t often have titles so this field should not be filled out for journal issues titles are not allowed to be null but can be the empty string instead i am also assuming this means the title should not appear in the sip export for journal issues
1
264,874
20,036,663,322
IssuesEvent
2022-02-02 12:39:23
greenplum-db/sre-test
https://api.github.com/repos/greenplum-db/sre-test
opened
Binary data serialize/deserialize java class not well documented
enhancement PXF Documentation 6.2
Location: https://gpdb.docs.pivotal.io/pxf/6-2/using/hdfs_seqfile.html#write_binary Accessing Hadoop with PXF -> Reading and Writing HDFS SequenceFile Data -> Reading and Writing Binary Data -> Example: Writing Binary Data to HDFS The document provides a sample java class that is to be used with the schema specific to the example given in the document. However the sample code does not have any comments nor does the document provide any further instructions or links to additional information, that the user can use to modify the code for their own specific schema. Any errors in the code are only revealed when binary data is read/written using PXF making debugging difficult/tedious for someone not familiar with java.
1.0
Binary data serialize/deserialize java class not well documented - Location: https://gpdb.docs.pivotal.io/pxf/6-2/using/hdfs_seqfile.html#write_binary Accessing Hadoop with PXF -> Reading and Writing HDFS SequenceFile Data -> Reading and Writing Binary Data -> Example: Writing Binary Data to HDFS The document provides a sample java class that is to be used with the schema specific to the example given in the document. However the sample code does not have any comments nor does the document provide any further instructions or links to additional information, that the user can use to modify the code for their own specific schema. Any errors in the code are only revealed when binary data is read/written using PXF making debugging difficult/tedious for someone not familiar with java.
non_test
binary data serialize deserialize java class not well documented location accessing hadoop with pxf reading and writing hdfs sequencefile data reading and writing binary data example writing binary data to hdfs the document provides a sample java class that is to be used with the schema specific to the example given in the document however the sample code does not have any comments nor does the document provide any further instructions or links to additional information that the user can use to modify the code for their own specific schema any errors in the code are only revealed when binary data is read written using pxf making debugging difficult tedious for someone not familiar with java
0
737,516
25,519,502,396
IssuesEvent
2022-11-28 19:10:32
exercism/elm
https://api.github.com/repos/exercism/elm
closed
Concept: generics
x:action/create x:knowledge/elementary x:module/concept-exercise x:status/claimed x:type/content x:size/large x:priority/high
# Design ## Goal The goal of this exercise is to teach students how to consume and make generic code. ## Learning objectives - Understand what a type variable is. - Know how to define generic types. - Know how to define generic functions. ## Out of scope - Covariance and contravariance. - Higher order functions. ## Concepts - `generics`: Learn how to create and use Elm code that is generic. ## Prerequisites - `custom-types`: Learn how to use custom types in an Elm program.
1.0
Concept: generics - # Design ## Goal The goal of this exercise is to teach students how to consume and make generic code. ## Learning objectives - Understand what a type variable is. - Know how to define generic types. - Know how to define generic functions. ## Out of scope - Covariance and contravariance. - Higher order functions. ## Concepts - `generics`: Learn how to create and use Elm code that is generic. ## Prerequisites - `custom-types`: Learn how to use custom types in an Elm program.
non_test
concept generics design goal the goal of this exercise is to teach students how to consume and make generic code learning objectives understand what a type variable is know how to define generic types know how to define generic functions out of scope covariance and contravariance higher order functions concepts generics learn how to create and use elm code that is generic prerequisites custom types learn how to use custom types in an elm program
0
183,986
14,265,931,030
IssuesEvent
2020-11-20 17:53:05
dotnet/aspnetcore
https://api.github.com/repos/dotnet/aspnetcore
closed
QuarantinedTest: BlazorWasmStandaloneTemplate_IndividualAuth_Works
Done Working area-blazor task test-failure
https://dev.azure.com/dnceng/public/_build/results?buildId=714955&view=ms.vss-test-web.build-test-results-tab&runId=22159654&resultId=100007&paneView=debug ``` OpenQA.Selenium.BrowserAssertFailedException : Xunit.Sdk.NotEmptyException: Assert.NotEmpty() Failure at Xunit.Assert.NotEmpty(IEnumerable collection) in C:\\Dev\\xunit\\xunit\\src\\xunit.assert\\Asserts\\CollectionAsserts.cs:line 331 at Microsoft.AspNetCore.E2ETesting.WaitAssert.<>c__DisplayClass15_0.<Exists>b__0() in /_/src/Shared/E2ETesting/WaitAssert.cs:line 69 at Microsoft.AspNetCore.E2ETesting.WaitAssert.<>c__DisplayClass18_0`1.<WaitAssertCore>b__0(IWebDriver _) in /_/src/Shared/E2ETesting/WaitAssert.cs:line 106 Screen shot captured at 'F:\\workspace\\_work\\1\\s\\artifacts\\TestResults\\Release\\BlazorTemplates.Tests\\4d4d483d61f44d5f86142ca3643447bf.png' Encountered browser errors [2020-07-02T20:51:44Z] [Debug] http://localhost:52317/_framework/dotnet.5.0.0-preview.7.20326.1.js 0:124605 \"mono_wasm_runtime_ready\" \"fe00e07a-5519-4dfe-b35a-f867dbaf2e28\" [2020-07-02T20:51:44Z] [Info] https://login.microsoftonline.com/common/oauth2/v2.0/authorize?client_id=sample-client-id&redirect_uri=http%3A%2F%2Flocalhost%3A52317%2Fauthentication%2Flogin-callback&response_type=id_token&scope=openid%20profile&state=3a7e6662ac4e478e8ad7ed5d0b16258d&nonce=469d5890dd3845eeaf900ac8988ecd95&prompt=none 73:13657 \"BSSO Telemetry: {\\\"result\\\":\\\"Error\\\",\\\"error\\\":\\\"NoExtension\\\",\\\"type\\\":\\\"ChromeSsoTelemetry\\\",\\\"data\\\":{},\\\"traces\\\":[\\\"BrowserSSO Initialized\\\",\\\"Creating ChromeBrowserCore provider\\\",\\\"Sending message for method CreateProviderAsync\\\",\\\"Received message for method CreateProviderAsync\\\",\\\"Error: ChromeBrowserCore error NoExtension: Extension is not installed.\\\"]}\" [2020-07-02T20:51:44Z] [Severe] https://login.microsoftonline.com/common/oauth2/v2.0/authorize?client_id=sample-client-id&redirect_uri=http%3A%2F%2Flocalhost%3A52317%2Fauthentication%2Flogin-callback&response_type=id_token&scope=openid%20profile&state=3a7e6662ac4e478e8ad7ed5d0b16258d&nonce=469d5890dd3845eeaf900ac8988ecd95&prompt=none&sso_reload=true 137 Unsafe JavaScript attempt to initiate navigation for frame with URL 'http://localhost:52317/' from frame with URL 'https://login.microsoftonline.com/common/oauth2/v2.0/authorize?client_id=sample-client-id&redirect_uri=http%3A%2F%2Flocalhost%3A52317%2Fauthentication%2Flogin-callback&response_type=id_token&scope=openid%20profile&state=3a7e6662ac4e478e8ad7ed5d0b16258d&nonce=469d5890dd3845eeaf900ac8988ecd95&prompt=none&sso_reload=true'. The frame attempting navigation is targeting its top-level window, but is neither same-origin with its target nor has it received a user gesture. See https://www.chromestatus.com/features/5851021045661696. [2020-07-02T20:51:44Z] [Severe] https://login.microsoftonline.com/common/oauth2/v2.0/authorize?client_id=sample-client-id&redirect_uri=http%3A%2F%2Flocalhost%3A52317%2Fauthentication%2Flogin-callback&response_type=id_token&scope=openid%20profile&state=3a7e6662ac4e478e8ad7ed5d0b16258d&nonce=469d5890dd3845eeaf900ac8988ecd95&prompt=none&sso_reload=true 137:325 Uncaught DOMException: Failed to set the 'href' property on 'Location': Tโ€ฆrequest-id=5f269ce9-80c8-4cde-87a3-745d3ef76000'. [2020-07-02T20:51:53Z] [Info] http://localhost:52317/_framework/blazor.webassembly.js 0:2753 \"info: Microsoft.AspNetCore.Authorization.DefaultAuthorizationService[2]\ Authorization failed. These requirements were not met:\ DenyAnonymousAuthorizationRequirement: Requires an authenticated user.\" [2020-07-02T20:52:00Z] [Debug] https://localhost:52368/_framework/dotnet.5.0.0-preview.7.20326.1.js 0:124605 \"mono_wasm_runtime_ready\" \"fe00e07a-5519-4dfe-b35a-f867dbaf2e28\" [2020-07-02T20:52:00Z] [Severe] https://localhost:52368/_framework/blazor.webassembly.js 0:40764 RangeError: Maximum call stack size exceeded at mono_class_get_first_method_idx (https://localhost:52368/_framework/dotnet.wasm:wasm-function[603]:0x12964) at mono_find_method_in_metadata (https://localhost:52368/_framework/dotnet.wasm:wasm-function[2286]:0x605f2) at mono_class_get_method_from_name_checked (https://localhost:52368/_framework/dotnet.wasm:wasm-function[198]:0x50bd) at mono_class_get_cctor (https://localhost:52368/_framework/dotnet.wasm:wasm-function[1566]:0x42b21) at mono_runtime_class_init_full (https://localhost:52368/_framework/dotnet.wasm:wasm-function[233]:0x5c71) at mono_interp_transform_method (https://localhost:52368/_framework/dotnet.wasm:wasm-function[1144]:0x24a96) at do_transform_method (https://localhost:52368/_framework/dotnet.wasm:wasm-function[2694]:0x7e075) at interp_exec_method (https://localhost:52368/_framework/dotnet.wasm:wasm-function[1224]:0x353df) at interp_runtime_invoke (https://localhost:52368/_framework/dotnet.wasm:wasm-function[4966]:0xceb93) at mono_jit_runtime_invoke (https://localhost:52368/_framework/dotnet.wasm:wasm-function[4466]:0xb8932) Page content: <head> <meta charset=\"utf-8\"> <meta name=\"viewport\" content=\"width=device-width, initial-scale=1.0, maximum-scale=1.0, user-scalable=no\"> <title>AspNet.offwtaofawy</title> <base href=\"/\"> <link href=\"css/bootstrap/bootstrap.min.css\" rel=\"stylesheet\"> <link href=\"css/app.css\" rel=\"stylesheet\"> </head> <body> <app>Loading...</app> <div id=\"blazor-error-ui\" style=\"display: block;\"> An unhandled error has occurred. <a href=\"\" class=\"reload\">Reload</a> <a class=\"dismiss\">??</a> </div> <script src=\"_content/Microsoft.AspNetCore.Components.WebAssembly.Authentication/AuthenticationService.js\"></script> <script src=\"_framework/blazor.webassembly.js\"></script> <script type=\"text/javascript\">var Module; window.__wasmmodulecallback__(); delete window.__wasmmodulecallback__;</script><script src=\"_framework/dotnet.5.0.0-preview.7.20326.1.js\" defer=\"\" integrity=\"sha256-NInsB2Gm0V2B80TQrCgo8y9wzIddZvMysAFwSpIaejo=\" crossorigin=\"anonymous\"></script></body> ---- Assert.NotEmpty() Failure at Microsoft.AspNetCore.E2ETesting.WaitAssert.WaitAssertCore[TResult](IWebDriver driver, Func`1 assertion, TimeSpan timeout) in /_/src/Shared/E2ETesting/WaitAssert.cs:line 126 at Microsoft.AspNetCore.E2ETesting.WaitAssert.Exists(IWebDriver driver, By finder, TimeSpan timeout) in /_/src/Shared/E2ETesting/WaitAssert.cs:line 64 a ```
1.0
QuarantinedTest: BlazorWasmStandaloneTemplate_IndividualAuth_Works - https://dev.azure.com/dnceng/public/_build/results?buildId=714955&view=ms.vss-test-web.build-test-results-tab&runId=22159654&resultId=100007&paneView=debug ``` OpenQA.Selenium.BrowserAssertFailedException : Xunit.Sdk.NotEmptyException: Assert.NotEmpty() Failure at Xunit.Assert.NotEmpty(IEnumerable collection) in C:\\Dev\\xunit\\xunit\\src\\xunit.assert\\Asserts\\CollectionAsserts.cs:line 331 at Microsoft.AspNetCore.E2ETesting.WaitAssert.<>c__DisplayClass15_0.<Exists>b__0() in /_/src/Shared/E2ETesting/WaitAssert.cs:line 69 at Microsoft.AspNetCore.E2ETesting.WaitAssert.<>c__DisplayClass18_0`1.<WaitAssertCore>b__0(IWebDriver _) in /_/src/Shared/E2ETesting/WaitAssert.cs:line 106 Screen shot captured at 'F:\\workspace\\_work\\1\\s\\artifacts\\TestResults\\Release\\BlazorTemplates.Tests\\4d4d483d61f44d5f86142ca3643447bf.png' Encountered browser errors [2020-07-02T20:51:44Z] [Debug] http://localhost:52317/_framework/dotnet.5.0.0-preview.7.20326.1.js 0:124605 \"mono_wasm_runtime_ready\" \"fe00e07a-5519-4dfe-b35a-f867dbaf2e28\" [2020-07-02T20:51:44Z] [Info] https://login.microsoftonline.com/common/oauth2/v2.0/authorize?client_id=sample-client-id&redirect_uri=http%3A%2F%2Flocalhost%3A52317%2Fauthentication%2Flogin-callback&response_type=id_token&scope=openid%20profile&state=3a7e6662ac4e478e8ad7ed5d0b16258d&nonce=469d5890dd3845eeaf900ac8988ecd95&prompt=none 73:13657 \"BSSO Telemetry: {\\\"result\\\":\\\"Error\\\",\\\"error\\\":\\\"NoExtension\\\",\\\"type\\\":\\\"ChromeSsoTelemetry\\\",\\\"data\\\":{},\\\"traces\\\":[\\\"BrowserSSO Initialized\\\",\\\"Creating ChromeBrowserCore provider\\\",\\\"Sending message for method CreateProviderAsync\\\",\\\"Received message for method CreateProviderAsync\\\",\\\"Error: ChromeBrowserCore error NoExtension: Extension is not installed.\\\"]}\" [2020-07-02T20:51:44Z] [Severe] https://login.microsoftonline.com/common/oauth2/v2.0/authorize?client_id=sample-client-id&redirect_uri=http%3A%2F%2Flocalhost%3A52317%2Fauthentication%2Flogin-callback&response_type=id_token&scope=openid%20profile&state=3a7e6662ac4e478e8ad7ed5d0b16258d&nonce=469d5890dd3845eeaf900ac8988ecd95&prompt=none&sso_reload=true 137 Unsafe JavaScript attempt to initiate navigation for frame with URL 'http://localhost:52317/' from frame with URL 'https://login.microsoftonline.com/common/oauth2/v2.0/authorize?client_id=sample-client-id&redirect_uri=http%3A%2F%2Flocalhost%3A52317%2Fauthentication%2Flogin-callback&response_type=id_token&scope=openid%20profile&state=3a7e6662ac4e478e8ad7ed5d0b16258d&nonce=469d5890dd3845eeaf900ac8988ecd95&prompt=none&sso_reload=true'. The frame attempting navigation is targeting its top-level window, but is neither same-origin with its target nor has it received a user gesture. See https://www.chromestatus.com/features/5851021045661696. [2020-07-02T20:51:44Z] [Severe] https://login.microsoftonline.com/common/oauth2/v2.0/authorize?client_id=sample-client-id&redirect_uri=http%3A%2F%2Flocalhost%3A52317%2Fauthentication%2Flogin-callback&response_type=id_token&scope=openid%20profile&state=3a7e6662ac4e478e8ad7ed5d0b16258d&nonce=469d5890dd3845eeaf900ac8988ecd95&prompt=none&sso_reload=true 137:325 Uncaught DOMException: Failed to set the 'href' property on 'Location': Tโ€ฆrequest-id=5f269ce9-80c8-4cde-87a3-745d3ef76000'. [2020-07-02T20:51:53Z] [Info] http://localhost:52317/_framework/blazor.webassembly.js 0:2753 \"info: Microsoft.AspNetCore.Authorization.DefaultAuthorizationService[2]\ Authorization failed. These requirements were not met:\ DenyAnonymousAuthorizationRequirement: Requires an authenticated user.\" [2020-07-02T20:52:00Z] [Debug] https://localhost:52368/_framework/dotnet.5.0.0-preview.7.20326.1.js 0:124605 \"mono_wasm_runtime_ready\" \"fe00e07a-5519-4dfe-b35a-f867dbaf2e28\" [2020-07-02T20:52:00Z] [Severe] https://localhost:52368/_framework/blazor.webassembly.js 0:40764 RangeError: Maximum call stack size exceeded at mono_class_get_first_method_idx (https://localhost:52368/_framework/dotnet.wasm:wasm-function[603]:0x12964) at mono_find_method_in_metadata (https://localhost:52368/_framework/dotnet.wasm:wasm-function[2286]:0x605f2) at mono_class_get_method_from_name_checked (https://localhost:52368/_framework/dotnet.wasm:wasm-function[198]:0x50bd) at mono_class_get_cctor (https://localhost:52368/_framework/dotnet.wasm:wasm-function[1566]:0x42b21) at mono_runtime_class_init_full (https://localhost:52368/_framework/dotnet.wasm:wasm-function[233]:0x5c71) at mono_interp_transform_method (https://localhost:52368/_framework/dotnet.wasm:wasm-function[1144]:0x24a96) at do_transform_method (https://localhost:52368/_framework/dotnet.wasm:wasm-function[2694]:0x7e075) at interp_exec_method (https://localhost:52368/_framework/dotnet.wasm:wasm-function[1224]:0x353df) at interp_runtime_invoke (https://localhost:52368/_framework/dotnet.wasm:wasm-function[4966]:0xceb93) at mono_jit_runtime_invoke (https://localhost:52368/_framework/dotnet.wasm:wasm-function[4466]:0xb8932) Page content: <head> <meta charset=\"utf-8\"> <meta name=\"viewport\" content=\"width=device-width, initial-scale=1.0, maximum-scale=1.0, user-scalable=no\"> <title>AspNet.offwtaofawy</title> <base href=\"/\"> <link href=\"css/bootstrap/bootstrap.min.css\" rel=\"stylesheet\"> <link href=\"css/app.css\" rel=\"stylesheet\"> </head> <body> <app>Loading...</app> <div id=\"blazor-error-ui\" style=\"display: block;\"> An unhandled error has occurred. <a href=\"\" class=\"reload\">Reload</a> <a class=\"dismiss\">??</a> </div> <script src=\"_content/Microsoft.AspNetCore.Components.WebAssembly.Authentication/AuthenticationService.js\"></script> <script src=\"_framework/blazor.webassembly.js\"></script> <script type=\"text/javascript\">var Module; window.__wasmmodulecallback__(); delete window.__wasmmodulecallback__;</script><script src=\"_framework/dotnet.5.0.0-preview.7.20326.1.js\" defer=\"\" integrity=\"sha256-NInsB2Gm0V2B80TQrCgo8y9wzIddZvMysAFwSpIaejo=\" crossorigin=\"anonymous\"></script></body> ---- Assert.NotEmpty() Failure at Microsoft.AspNetCore.E2ETesting.WaitAssert.WaitAssertCore[TResult](IWebDriver driver, Func`1 assertion, TimeSpan timeout) in /_/src/Shared/E2ETesting/WaitAssert.cs:line 126 at Microsoft.AspNetCore.E2ETesting.WaitAssert.Exists(IWebDriver driver, By finder, TimeSpan timeout) in /_/src/Shared/E2ETesting/WaitAssert.cs:line 64 a ```
test
quarantinedtest blazorwasmstandalonetemplate individualauth works openqa selenium browserassertfailedexception xunit sdk notemptyexception assert notempty failure at xunit assert notempty ienumerable collection in c dev xunit xunit src xunit assert asserts collectionasserts cs line at microsoft aspnetcore waitassert c b in src shared waitassert cs line at microsoft aspnetcore waitassert c b iwebdriver in src shared waitassert cs line screen shot captured at f workspace work s artifacts testresults release blazortemplates tests png encountered browser errors mono wasm runtime ready bsso telemetry result error error noextension type chromessotelemetry data traces unsafe javascript attempt to initiate navigation for frame with url from frame with url the frame attempting navigation is targeting its top level window but is neither same origin with its target nor has it received a user gesture see uncaught domexception failed to set the href property on location tโ€ฆrequest id info microsoft aspnetcore authorization defaultauthorizationservice authorization failed these requirements were not met denyanonymousauthorizationrequirement requires an authenticated user mono wasm runtime ready rangeerror maximum call stack size exceeded at mono class get first method idx at mono find method in metadata at mono class get method from name checked at mono class get cctor at mono runtime class init full at mono interp transform method at do transform method at interp exec method at interp runtime invoke at mono jit runtime invoke page content aspnet offwtaofawy loading an unhandled error has occurred reload var module window wasmmodulecallback delete window wasmmodulecallback assert notempty failure at microsoft aspnetcore waitassert waitassertcore iwebdriver driver func assertion timespan timeout in src shared waitassert cs line at microsoft aspnetcore waitassert exists iwebdriver driver by finder timespan timeout in src shared waitassert cs line a
1
605,407
18,734,746,780
IssuesEvent
2021-11-04 05:08:18
opensrp/opensrp-server-web
https://api.github.com/repos/opensrp/opensrp-server-web
closed
Task Status for Thailand Productino not Updating Leading to blank WebUI Dashboard
Priority: High
BVBD has reported that there are mismatches between the app and web UI. For example, for plan ID 4daf4c34-3f38-495f-969f-dcfc5c2dd411, they see 27 tests and BCC on the android client, but on the web UI they only see BCC event. Checks have been made that there are blood screening events on OpenSRP with the following endpoint https://servermhealth.ddc.moph.go.th/opensrp/rest/event/sync?locationId=be847cbb-576b-4d76-b9da-f59175f74dcb&serverVersion=0&limit=10000 A check on OpenSRP Server DB showed there is no complete task: SELECT count(*) FROM raw_tasks WHERE full_json->>'planIdentifier' = '4daf4c34-3f38-495f-969f-dcfc5c2dd411' AND full_json->>'groupIdentifier' = 'be847cbb-576b-4d76-b9da-f59175f74dcb' AND full_json->>'code' = 'Blood Screening' AND full_json->>'businessStatus' = 'Complete'; This seems to be the same issue we had encountered on Thailand. Could we run a script to correct this for all affected plans?
1.0
Task Status for Thailand Productino not Updating Leading to blank WebUI Dashboard - BVBD has reported that there are mismatches between the app and web UI. For example, for plan ID 4daf4c34-3f38-495f-969f-dcfc5c2dd411, they see 27 tests and BCC on the android client, but on the web UI they only see BCC event. Checks have been made that there are blood screening events on OpenSRP with the following endpoint https://servermhealth.ddc.moph.go.th/opensrp/rest/event/sync?locationId=be847cbb-576b-4d76-b9da-f59175f74dcb&serverVersion=0&limit=10000 A check on OpenSRP Server DB showed there is no complete task: SELECT count(*) FROM raw_tasks WHERE full_json->>'planIdentifier' = '4daf4c34-3f38-495f-969f-dcfc5c2dd411' AND full_json->>'groupIdentifier' = 'be847cbb-576b-4d76-b9da-f59175f74dcb' AND full_json->>'code' = 'Blood Screening' AND full_json->>'businessStatus' = 'Complete'; This seems to be the same issue we had encountered on Thailand. Could we run a script to correct this for all affected plans?
non_test
task status for thailand productino not updating leading to blank webui dashboard bvbd has reported that there are mismatches between the app and web ui for example for plan id they see tests and bcc on the android client but on the web ui they only see bcc event checks have been made that there are blood screening events on opensrp with the following endpoint a check on opensrp server db showed there is no complete task select count from raw tasks where full json planidentifier and full json groupidentifier and full json code blood screening and full json businessstatus complete this seems to be the same issue we had encountered on thailand could we run a script to correct this for all affected plans
0
138,957
11,222,942,647
IssuesEvent
2020-01-07 21:24:08
IntellectualSites/PlotSquared
https://api.github.com/repos/IntellectualSites/PlotSquared
opened
Setting Border VIA Api cause error in console
[?] Testing Required
<!--- READ THIS BEFORE SUBMITTING AN ISSUE REPORT!!! --> <!--- ##### DO NOT REMOVE THIS TEMPLATE! YOUR ISSUE *WILL* FIT IN IT! ##### --> <!--- # NOTICE: ```diff ! PlotSquared for Minecraft Java Edition versions between 1.7 through to 1.12.2 are considered ! legacy, and will receive limited to no support. Please consider upgrading to 1.13+ for ! future support. Plugins exist for 1.13+ which bring back behaviors found in 1.8.8 ! All versions of PlotSquared for Sponge and Nukkit(X) will receive limited to no support ! due to lack of developer interest and time. Additionally, NukkitX has not had feature ! updates since the Better Together, which prevents some PlotSquared features from ever ! functioning. Contributions are always welcome however! ``` **Feature requests & Suggestions are to be submitted at the [PlotSquared Suggestions tracker](https://github.com/IntellectualSites/PlotSquaredSuggestions)** **Code contributions are to be done through [PRs](https://help.github.com/en/github/collaborating-with-issues-and-pull-requests/creating-a-pull-request), tagging the specific issue ticket(s) if applicable.** **[DISCORD INVITE LINK](https://discord.gg/cSMxtGn)** and please, for the love of the little sanity we have left, use the correct channels! --> # Bug Report Template: <!--- Incomplete reports will most likely be marked as invalid, and closed, with few exceptions.--> ## Required Information section: > ALL FIELDS IN THIS SECTION ARE REQUIRED, and must contain appropriate information ### Server config info (/plot debugpaste / file paste links): <!--- Issue /plot debugpaste in game or in your console and copy the supplied URL here --> <!--- If you cannot perform the above, we require logs/latest.log; settings.yml; worlds.yml and possibly PlotSquared.use_THIS.yml --> <!--- If you are unwilling to supply the information we need, we reserve the right to not assist you. Redact IP addresses if you need to. --> https://athion.net/ISPaster/paste/view/489737d9c83b4087ab62c20d2f59cc09 ### Server type: **Select one** <!-- Select the type you are reporting the issue for (put an "X" between of brackets): --> - [x] Spigot / Paper *(CraftBukkit should not be used, re-test with Spigot first!)* - [] Sponge *- NOTE: NOT ACTIVELY MAINTAINED* - [] NukkitX *- NOTE: NOT ACTIVELY MAINTAINED* ### Minecraft Version: **Select one** <!-- Select the type you are reporting the issue for (put an "X" between of brackets): The maintained versions are 1.14.4 and 1.15.x --> - [] Minecraft 1.15 - [x] Minecraft 1.14.4 - [] Minecraft 1.13.2 - [] Minecraft Java Edition *other versions, please specify*: - [] Minecraft Bedrock Edition *specify version*: ### Server build info: <!--- Run /version in-game or in console & paste the full output here: --> ``` This server is running Paper version git-Paper-226 (MC: 1.14.4) (Implementing API version 1.14.4-R0.1-SNAPSHOT) ``` ### WorldEdit/FAWE versions: <!--- Specify which plugin you are using, and add its version --> - [] FAWE *version*: - [x] WorldEdit *version*: 7.0.1;61bc012 ### Description of the problem: <!--- Be as specific as possible. Don't lie, redact information, or use false names/situations. --> <!--- Who, What, When, Where, Why, How, Expected behavior, Resultant behavior, etc --> If i set the Border via API to beacon and change it to a other material i getting my console spammed with errors ### How to replicate: <!--- If you can reproduce the issue please tell us as detailed as possible step by step how to do that --> Write a plugin that replaces the border and change to beacon and than back zu a other material ## Additional Information: > The information here is optional for you to provide, however it may help us to more readily diagnose any compatibility and bug issues. ### Other plugins being used on the server: <!--- Optional but recommended - issue "/plugins" in-game or in console and copy/paste the list --> ProxySockeSystem, WorldGuard, WorldEdit ### Relevant console output, log lines, and/or screenshots: <!--- Please use in-line code insertion ``` like this ``` for short (20 lines or less) text blobs, or a paste service for large blobs --> https://hasteb.in/ragudeza.css ### Additional relevant comments/remarks: <!--- Use this space to give us any additional information which may be relevant to this issue, such as: if you are using a Minecraft hosting provider; unusual installation environment; etc --> # AFFIRMATION OF COMPLETION: <!-- Make sure you have completed the following steps (put an "X" between of brackets): --> - [x] I included all information required in the sections above - [x] I made sure there are no duplicates of this report [(Use Search)](https://github.com/IntellectualSites/PlotSquared/issues?utf8=%E2%9C%93&q=is%3Aissue) - [x] I made sure I am using an up-to-date version of PlotSquared - [x] I made sure the bug/error is not caused by any other plugin - [] I didn't read but checked everything above.
1.0
Setting Border VIA Api cause error in console - <!--- READ THIS BEFORE SUBMITTING AN ISSUE REPORT!!! --> <!--- ##### DO NOT REMOVE THIS TEMPLATE! YOUR ISSUE *WILL* FIT IN IT! ##### --> <!--- # NOTICE: ```diff ! PlotSquared for Minecraft Java Edition versions between 1.7 through to 1.12.2 are considered ! legacy, and will receive limited to no support. Please consider upgrading to 1.13+ for ! future support. Plugins exist for 1.13+ which bring back behaviors found in 1.8.8 ! All versions of PlotSquared for Sponge and Nukkit(X) will receive limited to no support ! due to lack of developer interest and time. Additionally, NukkitX has not had feature ! updates since the Better Together, which prevents some PlotSquared features from ever ! functioning. Contributions are always welcome however! ``` **Feature requests & Suggestions are to be submitted at the [PlotSquared Suggestions tracker](https://github.com/IntellectualSites/PlotSquaredSuggestions)** **Code contributions are to be done through [PRs](https://help.github.com/en/github/collaborating-with-issues-and-pull-requests/creating-a-pull-request), tagging the specific issue ticket(s) if applicable.** **[DISCORD INVITE LINK](https://discord.gg/cSMxtGn)** and please, for the love of the little sanity we have left, use the correct channels! --> # Bug Report Template: <!--- Incomplete reports will most likely be marked as invalid, and closed, with few exceptions.--> ## Required Information section: > ALL FIELDS IN THIS SECTION ARE REQUIRED, and must contain appropriate information ### Server config info (/plot debugpaste / file paste links): <!--- Issue /plot debugpaste in game or in your console and copy the supplied URL here --> <!--- If you cannot perform the above, we require logs/latest.log; settings.yml; worlds.yml and possibly PlotSquared.use_THIS.yml --> <!--- If you are unwilling to supply the information we need, we reserve the right to not assist you. Redact IP addresses if you need to. --> https://athion.net/ISPaster/paste/view/489737d9c83b4087ab62c20d2f59cc09 ### Server type: **Select one** <!-- Select the type you are reporting the issue for (put an "X" between of brackets): --> - [x] Spigot / Paper *(CraftBukkit should not be used, re-test with Spigot first!)* - [] Sponge *- NOTE: NOT ACTIVELY MAINTAINED* - [] NukkitX *- NOTE: NOT ACTIVELY MAINTAINED* ### Minecraft Version: **Select one** <!-- Select the type you are reporting the issue for (put an "X" between of brackets): The maintained versions are 1.14.4 and 1.15.x --> - [] Minecraft 1.15 - [x] Minecraft 1.14.4 - [] Minecraft 1.13.2 - [] Minecraft Java Edition *other versions, please specify*: - [] Minecraft Bedrock Edition *specify version*: ### Server build info: <!--- Run /version in-game or in console & paste the full output here: --> ``` This server is running Paper version git-Paper-226 (MC: 1.14.4) (Implementing API version 1.14.4-R0.1-SNAPSHOT) ``` ### WorldEdit/FAWE versions: <!--- Specify which plugin you are using, and add its version --> - [] FAWE *version*: - [x] WorldEdit *version*: 7.0.1;61bc012 ### Description of the problem: <!--- Be as specific as possible. Don't lie, redact information, or use false names/situations. --> <!--- Who, What, When, Where, Why, How, Expected behavior, Resultant behavior, etc --> If i set the Border via API to beacon and change it to a other material i getting my console spammed with errors ### How to replicate: <!--- If you can reproduce the issue please tell us as detailed as possible step by step how to do that --> Write a plugin that replaces the border and change to beacon and than back zu a other material ## Additional Information: > The information here is optional for you to provide, however it may help us to more readily diagnose any compatibility and bug issues. ### Other plugins being used on the server: <!--- Optional but recommended - issue "/plugins" in-game or in console and copy/paste the list --> ProxySockeSystem, WorldGuard, WorldEdit ### Relevant console output, log lines, and/or screenshots: <!--- Please use in-line code insertion ``` like this ``` for short (20 lines or less) text blobs, or a paste service for large blobs --> https://hasteb.in/ragudeza.css ### Additional relevant comments/remarks: <!--- Use this space to give us any additional information which may be relevant to this issue, such as: if you are using a Minecraft hosting provider; unusual installation environment; etc --> # AFFIRMATION OF COMPLETION: <!-- Make sure you have completed the following steps (put an "X" between of brackets): --> - [x] I included all information required in the sections above - [x] I made sure there are no duplicates of this report [(Use Search)](https://github.com/IntellectualSites/PlotSquared/issues?utf8=%E2%9C%93&q=is%3Aissue) - [x] I made sure I am using an up-to-date version of PlotSquared - [x] I made sure the bug/error is not caused by any other plugin - [] I didn't read but checked everything above.
test
setting border via api cause error in console notice diff plotsquared for minecraft java edition versions between through to are considered legacy and will receive limited to no support please consider upgrading to for future support plugins exist for which bring back behaviors found in all versions of plotsquared for sponge and nukkit x will receive limited to no support due to lack of developer interest and time additionally nukkitx has not had feature updates since the better together which prevents some plotsquared features from ever functioning contributions are always welcome however feature requests suggestions are to be submitted at the code contributions are to be done through tagging the specific issue ticket s if applicable and please for the love of the little sanity we have left use the correct channels bug report template required information section all fields in this section are required and must contain appropriate information server config info plot debugpaste file paste links server type select one spigot paper craftbukkit should not be used re test with spigot first sponge note not actively maintained nukkitx note not actively maintained minecraft version select one select the type you are reporting the issue for put an x between of brackets the maintained versions are and x minecraft minecraft minecraft minecraft java edition other versions please specify minecraft bedrock edition specify version server build info this server is running paper version git paper mc implementing api version snapshot worldedit fawe versions fawe version worldedit version description of the problem if i set the border via api to beacon and change it to a other material i getting my console spammed with errors how to replicate write a plugin that replaces the border and change to beacon and than back zu a other material additional information the information here is optional for you to provide however it may help us to more readily diagnose any compatibility and bug issues other plugins being used on the server proxysockesystem worldguard worldedit relevant console output log lines and or screenshots please use in line code insertion like this for short lines or less text blobs or a paste service for large blobs additional relevant comments remarks affirmation of completion i included all information required in the sections above i made sure there are no duplicates of this report i made sure i am using an up to date version of plotsquared i made sure the bug error is not caused by any other plugin i didn t read but checked everything above
1
659,906
21,944,841,416
IssuesEvent
2022-05-23 22:36:44
grpc/grpc
https://api.github.com/repos/grpc/grpc
closed
Server is stuck in CompletionQueue::Next when multiple server streaming request are sent from client
kind/bug lang/c++ priority/P2 untriaged
<!-- This form is for bug reports and feature requests ONLY! For general questions and troubleshooting, please ask/look for answers here: - grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io - StackOverflow, with "grpc" tag: https://stackoverflow.com/questions/tagged/grpc Issues specific to *grpc-java*, *grpc-go*, *grpc-node*, *grpc-dart*, *grpc-web* should be created in the repository they belong to (e.g. https://github.com/grpc/grpc-LANGUAGE/issues/new) --> ### What version of gRPC and what language are you using? https://github.com/grpc/grpc/archive/b54a5b338637f92bfcf4b0bc05e0f57a5fd8fadd.tar.gz The nearest tag is v1.27.0-pre1 and I'm using C++. ### What operating system (Linux, Windows,...) and version? Ubuntu 18.04 ### What runtime / compiler are you using (e.g. python version or version of gcc) Clang8.0 ### What did you do? I have one service(A) and the service has two server streaming RPC(A1 and A2). The main thread in client sends RPC request A1 then RPC request A2. If there is no delay between these 2 RPC requests, the client is able to receive both A1 and A2 responses from server. However, if I add a delay between 2 RPC requests, client only receives the response for the first sent RPC request. Besides the main thread, client has two threads to read the responses. The following class includes the code for the client to send RPC request A1 and the thread function to receive A1 response(I have another class ClientA2 for RPC A2). ``` class ClientA1 { public: explicit ClientA1(std::shared_ptr<Channel> channel) : stub_(helloworld::ServiceA::NewStub(channel)) {} void MethodA1() { helloworld::MethodA1Request request; call_.response_reader = stub_->AsyncMethodA1(&call_.context, request, &cq_, (void *)&call_); } // Loop while listening for completed responses. Run in a thread. void AsyncCompleteRpc() { void *got_tag; bool ok = false; // Block until the next result is available in the completion queue "cq". while (cq_.Next(&got_tag, &ok)) { // The tag in this example is the memory location of the call object AsyncClientCall *call = static_cast<AsyncClientCall *>(got_tag); if (ok) { call->response_reader->Read(&call->response, (void *)call); } else { call->response_reader->Finish(&call->status, (void *)call); break; } } // Deplete the completion queue while (cq_.Next(&got_tag, &ok)) { } } private: // struct for keeping state and data information struct AsyncClientCall { helloworld::MethodA1Response response; ClientContext context; Status status; std::unique_ptr<ClientAsyncReader<helloworld::MethodA1Response>> response_reader; }; std::unique_ptr<helloworld::ServiceA::Stub> stub_; CompletionQueue cq_; AsyncClientCall call_; }; ``` Main thread in client creates a ClientA1 object and a ClientA2 object, spawns threads to handle responses, and call MethodA1 and MethodA2 to send RPC requests to server. A 5-second delay is added between sending A1 request and A2 request. ``` threads.emplace_back(&ClientA1::AsyncCompleteRpc, &client_a1); client_a1.MethodA1(); // Add a delay. std::this_thread::sleep_for(std::chrono::milliseconds(5000)); threads.emplace_back(&ClientA2::AsyncCompleteRpc, &client_a2); client_a2.MethodA2(); ``` Server has two completion queues and spawns two threads to check the completion queue. The tag pushed to and pop from completion queue is defined as follows: ``` enum class Rpc { A1, A2}; struct CallData { CallData(const Rpc &rpc) : rpc_(rpc), response_cnt_(0), status_(CallStatus::CREATE) {} Rpc rpc_; // How many responses have been sent to the client. int response_cnt_; // State machine with the following states. enum class CallStatus { CREATE, PROCESS, FINISH }; CallStatus status_; // The current serving state. }; ``` Server run the following initialization to spawns two threads. ``` // Generate ServerAsyncWriter for A1 response and A2 response, respectively. std::shared_ptr<ServerAsyncWriter<helloworld::MethodA1Response>> a1_responder = std::make_shared<ServerAsyncWriter<helloworld::MethodA1Response>>( &ctx[0]); std::shared_ptr<ServerAsyncWriter<helloworld::MethodA2Response>> a2_responder = std::make_shared<ServerAsyncWriter<helloworld::MethodA2Response>>( &ctx[1]); // Spawn thread to check completion queue - cq[0] call_data.emplace_back(std::make_unique<CallData>(Rpc::A1)); threads_.emplace_back(&ServerImpl::HandleRpc, this, cq_[0].get(), a1_responder, a2_responder); helloworld::MethodA1Request request; call_data.back()->status_ = CallData::CallStatus::PROCESS; // Request RPC service_a_.RequestMethodA1(&ctx[0], &request, a1_responder.get(), cq_[0].get(), cq_[0].get(), (void *)call_data.back().get()); // Spawn another thread to check completion queue - cq[1] and then request RPC A2(same as above) ``` The function run in thread ServerImpl::HandleRpc basically checks the type of the tag then call ServerAsyncWriter::Write() to write responses. For example, if the type of the tag(CallData::rpc_) is Rpc::A1, a1_responder->Write() will be called. ### What did you expect to see? I expected client receives both server streaming responses. ### What did you see instead? Without the delay in client, client receives both server streaming responses. However, with delay, client only receives the first response. ### Anything else we should know about your project / environment? No.
1.0
Server is stuck in CompletionQueue::Next when multiple server streaming request are sent from client - <!-- This form is for bug reports and feature requests ONLY! For general questions and troubleshooting, please ask/look for answers here: - grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io - StackOverflow, with "grpc" tag: https://stackoverflow.com/questions/tagged/grpc Issues specific to *grpc-java*, *grpc-go*, *grpc-node*, *grpc-dart*, *grpc-web* should be created in the repository they belong to (e.g. https://github.com/grpc/grpc-LANGUAGE/issues/new) --> ### What version of gRPC and what language are you using? https://github.com/grpc/grpc/archive/b54a5b338637f92bfcf4b0bc05e0f57a5fd8fadd.tar.gz The nearest tag is v1.27.0-pre1 and I'm using C++. ### What operating system (Linux, Windows,...) and version? Ubuntu 18.04 ### What runtime / compiler are you using (e.g. python version or version of gcc) Clang8.0 ### What did you do? I have one service(A) and the service has two server streaming RPC(A1 and A2). The main thread in client sends RPC request A1 then RPC request A2. If there is no delay between these 2 RPC requests, the client is able to receive both A1 and A2 responses from server. However, if I add a delay between 2 RPC requests, client only receives the response for the first sent RPC request. Besides the main thread, client has two threads to read the responses. The following class includes the code for the client to send RPC request A1 and the thread function to receive A1 response(I have another class ClientA2 for RPC A2). ``` class ClientA1 { public: explicit ClientA1(std::shared_ptr<Channel> channel) : stub_(helloworld::ServiceA::NewStub(channel)) {} void MethodA1() { helloworld::MethodA1Request request; call_.response_reader = stub_->AsyncMethodA1(&call_.context, request, &cq_, (void *)&call_); } // Loop while listening for completed responses. Run in a thread. void AsyncCompleteRpc() { void *got_tag; bool ok = false; // Block until the next result is available in the completion queue "cq". while (cq_.Next(&got_tag, &ok)) { // The tag in this example is the memory location of the call object AsyncClientCall *call = static_cast<AsyncClientCall *>(got_tag); if (ok) { call->response_reader->Read(&call->response, (void *)call); } else { call->response_reader->Finish(&call->status, (void *)call); break; } } // Deplete the completion queue while (cq_.Next(&got_tag, &ok)) { } } private: // struct for keeping state and data information struct AsyncClientCall { helloworld::MethodA1Response response; ClientContext context; Status status; std::unique_ptr<ClientAsyncReader<helloworld::MethodA1Response>> response_reader; }; std::unique_ptr<helloworld::ServiceA::Stub> stub_; CompletionQueue cq_; AsyncClientCall call_; }; ``` Main thread in client creates a ClientA1 object and a ClientA2 object, spawns threads to handle responses, and call MethodA1 and MethodA2 to send RPC requests to server. A 5-second delay is added between sending A1 request and A2 request. ``` threads.emplace_back(&ClientA1::AsyncCompleteRpc, &client_a1); client_a1.MethodA1(); // Add a delay. std::this_thread::sleep_for(std::chrono::milliseconds(5000)); threads.emplace_back(&ClientA2::AsyncCompleteRpc, &client_a2); client_a2.MethodA2(); ``` Server has two completion queues and spawns two threads to check the completion queue. The tag pushed to and pop from completion queue is defined as follows: ``` enum class Rpc { A1, A2}; struct CallData { CallData(const Rpc &rpc) : rpc_(rpc), response_cnt_(0), status_(CallStatus::CREATE) {} Rpc rpc_; // How many responses have been sent to the client. int response_cnt_; // State machine with the following states. enum class CallStatus { CREATE, PROCESS, FINISH }; CallStatus status_; // The current serving state. }; ``` Server run the following initialization to spawns two threads. ``` // Generate ServerAsyncWriter for A1 response and A2 response, respectively. std::shared_ptr<ServerAsyncWriter<helloworld::MethodA1Response>> a1_responder = std::make_shared<ServerAsyncWriter<helloworld::MethodA1Response>>( &ctx[0]); std::shared_ptr<ServerAsyncWriter<helloworld::MethodA2Response>> a2_responder = std::make_shared<ServerAsyncWriter<helloworld::MethodA2Response>>( &ctx[1]); // Spawn thread to check completion queue - cq[0] call_data.emplace_back(std::make_unique<CallData>(Rpc::A1)); threads_.emplace_back(&ServerImpl::HandleRpc, this, cq_[0].get(), a1_responder, a2_responder); helloworld::MethodA1Request request; call_data.back()->status_ = CallData::CallStatus::PROCESS; // Request RPC service_a_.RequestMethodA1(&ctx[0], &request, a1_responder.get(), cq_[0].get(), cq_[0].get(), (void *)call_data.back().get()); // Spawn another thread to check completion queue - cq[1] and then request RPC A2(same as above) ``` The function run in thread ServerImpl::HandleRpc basically checks the type of the tag then call ServerAsyncWriter::Write() to write responses. For example, if the type of the tag(CallData::rpc_) is Rpc::A1, a1_responder->Write() will be called. ### What did you expect to see? I expected client receives both server streaming responses. ### What did you see instead? Without the delay in client, client receives both server streaming responses. However, with delay, client only receives the first response. ### Anything else we should know about your project / environment? No.
non_test
server is stuck in completionqueue next when multiple server streaming request are sent from client this form is for bug reports and feature requests only for general questions and troubleshooting please ask look for answers here grpc io mailing list stackoverflow with grpc tag issues specific to grpc java grpc go grpc node grpc dart grpc web should be created in the repository they belong to e g what version of grpc and what language are you using the nearest tag is and i m using c what operating system linux windows and version ubuntu what runtime compiler are you using e g python version or version of gcc what did you do i have one service a and the service has two server streaming rpc and the main thread in client sends rpc request then rpc request if there is no delay between these rpc requests the client is able to receive both and responses from server however if i add a delay between rpc requests client only receives the response for the first sent rpc request besides the main thread client has two threads to read the responses the following class includes the code for the client to send rpc request and the thread function to receive response i have another class for rpc class public explicit std shared ptr channel stub helloworld servicea newstub channel void helloworld request call response reader stub call context request cq void call loop while listening for completed responses run in a thread void asynccompleterpc void got tag bool ok false block until the next result is available in the completion queue cq while cq next got tag ok the tag in this example is the memory location of the call object asyncclientcall call static cast got tag if ok call response reader read call response void call else call response reader finish call status void call break deplete the completion queue while cq next got tag ok private struct for keeping state and data information struct asyncclientcall helloworld response clientcontext context status status std unique ptr response reader std unique ptr stub completionqueue cq asyncclientcall call main thread in client creates a object and a object spawns threads to handle responses and call and to send rpc requests to server a second delay is added between sending request and request threads emplace back asynccompleterpc client client add a delay std this thread sleep for std chrono milliseconds threads emplace back asynccompleterpc client client server has two completion queues and spawns two threads to check the completion queue the tag pushed to and pop from completion queue is defined as follows enum class rpc struct calldata calldata const rpc rpc rpc rpc response cnt status callstatus create rpc rpc how many responses have been sent to the client int response cnt state machine with the following states enum class callstatus create process finish callstatus status the current serving state server run the following initialization to spawns two threads generate serverasyncwriter for response and response respectively std shared ptr responder std make shared ctx std shared ptr responder std make shared ctx spawn thread to check completion queue cq call data emplace back std make unique rpc threads emplace back serverimpl handlerpc this cq get responder responder helloworld request call data back status calldata callstatus process request rpc service a ctx request responder get cq get cq get void call data back get spawn another thread to check completion queue cq and then request rpc same as above the function run in thread serverimpl handlerpc basically checks the type of the tag then call serverasyncwriter write to write responses for example if the type of the tag calldata rpc is rpc responder write will be called what did you expect to see i expected client receives both server streaming responses what did you see instead without the delay in client client receives both server streaming responses however with delay client only receives the first response anything else we should know about your project environment no
0
1,477
2,856,866,937
IssuesEvent
2015-06-02 16:54:10
CellProfiler/CellProfiler
https://api.github.com/repos/CellProfiler/CellProfiler
opened
For measurement modules that display a chart, make clear the fact that an aggregate is shown
For discussion Layout and usability
For the Nth time, someone misunderstood that the chart shown in by a measurement module shows the aggregate values and thought it was the actual result (in this case, MeasureObjectIntensity *and* CalculateMath). You'd think the the module name and the mean/median/std as headers would clear that up, but alas no... Any suggestions as to how to make this absolutely clear? For example, could we put text above the chart saying "Values shown are per-image aggregates of the per-object values)"?
True
For measurement modules that display a chart, make clear the fact that an aggregate is shown - For the Nth time, someone misunderstood that the chart shown in by a measurement module shows the aggregate values and thought it was the actual result (in this case, MeasureObjectIntensity *and* CalculateMath). You'd think the the module name and the mean/median/std as headers would clear that up, but alas no... Any suggestions as to how to make this absolutely clear? For example, could we put text above the chart saying "Values shown are per-image aggregates of the per-object values)"?
non_test
for measurement modules that display a chart make clear the fact that an aggregate is shown for the nth time someone misunderstood that the chart shown in by a measurement module shows the aggregate values and thought it was the actual result in this case measureobjectintensity and calculatemath you d think the the module name and the mean median std as headers would clear that up but alas no any suggestions as to how to make this absolutely clear for example could we put text above the chart saying values shown are per image aggregates of the per object values
0
198,757
14,996,229,503
IssuesEvent
2021-01-29 15:19:40
umrover/mrover-workspace
https://api.github.com/repos/umrover/mrover-workspace
closed
Is the Voxel Filter Necessary
C/C++ enhancement percep testing
Currently we run a filter which down samples our points, by clustering together neighboring points. We feel that this may be an unnecessary step, and that we can get the same results, by reducing the camera resolution. We would like to test our processing speed if this were to be performed.
1.0
Is the Voxel Filter Necessary - Currently we run a filter which down samples our points, by clustering together neighboring points. We feel that this may be an unnecessary step, and that we can get the same results, by reducing the camera resolution. We would like to test our processing speed if this were to be performed.
test
is the voxel filter necessary currently we run a filter which down samples our points by clustering together neighboring points we feel that this may be an unnecessary step and that we can get the same results by reducing the camera resolution we would like to test our processing speed if this were to be performed
1
47,890
2,986,984,086
IssuesEvent
2015-07-20 09:21:59
52North/SOS
https://api.github.com/repos/52North/SOS
opened
Add JSON GetObservation example with MergeObservationsIntoDataArray extension
enhancement low priority
Add JSON GetObservation example with MergeObservationsIntoDataArray extension like this: ``` "extensions": [ { "definition": "MergeObservationsIntoDataArray", "value": true } ] ``` https://wiki.52north.org/bin/view/SensorWeb/SensorObservationServiceIVDocumentation#MergeObservationsIntoDataArray
1.0
Add JSON GetObservation example with MergeObservationsIntoDataArray extension - Add JSON GetObservation example with MergeObservationsIntoDataArray extension like this: ``` "extensions": [ { "definition": "MergeObservationsIntoDataArray", "value": true } ] ``` https://wiki.52north.org/bin/view/SensorWeb/SensorObservationServiceIVDocumentation#MergeObservationsIntoDataArray
non_test
add json getobservation example with mergeobservationsintodataarray extension add json getobservation example with mergeobservationsintodataarray extension like this extensions definition mergeobservationsintodataarray value true
0
36,699
2,811,733,196
IssuesEvent
2015-05-18 00:23:10
openshift/origin
https://api.github.com/repos/openshift/origin
closed
self-subject access reviews (IsPersonalSubjectAccessReview) doesn't work
component/auth kind/bug priority/P2
Please see following output. ~~~ $ _output/local/go/bin/osc --v=4 get IsPersonalSubjectAccessReview I0517 21:23:47.755036 6852 selector.go:53] Unable to list "ispersonalsubjectaccessreviews": the server could not find the requested resource ~~~ Even though `printIsPersonalSubjectAccessReview()` is implemented in [`pkg/cmd/cli/describe/printer.go`](https://github.com/openshift/origin/blob/53e8a0d27c6a7e8ee663898a437fde6908c368c5/pkg/cmd/cli/describe/printer.go#L449-L452), its rest API is not implemented and `master.go` doesn't register the API during the start. However, according to this discussion [pull/139](https://github.com/openshift/origin/pull/1392), it is not clear if `IsPersonalSubjectAccessReview` will be obsolete. I open here as a new issue to track.
1.0
self-subject access reviews (IsPersonalSubjectAccessReview) doesn't work - Please see following output. ~~~ $ _output/local/go/bin/osc --v=4 get IsPersonalSubjectAccessReview I0517 21:23:47.755036 6852 selector.go:53] Unable to list "ispersonalsubjectaccessreviews": the server could not find the requested resource ~~~ Even though `printIsPersonalSubjectAccessReview()` is implemented in [`pkg/cmd/cli/describe/printer.go`](https://github.com/openshift/origin/blob/53e8a0d27c6a7e8ee663898a437fde6908c368c5/pkg/cmd/cli/describe/printer.go#L449-L452), its rest API is not implemented and `master.go` doesn't register the API during the start. However, according to this discussion [pull/139](https://github.com/openshift/origin/pull/1392), it is not clear if `IsPersonalSubjectAccessReview` will be obsolete. I open here as a new issue to track.
non_test
self subject access reviews ispersonalsubjectaccessreview doesn t work please see following output output local go bin osc v get ispersonalsubjectaccessreview selector go unable to list ispersonalsubjectaccessreviews the server could not find the requested resource even though printispersonalsubjectaccessreview is implemented in its rest api is not implemented and master go doesn t register the api during the start however according to this discussion it is not clear if ispersonalsubjectaccessreview will be obsolete i open here as a new issue to track
0
282,460
24,477,952,701
IssuesEvent
2022-10-08 12:26:47
windmill-labs/windmill
https://api.github.com/repos/windmill-labs/windmill
closed
e2e tests for backend flow executions
test
No UI, only backend + postgres up, run a dozen different flows in a few different scenarii. E.g: - Single worker + no forloop - Multiple worker + no forloop - Multiple worker with one worker going down midway + no forloop + all the above with forloop cc @sqwishy
1.0
e2e tests for backend flow executions - No UI, only backend + postgres up, run a dozen different flows in a few different scenarii. E.g: - Single worker + no forloop - Multiple worker + no forloop - Multiple worker with one worker going down midway + no forloop + all the above with forloop cc @sqwishy
test
tests for backend flow executions no ui only backend postgres up run a dozen different flows in a few different scenarii e g single worker no forloop multiple worker no forloop multiple worker with one worker going down midway no forloop all the above with forloop cc sqwishy
1
119,865
10,076,107,311
IssuesEvent
2019-07-24 15:32:49
Th3-Fr3d/pmdbs
https://api.github.com/repos/Th3-Fr3d/pmdbs
closed
Server Pen Testing
high priority testing
- Verify that all SQLite queries using user input have **DatabaseManagement.Security()** Methods implemented to check for SQL Injection. - Look for arbitrary code execution vulnerabilities --> **Config file validation**
1.0
Server Pen Testing - - Verify that all SQLite queries using user input have **DatabaseManagement.Security()** Methods implemented to check for SQL Injection. - Look for arbitrary code execution vulnerabilities --> **Config file validation**
test
server pen testing verify that all sqlite queries using user input have databasemanagement security methods implemented to check for sql injection look for arbitrary code execution vulnerabilities config file validation
1
109,233
9,371,727,802
IssuesEvent
2019-04-03 15:58:54
medialab/fonio
https://api.github.com/repos/medialab/fonio
closed
When pasting content containing a note containing a contextualization to a note, the related contextualizations should be filtered out
bug priority:high unit test
This causes related unit tests to fail (which is legit) This requires to : * filter out related contextualizations to create * filter out related contextualizers * not create resources implied only by those contextualizations
1.0
When pasting content containing a note containing a contextualization to a note, the related contextualizations should be filtered out - This causes related unit tests to fail (which is legit) This requires to : * filter out related contextualizations to create * filter out related contextualizers * not create resources implied only by those contextualizations
test
when pasting content containing a note containing a contextualization to a note the related contextualizations should be filtered out this causes related unit tests to fail which is legit this requires to filter out related contextualizations to create filter out related contextualizers not create resources implied only by those contextualizations
1
290,733
21,897,647,878
IssuesEvent
2022-05-20 10:11:15
r5py/r5py
https://api.github.com/repos/r5py/r5py
closed
Setup Binder environment for r5py
documentation enhancement
With the new theme, it would be very easy to launch a binder and test the tool immediately in a browser. We might want to try and setup that at some point. This requires a bit of testing because there are quite specific dependencies with this package (namely Java DK and r5)
1.0
Setup Binder environment for r5py - With the new theme, it would be very easy to launch a binder and test the tool immediately in a browser. We might want to try and setup that at some point. This requires a bit of testing because there are quite specific dependencies with this package (namely Java DK and r5)
non_test
setup binder environment for with the new theme it would be very easy to launch a binder and test the tool immediately in a browser we might want to try and setup that at some point this requires a bit of testing because there are quite specific dependencies with this package namely java dk and
0
15,549
10,318,844,627
IssuesEvent
2019-08-30 15:53:00
dockstore/dockstore
https://api.github.com/repos/dockstore/dockstore
closed
Workflow and Tool Versions Should be Immutable
enhancement web-service
## Feature Request Workflow and tool versions should be immutable. While they currently are for hosted entries, they are not for entries that come from GitHub and other source control systems -- Dockstore versions are sourced from from GitHub tags, and GitHub tags can be moved, deleted, etc. When you refresh a workflow, if the GitHub tag was moved, the Dockstore version will reflect the new tag. It would also be nice to be aware of Docker containers, e.g., a workflow refers to a Docker image and the Docker image changes (in that case we could issue a warning). The motivation for this is so that papers/experiments can refer to a specific version of a Dockstore workflow/tool with confidence that the workflow/tool does not change. A lot of details need to ironed out on how to implement this. โ”†Issue is synchronized with this [Jira Epic](https://ucsc-cgl.atlassian.net/browse/DOCK-274) โ”†Issue Type: Epic โ”†Fix Versions: Dockstore 2.X โ”†Issue Number: DOCK-274
1.0
Workflow and Tool Versions Should be Immutable - ## Feature Request Workflow and tool versions should be immutable. While they currently are for hosted entries, they are not for entries that come from GitHub and other source control systems -- Dockstore versions are sourced from from GitHub tags, and GitHub tags can be moved, deleted, etc. When you refresh a workflow, if the GitHub tag was moved, the Dockstore version will reflect the new tag. It would also be nice to be aware of Docker containers, e.g., a workflow refers to a Docker image and the Docker image changes (in that case we could issue a warning). The motivation for this is so that papers/experiments can refer to a specific version of a Dockstore workflow/tool with confidence that the workflow/tool does not change. A lot of details need to ironed out on how to implement this. โ”†Issue is synchronized with this [Jira Epic](https://ucsc-cgl.atlassian.net/browse/DOCK-274) โ”†Issue Type: Epic โ”†Fix Versions: Dockstore 2.X โ”†Issue Number: DOCK-274
non_test
workflow and tool versions should be immutable feature request workflow and tool versions should be immutable while they currently are for hosted entries they are not for entries that come from github and other source control systems dockstore versions are sourced from from github tags and github tags can be moved deleted etc when you refresh a workflow if the github tag was moved the dockstore version will reflect the new tag it would also be nice to be aware of docker containers e g a workflow refers to a docker image and the docker image changes in that case we could issue a warning the motivation for this is so that papers experiments can refer to a specific version of a dockstore workflow tool with confidence that the workflow tool does not change a lot of details need to ironed out on how to implement this โ”†issue is synchronized with this โ”†issue type epic โ”†fix versions dockstore x โ”†issue number dock
0
226,862
18,045,614,856
IssuesEvent
2021-09-18 21:05:07
ceph/ceph-csi
https://api.github.com/repos/ceph/ceph-csi
closed
rbd-nbd does not support `--options notrim`, fails mapping thick-provisioned volumes
bug wontfix need test case component/rbd
# Describe the bug # Enabling both thick-provisioning and the `rbd-nbd` mounter in a StorageClass causes mapping RBD images to fail: ``` rbd: map failed with error an error (exit status 22) occurred while running rbd args: [--id cephcsi-rbd-node -m rook-ceph-mon-a.rook-ceph.svc.cluster.local:6789 --keyfile=***stripped*** map replicapool/csi-vol-2ff8537d-ce8b-11eb-8cea-f696da6dd1c4 --device-type nbd --options notrim], rbd error output: rbd-nbd: unknown args: --notrim ``` It seems that the `--options notrim` is not accepted when using nbd. There might be an alternative option, or we should return a failure while provisioning when both thick-provisioning and nbd are enabled. cc: @pkalever
1.0
rbd-nbd does not support `--options notrim`, fails mapping thick-provisioned volumes - # Describe the bug # Enabling both thick-provisioning and the `rbd-nbd` mounter in a StorageClass causes mapping RBD images to fail: ``` rbd: map failed with error an error (exit status 22) occurred while running rbd args: [--id cephcsi-rbd-node -m rook-ceph-mon-a.rook-ceph.svc.cluster.local:6789 --keyfile=***stripped*** map replicapool/csi-vol-2ff8537d-ce8b-11eb-8cea-f696da6dd1c4 --device-type nbd --options notrim], rbd error output: rbd-nbd: unknown args: --notrim ``` It seems that the `--options notrim` is not accepted when using nbd. There might be an alternative option, or we should return a failure while provisioning when both thick-provisioning and nbd are enabled. cc: @pkalever
test
rbd nbd does not support options notrim fails mapping thick provisioned volumes describe the bug enabling both thick provisioning and the rbd nbd mounter in a storageclass causes mapping rbd images to fail rbd map failed with error an error exit status occurred while running rbd args rbd error output rbd nbd unknown args notrim it seems that the options notrim is not accepted when using nbd there might be an alternative option or we should return a failure while provisioning when both thick provisioning and nbd are enabled cc pkalever
1
46,321
7,247,608,937
IssuesEvent
2018-02-15 04:18:03
datatogether/roadmap
https://api.github.com/repos/datatogether/roadmap
closed
Create Service Access doc for DataTogether
documentation
A public HackMD/markdown inventory file would be what I'd love us to do for all our assets :) EDGI has this (restricted): https://paper.dropbox.com/doc/Services-Deployment-pTsHUVIb2JuFHo8KVfpq2 Will be stubbing one out here in the meantime, but happy to make it go away if we don't want one: https://hackmd.io/s/r1Bj45GHZ Thoughts @b5?
1.0
Create Service Access doc for DataTogether - A public HackMD/markdown inventory file would be what I'd love us to do for all our assets :) EDGI has this (restricted): https://paper.dropbox.com/doc/Services-Deployment-pTsHUVIb2JuFHo8KVfpq2 Will be stubbing one out here in the meantime, but happy to make it go away if we don't want one: https://hackmd.io/s/r1Bj45GHZ Thoughts @b5?
non_test
create service access doc for datatogether a public hackmd markdown inventory file would be what i d love us to do for all our assets edgi has this restricted will be stubbing one out here in the meantime but happy to make it go away if we don t want one thoughts
0
429,038
12,420,197,209
IssuesEvent
2020-05-23 10:41:44
nf-core/proteomicslfq
https://api.github.com/repos/nf-core/proteomicslfq
closed
Luciphor not working
bug high-priority
May-12 01:17:05.539 [Task monitor] ERROR nextflow.processor.TaskProcessor - Error executing process > 'luciphor (1)' Caused by: Process `luciphor (1)` terminated with an error exit status (1) Command executed: LuciphorAdapter -id 020320_FeNTA_phospho_idx_feat_perc_filter_switched.idXML \ -in 020320_FeNTA_phospho.mzML \ -out 020320_FeNTA_phospho_idx_feat_perc_filter_switched_luciphor.idXML \ -threads 2 \ -num_threads 2 \ -target_modifications 'Phospho (S)' 'Phospho (T)' 'Phospho (Y)' \ -fragment_method HCD \ \ \ \ -max_charge_state 4 \ -max_peptide_length 40 \ > 020320_FeNTA_phospho_idx_feat_perc_filter_switched_scoreswitcher.log Command exit status: 1 Command output: (empty) Command wrapper: pstat() { local x_pid=$1 local STATUS=$(2> /dev/null < /proc/$1/status egrep 'Vm|ctxt') if [ $? = 0 ]; then local x_vsz=$(echo "$STATUS" | grep VmSize | awk '{print $2}' || echo -n '0') local x_rss=$(echo "$STATUS" | grep VmRSS | awk '{print $2}' || echo -n '0') local x_peak=$(echo "$STATUS" | egrep 'VmPeak|VmHWM' | sed 's/^.*:\s*//' | sed 's/[\sa-zA-Z]*$//' | tr '\n' ' ' || echo -n '0 0') local x_pmem=$(awk -v rss=$x_rss -v mem_tot=$mem_tot 'BEGIN {printf "%.0f", rss/mem_tot*100*10}' || echo -n '0') local vol_ctxt=$(echo "$STATUS" | grep '\bvoluntary_ctxt_switches' | awk '{print $2}' || echo -n '0') local inv_ctxt=$(echo "$STATUS" | grep '\bnonvoluntary_ctxt_switches' | awk '{print $2}' || echo -n '0') cpu_stat[x_pid]="$x_pid $x_pmem $x_vsz $x_rss $x_peak $vol_ctxt $inv_ctxt" fi } pwalk() { pstat $1 for i in ${ALL_CHILDREN[$1]:=}; do pwalk $i; done } pwalk $1 } nxf_stat() { cpu_stat=() nxf_tree $1 declare -a sum=(0 0 0 0 0 0 0 0) (... more ...) ------------------------------------------------------------ Exited with exit code 1. Resource usage summary: CPU time : 349.28 sec. Max Memory : 6949 MB Average Memory : 2848.86 MB Total Requested Memory : 8192.00 MB Delta Memory : 1243.00 MB Max Swap : 8364 MB Max Processes : 10 Max Threads : 13 Run time : 211 sec. Turnaround time : 198 sec. The output (if any) is above this job summary. Work dir: /nfs/production3/pride/prod/cluster/reanalysis/PXD018241-phospho/work/1b/3e5863cfaac78d205ac0dec563350e Tip: you can replicate the issue by changing to the process work dir and entering the command `bash .command.run`
1.0
Luciphor not working - May-12 01:17:05.539 [Task monitor] ERROR nextflow.processor.TaskProcessor - Error executing process > 'luciphor (1)' Caused by: Process `luciphor (1)` terminated with an error exit status (1) Command executed: LuciphorAdapter -id 020320_FeNTA_phospho_idx_feat_perc_filter_switched.idXML \ -in 020320_FeNTA_phospho.mzML \ -out 020320_FeNTA_phospho_idx_feat_perc_filter_switched_luciphor.idXML \ -threads 2 \ -num_threads 2 \ -target_modifications 'Phospho (S)' 'Phospho (T)' 'Phospho (Y)' \ -fragment_method HCD \ \ \ \ -max_charge_state 4 \ -max_peptide_length 40 \ > 020320_FeNTA_phospho_idx_feat_perc_filter_switched_scoreswitcher.log Command exit status: 1 Command output: (empty) Command wrapper: pstat() { local x_pid=$1 local STATUS=$(2> /dev/null < /proc/$1/status egrep 'Vm|ctxt') if [ $? = 0 ]; then local x_vsz=$(echo "$STATUS" | grep VmSize | awk '{print $2}' || echo -n '0') local x_rss=$(echo "$STATUS" | grep VmRSS | awk '{print $2}' || echo -n '0') local x_peak=$(echo "$STATUS" | egrep 'VmPeak|VmHWM' | sed 's/^.*:\s*//' | sed 's/[\sa-zA-Z]*$//' | tr '\n' ' ' || echo -n '0 0') local x_pmem=$(awk -v rss=$x_rss -v mem_tot=$mem_tot 'BEGIN {printf "%.0f", rss/mem_tot*100*10}' || echo -n '0') local vol_ctxt=$(echo "$STATUS" | grep '\bvoluntary_ctxt_switches' | awk '{print $2}' || echo -n '0') local inv_ctxt=$(echo "$STATUS" | grep '\bnonvoluntary_ctxt_switches' | awk '{print $2}' || echo -n '0') cpu_stat[x_pid]="$x_pid $x_pmem $x_vsz $x_rss $x_peak $vol_ctxt $inv_ctxt" fi } pwalk() { pstat $1 for i in ${ALL_CHILDREN[$1]:=}; do pwalk $i; done } pwalk $1 } nxf_stat() { cpu_stat=() nxf_tree $1 declare -a sum=(0 0 0 0 0 0 0 0) (... more ...) ------------------------------------------------------------ Exited with exit code 1. Resource usage summary: CPU time : 349.28 sec. Max Memory : 6949 MB Average Memory : 2848.86 MB Total Requested Memory : 8192.00 MB Delta Memory : 1243.00 MB Max Swap : 8364 MB Max Processes : 10 Max Threads : 13 Run time : 211 sec. Turnaround time : 198 sec. The output (if any) is above this job summary. Work dir: /nfs/production3/pride/prod/cluster/reanalysis/PXD018241-phospho/work/1b/3e5863cfaac78d205ac0dec563350e Tip: you can replicate the issue by changing to the process work dir and entering the command `bash .command.run`
non_test
luciphor not working may error nextflow processor taskprocessor error executing process luciphor caused by process luciphor terminated with an error exit status command executed luciphoradapter id fenta phospho idx feat perc filter switched idxml in fenta phospho mzml out fenta phospho idx feat perc filter switched luciphor idxml threads num threads target modifications phospho s phospho t phospho y fragment method hcd max charge state max peptide length fenta phospho idx feat perc filter switched scoreswitcher log command exit status command output empty command wrapper pstat local x pid local status dev null proc status egrep vm ctxt if then local x vsz echo status grep vmsize awk print echo n local x rss echo status grep vmrss awk print echo n local x peak echo status egrep vmpeak vmhwm sed s s sed s tr n echo n local x pmem awk v rss x rss v mem tot mem tot begin printf rss mem tot echo n local vol ctxt echo status grep bvoluntary ctxt switches awk print echo n local inv ctxt echo status grep bnonvoluntary ctxt switches awk print echo n cpu stat x pid x pmem x vsz x rss x peak vol ctxt inv ctxt fi pwalk pstat for i in all children do pwalk i done pwalk nxf stat cpu stat nxf tree declare a sum more exited with exit code resource usage summary cpu time sec max memory mb average memory mb total requested memory mb delta memory mb max swap mb max processes max threads run time sec turnaround time sec the output if any is above this job summary work dir nfs pride prod cluster reanalysis phospho work tip you can replicate the issue by changing to the process work dir and entering the command bash command run
0