Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
1k
labels
stringlengths
4
1.38k
body
stringlengths
1
262k
index
stringclasses
16 values
text_combine
stringlengths
96
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
307,925
23,223,035,032
IssuesEvent
2022-08-02 20:14:40
gravitational/teleport
https://api.github.com/repos/gravitational/teleport
closed
Add calls to action with links to the Teleport Cloud signup page
documentation time-to-value
## Details Teleport Cloud simplifies a lot of setup work that we describe in the docs, so we can add links to the Teleport Cloud signup page (`https://goteleport.com/signup?source=docs`) in pages within the docs that describe some complex setup process. The `source` param indicates that traffic is coming from the docs so we can gauge the effectiveness of our call to action links. ### Category - Improve Existing
1.0
Add calls to action with links to the Teleport Cloud signup page - ## Details Teleport Cloud simplifies a lot of setup work that we describe in the docs, so we can add links to the Teleport Cloud signup page (`https://goteleport.com/signup?source=docs`) in pages within the docs that describe some complex setup process. The `source` param indicates that traffic is coming from the docs so we can gauge the effectiveness of our call to action links. ### Category - Improve Existing
non_priority
add calls to action with links to the teleport cloud signup page details teleport cloud simplifies a lot of setup work that we describe in the docs so we can add links to the teleport cloud signup page in pages within the docs that describe some complex setup process the source param indicates that traffic is coming from the docs so we can gauge the effectiveness of our call to action links category improve existing
0
238,582
19,726,479,159
IssuesEvent
2022-01-13 20:30:33
urapadmin/kiosk
https://api.github.com/repos/urapadmin/kiosk
reopened
"Master Control not up" message pops up like a bad penny on arch1900
bug kiosk test-stage Beset
but I have seen it on other machines, as well. I wondered this time, whether or not it has anything to do with user privileges. Needs a closer look at the mechanism.
1.0
"Master Control not up" message pops up like a bad penny on arch1900 - but I have seen it on other machines, as well. I wondered this time, whether or not it has anything to do with user privileges. Needs a closer look at the mechanism.
non_priority
master control not up message pops up like a bad penny on but i have seen it on other machines as well i wondered this time whether or not it has anything to do with user privileges needs a closer look at the mechanism
0
114,961
9,777,099,463
IssuesEvent
2019-06-07 08:08:35
istio/istio
https://api.github.com/repos/istio/istio
closed
Add golden master tests for istio-iptables.sh
area/networking area/test and release
To ensure that the istio-iptables.sh-script remains stable, add golden master tests to istio. [ ] Configuration Infrastructure [ ] Docs [ ] Installation [X] Networking [ ] Performance and Scalability [ ] Policies and Telemetry [ ] Security [X] Test and Release [ ] User Experience **Additional context** These tests can also be used to compare different implementations of istio-iptables, e.g. in golang. Related: #14355
1.0
Add golden master tests for istio-iptables.sh - To ensure that the istio-iptables.sh-script remains stable, add golden master tests to istio. [ ] Configuration Infrastructure [ ] Docs [ ] Installation [X] Networking [ ] Performance and Scalability [ ] Policies and Telemetry [ ] Security [X] Test and Release [ ] User Experience **Additional context** These tests can also be used to compare different implementations of istio-iptables, e.g. in golang. Related: #14355
non_priority
add golden master tests for istio iptables sh to ensure that the istio iptables sh script remains stable add golden master tests to istio configuration infrastructure docs installation networking performance and scalability policies and telemetry security test and release user experience additional context these tests can also be used to compare different implementations of istio iptables e g in golang related
0
129,546
18,103,311,332
IssuesEvent
2021-09-22 16:18:37
gms-ws-demo/JS-Demo-Sep2021
https://api.github.com/repos/gms-ws-demo/JS-Demo-Sep2021
closed
CVE-2016-10540 (High) detected in minimatch-0.3.0.tgz - autoclosed
security vulnerability
## CVE-2016-10540 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimatch-0.3.0.tgz</b></p></summary> <p>a glob matcher in javascript</p> <p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-0.3.0.tgz">https://registry.npmjs.org/minimatch/-/minimatch-0.3.0.tgz</a></p> <p>Path to dependency file: JS-Demo-Sep2021/package.json</p> <p>Path to vulnerable library: JS-Demo-Sep2021/node_modules/mocha/node_modules/minimatch/package.json</p> <p> Dependency Hierarchy: - mocha-2.5.3.tgz (Root Library) - glob-3.2.11.tgz - :x: **minimatch-0.3.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/gms-ws-demo/JS-Demo-Sep2021/commit/e8cd219daa23fb09c60a7e7095b13c9e8372f529">e8cd219daa23fb09c60a7e7095b13c9e8372f529</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Minimatch is a minimal matching utility that works by converting glob expressions into JavaScript `RegExp` objects. The primary function, `minimatch(path, pattern)` in Minimatch 3.0.1 and earlier is vulnerable to ReDoS in the `pattern` parameter. <p>Publish Date: 2018-05-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10540>CVE-2016-10540</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nodesecurity.io/advisories/118">https://nodesecurity.io/advisories/118</a></p> <p>Release Date: 2016-06-20</p> <p>Fix Resolution: Update to version 3.0.2 or later.</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"minimatch","packageVersion":"0.3.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"mocha:2.5.3;glob:3.2.11;minimatch:0.3.0","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2016-10540","vulnerabilityDetails":"Minimatch is a minimal matching utility that works by converting glob expressions into JavaScript `RegExp` objects. The primary function, `minimatch(path, pattern)` in Minimatch 3.0.1 and earlier is vulnerable to ReDoS in the `pattern` parameter.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10540","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2016-10540 (High) detected in minimatch-0.3.0.tgz - autoclosed - ## CVE-2016-10540 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimatch-0.3.0.tgz</b></p></summary> <p>a glob matcher in javascript</p> <p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-0.3.0.tgz">https://registry.npmjs.org/minimatch/-/minimatch-0.3.0.tgz</a></p> <p>Path to dependency file: JS-Demo-Sep2021/package.json</p> <p>Path to vulnerable library: JS-Demo-Sep2021/node_modules/mocha/node_modules/minimatch/package.json</p> <p> Dependency Hierarchy: - mocha-2.5.3.tgz (Root Library) - glob-3.2.11.tgz - :x: **minimatch-0.3.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/gms-ws-demo/JS-Demo-Sep2021/commit/e8cd219daa23fb09c60a7e7095b13c9e8372f529">e8cd219daa23fb09c60a7e7095b13c9e8372f529</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Minimatch is a minimal matching utility that works by converting glob expressions into JavaScript `RegExp` objects. The primary function, `minimatch(path, pattern)` in Minimatch 3.0.1 and earlier is vulnerable to ReDoS in the `pattern` parameter. <p>Publish Date: 2018-05-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10540>CVE-2016-10540</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nodesecurity.io/advisories/118">https://nodesecurity.io/advisories/118</a></p> <p>Release Date: 2016-06-20</p> <p>Fix Resolution: Update to version 3.0.2 or later.</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"minimatch","packageVersion":"0.3.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"mocha:2.5.3;glob:3.2.11;minimatch:0.3.0","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2016-10540","vulnerabilityDetails":"Minimatch is a minimal matching utility that works by converting glob expressions into JavaScript `RegExp` objects. The primary function, `minimatch(path, pattern)` in Minimatch 3.0.1 and earlier is vulnerable to ReDoS in the `pattern` parameter.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10540","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_priority
cve high detected in minimatch tgz autoclosed cve high severity vulnerability vulnerable library minimatch tgz a glob matcher in javascript library home page a href path to dependency file js demo package json path to vulnerable library js demo node modules mocha node modules minimatch package json dependency hierarchy mocha tgz root library glob tgz x minimatch tgz vulnerable library found in head commit a href found in base branch master vulnerability details minimatch is a minimal matching utility that works by converting glob expressions into javascript regexp objects the primary function minimatch path pattern in minimatch and earlier is vulnerable to redos in the pattern parameter publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution update to version or later isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree mocha glob minimatch isminimumfixversionavailable false basebranches vulnerabilityidentifier cve vulnerabilitydetails minimatch is a minimal matching utility that works by converting glob expressions into javascript regexp objects the primary function minimatch path pattern in minimatch and earlier is vulnerable to redos in the pattern parameter vulnerabilityurl
0
677,078
23,149,350,869
IssuesEvent
2022-07-29 06:35:35
trustwallet/wallet-core
https://api.github.com/repos/trustwallet/wallet-core
closed
[Wasm] Can't import module error
bug priority:medium size:small
Reported from https://github.com/robot-ux/wallet-core-example/issues/3 ``` Uncaught (in promise) Error: Cannot find module './generated/core_proto' at webpackEmptyContext (dist|sync:2:1) at index.js:18:1 at index.js:8:1 at ./node_modules/@trustwallet/wallet-core/dist/index.js (index.js:14:1) at options.factory (react refresh:6:1) at __webpack_require__ (bootstrap:24:1) at fn (hot module replacement:62:1) at initWasm.ts:15:1 at new Promise (<anonymous>) at initWasm (initWasm.ts:3:1) ```
1.0
[Wasm] Can't import module error - Reported from https://github.com/robot-ux/wallet-core-example/issues/3 ``` Uncaught (in promise) Error: Cannot find module './generated/core_proto' at webpackEmptyContext (dist|sync:2:1) at index.js:18:1 at index.js:8:1 at ./node_modules/@trustwallet/wallet-core/dist/index.js (index.js:14:1) at options.factory (react refresh:6:1) at __webpack_require__ (bootstrap:24:1) at fn (hot module replacement:62:1) at initWasm.ts:15:1 at new Promise (<anonymous>) at initWasm (initWasm.ts:3:1) ```
priority
can t import module error reported from uncaught in promise error cannot find module generated core proto at webpackemptycontext dist sync at index js at index js at node modules trustwallet wallet core dist index js index js at options factory react refresh at webpack require bootstrap at fn hot module replacement at initwasm ts at new promise at initwasm initwasm ts
1
302,655
9,285,098,806
IssuesEvent
2019-03-21 05:18:22
zulip/zulip
https://api.github.com/repos/zulip/zulip
closed
Regression 1.9 -> 2.0: linkifiers with dashes no longer work
area: markdown bug in progress priority: high
After the upgrade to 2.0, we've noticed that all of our linkifiers with dashes in the regex pattern have stopped working. The `-` might be a red herring, but that's what we've noticed between working linkifiers and non-working linkifiers. Not working: ``` (?P<issue>[A-Z][A-Z_0-9]*-\d+) # HELP-4568 http://jira.example.com/browse/%(issue)s (?P<id>[0-9a-f-]{1,64}@[a-zA-Z]{1,32}) # c9f4fb80-a9df-4e36-a202-5a2aa0853681@Thing http://github.example.com/pages/swallitsch/resourceidentifier?id=%(id)s ``` Working: ``` (?P<subreddit>\/r\/[a-zA-Z]+) # /r/programming http://www.reddit.com/%(subreddit)s (?P<word>[tT]hanks|[tT]hank[- ]you|:thx:|thx) # Thanks https://thx.youearnedit.com/?%(word)s ``` We've tried escaping the dashes in the regex and deleting and re-adding the regexes.
1.0
Regression 1.9 -> 2.0: linkifiers with dashes no longer work - After the upgrade to 2.0, we've noticed that all of our linkifiers with dashes in the regex pattern have stopped working. The `-` might be a red herring, but that's what we've noticed between working linkifiers and non-working linkifiers. Not working: ``` (?P<issue>[A-Z][A-Z_0-9]*-\d+) # HELP-4568 http://jira.example.com/browse/%(issue)s (?P<id>[0-9a-f-]{1,64}@[a-zA-Z]{1,32}) # c9f4fb80-a9df-4e36-a202-5a2aa0853681@Thing http://github.example.com/pages/swallitsch/resourceidentifier?id=%(id)s ``` Working: ``` (?P<subreddit>\/r\/[a-zA-Z]+) # /r/programming http://www.reddit.com/%(subreddit)s (?P<word>[tT]hanks|[tT]hank[- ]you|:thx:|thx) # Thanks https://thx.youearnedit.com/?%(word)s ``` We've tried escaping the dashes in the regex and deleting and re-adding the regexes.
priority
regression linkifiers with dashes no longer work after the upgrade to we ve noticed that all of our linkifiers with dashes in the regex pattern have stopped working the might be a red herring but that s what we ve noticed between working linkifiers and non working linkifiers not working p d help p thing working p r r programming p hanks hank you thx thx thanks we ve tried escaping the dashes in the regex and deleting and re adding the regexes
1
441,864
12,733,739,241
IssuesEvent
2020-06-25 12:49:46
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
Current elections not showing up
Category: Web Priority: Medium Status: Fixed
Run 'generate test elections' in swagger, then this is empty: ![image](https://user-images.githubusercontent.com/3536496/80041645-cd7fbd00-84b1-11ea-8095-e3ff7e977a92.png) Once I click the arrow, they appear: ![image](https://user-images.githubusercontent.com/3536496/80041682-e7210480-84b1-11ea-8a06-972c00a14f43.png) Check other lists getting displayed here too.
1.0
Current elections not showing up - Run 'generate test elections' in swagger, then this is empty: ![image](https://user-images.githubusercontent.com/3536496/80041645-cd7fbd00-84b1-11ea-8095-e3ff7e977a92.png) Once I click the arrow, they appear: ![image](https://user-images.githubusercontent.com/3536496/80041682-e7210480-84b1-11ea-8a06-972c00a14f43.png) Check other lists getting displayed here too.
priority
current elections not showing up run generate test elections in swagger then this is empty once i click the arrow they appear check other lists getting displayed here too
1
455,743
13,132,147,687
IssuesEvent
2020-08-06 18:21:03
googleapis/google-auth-library-nodejs
https://api.github.com/repos/googleapis/google-auth-library-nodejs
closed
Synthesis failed for google-auth-library-nodejs
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate google-auth-library-nodejs. :broken_heart: Here's the output from running `synth.py`: ``` B, cookie=0, name=docs-devsite.sh> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/release/publish.cfg', wd=41, mask=IN_MODIFY, cookie=0, name=publish.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/release/publish.cfg', wd=41, mask=IN_MODIFY, cookie=0, name=publish.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/release/publish.cfg', wd=41, mask=IN_ATTRIB, cookie=0, name=publish.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/release/docs.cfg', wd=41, mask=IN_MODIFY, cookie=0, name=docs.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/release/docs.cfg', wd=41, mask=IN_MODIFY, cookie=0, name=docs.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/release/docs.cfg', wd=41, mask=IN_ATTRIB, cookie=0, name=docs.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/samples-test.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=samples-test.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/samples-test.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=samples-test.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/samples-test.cfg', wd=45, mask=IN_ATTRIB, cookie=0, name=samples-test.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/lint.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=lint.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/lint.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=lint.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/lint.cfg', wd=45, mask=IN_ATTRIB, cookie=0, name=lint.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/test.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=test.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/test.cfg', wd=45, mask=IN_ATTRIB, cookie=0, name=test.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/common.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=common.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/common.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=common.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/common.cfg', wd=45, mask=IN_ATTRIB, cookie=0, name=common.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/system-test.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=system-test.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/system-test.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=system-test.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/system-test.cfg', wd=45, mask=IN_ATTRIB, cookie=0, name=system-test.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/docs.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=docs.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/docs.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=docs.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/docs.cfg', wd=45, mask=IN_ATTRIB, cookie=0, name=docs.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/docs.cfg', wd=45, mask=IN_ATTRIB, cookie=0, name=docs.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node12/test.cfg', wd=47, mask=IN_MODIFY, cookie=0, name=test.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node12/test.cfg', wd=47, mask=IN_ATTRIB, cookie=0, name=test.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node12/common.cfg', wd=47, mask=IN_MODIFY, cookie=0, name=common.cfg> 2020-08-06 04:09:57,577 synthtool [DEBUG] > Installing dependencies... DEBUG:synthtool:Installing dependencies... DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node12/common.cfg', wd=47, mask=IN_ATTRIB, cookie=0, name=common.cfg> npm WARN deprecated istanbul@0.4.5: This module is no longer maintained, try this instead: npm WARN deprecated npm i nyc npm WARN deprecated Visit https://istanbul.js.org/integrations for other alternatives. npm WARN deprecated chokidar@2.1.8: Chokidar 2 will break on node v14+. Upgrade to chokidar 3 with 15x less dependencies. npm WARN deprecated resolve-url@0.2.1: https://github.com/lydell/resolve-url#deprecated npm WARN deprecated urix@0.1.0: Please see https://github.com/lydell/urix#deprecated npm WARN deprecated fsevents@1.2.13: fsevents 1 will break on node v14+ and could be using insecure binaries. Upgrade to fsevents 2. npm ERR! code E404 npm ERR! 404 Not Found - GET https://registry.npmjs.org/@compodoc%2fcompodoc - Not found npm ERR! 404 npm ERR! 404 '@compodoc/compodoc@^1.1.7' is not in the npm registry. npm ERR! 404 You should bug the author to publish it (or use the name yourself!) npm ERR! 404 It was specified as a dependency of 'google-auth-library-nodejs' npm ERR! 404 npm ERR! 404 Note that you can also install from a npm ERR! 404 tarball, folder, http url, or git url. npm ERR! A complete log of this run can be found in: npm ERR! /home/kbuilder/.npm/_logs/2020-08-06T11_10_04_702Z-debug.log 2020-08-06 04:10:04,718 synthtool [ERROR] > Failed executing npm install: None ERROR:synthtool:Failed executing npm install: None 2020-08-06 04:10:04,733 synthtool [DEBUG] > Wrote metadata to synth.metadata. DEBUG:synthtool:Wrote metadata to synth.metadata. Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/home/kbuilder/.cache/synthtool/google-auth-library-nodejs/synth.py", line 12, in <module> node.install() File "/tmpfs/src/github/synthtool/synthtool/languages/node.py", line 167, in install shell.run(["npm", "install"], hide_output=hide_output) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['npm', 'install']' returned non-zero exit status 1. 2020-08-06 04:10:04,785 autosynth [ERROR] > Synthesis failed 2020-08-06 04:10:04,785 autosynth [DEBUG] > Running: git reset --hard HEAD HEAD is now at a7e5701 fix: migrate token info API to not pass token in query string (#991) 2020-08-06 04:10:04,795 autosynth [DEBUG] > Running: git checkout autosynth Switched to branch 'autosynth' 2020-08-06 04:10:04,801 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 690, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 539, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 670, in _inner_main commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 375, in synthesize_loop has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 273, in synthesize_version_in_new_branch synthesizer.synthesize(synth_log_path, self.environ) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/76bb7f6f-4d47-4888-b0d2-a8761a276cc8/targets/github%2Fsynthtool;config=default/tests;query=google-auth-library-nodejs;failed=false).
1.0
Synthesis failed for google-auth-library-nodejs - Hello! Autosynth couldn't regenerate google-auth-library-nodejs. :broken_heart: Here's the output from running `synth.py`: ``` B, cookie=0, name=docs-devsite.sh> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/release/publish.cfg', wd=41, mask=IN_MODIFY, cookie=0, name=publish.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/release/publish.cfg', wd=41, mask=IN_MODIFY, cookie=0, name=publish.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/release/publish.cfg', wd=41, mask=IN_ATTRIB, cookie=0, name=publish.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/release/docs.cfg', wd=41, mask=IN_MODIFY, cookie=0, name=docs.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/release/docs.cfg', wd=41, mask=IN_MODIFY, cookie=0, name=docs.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/release/docs.cfg', wd=41, mask=IN_ATTRIB, cookie=0, name=docs.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/samples-test.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=samples-test.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/samples-test.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=samples-test.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/samples-test.cfg', wd=45, mask=IN_ATTRIB, cookie=0, name=samples-test.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/lint.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=lint.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/lint.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=lint.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/lint.cfg', wd=45, mask=IN_ATTRIB, cookie=0, name=lint.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/test.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=test.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/test.cfg', wd=45, mask=IN_ATTRIB, cookie=0, name=test.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/common.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=common.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/common.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=common.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/common.cfg', wd=45, mask=IN_ATTRIB, cookie=0, name=common.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/system-test.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=system-test.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/system-test.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=system-test.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/system-test.cfg', wd=45, mask=IN_ATTRIB, cookie=0, name=system-test.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/docs.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=docs.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/docs.cfg', wd=45, mask=IN_MODIFY, cookie=0, name=docs.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/docs.cfg', wd=45, mask=IN_ATTRIB, cookie=0, name=docs.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node10/docs.cfg', wd=45, mask=IN_ATTRIB, cookie=0, name=docs.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node12/test.cfg', wd=47, mask=IN_MODIFY, cookie=0, name=test.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node12/test.cfg', wd=47, mask=IN_ATTRIB, cookie=0, name=test.cfg> DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node12/common.cfg', wd=47, mask=IN_MODIFY, cookie=0, name=common.cfg> 2020-08-06 04:09:57,577 synthtool [DEBUG] > Installing dependencies... DEBUG:synthtool:Installing dependencies... DEBUG:watchdog.observers.inotify_buffer:in-event <InotifyEvent: src_path=b'./.kokoro/continuous/node12/common.cfg', wd=47, mask=IN_ATTRIB, cookie=0, name=common.cfg> npm WARN deprecated istanbul@0.4.5: This module is no longer maintained, try this instead: npm WARN deprecated npm i nyc npm WARN deprecated Visit https://istanbul.js.org/integrations for other alternatives. npm WARN deprecated chokidar@2.1.8: Chokidar 2 will break on node v14+. Upgrade to chokidar 3 with 15x less dependencies. npm WARN deprecated resolve-url@0.2.1: https://github.com/lydell/resolve-url#deprecated npm WARN deprecated urix@0.1.0: Please see https://github.com/lydell/urix#deprecated npm WARN deprecated fsevents@1.2.13: fsevents 1 will break on node v14+ and could be using insecure binaries. Upgrade to fsevents 2. npm ERR! code E404 npm ERR! 404 Not Found - GET https://registry.npmjs.org/@compodoc%2fcompodoc - Not found npm ERR! 404 npm ERR! 404 '@compodoc/compodoc@^1.1.7' is not in the npm registry. npm ERR! 404 You should bug the author to publish it (or use the name yourself!) npm ERR! 404 It was specified as a dependency of 'google-auth-library-nodejs' npm ERR! 404 npm ERR! 404 Note that you can also install from a npm ERR! 404 tarball, folder, http url, or git url. npm ERR! A complete log of this run can be found in: npm ERR! /home/kbuilder/.npm/_logs/2020-08-06T11_10_04_702Z-debug.log 2020-08-06 04:10:04,718 synthtool [ERROR] > Failed executing npm install: None ERROR:synthtool:Failed executing npm install: None 2020-08-06 04:10:04,733 synthtool [DEBUG] > Wrote metadata to synth.metadata. DEBUG:synthtool:Wrote metadata to synth.metadata. Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/home/kbuilder/.cache/synthtool/google-auth-library-nodejs/synth.py", line 12, in <module> node.install() File "/tmpfs/src/github/synthtool/synthtool/languages/node.py", line 167, in install shell.run(["npm", "install"], hide_output=hide_output) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['npm', 'install']' returned non-zero exit status 1. 2020-08-06 04:10:04,785 autosynth [ERROR] > Synthesis failed 2020-08-06 04:10:04,785 autosynth [DEBUG] > Running: git reset --hard HEAD HEAD is now at a7e5701 fix: migrate token info API to not pass token in query string (#991) 2020-08-06 04:10:04,795 autosynth [DEBUG] > Running: git checkout autosynth Switched to branch 'autosynth' 2020-08-06 04:10:04,801 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 690, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 539, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 670, in _inner_main commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 375, in synthesize_loop has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 273, in synthesize_version_in_new_branch synthesizer.synthesize(synth_log_path, self.environ) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/76bb7f6f-4d47-4888-b0d2-a8761a276cc8/targets/github%2Fsynthtool;config=default/tests;query=google-auth-library-nodejs;failed=false).
priority
synthesis failed for google auth library nodejs hello autosynth couldn t regenerate google auth library nodejs broken heart here s the output from running synth py b cookie name docs devsite sh debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event debug watchdog observers inotify buffer in event synthtool installing dependencies debug synthtool installing dependencies debug watchdog observers inotify buffer in event npm warn deprecated istanbul this module is no longer maintained try this instead npm warn deprecated npm i nyc npm warn deprecated visit for other alternatives npm warn deprecated chokidar chokidar will break on node upgrade to chokidar with less dependencies npm warn deprecated resolve url npm warn deprecated urix please see npm warn deprecated fsevents fsevents will break on node and could be using insecure binaries upgrade to fsevents npm err code npm err not found get not found npm err npm err compodoc compodoc is not in the npm registry npm err you should bug the author to publish it or use the name yourself npm err it was specified as a dependency of google auth library nodejs npm err npm err note that you can also install from a npm err tarball folder http url or git url npm err a complete log of this run can be found in npm err home kbuilder npm logs debug log synthtool failed executing npm install none error synthtool failed executing npm install none synthtool wrote metadata to synth metadata debug synthtool wrote metadata to synth metadata traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file home kbuilder cache synthtool google auth library nodejs synth py line in node install file tmpfs src github synthtool synthtool languages node py line in install shell run hide output hide output file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status autosynth synthesis failed autosynth running git reset hard head head is now at fix migrate token info api to not pass token in query string autosynth running git checkout autosynth switched to branch autosynth autosynth running git clean fdx removing pycache traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main commit count synthesize loop x multiple prs change pusher synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize loop has changes toolbox synthesize version in new branch synthesizer youngest file tmpfs src github synthtool autosynth synth py line in synthesize version in new branch synthesizer synthesize synth log path self environ file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
1
446,238
12,842,866,536
IssuesEvent
2020-07-08 03:09:19
code4nagoya/covid19
https://api.github.com/repos/code4nagoya/covid19
opened
愛知県HPの検査件数表に「抗原検査」が追加されたことによる対応
improve priority high
## 改善詳細 / Details of Improvement [愛知県HP の 3.愛知県内の検査件数](https://www.pref.aichi.jp/site/covid19-aichi/kansensya-kensa.html) に「抗原検査」の列が追加されました。 それにより以下の事象の影響が発生します。 1. [優先度:高]表形式が発生したことにより(?)、定期自動スクレイピングが失敗している ※1 3. [優先度:低]「検査実施件数」グラフに、抗原検査の検査件数も表現する(「件数」としてなら、PCRと抗原検査の積み上げ防棒グラフでよい?) ※2 元々、検査件数表の情報は「件数」しか使っていませんでした(陽性者数は未使用)。そのため他の影響は無いと思います。 ### ※1 https://github.com/code4nagoya/covid19/pull/712 ### ※2 > 「抗原検査」ですね。「PCR検査」を代替するもので、短時間で判定出来ますが、感度が劣る(一定のウイルス量が必要)ようです。 > 「抗原検査」で陽性なら感染確定とし、陰性となっても症状から感染を疑う場合「PCR検査」を引き続き行うようです。 > https://news.yahoo.co.jp/byline/kutsunasatoshi/20200517-00178720/ > 「検査件数」は、情報として特に分けて見る必要は無いと考えます。重要なのは「検査人数」です。前述の例のように「抗原検査」→「PCR検査」とした場合、「検査人数」は1人なので、『愛知県は日毎の「検査人数」を公表して下さい』です。 ## スクリーンショット / Screenshot ![image](https://user-images.githubusercontent.com/401369/86869969-7d101700-c112-11ea-890b-808459091f43.png) ## 動作環境・ブラウザ / Environment - macOS / Windows / Linux / iOS / Android - Chrome / Safari / Firefox / Edge / Internet Explorer
1.0
愛知県HPの検査件数表に「抗原検査」が追加されたことによる対応 - ## 改善詳細 / Details of Improvement [愛知県HP の 3.愛知県内の検査件数](https://www.pref.aichi.jp/site/covid19-aichi/kansensya-kensa.html) に「抗原検査」の列が追加されました。 それにより以下の事象の影響が発生します。 1. [優先度:高]表形式が発生したことにより(?)、定期自動スクレイピングが失敗している ※1 3. [優先度:低]「検査実施件数」グラフに、抗原検査の検査件数も表現する(「件数」としてなら、PCRと抗原検査の積み上げ防棒グラフでよい?) ※2 元々、検査件数表の情報は「件数」しか使っていませんでした(陽性者数は未使用)。そのため他の影響は無いと思います。 ### ※1 https://github.com/code4nagoya/covid19/pull/712 ### ※2 > 「抗原検査」ですね。「PCR検査」を代替するもので、短時間で判定出来ますが、感度が劣る(一定のウイルス量が必要)ようです。 > 「抗原検査」で陽性なら感染確定とし、陰性となっても症状から感染を疑う場合「PCR検査」を引き続き行うようです。 > https://news.yahoo.co.jp/byline/kutsunasatoshi/20200517-00178720/ > 「検査件数」は、情報として特に分けて見る必要は無いと考えます。重要なのは「検査人数」です。前述の例のように「抗原検査」→「PCR検査」とした場合、「検査人数」は1人なので、『愛知県は日毎の「検査人数」を公表して下さい』です。 ## スクリーンショット / Screenshot ![image](https://user-images.githubusercontent.com/401369/86869969-7d101700-c112-11ea-890b-808459091f43.png) ## 動作環境・ブラウザ / Environment - macOS / Windows / Linux / iOS / Android - Chrome / Safari / Firefox / Edge / Internet Explorer
priority
愛知県hpの検査件数表に「抗原検査」が追加されたことによる対応 改善詳細 details of improvement に「抗原検査」の列が追加されました。 それにより以下の事象の影響が発生します。 表形式が発生したことにより 、定期自動スクレイピングが失敗している ※ 「検査実施件数」グラフに、抗原検査の検査件数も表現する(「件数」としてなら、pcrと抗原検査の積み上げ防棒グラフでよい?) ※ 元々、検査件数表の情報は「件数」しか使っていませんでした(陽性者数は未使用)。そのため他の影響は無いと思います。 ※ ※ 「抗原検査」ですね。「pcr検査」を代替するもので、短時間で判定出来ますが、感度が劣る 一定のウイルス量が必要 ようです。 「抗原検査」で陽性なら感染確定とし、陰性となっても症状から感染を疑う場合「pcr検査」を引き続き行うようです。 「検査件数」は、情報として特に分けて見る必要は無いと考えます。重要なのは「検査人数」です。前述の例のように「抗原検査」→「pcr検査」とした場合、「検査人数」 、『愛知県は日毎の「検査人数」を公表して下さい』です。 スクリーンショット screenshot 動作環境・ブラウザ environment macos windows linux ios android chrome safari firefox edge internet explorer
1
323,591
23,955,876,035
IssuesEvent
2022-09-12 14:55:25
Dasharo/dasharo-issues
https://api.github.com/repos/Dasharo/dasharo-issues
closed
Talos II - "Testing firmware images without flashing" instructions are not working
documentation raptor-cs_talos-2
At https://docs.dasharo.com/variants/talos_2/installation-manual/#testing-firmware-images-without-flashing 3. Mount the file as flash device: 1. mboxctl --backend file:/tmp/flash.pnor 1. root@talos:~# mboxctl --backend file:/tmp/flash.pnor 1. Failed to resolve path: No such file or directory 2. root@talos:~# mboxctl --version 3. Mailbox Control V2.1.1 2. Those instructions are not working. 1. As documented from https://wiki.raptorcs.com/wiki/Compiling_Firmware#Running_the_firmware_temporarily 1. systemctl stop mboxd 2. Point mboxd to prepared talos.pnor image 1. mboxd -f 64M -w 1M -b file:/tmp/talos.pnor -v 3. Open another ssh to BMC 1. mboxctl --lpc-state 1. should show “LPC Bus Maps: BMC Memory” 2. obmcutil poweron 3. When done testing 1. obmcutil poweroff 2. systemctl start mboxd 3. mboxctl --lpc-state 1. Should show: “LPC Bus Maps: Flash Device”
1.0
Talos II - "Testing firmware images without flashing" instructions are not working - At https://docs.dasharo.com/variants/talos_2/installation-manual/#testing-firmware-images-without-flashing 3. Mount the file as flash device: 1. mboxctl --backend file:/tmp/flash.pnor 1. root@talos:~# mboxctl --backend file:/tmp/flash.pnor 1. Failed to resolve path: No such file or directory 2. root@talos:~# mboxctl --version 3. Mailbox Control V2.1.1 2. Those instructions are not working. 1. As documented from https://wiki.raptorcs.com/wiki/Compiling_Firmware#Running_the_firmware_temporarily 1. systemctl stop mboxd 2. Point mboxd to prepared talos.pnor image 1. mboxd -f 64M -w 1M -b file:/tmp/talos.pnor -v 3. Open another ssh to BMC 1. mboxctl --lpc-state 1. should show “LPC Bus Maps: BMC Memory” 2. obmcutil poweron 3. When done testing 1. obmcutil poweroff 2. systemctl start mboxd 3. mboxctl --lpc-state 1. Should show: “LPC Bus Maps: Flash Device”
non_priority
talos ii testing firmware images without flashing instructions are not working at mount the file as flash device mboxctl backend file tmp flash pnor root talos mboxctl backend file tmp flash pnor failed to resolve path no such file or directory root talos mboxctl version mailbox control those instructions are not working as documented from systemctl stop mboxd point mboxd to prepared talos pnor image mboxd f w b file tmp talos pnor v open another ssh to bmc mboxctl lpc state should show “lpc bus maps bmc memory” obmcutil poweron when done testing obmcutil poweroff systemctl start mboxd mboxctl lpc state should show “lpc bus maps flash device”
0
31,087
4,231,119,215
IssuesEvent
2016-07-04 14:43:19
governmentbg/opendata-cms
https://api.github.com/repos/governmentbg/opendata-cms
closed
Прецизиране на размера на шрифтовете в хедъра
design low priority question
Казахме, че не гоним pixel-perfect matching на дизайна с този на портала за отворени данни и това все още е валидно, но шрифтовете в хедъра се изрисуват по видимо различен начин на двете места: Портал: <img width="985" alt="screenshot 2016-06-29 17 47 12" src="https://cloud.githubusercontent.com/assets/129307/16456720/8d8ad26e-3e21-11e6-887a-3ab2172b7c77.png"> CMS: <img width="980" alt="screenshot 2016-06-29 17 47 19" src="https://cloud.githubusercontent.com/assets/129307/16456722/8fa75cca-3e21-11e6-9c32-4bdab1619774.png"> Приемам предложения дали лесно можем да направим нещо, така че да сближим двете визии още малко. Неща, които на мен ми хрумват след бърза проверка: - `font-weight: 900;` вместо `bold` на почернената част от заглавния текст в хедъра. - Донагласяне на `line-height`, `font-size`, може би и `font-weight` или опциите за изчертаване и antialiasing. Виждам, че почти навсякъде се ползва `rem`. Този `rem` стъпва на размера на шрифта на `<html>`, ако не се лъжа. Там е зададено `110%`. Това променя ли се някога, при някакви условия? Дали не е по-добре да се стъпи на фиксиран размер на шрифта в `<html>`, който да бъде увеличаван/намаляван с media queries, в зависимост от размера на екрана?
1.0
Прецизиране на размера на шрифтовете в хедъра - Казахме, че не гоним pixel-perfect matching на дизайна с този на портала за отворени данни и това все още е валидно, но шрифтовете в хедъра се изрисуват по видимо различен начин на двете места: Портал: <img width="985" alt="screenshot 2016-06-29 17 47 12" src="https://cloud.githubusercontent.com/assets/129307/16456720/8d8ad26e-3e21-11e6-887a-3ab2172b7c77.png"> CMS: <img width="980" alt="screenshot 2016-06-29 17 47 19" src="https://cloud.githubusercontent.com/assets/129307/16456722/8fa75cca-3e21-11e6-9c32-4bdab1619774.png"> Приемам предложения дали лесно можем да направим нещо, така че да сближим двете визии още малко. Неща, които на мен ми хрумват след бърза проверка: - `font-weight: 900;` вместо `bold` на почернената част от заглавния текст в хедъра. - Донагласяне на `line-height`, `font-size`, може би и `font-weight` или опциите за изчертаване и antialiasing. Виждам, че почти навсякъде се ползва `rem`. Този `rem` стъпва на размера на шрифта на `<html>`, ако не се лъжа. Там е зададено `110%`. Това променя ли се някога, при някакви условия? Дали не е по-добре да се стъпи на фиксиран размер на шрифта в `<html>`, който да бъде увеличаван/намаляван с media queries, в зависимост от размера на екрана?
non_priority
прецизиране на размера на шрифтовете в хедъра казахме че не гоним pixel perfect matching на дизайна с този на портала за отворени данни и това все още е валидно но шрифтовете в хедъра се изрисуват по видимо различен начин на двете места портал img width alt screenshot src cms img width alt screenshot src приемам предложения дали лесно можем да направим нещо така че да сближим двете визии още малко неща които на мен ми хрумват след бърза проверка font weight вместо bold на почернената част от заглавния текст в хедъра донагласяне на line height font size може би и font weight или опциите за изчертаване и antialiasing виждам че почти навсякъде се ползва rem този rem стъпва на размера на шрифта на ако не се лъжа там е зададено това променя ли се някога при някакви условия дали не е по добре да се стъпи на фиксиран размер на шрифта в който да бъде увеличаван намаляван с media queries в зависимост от размера на екрана
0
635,627
20,423,849,592
IssuesEvent
2022-02-24 00:15:29
aws/aws-node-termination-handler
https://api.github.com/repos/aws/aws-node-termination-handler
closed
Incorrectly pulls amd64 image on arm64 machine.
Type: Bug Priority: Medium
**Describe the bug** On an arm machine when you pull the multi-arch image, it pulls the amd64 image instead of the arm64 image. **Steps to reproduce** Some weird behavior I'm noticing with upstream amazon/aws-node-termination-handler On inspecting the manifest it's clearly a multi-arch image and does have an arm64 arch version available. ``` $ docker manifest inspect amazon/aws-node-termination-handler:v1.6.1 { "schemaVersion": 2, "mediaType": "application/vnd.docker.distribution.manifest.list.v2+json", "manifests": [ { "mediaType": "application/vnd.docker.distribution.manifest.v2+json", "size": 947, "digest": "sha256:7e91ba3ff76e3c540f8e2f1d3935b61fb05f831a9c36f42961a5d0e878e7c8a4", "platform": { "architecture": "amd64", "os": "linux" } }, { "mediaType": "application/vnd.docker.distribution.manifest.v2+json", "size": 947, "digest": "sha256:f83038b59db9cebe1b1904fc6f351e82f8aa9e6cdb09bd89d34996998ad5ab16", "platform": { "architecture": "arm", "os": "linux" } }, { "mediaType": "application/vnd.docker.distribution.manifest.v2+json", "size": 947, "digest": "sha256:c3e1c2d17a05c8ecb661658bd179be3784fe605da58ab7940e1303c0fb8eda70", "platform": { "architecture": "arm64", "os": "linux" } }, { "mediaType": "application/vnd.docker.distribution.manifest.v2+json", "size": 1752, "digest": "sha256:2a72849bcd7c46518523072f8ad3453c451030a1ddfd2dd9c95bdae3ffdfdc8e", "platform": { "architecture": "amd64", "os": "windows", "os.version": "10.0.17763.1282" } } ] } ``` But when I pull the image on my arm64 machine, it pulls the amd64 image instead of the arm64 image. ``` $ docker pull amazon/aws-node-termination-handler:v1.6.1 $ docker image inspect amazon/aws-node-termination-handler:v1.6.1 ..... ], "OnBuild": null, "Labels": null }, "Architecture": "amd64", "Os": "linux", "Size": 36635931, "VirtualSize": 36635931, "GraphDriver": { .... ``` **Expected outcome** Architecture should be arm64 instead of amd64. **Application Logs** The log output when experiencing the issue. **Environment** * NTH App Version: * NTH Mode (IMDS/Queue processor): * OS/Arch: MacOS/arm64 * Kubernetes version: * Installation method:
1.0
Incorrectly pulls amd64 image on arm64 machine. - **Describe the bug** On an arm machine when you pull the multi-arch image, it pulls the amd64 image instead of the arm64 image. **Steps to reproduce** Some weird behavior I'm noticing with upstream amazon/aws-node-termination-handler On inspecting the manifest it's clearly a multi-arch image and does have an arm64 arch version available. ``` $ docker manifest inspect amazon/aws-node-termination-handler:v1.6.1 { "schemaVersion": 2, "mediaType": "application/vnd.docker.distribution.manifest.list.v2+json", "manifests": [ { "mediaType": "application/vnd.docker.distribution.manifest.v2+json", "size": 947, "digest": "sha256:7e91ba3ff76e3c540f8e2f1d3935b61fb05f831a9c36f42961a5d0e878e7c8a4", "platform": { "architecture": "amd64", "os": "linux" } }, { "mediaType": "application/vnd.docker.distribution.manifest.v2+json", "size": 947, "digest": "sha256:f83038b59db9cebe1b1904fc6f351e82f8aa9e6cdb09bd89d34996998ad5ab16", "platform": { "architecture": "arm", "os": "linux" } }, { "mediaType": "application/vnd.docker.distribution.manifest.v2+json", "size": 947, "digest": "sha256:c3e1c2d17a05c8ecb661658bd179be3784fe605da58ab7940e1303c0fb8eda70", "platform": { "architecture": "arm64", "os": "linux" } }, { "mediaType": "application/vnd.docker.distribution.manifest.v2+json", "size": 1752, "digest": "sha256:2a72849bcd7c46518523072f8ad3453c451030a1ddfd2dd9c95bdae3ffdfdc8e", "platform": { "architecture": "amd64", "os": "windows", "os.version": "10.0.17763.1282" } } ] } ``` But when I pull the image on my arm64 machine, it pulls the amd64 image instead of the arm64 image. ``` $ docker pull amazon/aws-node-termination-handler:v1.6.1 $ docker image inspect amazon/aws-node-termination-handler:v1.6.1 ..... ], "OnBuild": null, "Labels": null }, "Architecture": "amd64", "Os": "linux", "Size": 36635931, "VirtualSize": 36635931, "GraphDriver": { .... ``` **Expected outcome** Architecture should be arm64 instead of amd64. **Application Logs** The log output when experiencing the issue. **Environment** * NTH App Version: * NTH Mode (IMDS/Queue processor): * OS/Arch: MacOS/arm64 * Kubernetes version: * Installation method:
priority
incorrectly pulls image on machine describe the bug on an arm machine when you pull the multi arch image it pulls the image instead of the image steps to reproduce some weird behavior i m noticing with upstream amazon aws node termination handler on inspecting the manifest it s clearly a multi arch image and does have an arch version available docker manifest inspect amazon aws node termination handler schemaversion mediatype application vnd docker distribution manifest list json manifests mediatype application vnd docker distribution manifest json size digest platform architecture os linux mediatype application vnd docker distribution manifest json size digest platform architecture arm os linux mediatype application vnd docker distribution manifest json size digest platform architecture os linux mediatype application vnd docker distribution manifest json size digest platform architecture os windows os version but when i pull the image on my machine it pulls the image instead of the image docker pull amazon aws node termination handler docker image inspect amazon aws node termination handler onbuild null labels null architecture os linux size virtualsize graphdriver expected outcome architecture should be instead of application logs the log output when experiencing the issue environment nth app version nth mode imds queue processor os arch macos kubernetes version installation method
1
27,715
2,695,320,650
IssuesEvent
2015-04-02 03:59:36
cs2103jan2015-t15-4j/main
https://api.github.com/repos/cs2103jan2015-t15-4j/main
closed
A user can view a list of already completed tasks
priority.medium type.story
...so that the user can review his/her completed tasks when necessary.
1.0
A user can view a list of already completed tasks - ...so that the user can review his/her completed tasks when necessary.
priority
a user can view a list of already completed tasks so that the user can review his her completed tasks when necessary
1
283,223
8,717,913,612
IssuesEvent
2018-12-07 18:42:11
rubykube/peatio
https://api.github.com/repos/rubykube/peatio
closed
Sessions do not delete when DELETE /api/v2/sessions
Priority: High Type: Bug v1.9
After send DELETE /api/v2/sessions I can use old JWToken for creating new order and other actions.
1.0
Sessions do not delete when DELETE /api/v2/sessions - After send DELETE /api/v2/sessions I can use old JWToken for creating new order and other actions.
priority
sessions do not delete when delete api sessions after send delete api sessions i can use old jwtoken for creating new order and other actions
1
538,187
15,764,592,414
IssuesEvent
2021-03-31 13:23:12
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
[Coverity CID :219484] Out-of-bounds access in tests/drivers/timer/nrf_rtc_timer/src/main.c
Coverity bug priority: low
Static code scan issues found in file: https://github.com/zephyrproject-rtos/zephyr/tree/bd97359a5338b2542d19011b6d6aa1d8d1b9cc3f/tests/drivers/timer/nrf_rtc_timer/src/main.c Category: Memory - corruptions Function: `test_int_disable_enabled` Component: Tests CID: [219484](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=219484) Details: https://github.com/zephyrproject-rtos/zephyr/blob/bd97359a5338b2542d19011b6d6aa1d8d1b9cc3f/tests/drivers/timer/nrf_rtc_timer/src/main.c#L147 Please fix or provide comments in coverity using the link: https://scan9.coverity.com/reports.htm#v32951/p12996. Note: This issue was created automatically. Priority was set based on classification of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
1.0
[Coverity CID :219484] Out-of-bounds access in tests/drivers/timer/nrf_rtc_timer/src/main.c - Static code scan issues found in file: https://github.com/zephyrproject-rtos/zephyr/tree/bd97359a5338b2542d19011b6d6aa1d8d1b9cc3f/tests/drivers/timer/nrf_rtc_timer/src/main.c Category: Memory - corruptions Function: `test_int_disable_enabled` Component: Tests CID: [219484](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=219484) Details: https://github.com/zephyrproject-rtos/zephyr/blob/bd97359a5338b2542d19011b6d6aa1d8d1b9cc3f/tests/drivers/timer/nrf_rtc_timer/src/main.c#L147 Please fix or provide comments in coverity using the link: https://scan9.coverity.com/reports.htm#v32951/p12996. Note: This issue was created automatically. Priority was set based on classification of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
priority
out of bounds access in tests drivers timer nrf rtc timer src main c static code scan issues found in file category memory corruptions function test int disable enabled component tests cid details please fix or provide comments in coverity using the link note this issue was created automatically priority was set based on classification of the file affected and the impact field in coverity assignees were set using the codeowners file
1
742,450
25,855,464,409
IssuesEvent
2022-12-13 13:27:12
hengband/hengband
https://api.github.com/repos/hengband/hengband
closed
BasitemInfo::locale/chance の取り扱い改善
refactor Priority:MIDDLE
主に以下: ・「常に両方の要素数は同じ」という性質を活かして構造体化 ・生配列からstd::array への転換 ・関連コードの整備
1.0
BasitemInfo::locale/chance の取り扱い改善 - 主に以下: ・「常に両方の要素数は同じ」という性質を活かして構造体化 ・生配列からstd::array への転換 ・関連コードの整備
priority
basiteminfo locale chance の取り扱い改善 主に以下: ・「常に両方の要素数は同じ」という性質を活かして構造体化 ・生配列からstd array への転換 ・関連コードの整備
1
282,028
21,315,456,816
IssuesEvent
2022-04-16 07:31:43
channne/pe
https://api.github.com/repos/channne/pe
opened
UC06 and UC10
severity.Low type.DocumentationBug
UC06 and UC10 seem very similar, but also describe behavior that does not seem to exist - `Tracey` does not summarise over the entire hall, just by block and faculty ![Screenshot 2022-04-16 at 3.30.27 PM.png](https://raw.githubusercontent.com/channne/pe/main/files/7972cc11-de55-4028-b23c-7578cd1c3634.png) ![Screenshot 2022-04-16 at 3.30.42 PM.png](https://raw.githubusercontent.com/channne/pe/main/files/d93db15c-bba2-4860-b3d2-72818c2fd745.png) <!--session: 1650086993451-197aa2c0-ab17-433e-a4ec-44761c99441f--> <!--Version: Web v3.4.2-->
1.0
UC06 and UC10 - UC06 and UC10 seem very similar, but also describe behavior that does not seem to exist - `Tracey` does not summarise over the entire hall, just by block and faculty ![Screenshot 2022-04-16 at 3.30.27 PM.png](https://raw.githubusercontent.com/channne/pe/main/files/7972cc11-de55-4028-b23c-7578cd1c3634.png) ![Screenshot 2022-04-16 at 3.30.42 PM.png](https://raw.githubusercontent.com/channne/pe/main/files/d93db15c-bba2-4860-b3d2-72818c2fd745.png) <!--session: 1650086993451-197aa2c0-ab17-433e-a4ec-44761c99441f--> <!--Version: Web v3.4.2-->
non_priority
and and seem very similar but also describe behavior that does not seem to exist tracey does not summarise over the entire hall just by block and faculty
0
240,773
20,073,542,171
IssuesEvent
2022-02-04 10:06:17
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: jepsen/g2/start-stop-2 failed
C-test-failure O-robot O-roachtest release-blocker branch-release-20.2
[(roachtest).jepsen/g2/start-stop-2 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4299094&tab=buildLog) on [release-20.2@09707aabb12e50f6e7345b5c9664c0745bb7d742](https://github.com/cockroachdb/cockroach/commits/09707aabb12e50f6e7345b5c9664c0745bb7d742): ``` | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2291 | main.runJepsen.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/jepsen.go:160 | main.runJepsen.func3 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/jepsen.go:195 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1581 Wraps: (2) output in run_100536.122_n6_bash Wraps: (3) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-4299094-1643958549-63-n6cpu4:6 -- bash -e -c "\ | cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \ | ~/lein run test \ | --tarball file://${PWD}/cockroach.tgz \ | --username ${USER} \ | --ssh-private-key ~/.ssh/id_rsa \ | --os ubuntu \ | --time-limit 300 \ | --concurrency 30 \ | --recovery-time 25 \ | --test-count 1 \ | -n 10.128.0.25 -n 10.128.0.37 -n 10.128.0.23 -n 10.128.0.33 -n 10.128.0.21 \ | --test g2 --nemesis start-stop-2 \ | > invoke.log 2>&1 \ | " returned | stderr: | Error: SSH_PROBLEM: exit status 255 | (1) SSH_PROBLEM | Wraps: (2) Node 6. Command with error: | | ``` | | bash -e -c "\ | | cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \ | | ~/lein run test \ | | --tarball file://${PWD}/cockroach.tgz \ | | --username ${USER} \ | | --ssh-private-key ~/.ssh/id_rsa \ | | --os ubuntu \ | | --time-limit 300 \ | | --concurrency 30 \ | | --recovery-time 25 \ | | --test-count 1 \ | | -n 10.128.0.25 -n 10.128.0.37 -n 10.128.0.23 -n 10.128.0.33 -n 10.128.0.21 \ | | --test g2 --nemesis start-stop-2 \ | | > invoke.log 2>&1 \ | | " | | ``` | Wraps: (3) exit status 255 | Error types: (1) errors.SSH (2) *hintdetail.withDetail (3) *exec.ExitError | | stdout: Wraps: (4) exit status 10 Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *main.withCommandDetails (4) *exec.ExitError ``` <details><summary>More</summary><p> Artifacts: [/jepsen/g2/start-stop-2](https://teamcity.cockroachdb.com/viewLog.html?buildId=4299094&tab=artifacts#/jepsen/g2/start-stop-2) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Ajepsen%2Fg2%2Fstart-stop-2.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
2.0
roachtest: jepsen/g2/start-stop-2 failed - [(roachtest).jepsen/g2/start-stop-2 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4299094&tab=buildLog) on [release-20.2@09707aabb12e50f6e7345b5c9664c0745bb7d742](https://github.com/cockroachdb/cockroach/commits/09707aabb12e50f6e7345b5c9664c0745bb7d742): ``` | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2291 | main.runJepsen.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/jepsen.go:160 | main.runJepsen.func3 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/jepsen.go:195 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1581 Wraps: (2) output in run_100536.122_n6_bash Wraps: (3) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-4299094-1643958549-63-n6cpu4:6 -- bash -e -c "\ | cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \ | ~/lein run test \ | --tarball file://${PWD}/cockroach.tgz \ | --username ${USER} \ | --ssh-private-key ~/.ssh/id_rsa \ | --os ubuntu \ | --time-limit 300 \ | --concurrency 30 \ | --recovery-time 25 \ | --test-count 1 \ | -n 10.128.0.25 -n 10.128.0.37 -n 10.128.0.23 -n 10.128.0.33 -n 10.128.0.21 \ | --test g2 --nemesis start-stop-2 \ | > invoke.log 2>&1 \ | " returned | stderr: | Error: SSH_PROBLEM: exit status 255 | (1) SSH_PROBLEM | Wraps: (2) Node 6. Command with error: | | ``` | | bash -e -c "\ | | cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \ | | ~/lein run test \ | | --tarball file://${PWD}/cockroach.tgz \ | | --username ${USER} \ | | --ssh-private-key ~/.ssh/id_rsa \ | | --os ubuntu \ | | --time-limit 300 \ | | --concurrency 30 \ | | --recovery-time 25 \ | | --test-count 1 \ | | -n 10.128.0.25 -n 10.128.0.37 -n 10.128.0.23 -n 10.128.0.33 -n 10.128.0.21 \ | | --test g2 --nemesis start-stop-2 \ | | > invoke.log 2>&1 \ | | " | | ``` | Wraps: (3) exit status 255 | Error types: (1) errors.SSH (2) *hintdetail.withDetail (3) *exec.ExitError | | stdout: Wraps: (4) exit status 10 Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *main.withCommandDetails (4) *exec.ExitError ``` <details><summary>More</summary><p> Artifacts: [/jepsen/g2/start-stop-2](https://teamcity.cockroachdb.com/viewLog.html?buildId=4299094&tab=artifacts#/jepsen/g2/start-stop-2) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Ajepsen%2Fg2%2Fstart-stop-2.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
non_priority
roachtest jepsen start stop failed on home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main runjepsen home agent work go src github com cockroachdb cockroach pkg cmd roachtest jepsen go main runjepsen home agent work go src github com cockroachdb cockroach pkg cmd roachtest jepsen go runtime goexit usr local go src runtime asm s wraps output in run bash wraps home agent work go src github com cockroachdb cockroach bin roachprod run teamcity bash e c cd mnt jepsen cockroachdb set eo pipefail lein run test tarball file pwd cockroach tgz username user ssh private key ssh id rsa os ubuntu time limit concurrency recovery time test count n n n n n test nemesis start stop invoke log returned stderr error ssh problem exit status ssh problem wraps node command with error bash e c cd mnt jepsen cockroachdb set eo pipefail lein run test tarball file pwd cockroach tgz username user ssh private key ssh id rsa os ubuntu time limit concurrency recovery time test count n n n n n test nemesis start stop invoke log wraps exit status error types errors ssh hintdetail withdetail exec exiterror stdout wraps exit status error types withstack withstack errutil withprefix main withcommanddetails exec exiterror more artifacts powered by
0
405,895
11,884,051,757
IssuesEvent
2020-03-27 16:58:30
OpenFAM/OpenFAM
https://api.github.com/repos/OpenFAM/OpenFAM
closed
Add support to libfabric provider for Infiniband
<PRIORITY>- P0 enhancement
OpenFAM API need to support libfabric verbs provider for Inifiniband.
1.0
Add support to libfabric provider for Infiniband - OpenFAM API need to support libfabric verbs provider for Inifiniband.
priority
add support to libfabric provider for infiniband openfam api need to support libfabric verbs provider for inifiniband
1
333,519
10,127,604,686
IssuesEvent
2019-08-01 10:37:36
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.apple.com - site is not usable
browser-firefox engine-gecko priority-critical
<!-- @browser: Firefox 69.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:69.0) Gecko/20100101 Firefox/69.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://www.apple.com/fr/itunes/download/thank-you/ **Browser / Version**: Firefox 69.0 **Operating System**: Windows 7 **Tested Another Browser**: Unknown **Problem type**: Site is not usable **Description**: update doesn't charge **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2019/7/3bd85b88-0186-4faa-8a7d-8644ca7178dd-thumb.jpeg)](https://webcompat.com/uploads/2019/7/3bd85b88-0186-4faa-8a7d-8644ca7178dd.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190722201635</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: aurora</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.apple.com - site is not usable - <!-- @browser: Firefox 69.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:69.0) Gecko/20100101 Firefox/69.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://www.apple.com/fr/itunes/download/thank-you/ **Browser / Version**: Firefox 69.0 **Operating System**: Windows 7 **Tested Another Browser**: Unknown **Problem type**: Site is not usable **Description**: update doesn't charge **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2019/7/3bd85b88-0186-4faa-8a7d-8644ca7178dd-thumb.jpeg)](https://webcompat.com/uploads/2019/7/3bd85b88-0186-4faa-8a7d-8644ca7178dd.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190722201635</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: aurora</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
site is not usable url browser version firefox operating system windows tested another browser unknown problem type site is not usable description update doesn t charge steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen true mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel aurora from with ❤️
1
111,993
4,500,298,112
IssuesEvent
2016-09-01 03:42:14
anishathalye/gavel
https://api.github.com/repos/anishathalye/gavel
closed
Allow deletion of projects that have been judged / judges who have judged
bug high priority
Yeah, this is broken right now. It's cause the system keeps track of all decisions (why not store more data), and so if we delete a project, it would be bad, cause the decision would point to a deleted project. It's possible to delete projects before they're judged, though. The right way to fix this would be to implement some sort of "deactivated" flag for projects and not have them show up anywhere (but still live in the database). This is somewhat involved, so post here before you start working if you're interested in working on the issue.
1.0
Allow deletion of projects that have been judged / judges who have judged - Yeah, this is broken right now. It's cause the system keeps track of all decisions (why not store more data), and so if we delete a project, it would be bad, cause the decision would point to a deleted project. It's possible to delete projects before they're judged, though. The right way to fix this would be to implement some sort of "deactivated" flag for projects and not have them show up anywhere (but still live in the database). This is somewhat involved, so post here before you start working if you're interested in working on the issue.
priority
allow deletion of projects that have been judged judges who have judged yeah this is broken right now it s cause the system keeps track of all decisions why not store more data and so if we delete a project it would be bad cause the decision would point to a deleted project it s possible to delete projects before they re judged though the right way to fix this would be to implement some sort of deactivated flag for projects and not have them show up anywhere but still live in the database this is somewhat involved so post here before you start working if you re interested in working on the issue
1
368,958
10,886,590,297
IssuesEvent
2019-11-18 12:54:38
palantir/atlasdb
https://api.github.com/repos/palantir/atlasdb
opened
Conjure Java Runtime Stuck On Internal Butterfly Product
priority: P1
@gsheasby reported an issue where a CJR client in the `TimestampManagementService` on atlasdb `0.173.2` in conjunction with a latest TimeLock could get stuck during initialization. It seems that we are getting a `RetryOther` exception thrown out. ``` com.palantir.async.initializer.AsyncInitializer: Failed to initialize {} on the attempt {} (className: TimeLockMigrator, numberOfAttempts: 4, throwable3_redirectTo: BOX_6/timelock/api) (throwable0_message: Could not contact the Timelock Server., throwable1_message: Exceeded the maximum number of allowed redirects: {requestUrl=BOX_8/timelock/api/SERVICE/timestamp-management/ping} executing GET BOX_6/timelock/api/SERVICE/timestamp-management/ping, throwable2_requestUrl: BOX_8/timelock/api/SERVICE/timestamp-management/ping) com.palantir.common.exception.AtlasDbDependencyException: Could not contact the Timelock Server. at com.palantir.atlasdb.factory.startup.TimeLockMigrator.tryInitialize(TimeLockMigrator.java:77) /* snip */ Caused by: feign.RetryableException: Exceeded the maximum number of allowed redirects: {requestUrl=BOX_8:8421/timelock/api/SERVICE/timestamp-management/ping} executing GET BOX_6/timelock/api/SERVICE/timestamp-management/ping at feign.FeignException.errorExecuting(FeignException.java:67) at feign.SynchronousMethodHandler.executeAndDecode(SynchronousMethodHandler.java:102) at feign.SynchronousMethodHandler.invoke(SynchronousMethodHandler.java:76) at feign.ReflectiveFeign$FeignInvocationHandler.invoke(ReflectiveFeign.java:103) at com.sun.proxy.$Proxy73.ping(Unknown Source) at jdk.internal.reflect.GeneratedMethodAccessor67.invoke(Unknown Source) at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:566) at com.palantir.conjure.java.ext.refresh.RefreshableProxyInvocationHandler.handleInvocation(RefreshableProxyInvocationHandler.java:61) at com.google.common.reflect.AbstractInvocationHandler.invoke(AbstractInvocationHandler.java:86) at com.sun.proxy.$Proxy73.ping(Unknown Source) at jdk.internal.reflect.GeneratedMethodAccessor67.invoke(Unknown Source) at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:566) at com.palantir.atlasdb.http.v2.FastFailoverProxy.singleInvocation(FastFailoverProxy.java:88) at com.palantir.atlasdb.http.v2.FastFailoverProxy.handleInvocation(FastFailoverProxy.java:78) at com.google.common.reflect.AbstractInvocationHandler.invoke(AbstractInvocationHandler.java:86) at com.sun.proxy.$Proxy73.ping(Unknown Source) at com.palantir.tritium.proxy.InstrumentedTimestampManagementRpcClient$7.ping(Unknown Source) at jdk.internal.reflect.GeneratedMethodAccessor67.invoke(Unknown Source) at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:566) at com.palantir.common.proxy.ExperimentRunningProxy.handleInvocation(ExperimentRunningProxy.java:91) at com.google.common.reflect.AbstractInvocationHandler.invoke(AbstractInvocationHandler.java:86) at com.sun.proxy.$Proxy73.ping(Unknown Source) at com.palantir.timestamp.RemoteTimestampManagementAdapter.ping(RemoteTimestampManagementAdapter.java:35) at com.palantir.timestamp.DelegatingManagedTimestampService.ping(DelegatingManagedTimestampService.java:57) at com.palantir.atlasdb.factory.startup.TimeLockMigrator.tryInitialize(TimeLockMigrator.java:75) ... 11 more Caused by: com.palantir.logsafe.exceptions.SafeIoException: Exceeded the maximum number of allowed redirects at com.palantir.conjure.java.okhttp.RemotingOkHttpCall$5.visit(RemotingOkHttpCall.java:401) at com.palantir.conjure.java.okhttp.RemotingOkHttpCall$5.visit(RemotingOkHttpCall.java:359) at com.palantir.conjure.java.api.errors.QosException$RetryOther.accept(QosException.java:118) at com.palantir.conjure.java.okhttp.RemotingOkHttpCall$4.onResponse(RemotingOkHttpCall.java:292) at okhttp3.RealCall$AsyncCall.execute(RealCall.java:216) at okhttp3.internal.NamedRunnable.run(NamedRunnable.java:32) ... 3 more Caused by: com.palantir.conjure.java.api.errors.QosException$RetryOther: RetryOther: Requesting retry at com.palantir.conjure.java.api.errors.QosException.retryOther(QosException.java:72) at com.palantir.conjure.java.QosExceptionResponseMapper.map308(QosExceptionResponseMapper.java:65) at com.palantir.conjure.java.QosExceptionResponseMapper.mapResponseCode(QosExceptionResponseMapper.java:46) at com.palantir.conjure.java.okhttp.QosExceptionResponseHandler.handle(QosExceptionResponseHandler.java:32) at com.palantir.conjure.java.okhttp.RemotingOkHttpCall$4.onResponse(RemotingOkHttpCall.java:290) ... 5 more```
1.0
Conjure Java Runtime Stuck On Internal Butterfly Product - @gsheasby reported an issue where a CJR client in the `TimestampManagementService` on atlasdb `0.173.2` in conjunction with a latest TimeLock could get stuck during initialization. It seems that we are getting a `RetryOther` exception thrown out. ``` com.palantir.async.initializer.AsyncInitializer: Failed to initialize {} on the attempt {} (className: TimeLockMigrator, numberOfAttempts: 4, throwable3_redirectTo: BOX_6/timelock/api) (throwable0_message: Could not contact the Timelock Server., throwable1_message: Exceeded the maximum number of allowed redirects: {requestUrl=BOX_8/timelock/api/SERVICE/timestamp-management/ping} executing GET BOX_6/timelock/api/SERVICE/timestamp-management/ping, throwable2_requestUrl: BOX_8/timelock/api/SERVICE/timestamp-management/ping) com.palantir.common.exception.AtlasDbDependencyException: Could not contact the Timelock Server. at com.palantir.atlasdb.factory.startup.TimeLockMigrator.tryInitialize(TimeLockMigrator.java:77) /* snip */ Caused by: feign.RetryableException: Exceeded the maximum number of allowed redirects: {requestUrl=BOX_8:8421/timelock/api/SERVICE/timestamp-management/ping} executing GET BOX_6/timelock/api/SERVICE/timestamp-management/ping at feign.FeignException.errorExecuting(FeignException.java:67) at feign.SynchronousMethodHandler.executeAndDecode(SynchronousMethodHandler.java:102) at feign.SynchronousMethodHandler.invoke(SynchronousMethodHandler.java:76) at feign.ReflectiveFeign$FeignInvocationHandler.invoke(ReflectiveFeign.java:103) at com.sun.proxy.$Proxy73.ping(Unknown Source) at jdk.internal.reflect.GeneratedMethodAccessor67.invoke(Unknown Source) at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:566) at com.palantir.conjure.java.ext.refresh.RefreshableProxyInvocationHandler.handleInvocation(RefreshableProxyInvocationHandler.java:61) at com.google.common.reflect.AbstractInvocationHandler.invoke(AbstractInvocationHandler.java:86) at com.sun.proxy.$Proxy73.ping(Unknown Source) at jdk.internal.reflect.GeneratedMethodAccessor67.invoke(Unknown Source) at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:566) at com.palantir.atlasdb.http.v2.FastFailoverProxy.singleInvocation(FastFailoverProxy.java:88) at com.palantir.atlasdb.http.v2.FastFailoverProxy.handleInvocation(FastFailoverProxy.java:78) at com.google.common.reflect.AbstractInvocationHandler.invoke(AbstractInvocationHandler.java:86) at com.sun.proxy.$Proxy73.ping(Unknown Source) at com.palantir.tritium.proxy.InstrumentedTimestampManagementRpcClient$7.ping(Unknown Source) at jdk.internal.reflect.GeneratedMethodAccessor67.invoke(Unknown Source) at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:566) at com.palantir.common.proxy.ExperimentRunningProxy.handleInvocation(ExperimentRunningProxy.java:91) at com.google.common.reflect.AbstractInvocationHandler.invoke(AbstractInvocationHandler.java:86) at com.sun.proxy.$Proxy73.ping(Unknown Source) at com.palantir.timestamp.RemoteTimestampManagementAdapter.ping(RemoteTimestampManagementAdapter.java:35) at com.palantir.timestamp.DelegatingManagedTimestampService.ping(DelegatingManagedTimestampService.java:57) at com.palantir.atlasdb.factory.startup.TimeLockMigrator.tryInitialize(TimeLockMigrator.java:75) ... 11 more Caused by: com.palantir.logsafe.exceptions.SafeIoException: Exceeded the maximum number of allowed redirects at com.palantir.conjure.java.okhttp.RemotingOkHttpCall$5.visit(RemotingOkHttpCall.java:401) at com.palantir.conjure.java.okhttp.RemotingOkHttpCall$5.visit(RemotingOkHttpCall.java:359) at com.palantir.conjure.java.api.errors.QosException$RetryOther.accept(QosException.java:118) at com.palantir.conjure.java.okhttp.RemotingOkHttpCall$4.onResponse(RemotingOkHttpCall.java:292) at okhttp3.RealCall$AsyncCall.execute(RealCall.java:216) at okhttp3.internal.NamedRunnable.run(NamedRunnable.java:32) ... 3 more Caused by: com.palantir.conjure.java.api.errors.QosException$RetryOther: RetryOther: Requesting retry at com.palantir.conjure.java.api.errors.QosException.retryOther(QosException.java:72) at com.palantir.conjure.java.QosExceptionResponseMapper.map308(QosExceptionResponseMapper.java:65) at com.palantir.conjure.java.QosExceptionResponseMapper.mapResponseCode(QosExceptionResponseMapper.java:46) at com.palantir.conjure.java.okhttp.QosExceptionResponseHandler.handle(QosExceptionResponseHandler.java:32) at com.palantir.conjure.java.okhttp.RemotingOkHttpCall$4.onResponse(RemotingOkHttpCall.java:290) ... 5 more```
priority
conjure java runtime stuck on internal butterfly product gsheasby reported an issue where a cjr client in the timestampmanagementservice on atlasdb in conjunction with a latest timelock could get stuck during initialization it seems that we are getting a retryother exception thrown out com palantir async initializer asyncinitializer failed to initialize on the attempt classname timelockmigrator numberofattempts redirectto box timelock api message could not contact the timelock server message exceeded the maximum number of allowed redirects requesturl box timelock api service timestamp management ping executing get box timelock api service timestamp management ping requesturl box timelock api service timestamp management ping com palantir common exception atlasdbdependencyexception could not contact the timelock server at com palantir atlasdb factory startup timelockmigrator tryinitialize timelockmigrator java snip caused by feign retryableexception exceeded the maximum number of allowed redirects requesturl box timelock api service timestamp management ping executing get box timelock api service timestamp management ping at feign feignexception errorexecuting feignexception java at feign synchronousmethodhandler executeanddecode synchronousmethodhandler java at feign synchronousmethodhandler invoke synchronousmethodhandler java at feign reflectivefeign feigninvocationhandler invoke reflectivefeign java at com sun proxy ping unknown source at jdk internal reflect invoke unknown source at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com palantir conjure java ext refresh refreshableproxyinvocationhandler handleinvocation refreshableproxyinvocationhandler java at com google common reflect abstractinvocationhandler invoke abstractinvocationhandler java at com sun proxy ping unknown source at jdk internal reflect invoke unknown source at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com palantir atlasdb http fastfailoverproxy singleinvocation fastfailoverproxy java at com palantir atlasdb http fastfailoverproxy handleinvocation fastfailoverproxy java at com google common reflect abstractinvocationhandler invoke abstractinvocationhandler java at com sun proxy ping unknown source at com palantir tritium proxy instrumentedtimestampmanagementrpcclient ping unknown source at jdk internal reflect invoke unknown source at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com palantir common proxy experimentrunningproxy handleinvocation experimentrunningproxy java at com google common reflect abstractinvocationhandler invoke abstractinvocationhandler java at com sun proxy ping unknown source at com palantir timestamp remotetimestampmanagementadapter ping remotetimestampmanagementadapter java at com palantir timestamp delegatingmanagedtimestampservice ping delegatingmanagedtimestampservice java at com palantir atlasdb factory startup timelockmigrator tryinitialize timelockmigrator java more caused by com palantir logsafe exceptions safeioexception exceeded the maximum number of allowed redirects at com palantir conjure java okhttp remotingokhttpcall visit remotingokhttpcall java at com palantir conjure java okhttp remotingokhttpcall visit remotingokhttpcall java at com palantir conjure java api errors qosexception retryother accept qosexception java at com palantir conjure java okhttp remotingokhttpcall onresponse remotingokhttpcall java at realcall asynccall execute realcall java at internal namedrunnable run namedrunnable java more caused by com palantir conjure java api errors qosexception retryother retryother requesting retry at com palantir conjure java api errors qosexception retryother qosexception java at com palantir conjure java qosexceptionresponsemapper qosexceptionresponsemapper java at com palantir conjure java qosexceptionresponsemapper mapresponsecode qosexceptionresponsemapper java at com palantir conjure java okhttp qosexceptionresponsehandler handle qosexceptionresponsehandler java at com palantir conjure java okhttp remotingokhttpcall onresponse remotingokhttpcall java more
1
556,350
16,482,422,059
IssuesEvent
2021-05-24 13:32:00
Redocly/openapi-cli
https://api.github.com/repos/Redocly/openapi-cli
closed
paths-kebab-case rule isn't checking for snake_case
Cat Priority: Low Type: Bug
**Describe the bug** The `paths-kebab-case` rule doesn't forbid snake_case paths, it only seems to detect camelCase. **To Reproduce** Steps to reproduce the behavior: 1. Add snake_case paths. 2. Lint the spec. **Expected behavior** The `paths-kebab-case` rule detects the snake_case paths and shows an error or warning. **Logs** N/a **OpenAPI definition** N/a **`openapi-cli` Version(s)** 1.0.0-beta.39 **`Node.js` Version(s)** v12.19.0 **Additional context** None.
1.0
paths-kebab-case rule isn't checking for snake_case - **Describe the bug** The `paths-kebab-case` rule doesn't forbid snake_case paths, it only seems to detect camelCase. **To Reproduce** Steps to reproduce the behavior: 1. Add snake_case paths. 2. Lint the spec. **Expected behavior** The `paths-kebab-case` rule detects the snake_case paths and shows an error or warning. **Logs** N/a **OpenAPI definition** N/a **`openapi-cli` Version(s)** 1.0.0-beta.39 **`Node.js` Version(s)** v12.19.0 **Additional context** None.
priority
paths kebab case rule isn t checking for snake case describe the bug the paths kebab case rule doesn t forbid snake case paths it only seems to detect camelcase to reproduce steps to reproduce the behavior add snake case paths lint the spec expected behavior the paths kebab case rule detects the snake case paths and shows an error or warning logs n a openapi definition n a openapi cli version s beta node js version s additional context none
1
179,552
14,705,558,537
IssuesEvent
2021-01-04 18:20:17
emmysteven/restaurant
https://api.github.com/repos/emmysteven/restaurant
closed
Add badge to indicate project stage
documentation enhancement
Hi, Pleasing add a developing stats badge to get visitors to know that this project is still in it's development stage Example: ![status](https://img.shields.io/badge/Status-developing-brightgree) Thanks!
1.0
Add badge to indicate project stage - Hi, Pleasing add a developing stats badge to get visitors to know that this project is still in it's development stage Example: ![status](https://img.shields.io/badge/Status-developing-brightgree) Thanks!
non_priority
add badge to indicate project stage hi pleasing add a developing stats badge to get visitors to know that this project is still in it s development stage example thanks
0
219,936
24,539,589,816
IssuesEvent
2022-10-12 01:36:32
RG4421/react-credit-card-input
https://api.github.com/repos/RG4421/react-credit-card-input
closed
CVE-2021-33587 (High) detected in css-what-3.4.2.tgz - autoclosed
security vulnerability
## CVE-2021-33587 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>css-what-3.4.2.tgz</b></p></summary> <p>a CSS selector parser</p> <p>Library home page: <a href="https://registry.npmjs.org/css-what/-/css-what-3.4.2.tgz">https://registry.npmjs.org/css-what/-/css-what-3.4.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/css-what/package.json</p> <p> Dependency Hierarchy: - react-3.4.12.tgz (Root Library) - html-webpack-plugin-2.30.1.tgz - pretty-error-2.1.2.tgz - renderkid-2.0.5.tgz - css-select-2.1.0.tgz - :x: **css-what-3.4.2.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The css-what package 4.0.0 through 5.0.0 for Node.js does not ensure that attribute parsing has Linear Time Complexity relative to the size of the input. <p>Publish Date: 2021-05-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33587>CVE-2021-33587</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33587">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33587</a></p> <p>Release Date: 2021-05-28</p> <p>Fix Resolution (css-what): 5.0.1</p> <p>Direct dependency fix Resolution (@storybook/react): 4.0.0</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
True
CVE-2021-33587 (High) detected in css-what-3.4.2.tgz - autoclosed - ## CVE-2021-33587 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>css-what-3.4.2.tgz</b></p></summary> <p>a CSS selector parser</p> <p>Library home page: <a href="https://registry.npmjs.org/css-what/-/css-what-3.4.2.tgz">https://registry.npmjs.org/css-what/-/css-what-3.4.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/css-what/package.json</p> <p> Dependency Hierarchy: - react-3.4.12.tgz (Root Library) - html-webpack-plugin-2.30.1.tgz - pretty-error-2.1.2.tgz - renderkid-2.0.5.tgz - css-select-2.1.0.tgz - :x: **css-what-3.4.2.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The css-what package 4.0.0 through 5.0.0 for Node.js does not ensure that attribute parsing has Linear Time Complexity relative to the size of the input. <p>Publish Date: 2021-05-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33587>CVE-2021-33587</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33587">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33587</a></p> <p>Release Date: 2021-05-28</p> <p>Fix Resolution (css-what): 5.0.1</p> <p>Direct dependency fix Resolution (@storybook/react): 4.0.0</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
non_priority
cve high detected in css what tgz autoclosed cve high severity vulnerability vulnerable library css what tgz a css selector parser library home page a href path to dependency file package json path to vulnerable library node modules css what package json dependency hierarchy react tgz root library html webpack plugin tgz pretty error tgz renderkid tgz css select tgz x css what tgz vulnerable library found in base branch master vulnerability details the css what package through for node js does not ensure that attribute parsing has linear time complexity relative to the size of the input publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution css what direct dependency fix resolution storybook react rescue worker helmet automatic remediation is available for this issue
0
362,919
10,735,156,077
IssuesEvent
2019-10-29 08:03:28
k8smeetup/website-tasks
https://api.github.com/repos/k8smeetup/website-tasks
closed
/docs/reference/setup-tools/kubeadm/generated/kubeadm_init_phase_certs_etcd-ca.md
finished lang/zh priority/P0 sync/update version/1.16
Source File: [/docs/reference/setup-tools/kubeadm/generated/kubeadm_init_phase_certs_etcd-ca.md](https://github.com/kubernetes/website/blob/release-1.16/content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_init_phase_certs_etcd-ca.md) Diff 查看原始文档更新差异命令: ```bash git diff release-1.14 release-1.16 -- content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_init_phase_certs_etcd-ca.md ```
1.0
/docs/reference/setup-tools/kubeadm/generated/kubeadm_init_phase_certs_etcd-ca.md - Source File: [/docs/reference/setup-tools/kubeadm/generated/kubeadm_init_phase_certs_etcd-ca.md](https://github.com/kubernetes/website/blob/release-1.16/content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_init_phase_certs_etcd-ca.md) Diff 查看原始文档更新差异命令: ```bash git diff release-1.14 release-1.16 -- content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_init_phase_certs_etcd-ca.md ```
priority
docs reference setup tools kubeadm generated kubeadm init phase certs etcd ca md source file diff 查看原始文档更新差异命令 bash git diff release release content en docs reference setup tools kubeadm generated kubeadm init phase certs etcd ca md
1
602,453
18,469,537,898
IssuesEvent
2021-10-17 13:52:45
cu-coders/cucoders
https://api.github.com/repos/cu-coders/cucoders
opened
[Bug]: CSRF vulnerability in forms
bug Done priority/critical
### Contact Details rajeswar.sh47@gmail.com ### What happened? Forms are vulnerable to cross-site forgery attacks. ### Version 1.0.0 (Default) ### What browsers are you seeing the problem on? Firefox, Chrome, Safari, Microsoft Edge ### What OS are you seeing the problem on? Windows, MacOS, Linux ### Relevant log output _No response_ ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
1.0
[Bug]: CSRF vulnerability in forms - ### Contact Details rajeswar.sh47@gmail.com ### What happened? Forms are vulnerable to cross-site forgery attacks. ### Version 1.0.0 (Default) ### What browsers are you seeing the problem on? Firefox, Chrome, Safari, Microsoft Edge ### What OS are you seeing the problem on? Windows, MacOS, Linux ### Relevant log output _No response_ ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
priority
csrf vulnerability in forms contact details rajeswar gmail com what happened forms are vulnerable to cross site forgery attacks version default what browsers are you seeing the problem on firefox chrome safari microsoft edge what os are you seeing the problem on windows macos linux relevant log output no response code of conduct i agree to follow this project s code of conduct
1
4,040
6,972,783,515
IssuesEvent
2017-12-11 18:10:45
triplea-game/triplea
https://api.github.com/repos/triplea-game/triplea
reopened
Move install4j bundled JREs from GitHub to Linode
category: dev & admin process discussion type: process
We bundle JREs with our installer for users that do not have a Java 8 JRE installed on their machine. We currently host these bundled JREs in the triplea-game/assets repo, and the installer downloads them, if needed, directly from GitHub. There are a few problems hosting these files on GitHub: * We can't just provide the "latest" JRE because older installers are built with a hard-coded link to a specific path in the repo. As long as those older builds are considered compatible, we have to keep multiple JRE versions on `HEAD` in the repo (we currently have two versions for three platforms). Each JRE is approximately 350 MiB. * Adding additional JREs increases the time to run the Gradle build on Travis due to the additional download requirements. * Adding additional JREs increases the size of the triplea-game/assets repo significantly. Even when we can remove an older bundled JRE from `HEAD`, it stays in the repo history unless we rewrite history. A static file server is probably a better place to host these resources rather than using a Git repo. **The purpose of this issue is to discuss possibly moving the bundled JREs to one of our Linode servers and serve them from there using Nginx (or something equivalent).** Some issues with self-hosting include: 1. Increased bandwidth possibly leading to an increased Linode bill. It would be great if we could get some metrics from GitHub to see how often the bundled JREs are downloaded so we can predict how much traffic we'll see. 1. The host should be available as much as possible. It should not be taken down except for maintenance. 1. The host name should not change over time. We won't be able to go back and modify installers from older releases to point them to a new host. I'm not familiar with this aspect of our Linode setup, so if changing host names is common, we might have to consider having a reverse proxy or something to redirect from the old name to the new name. @DanVanAtta @prastle @RoiEXLab @ron-murhammer Thoughts?
2.0
Move install4j bundled JREs from GitHub to Linode - We bundle JREs with our installer for users that do not have a Java 8 JRE installed on their machine. We currently host these bundled JREs in the triplea-game/assets repo, and the installer downloads them, if needed, directly from GitHub. There are a few problems hosting these files on GitHub: * We can't just provide the "latest" JRE because older installers are built with a hard-coded link to a specific path in the repo. As long as those older builds are considered compatible, we have to keep multiple JRE versions on `HEAD` in the repo (we currently have two versions for three platforms). Each JRE is approximately 350 MiB. * Adding additional JREs increases the time to run the Gradle build on Travis due to the additional download requirements. * Adding additional JREs increases the size of the triplea-game/assets repo significantly. Even when we can remove an older bundled JRE from `HEAD`, it stays in the repo history unless we rewrite history. A static file server is probably a better place to host these resources rather than using a Git repo. **The purpose of this issue is to discuss possibly moving the bundled JREs to one of our Linode servers and serve them from there using Nginx (or something equivalent).** Some issues with self-hosting include: 1. Increased bandwidth possibly leading to an increased Linode bill. It would be great if we could get some metrics from GitHub to see how often the bundled JREs are downloaded so we can predict how much traffic we'll see. 1. The host should be available as much as possible. It should not be taken down except for maintenance. 1. The host name should not change over time. We won't be able to go back and modify installers from older releases to point them to a new host. I'm not familiar with this aspect of our Linode setup, so if changing host names is common, we might have to consider having a reverse proxy or something to redirect from the old name to the new name. @DanVanAtta @prastle @RoiEXLab @ron-murhammer Thoughts?
non_priority
move bundled jres from github to linode we bundle jres with our installer for users that do not have a java jre installed on their machine we currently host these bundled jres in the triplea game assets repo and the installer downloads them if needed directly from github there are a few problems hosting these files on github we can t just provide the latest jre because older installers are built with a hard coded link to a specific path in the repo as long as those older builds are considered compatible we have to keep multiple jre versions on head in the repo we currently have two versions for three platforms each jre is approximately mib adding additional jres increases the time to run the gradle build on travis due to the additional download requirements adding additional jres increases the size of the triplea game assets repo significantly even when we can remove an older bundled jre from head it stays in the repo history unless we rewrite history a static file server is probably a better place to host these resources rather than using a git repo the purpose of this issue is to discuss possibly moving the bundled jres to one of our linode servers and serve them from there using nginx or something equivalent some issues with self hosting include increased bandwidth possibly leading to an increased linode bill it would be great if we could get some metrics from github to see how often the bundled jres are downloaded so we can predict how much traffic we ll see the host should be available as much as possible it should not be taken down except for maintenance the host name should not change over time we won t be able to go back and modify installers from older releases to point them to a new host i m not familiar with this aspect of our linode setup so if changing host names is common we might have to consider having a reverse proxy or something to redirect from the old name to the new name danvanatta prastle roiexlab ron murhammer thoughts
0
451,966
13,044,211,866
IssuesEvent
2020-07-29 03:57:38
kubesphere/console
https://api.github.com/repos/kubesphere/console
closed
multi cluster deployment UI problem
area/console kind/bug kind/need-to-verify priority/high
**Describe the bug** A self provisioner creates a multi-cluster project into two clusters. There are several problems 1. The project list shows zero in member cluster(cluster-01) which is incorrect since it is created in the cluster as you can see in the third screenshot. 2. The detailed project page shows only host info 3. the cluster dropdown menu does not work in the detailed project page If log in using ws admin, everything is fine. <img width="966" alt="Screen Shot 2020-07-25 at 8 19 32 PM" src="https://user-images.githubusercontent.com/28859385/88456931-d028ee80-ceb4-11ea-929b-7e129a9fc16c.png"> <img width="1237" alt="Screen Shot 2020-07-25 at 8 20 08 PM" src="https://user-images.githubusercontent.com/28859385/88456932-d61ecf80-ceb4-11ea-8fe1-895523578b23.png"> <img width="928" alt="Screen Shot 2020-07-25 at 8 20 53 PM" src="https://user-images.githubusercontent.com/28859385/88456933-db7c1a00-ceb4-11ea-8a58-8a7858314021.png"> **Versions used(KubeSphere/Kubernetes)** KubeSphere: 3.0.0-dev Kubernetes: 1.15.12(host), 1.18.6(cluster-01)
1.0
multi cluster deployment UI problem - **Describe the bug** A self provisioner creates a multi-cluster project into two clusters. There are several problems 1. The project list shows zero in member cluster(cluster-01) which is incorrect since it is created in the cluster as you can see in the third screenshot. 2. The detailed project page shows only host info 3. the cluster dropdown menu does not work in the detailed project page If log in using ws admin, everything is fine. <img width="966" alt="Screen Shot 2020-07-25 at 8 19 32 PM" src="https://user-images.githubusercontent.com/28859385/88456931-d028ee80-ceb4-11ea-929b-7e129a9fc16c.png"> <img width="1237" alt="Screen Shot 2020-07-25 at 8 20 08 PM" src="https://user-images.githubusercontent.com/28859385/88456932-d61ecf80-ceb4-11ea-8fe1-895523578b23.png"> <img width="928" alt="Screen Shot 2020-07-25 at 8 20 53 PM" src="https://user-images.githubusercontent.com/28859385/88456933-db7c1a00-ceb4-11ea-8a58-8a7858314021.png"> **Versions used(KubeSphere/Kubernetes)** KubeSphere: 3.0.0-dev Kubernetes: 1.15.12(host), 1.18.6(cluster-01)
priority
multi cluster deployment ui problem describe the bug a self provisioner creates a multi cluster project into two clusters there are several problems the project list shows zero in member cluster cluster which is incorrect since it is created in the cluster as you can see in the third screenshot the detailed project page shows only host info the cluster dropdown menu does not work in the detailed project page if log in using ws admin everything is fine img width alt screen shot at pm src img width alt screen shot at pm src img width alt screen shot at pm src versions used kubesphere kubernetes kubesphere dev kubernetes host cluster
1
68,771
3,292,575,895
IssuesEvent
2015-10-30 15:15:14
thesgc/chembiohub_helpdesk
https://api.github.com/repos/thesgc/chembiohub_helpdesk
closed
Batch uploads list per user doesn't have the project name truncated - so the list does that untidy j
app: ChemReg enhancement name: Karen priority: Low status: New
Batch uploads list per user doesn't have the project name truncated - so the list does that untidy jumbled up thing that has been fixed elsewhere wherever projects are listed. Can the name be truncated here too please
1.0
Batch uploads list per user doesn't have the project name truncated - so the list does that untidy j - Batch uploads list per user doesn't have the project name truncated - so the list does that untidy jumbled up thing that has been fixed elsewhere wherever projects are listed. Can the name be truncated here too please
priority
batch uploads list per user doesn t have the project name truncated so the list does that untidy j batch uploads list per user doesn t have the project name truncated so the list does that untidy jumbled up thing that has been fixed elsewhere wherever projects are listed can the name be truncated here too please
1
470,447
13,537,838,280
IssuesEvent
2020-09-16 11:08:51
buddyboss/buddyboss-platform
https://api.github.com/repos/buddyboss/buddyboss-platform
closed
Mention not working properly on Mobile - Android 9
bug component: activity priority: high priority: medium
**Describe the bug** On mobile, with Android v9. The @mention is not working on replying comments on Activity feed or posting directly on Activity feed **To Reproduce** Steps to reproduce the behavior: Issue can be replicated on demo 1. Use an Android 9 phone 2. Go to Activity feed in demo 3. Reply to some comments by @mention someone and add text or @mention someone then click spacebar 4. See error **Expected behavior** The issue happens when using Android 9. When tagging someone, after you click their name and hit spacebar, this eaither deletes the TAG or doubles the TAG The behavior is randomly happening after tagging. On some devices, @mention text @mention the "2nd @mention" will not work also. On Android 10, this is randomly happening, after tagging someone in @mention, this creates a spacebar automatically, but sometimes when you hit spacebar, this also deletes the tag **Notes** This issue is tested on our demo using actual device and Browserstack **Screenshots** Android 9 browserstack - https://drive.google.com/file/d/1u-trk3nxBWa4gXovzB0I_yPZvYrt4F3E/view?usp=sharing Android 10 browserstack - https://drive.google.com/file/d/1dH_6zPM4PM3oc9UvqjpzTZKn65r1fCdK/view?usp=sharing Actual Android 9 - **Support ticket links** https://secure.helpscout.net/conversation/1237123523/85258
2.0
Mention not working properly on Mobile - Android 9 - **Describe the bug** On mobile, with Android v9. The @mention is not working on replying comments on Activity feed or posting directly on Activity feed **To Reproduce** Steps to reproduce the behavior: Issue can be replicated on demo 1. Use an Android 9 phone 2. Go to Activity feed in demo 3. Reply to some comments by @mention someone and add text or @mention someone then click spacebar 4. See error **Expected behavior** The issue happens when using Android 9. When tagging someone, after you click their name and hit spacebar, this eaither deletes the TAG or doubles the TAG The behavior is randomly happening after tagging. On some devices, @mention text @mention the "2nd @mention" will not work also. On Android 10, this is randomly happening, after tagging someone in @mention, this creates a spacebar automatically, but sometimes when you hit spacebar, this also deletes the tag **Notes** This issue is tested on our demo using actual device and Browserstack **Screenshots** Android 9 browserstack - https://drive.google.com/file/d/1u-trk3nxBWa4gXovzB0I_yPZvYrt4F3E/view?usp=sharing Android 10 browserstack - https://drive.google.com/file/d/1dH_6zPM4PM3oc9UvqjpzTZKn65r1fCdK/view?usp=sharing Actual Android 9 - **Support ticket links** https://secure.helpscout.net/conversation/1237123523/85258
priority
mention not working properly on mobile android describe the bug on mobile with android the mention is not working on replying comments on activity feed or posting directly on activity feed to reproduce steps to reproduce the behavior issue can be replicated on demo use an android phone go to activity feed in demo reply to some comments by mention someone and add text or mention someone then click spacebar see error expected behavior the issue happens when using android when tagging someone after you click their name and hit spacebar this eaither deletes the tag or doubles the tag the behavior is randomly happening after tagging on some devices mention text mention the mention will not work also on android this is randomly happening after tagging someone in mention this creates a spacebar automatically but sometimes when you hit spacebar this also deletes the tag notes this issue is tested on our demo using actual device and browserstack screenshots android browserstack android browserstack actual android support ticket links
1
15,521
9,559,157,443
IssuesEvent
2019-05-03 15:55:20
manusa/isotope-mail
https://api.github.com/repos/manusa/isotope-mail
opened
Security alert: tar v2.2.1 (node-sass)
security technical
Node-sass v4.9.3 depends on tar v2.2.0 which has a security vulnerability ``` +-- node-sass@4.9.3 `-- node-gyp@3.8.0 `-- tar@2.2.1 ``` > CVE-2018-20834 More information > high severity > Vulnerable versions: < 4.4.2 > Patched version: 4.4.2 > A vulnerability was found in node-tar before version 4.4.2. An Arbitrary File Overwrite issue exists when extracting a tarball containing a hardlink to a file that already exists on the system, in conjunction with a later plain file with the same name as the hardlink. This plain file content replaces the existing file content.
True
Security alert: tar v2.2.1 (node-sass) - Node-sass v4.9.3 depends on tar v2.2.0 which has a security vulnerability ``` +-- node-sass@4.9.3 `-- node-gyp@3.8.0 `-- tar@2.2.1 ``` > CVE-2018-20834 More information > high severity > Vulnerable versions: < 4.4.2 > Patched version: 4.4.2 > A vulnerability was found in node-tar before version 4.4.2. An Arbitrary File Overwrite issue exists when extracting a tarball containing a hardlink to a file that already exists on the system, in conjunction with a later plain file with the same name as the hardlink. This plain file content replaces the existing file content.
non_priority
security alert tar node sass node sass depends on tar which has a security vulnerability node sass node gyp tar cve more information high severity vulnerable versions patched version a vulnerability was found in node tar before version an arbitrary file overwrite issue exists when extracting a tarball containing a hardlink to a file that already exists on the system in conjunction with a later plain file with the same name as the hardlink this plain file content replaces the existing file content
0
720
2,587,684,868
IssuesEvent
2015-02-17 20:01:08
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
TransactionalMap.size() returns inconsistent results during a transaction
Team: Core Type: Defect
Hazelcast version: 3.4 Node type: in-cluster Configuration: a basic unit-test setup The problem is probably best described by the following JUnit example: @Test public void transactionalMapCanHaveNegativeSize() throws Exception { final TransactionContext transactionContext = hazelcastInstance.newTransactionContext(new TransactionOptions().setTransactionType(TransactionOptions.TransactionType.LOCAL)); transactionContext.beginTransaction(); final TransactionalMap<String, String> map = transactionContext.getMap("test-map"); map.put("k", "v"); map.remove("k"); assertEquals(0, map.size()); // failure: expected 0 got -1 } I think the problem is in `TransactionalMap#remove(Object)`, where it overwrites the "insert" operation in its internal transaction map with a "remove" - so you end up with a remove for an object that never existed. Then when it comes to decide size, it counts the removal with negative size. I think the right thing to do here is to remove the insertion `TxValueWrapper` from `txMap`. I'm happy to write up a PR if the above is agreed - the reason I haven't already is that I'm surprised this hasn't been stumbled on before, so I'm wondering if there's some reason it would be intentional. The work-around we're using is to get the size of the keyset (which returns the expected result).
1.0
TransactionalMap.size() returns inconsistent results during a transaction - Hazelcast version: 3.4 Node type: in-cluster Configuration: a basic unit-test setup The problem is probably best described by the following JUnit example: @Test public void transactionalMapCanHaveNegativeSize() throws Exception { final TransactionContext transactionContext = hazelcastInstance.newTransactionContext(new TransactionOptions().setTransactionType(TransactionOptions.TransactionType.LOCAL)); transactionContext.beginTransaction(); final TransactionalMap<String, String> map = transactionContext.getMap("test-map"); map.put("k", "v"); map.remove("k"); assertEquals(0, map.size()); // failure: expected 0 got -1 } I think the problem is in `TransactionalMap#remove(Object)`, where it overwrites the "insert" operation in its internal transaction map with a "remove" - so you end up with a remove for an object that never existed. Then when it comes to decide size, it counts the removal with negative size. I think the right thing to do here is to remove the insertion `TxValueWrapper` from `txMap`. I'm happy to write up a PR if the above is agreed - the reason I haven't already is that I'm surprised this hasn't been stumbled on before, so I'm wondering if there's some reason it would be intentional. The work-around we're using is to get the size of the keyset (which returns the expected result).
non_priority
transactionalmap size returns inconsistent results during a transaction hazelcast version node type in cluster configuration a basic unit test setup the problem is probably best described by the following junit example test public void transactionalmapcanhavenegativesize throws exception final transactioncontext transactioncontext hazelcastinstance newtransactioncontext new transactionoptions settransactiontype transactionoptions transactiontype local transactioncontext begintransaction final transactionalmap map transactioncontext getmap test map map put k v map remove k assertequals map size failure expected got i think the problem is in transactionalmap remove object where it overwrites the insert operation in its internal transaction map with a remove so you end up with a remove for an object that never existed then when it comes to decide size it counts the removal with negative size i think the right thing to do here is to remove the insertion txvaluewrapper from txmap i m happy to write up a pr if the above is agreed the reason i haven t already is that i m surprised this hasn t been stumbled on before so i m wondering if there s some reason it would be intentional the work around we re using is to get the size of the keyset which returns the expected result
0
97,842
12,264,389,529
IssuesEvent
2020-05-07 04:08:28
dotnet/winforms
https://api.github.com/repos/dotnet/winforms
closed
System.Windows.Forms.Design.ImageListCodeDomSerializer appears to be missing in the list of System.Design type forwards
area: designer support tenet-compatibility
https://github.com/dotnet/winforms/blob/139ad6f018b334a12fdd30104bc9776b79466295/src/System.Design/src/System.Design.Forwards.cs#L28-L33
1.0
System.Windows.Forms.Design.ImageListCodeDomSerializer appears to be missing in the list of System.Design type forwards - https://github.com/dotnet/winforms/blob/139ad6f018b334a12fdd30104bc9776b79466295/src/System.Design/src/System.Design.Forwards.cs#L28-L33
non_priority
system windows forms design imagelistcodedomserializer appears to be missing in the list of system design type forwards
0
230,120
7,604,386,611
IssuesEvent
2018-04-30 00:35:05
kubeflow/kubeflow
https://api.github.com/repos/kubeflow/kubeflow
closed
Add resource request and limit fields to tf-job ksonnet prototype
priority/p2
When jobs are run with default or incorrect resource requests they will be evicted if resources cannot be allocated when needed, e.g. ```bash > kubectl get pods -n rl --show-all agents-ppo-ant-feedforward-93168c9e-master-... 0/1 Evicted 0 19h agents-ppo-ant-feedforward-93168c9e-master-... 0/1 Evicted 0 18h agents-ppo-ant-feedforward-93168c9e-master-... 1/1 Running 0 9h agents-ppo-ant-feedforward-93168c9e-master-... 0/1 Evicted 0 18h agents-ppo-ant-feedforward-93168c9e-tensorboard-... 1/1 Running 0 19h agents-ppo-ant-ee3e499e-master-aiod-0-768ls 1/1 Running 0 18h agents-ppo-ant-ee3e499e-tensorboard-aiod-76dbf4brlz 1/1 Running 0 18h ... > kubectl describe pod -n rl agents-ppo-ant-feedforward-93168c9e-master-3ti2-0-9st9r ... Status: Failed Reason: Evicted Message: The node was low on resource: memory. ... ``` Proposing to add those fields to tf-job prototype.
1.0
Add resource request and limit fields to tf-job ksonnet prototype - When jobs are run with default or incorrect resource requests they will be evicted if resources cannot be allocated when needed, e.g. ```bash > kubectl get pods -n rl --show-all agents-ppo-ant-feedforward-93168c9e-master-... 0/1 Evicted 0 19h agents-ppo-ant-feedforward-93168c9e-master-... 0/1 Evicted 0 18h agents-ppo-ant-feedforward-93168c9e-master-... 1/1 Running 0 9h agents-ppo-ant-feedforward-93168c9e-master-... 0/1 Evicted 0 18h agents-ppo-ant-feedforward-93168c9e-tensorboard-... 1/1 Running 0 19h agents-ppo-ant-ee3e499e-master-aiod-0-768ls 1/1 Running 0 18h agents-ppo-ant-ee3e499e-tensorboard-aiod-76dbf4brlz 1/1 Running 0 18h ... > kubectl describe pod -n rl agents-ppo-ant-feedforward-93168c9e-master-3ti2-0-9st9r ... Status: Failed Reason: Evicted Message: The node was low on resource: memory. ... ``` Proposing to add those fields to tf-job prototype.
priority
add resource request and limit fields to tf job ksonnet prototype when jobs are run with default or incorrect resource requests they will be evicted if resources cannot be allocated when needed e g bash kubectl get pods n rl show all agents ppo ant feedforward master evicted agents ppo ant feedforward master evicted agents ppo ant feedforward master running agents ppo ant feedforward master evicted agents ppo ant feedforward tensorboard running agents ppo ant master aiod running agents ppo ant tensorboard aiod running kubectl describe pod n rl agents ppo ant feedforward master status failed reason evicted message the node was low on resource memory proposing to add those fields to tf job prototype
1
694,057
23,800,331,491
IssuesEvent
2022-09-03 07:07:55
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
crash when disabling ads
bug crash feature/rewards priority/P2 QA/Yes release-notes/include feature/ads OS/Android
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!-- Provide a brief description of the issue --> Brave is crashing when attempting to disable `Ads` via the `Brave Rewards` settings page. We'll need to get this one fixed ASAP as disabling/enabling ads is an important feature that a lot of users use. Crashing at https://github.com/brave/brave-core/blob/master/components/brave_ads/browser/ads_service_impl.cc#L747 as per @SergeyZhukovsky. ## Steps to reproduce <!-- Please add a series of steps to reproduce the issue --> 1. install/launch `1.43.88 Chromium: 105.0.5195.68` 2. enable rewards via the panel 3. once enabled, go into the rewards settings (`brave://rewards`) and attempt to disable `Brave Private Ads` via the toggle ## Actual result <!-- Please add screenshots if needed --> https://user-images.githubusercontent.com/2602313/188259073-ce2131f2-f3cc-4d4d-9235-0292dfccd448.mp4 ## Expected result Brave shouldn't be crashing when users attempt to disable `Ads` via `Brave Rewards` settings. ## Issue reproduces how often <!-- [Easily reproduced/Intermittent issue/No steps to reproduce] --> 100% reproducible using the STR/Cases outlined above. ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current Play Store version? `Yes` - Can you reproduce this issue with the current Play Store Beta version? `Yes` - Can you reproduce this issue with the current Play Store Nightly version? `Yes` ## Device details - Install type (ARM, x86): `ARM` - Device type (Phone, Tablet, Phablet): `Samsung S10+` (Phone) - Android version: `Android 12` ## Brave version ``` Brave | 1.43.88 Chromium: 105.0.5195.68 (Official Build) (32-bit) --- | --- Revision | ad13e82529051bac6a0e65f455e6d7a1e5fd7938-refs/branch-heads/5195@{#903} OS | Android 12; Build/SP1A.210812.016 ``` ### Website problems only - Does the issue resolve itself when disabling Brave Shields? `N/A` - Does the issue resolve itself when disabling Brave Rewards? `N/A` - Is the issue reproducible on the latest version of Chrome? `N/A` ### Additional information <!-- Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue -->
1.0
crash when disabling ads - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!-- Provide a brief description of the issue --> Brave is crashing when attempting to disable `Ads` via the `Brave Rewards` settings page. We'll need to get this one fixed ASAP as disabling/enabling ads is an important feature that a lot of users use. Crashing at https://github.com/brave/brave-core/blob/master/components/brave_ads/browser/ads_service_impl.cc#L747 as per @SergeyZhukovsky. ## Steps to reproduce <!-- Please add a series of steps to reproduce the issue --> 1. install/launch `1.43.88 Chromium: 105.0.5195.68` 2. enable rewards via the panel 3. once enabled, go into the rewards settings (`brave://rewards`) and attempt to disable `Brave Private Ads` via the toggle ## Actual result <!-- Please add screenshots if needed --> https://user-images.githubusercontent.com/2602313/188259073-ce2131f2-f3cc-4d4d-9235-0292dfccd448.mp4 ## Expected result Brave shouldn't be crashing when users attempt to disable `Ads` via `Brave Rewards` settings. ## Issue reproduces how often <!-- [Easily reproduced/Intermittent issue/No steps to reproduce] --> 100% reproducible using the STR/Cases outlined above. ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current Play Store version? `Yes` - Can you reproduce this issue with the current Play Store Beta version? `Yes` - Can you reproduce this issue with the current Play Store Nightly version? `Yes` ## Device details - Install type (ARM, x86): `ARM` - Device type (Phone, Tablet, Phablet): `Samsung S10+` (Phone) - Android version: `Android 12` ## Brave version ``` Brave | 1.43.88 Chromium: 105.0.5195.68 (Official Build) (32-bit) --- | --- Revision | ad13e82529051bac6a0e65f455e6d7a1e5fd7938-refs/branch-heads/5195@{#903} OS | Android 12; Build/SP1A.210812.016 ``` ### Website problems only - Does the issue resolve itself when disabling Brave Shields? `N/A` - Does the issue resolve itself when disabling Brave Rewards? `N/A` - Is the issue reproducible on the latest version of Chrome? `N/A` ### Additional information <!-- Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue -->
priority
crash when disabling ads have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description brave is crashing when attempting to disable ads via the brave rewards settings page we ll need to get this one fixed asap as disabling enabling ads is an important feature that a lot of users use crashing at as per sergeyzhukovsky steps to reproduce install launch chromium enable rewards via the panel once enabled go into the rewards settings brave rewards and attempt to disable brave private ads via the toggle actual result expected result brave shouldn t be crashing when users attempt to disable ads via brave rewards settings issue reproduces how often reproducible using the str cases outlined above version channel information can you reproduce this issue with the current play store version yes can you reproduce this issue with the current play store beta version yes can you reproduce this issue with the current play store nightly version yes device details install type arm arm device type phone tablet phablet samsung phone android version android brave version brave chromium official build bit revision refs branch heads os android build website problems only does the issue resolve itself when disabling brave shields n a does the issue resolve itself when disabling brave rewards n a is the issue reproducible on the latest version of chrome n a additional information
1
307,192
23,188,814,336
IssuesEvent
2022-08-01 10:45:48
0x192/universal-android-debloater
https://api.github.com/repos/0x192/universal-android-debloater
opened
com.miui.home breaks menu button
package::documentation
My phone model: Mi 11 Lite 5G **Packages documentation to update:** ``` com.miui.home ... ``` ## Documentation change Change It from Advanced to Expert There was a lot of stuff to fill out here that i didn't know... So let me just get it straight: Deleting this package breaks the menu button (i only tested it on my phone sooo i don't know if other people have this issue)
1.0
com.miui.home breaks menu button - My phone model: Mi 11 Lite 5G **Packages documentation to update:** ``` com.miui.home ... ``` ## Documentation change Change It from Advanced to Expert There was a lot of stuff to fill out here that i didn't know... So let me just get it straight: Deleting this package breaks the menu button (i only tested it on my phone sooo i don't know if other people have this issue)
non_priority
com miui home breaks menu button my phone model mi lite packages documentation to update com miui home documentation change change it from advanced to expert there was a lot of stuff to fill out here that i didn t know so let me just get it straight deleting this package breaks the menu button i only tested it on my phone sooo i don t know if other people have this issue
0
194,676
22,262,070,538
IssuesEvent
2022-06-10 02:04:21
Nivaskumark/kernel_v4.19.72_old
https://api.github.com/repos/Nivaskumark/kernel_v4.19.72_old
reopened
CVE-2022-24958 (High) detected in linux-yoctov5.4.51
security vulnerability
## CVE-2022-24958 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary> <p> <p>Yocto Linux Embedded kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p> <p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/kernel_v4.19.72/commit/ce49083a1c14be2d13cb5e878257d293e6c748bc">ce49083a1c14be2d13cb5e878257d293e6c748bc</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> drivers/usb/gadget/legacy/inode.c in the Linux kernel through 5.16.8 mishandles dev->buf release. <p>Publish Date: 2022-02-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24958>CVE-2022-24958</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24958">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24958</a></p> <p>Release Date: 2022-02-11</p> <p>Fix Resolution: v5.16</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-24958 (High) detected in linux-yoctov5.4.51 - ## CVE-2022-24958 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary> <p> <p>Yocto Linux Embedded kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p> <p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/kernel_v4.19.72/commit/ce49083a1c14be2d13cb5e878257d293e6c748bc">ce49083a1c14be2d13cb5e878257d293e6c748bc</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> drivers/usb/gadget/legacy/inode.c in the Linux kernel through 5.16.8 mishandles dev->buf release. <p>Publish Date: 2022-02-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24958>CVE-2022-24958</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24958">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24958</a></p> <p>Release Date: 2022-02-11</p> <p>Fix Resolution: v5.16</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in linux cve high severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details drivers usb gadget legacy inode c in the linux kernel through mishandles dev buf release publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
12,846
21,005,576,820
IssuesEvent
2022-03-29 22:14:53
NASA-PDS/pds-registry-app
https://api.github.com/repos/NASA-PDS/pds-registry-app
opened
Update the user documentation of the registry with the docker compose deployments
requirement needs:triage
<!-- For more information on how to populate this new feature request, see the PDS Wiki on User Story Development: https://github.com/NASA-PDS/nasa-pds.github.io/wiki/Issue-Tracking#user-story-development --> ## 💪 Motivation ...so that a user can deploy the scalable harvest service easily, as well as the rest of the registry. - [ ] harvest service - [ ] registry core (opensearch + api) - [ ] registry-loader ## 📖 Additional Details <!-- Please prove any additional details or information that could help provide some context for the user story. --> ## ⚖️ Acceptance Criteria **Given** the published documentation of the registry **When I perform** reading of the doc **Then I expect** I can easily find the guide to deploy each components of the registry with docker compose <!-- For Internal Dev Team Use --> ## ⚙️ Engineering Details <!-- Provide some design / implementation details and/or a sub-task checklist as needed. Convert issue to Epic if estimate is outside the scope of 1 sprint. -->
1.0
Update the user documentation of the registry with the docker compose deployments - <!-- For more information on how to populate this new feature request, see the PDS Wiki on User Story Development: https://github.com/NASA-PDS/nasa-pds.github.io/wiki/Issue-Tracking#user-story-development --> ## 💪 Motivation ...so that a user can deploy the scalable harvest service easily, as well as the rest of the registry. - [ ] harvest service - [ ] registry core (opensearch + api) - [ ] registry-loader ## 📖 Additional Details <!-- Please prove any additional details or information that could help provide some context for the user story. --> ## ⚖️ Acceptance Criteria **Given** the published documentation of the registry **When I perform** reading of the doc **Then I expect** I can easily find the guide to deploy each components of the registry with docker compose <!-- For Internal Dev Team Use --> ## ⚙️ Engineering Details <!-- Provide some design / implementation details and/or a sub-task checklist as needed. Convert issue to Epic if estimate is outside the scope of 1 sprint. -->
non_priority
update the user documentation of the registry with the docker compose deployments for more information on how to populate this new feature request see the pds wiki on user story development 💪 motivation so that a user can deploy the scalable harvest service easily as well as the rest of the registry harvest service registry core opensearch api registry loader 📖 additional details ⚖️ acceptance criteria given the published documentation of the registry when i perform reading of the doc then i expect i can easily find the guide to deploy each components of the registry with docker compose ⚙️ engineering details provide some design implementation details and or a sub task checklist as needed convert issue to epic if estimate is outside the scope of sprint
0
8,142
11,438,101,161
IssuesEvent
2020-02-05 02:14:11
rssab/library
https://api.github.com/repos/rssab/library
opened
Unreturned Book Flow
must have requirement
Administrators should have the ability to check for and be notified for books that are late within the system.
1.0
Unreturned Book Flow - Administrators should have the ability to check for and be notified for books that are late within the system.
non_priority
unreturned book flow administrators should have the ability to check for and be notified for books that are late within the system
0
496,197
14,342,433,060
IssuesEvent
2020-11-28 03:07:13
metasfresh/metasfresh
https://api.github.com/repos/metasfresh/metasfresh
closed
Implement support for MSV3 version 1
branch:release bug_please_use_type:bug_instead priority:high type:enhancement
### Is this a bug or feature request? ### What is the current behavior? #### Which are the steps to reproduce? ### What is the expected or desired behavior?
1.0
Implement support for MSV3 version 1 - ### Is this a bug or feature request? ### What is the current behavior? #### Which are the steps to reproduce? ### What is the expected or desired behavior?
priority
implement support for version is this a bug or feature request what is the current behavior which are the steps to reproduce what is the expected or desired behavior
1
821,584
30,827,592,419
IssuesEvent
2023-08-01 21:28:33
Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2
https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2
opened
Event to trigger quel'thalas culture conversion option for scourge survivors
suggestion :question: priority low :grey_exclamation:
------------------------------------------------------------------------------------------------------------ **Describe your suggestion in full detail below:** Title
1.0
Event to trigger quel'thalas culture conversion option for scourge survivors - ------------------------------------------------------------------------------------------------------------ **Describe your suggestion in full detail below:** Title
priority
event to trigger quel thalas culture conversion option for scourge survivors describe your suggestion in full detail below title
1
683,306
23,376,276,892
IssuesEvent
2022-08-11 03:40:12
kubevela/kubevela
https://api.github.com/repos/kubevela/kubevela
closed
[Feature] add policy definitions for advanced internal policy
type/enhancement good first issue help wanted priority/nice-to-have effort/small area/policy
**Is your feature request related to a problem? Please describe.** <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> Previous PR(https://github.com/kubevela/kubevela/pull/3894) adds topology and override policy definition for schema. Advanced policies' are also welcomed, including apply-once, garbage-collect, shared-resouece. **Describe the solution you'd like** <!-- A clear and concise description of what you want to happen. --> **Describe alternatives you've considered** <!-- A clear and concise description of any alternative solutions or features you've considered. --> **Additional context** <!-- Add any other context or screenshots about the feature request here. -->
1.0
[Feature] add policy definitions for advanced internal policy - **Is your feature request related to a problem? Please describe.** <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> Previous PR(https://github.com/kubevela/kubevela/pull/3894) adds topology and override policy definition for schema. Advanced policies' are also welcomed, including apply-once, garbage-collect, shared-resouece. **Describe the solution you'd like** <!-- A clear and concise description of what you want to happen. --> **Describe alternatives you've considered** <!-- A clear and concise description of any alternative solutions or features you've considered. --> **Additional context** <!-- Add any other context or screenshots about the feature request here. -->
priority
add policy definitions for advanced internal policy is your feature request related to a problem please describe a clear and concise description of what the problem is ex i m always frustrated when previous pr adds topology and override policy definition for schema advanced policies are also welcomed including apply once garbage collect shared resouece describe the solution you d like a clear and concise description of what you want to happen describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here
1
139,366
31,469,203,877
IssuesEvent
2023-08-30 06:06:30
FerretDB/FerretDB
https://api.github.com/repos/FerretDB/FerretDB
closed
Support `ordered` `insert`s for SQLite
code/feature backend/sqlite
### What should be done? * The handler should set `_id`s in documents and validate them. The backend should receive only valid documents. * The handler should support `ordered` `insert`s. * Tests should pass. Batching will be implemented in #3271. ### Where? See #3223. ### Definition of Done - the handler updated; - integration/compatibility tests un-skipped.
1.0
Support `ordered` `insert`s for SQLite - ### What should be done? * The handler should set `_id`s in documents and validate them. The backend should receive only valid documents. * The handler should support `ordered` `insert`s. * Tests should pass. Batching will be implemented in #3271. ### Where? See #3223. ### Definition of Done - the handler updated; - integration/compatibility tests un-skipped.
non_priority
support ordered insert s for sqlite what should be done the handler should set id s in documents and validate them the backend should receive only valid documents the handler should support ordered insert s tests should pass batching will be implemented in where see definition of done the handler updated integration compatibility tests un skipped
0
80,242
15,586,273,004
IssuesEvent
2021-03-18 01:33:52
peterwkc85/Spring_Rest
https://api.github.com/repos/peterwkc85/Spring_Rest
opened
CVE-2020-36188 (High) detected in jackson-databind-2.8.6.jar
security vulnerability
## CVE-2020-36188 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: Spring_Rest/spring-restbucks-master/spring-restbucks-master/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.6/jackson-databind-2.8.6.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-data-rest-1.5.0.RELEASE.jar (Root Library) - :x: **jackson-databind-2.8.6.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.JNDIConnectionSource. <p>Publish Date: 2021-01-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36188>CVE-2020-36188</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2996">https://github.com/FasterXML/jackson-databind/issues/2996</a></p> <p>Release Date: 2021-01-06</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-36188 (High) detected in jackson-databind-2.8.6.jar - ## CVE-2020-36188 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: Spring_Rest/spring-restbucks-master/spring-restbucks-master/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.6/jackson-databind-2.8.6.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-data-rest-1.5.0.RELEASE.jar (Root Library) - :x: **jackson-databind-2.8.6.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.JNDIConnectionSource. <p>Publish Date: 2021-01-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36188>CVE-2020-36188</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2996">https://github.com/FasterXML/jackson-databind/issues/2996</a></p> <p>Release Date: 2021-01-06</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file spring rest spring restbucks master spring restbucks master pom xml path to vulnerable library root repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter data rest release jar root library x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com newrelic agent deps ch qos logback core db jndiconnectionsource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
0
663,046
22,160,467,022
IssuesEvent
2022-06-04 12:37:01
asrob-uc3m/robotDevastation-playground
https://api.github.com/repos/asrob-uc3m/robotDevastation-playground
closed
Orange Pi Lite robot
hardware priority: idea closed due to inactivity > 1 month
_From @David-Estevez on May 31, 2017 15:32_ Rd Ambassador was ok, but it has several flaws that are hard to solve, the main one being that there are too many things to put on the mobile platform, and not enough space (camera, QR, Raspberry Pi...). So I have started the development of a cool mobile robot to be the documented basic robot to be used in Robot Devastation. The basic features it will have are: * Instead of Raspberry Pi, it will use an Orange Pi Lite. The advantages of these board are the low cost, small size, and integrated WiFi card. We are also researching the possibility of using a camera connected to the camera bus of the board (does not require USB). * Camera (either USB webcam or camera connected to camera bus). * A shield for the Orange Pi with the following elements: * H-bridge or similar motor controller for 2 DC motors * Battery status feedback (related to #4 ) * Regulator (which might be placed outside the board if we find that it is a better solution). * LEDs for status feedback * Buttons for user input * It will use consumer camera batteries (or similar), that are easy to source as replacements, cheap and have their own safe and easy-to-use charger. _Copied from original issue: asrob-uc3m/robotDevastation-robots#35_
1.0
Orange Pi Lite robot - _From @David-Estevez on May 31, 2017 15:32_ Rd Ambassador was ok, but it has several flaws that are hard to solve, the main one being that there are too many things to put on the mobile platform, and not enough space (camera, QR, Raspberry Pi...). So I have started the development of a cool mobile robot to be the documented basic robot to be used in Robot Devastation. The basic features it will have are: * Instead of Raspberry Pi, it will use an Orange Pi Lite. The advantages of these board are the low cost, small size, and integrated WiFi card. We are also researching the possibility of using a camera connected to the camera bus of the board (does not require USB). * Camera (either USB webcam or camera connected to camera bus). * A shield for the Orange Pi with the following elements: * H-bridge or similar motor controller for 2 DC motors * Battery status feedback (related to #4 ) * Regulator (which might be placed outside the board if we find that it is a better solution). * LEDs for status feedback * Buttons for user input * It will use consumer camera batteries (or similar), that are easy to source as replacements, cheap and have their own safe and easy-to-use charger. _Copied from original issue: asrob-uc3m/robotDevastation-robots#35_
priority
orange pi lite robot from david estevez on may rd ambassador was ok but it has several flaws that are hard to solve the main one being that there are too many things to put on the mobile platform and not enough space camera qr raspberry pi so i have started the development of a cool mobile robot to be the documented basic robot to be used in robot devastation the basic features it will have are instead of raspberry pi it will use an orange pi lite the advantages of these board are the low cost small size and integrated wifi card we are also researching the possibility of using a camera connected to the camera bus of the board does not require usb camera either usb webcam or camera connected to camera bus a shield for the orange pi with the following elements h bridge or similar motor controller for dc motors battery status feedback related to regulator which might be placed outside the board if we find that it is a better solution leds for status feedback buttons for user input it will use consumer camera batteries or similar that are easy to source as replacements cheap and have their own safe and easy to use charger copied from original issue asrob robotdevastation robots
1
542,302
15,858,369,347
IssuesEvent
2021-04-08 06:36:23
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
opened
API publisher overview page improvements
API-M 4.0.0 Priority/Low Type/Improvement
### Describe your problem(s) In the current overview page even though it shows current configs, to change anything we need to navigate through side menu and find the corresponding page which is not intuitive plus can take sometime for a first time user ### Describe your solution Provide links from the overview page itself for the relevant config page in the same way we have done in deployment,test options. Ex: Infront of 'Business Plan' attribute there will be a link to take user to relevant tab
1.0
API publisher overview page improvements - ### Describe your problem(s) In the current overview page even though it shows current configs, to change anything we need to navigate through side menu and find the corresponding page which is not intuitive plus can take sometime for a first time user ### Describe your solution Provide links from the overview page itself for the relevant config page in the same way we have done in deployment,test options. Ex: Infront of 'Business Plan' attribute there will be a link to take user to relevant tab
priority
api publisher overview page improvements describe your problem s in the current overview page even though it shows current configs to change anything we need to navigate through side menu and find the corresponding page which is not intuitive plus can take sometime for a first time user describe your solution provide links from the overview page itself for the relevant config page in the same way we have done in deployment test options ex infront of business plan attribute there will be a link to take user to relevant tab
1
249,167
26,889,989,151
IssuesEvent
2023-02-06 08:08:30
valtech-ch/microservice-kubernetes-cluster
https://api.github.com/repos/valtech-ch/microservice-kubernetes-cluster
closed
CVE-2020-36187 (High) detected in jackson-databind-2.9.8.jar - autoclosed
security vulnerability
## CVE-2020-36187 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /functions/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.8/11283f21cc480aa86c4df7a0a3243ec508372ed2/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - spring-cloud-starter-function-web-4.0.1.jar (Root Library) - spring-boot-starter-web-2.7.8.jar - spring-boot-starter-json-2.7.8.jar - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/valtech-ch/microservice-kubernetes-cluster/commit/335a4047c89f52dfe860e93daefb32dc86a521a2">335a4047c89f52dfe860e93daefb32dc86a521a2</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.datasources.SharedPoolDataSource. <p>Publish Date: 2021-01-06 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-36187>CVE-2020-36187</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2021-01-06</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-36187 (High) detected in jackson-databind-2.9.8.jar - autoclosed - ## CVE-2020-36187 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /functions/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.8/11283f21cc480aa86c4df7a0a3243ec508372ed2/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - spring-cloud-starter-function-web-4.0.1.jar (Root Library) - spring-boot-starter-web-2.7.8.jar - spring-boot-starter-json-2.7.8.jar - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/valtech-ch/microservice-kubernetes-cluster/commit/335a4047c89f52dfe860e93daefb32dc86a521a2">335a4047c89f52dfe860e93daefb32dc86a521a2</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.datasources.SharedPoolDataSource. <p>Publish Date: 2021-01-06 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-36187>CVE-2020-36187</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2021-01-06</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file functions build gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring cloud starter function web jar root library spring boot starter web jar spring boot starter json jar x jackson databind jar vulnerable library found in head commit a href found in base branch develop vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp dbcp datasources sharedpooldatasource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with mend
0
44,913
23,821,570,522
IssuesEvent
2022-09-05 11:44:37
xamarin/xamarin-macios
https://api.github.com/repos/xamarin/xamarin-macios
closed
Document how to measure iOS startup times
performance
Just like the amazing doc here: * https://github.com/xamarin/xamarin-macios/wiki/Profiling I think we need one on measuring iOS startup times? And perhaps other Apple platforms? We have this document from Apple using instruments: * https://developer.apple.com/documentation/xcode/reducing-your-app-s-launch-time But I think there is some slowdown once you attach instruments to an iOS app (as seen in https://github.com/dotnet/maui/issues/7687). We did this for the measurements at https://aka.ms/dotnetperfstatus * https://github.com/dotnet/performance/pull/2355 * https://github.com/dotnet/performance/pull/2404 If we make a new doc, we can link to it from: * https://github.com/dotnet/maui/wiki/Profiling-.NET-MAUI-Apps
True
Document how to measure iOS startup times - Just like the amazing doc here: * https://github.com/xamarin/xamarin-macios/wiki/Profiling I think we need one on measuring iOS startup times? And perhaps other Apple platforms? We have this document from Apple using instruments: * https://developer.apple.com/documentation/xcode/reducing-your-app-s-launch-time But I think there is some slowdown once you attach instruments to an iOS app (as seen in https://github.com/dotnet/maui/issues/7687). We did this for the measurements at https://aka.ms/dotnetperfstatus * https://github.com/dotnet/performance/pull/2355 * https://github.com/dotnet/performance/pull/2404 If we make a new doc, we can link to it from: * https://github.com/dotnet/maui/wiki/Profiling-.NET-MAUI-Apps
non_priority
document how to measure ios startup times just like the amazing doc here i think we need one on measuring ios startup times and perhaps other apple platforms we have this document from apple using instruments but i think there is some slowdown once you attach instruments to an ios app as seen in we did this for the measurements at if we make a new doc we can link to it from
0
60,675
8,453,885,329
IssuesEvent
2018-10-20 20:01:31
donmccurdy/three-pathfinding
https://api.github.com/repos/donmccurdy/three-pathfinding
closed
Navmesh representation
documentation
"This library does not build navigation meshes for you — instead, create a navigation mesh using Blender, Recast (CLI), or another tool." I'd say that the original Recast writes the output to some binary structure which is more than just geometry. I don't expect it's that easy to just extract the geometry into the .gltf. The question is: are you sure of what you state above, I mean the Recast part? I am not commenting Blender or CLI outputs.
1.0
Navmesh representation - "This library does not build navigation meshes for you — instead, create a navigation mesh using Blender, Recast (CLI), or another tool." I'd say that the original Recast writes the output to some binary structure which is more than just geometry. I don't expect it's that easy to just extract the geometry into the .gltf. The question is: are you sure of what you state above, I mean the Recast part? I am not commenting Blender or CLI outputs.
non_priority
navmesh representation this library does not build navigation meshes for you — instead create a navigation mesh using blender recast cli or another tool i d say that the original recast writes the output to some binary structure which is more than just geometry i don t expect it s that easy to just extract the geometry into the gltf the question is are you sure of what you state above i mean the recast part i am not commenting blender or cli outputs
0
25,321
12,237,254,288
IssuesEvent
2020-05-04 17:42:48
terraform-providers/terraform-provider-aws
https://api.github.com/repos/terraform-providers/terraform-provider-aws
closed
TF plan & apply not seeing correct associate_public_ip_address
bug service/ec2 stale
_This issue was originally opened by @Gary-Armstrong as hashicorp/terraform#9811. It was migrated here as part of the [provider split](https://www.hashicorp.com/blog/upcoming-provider-changes-in-terraform-0-10/). The original body of the issue is below._ <hr> The symptom is that I get `-/+ aws_instance.stage` on each exec of plan and apply. Changing `associate_public_ip_address` from true to false did recreate the instance as expected, but subsequent plans and applies ALSO recreated the instance due to `associate_public_ip_address: "true" => "false" (forces new resource)`. Naturally I inspected state and find that it correctly contains `"associate_public_ip_address": "false"`. Perhaps a more pertinent detail is that I converted from 0.6.16 to 0.7.8 on this run, after which it began acting as described. I'm not sure if the conversion is related but since this same code does not behave this way in an existing 0.7.7 env, I'm going to start off looking at changes in the latest version. ### Terraform Version 0.7.8 & 0.8.1 ### Affected Resource(s) - aws_instance ### Terraform Configuration Files `associate_public_ip_address = false` The tf code has not changed, and in fact the same symlinked file performs as expected in 0.7.7 ### Debug Output Please provider a link to a GitHub Gist containing the complete debug output: https://gist.githubusercontent.com/Gary-Armstrong/abee399a98d8f297ad149b72d77eac3b/raw/f131e32dff6f13f4bd139a41a929fa32b88001f9/TF%25200.7.8%2520Adventure%2520in%2520Plan+Apply%2520Unexpectations ### Expected Behavior I expected after the initial recreation of the instance that it would not happen again. ### Actual Behavior TF will plan and apply a `-/+` action on the instance, citing the same change reason. ### Steps to Reproduce 1. `terraform apply` You might want to tell me what additional information you need. ### Important Factoids Moved from 0.6.16 to 0.7.8, while other envs in this repo use 0.7.7 instead.
1.0
TF plan & apply not seeing correct associate_public_ip_address - _This issue was originally opened by @Gary-Armstrong as hashicorp/terraform#9811. It was migrated here as part of the [provider split](https://www.hashicorp.com/blog/upcoming-provider-changes-in-terraform-0-10/). The original body of the issue is below._ <hr> The symptom is that I get `-/+ aws_instance.stage` on each exec of plan and apply. Changing `associate_public_ip_address` from true to false did recreate the instance as expected, but subsequent plans and applies ALSO recreated the instance due to `associate_public_ip_address: "true" => "false" (forces new resource)`. Naturally I inspected state and find that it correctly contains `"associate_public_ip_address": "false"`. Perhaps a more pertinent detail is that I converted from 0.6.16 to 0.7.8 on this run, after which it began acting as described. I'm not sure if the conversion is related but since this same code does not behave this way in an existing 0.7.7 env, I'm going to start off looking at changes in the latest version. ### Terraform Version 0.7.8 & 0.8.1 ### Affected Resource(s) - aws_instance ### Terraform Configuration Files `associate_public_ip_address = false` The tf code has not changed, and in fact the same symlinked file performs as expected in 0.7.7 ### Debug Output Please provider a link to a GitHub Gist containing the complete debug output: https://gist.githubusercontent.com/Gary-Armstrong/abee399a98d8f297ad149b72d77eac3b/raw/f131e32dff6f13f4bd139a41a929fa32b88001f9/TF%25200.7.8%2520Adventure%2520in%2520Plan+Apply%2520Unexpectations ### Expected Behavior I expected after the initial recreation of the instance that it would not happen again. ### Actual Behavior TF will plan and apply a `-/+` action on the instance, citing the same change reason. ### Steps to Reproduce 1. `terraform apply` You might want to tell me what additional information you need. ### Important Factoids Moved from 0.6.16 to 0.7.8, while other envs in this repo use 0.7.7 instead.
non_priority
tf plan apply not seeing correct associate public ip address this issue was originally opened by gary armstrong as hashicorp terraform it was migrated here as part of the the original body of the issue is below the symptom is that i get aws instance stage on each exec of plan and apply changing associate public ip address from true to false did recreate the instance as expected but subsequent plans and applies also recreated the instance due to associate public ip address true false forces new resource naturally i inspected state and find that it correctly contains associate public ip address false perhaps a more pertinent detail is that i converted from to on this run after which it began acting as described i m not sure if the conversion is related but since this same code does not behave this way in an existing env i m going to start off looking at changes in the latest version terraform version affected resource s aws instance terraform configuration files associate public ip address false the tf code has not changed and in fact the same symlinked file performs as expected in debug output please provider a link to a github gist containing the complete debug output expected behavior i expected after the initial recreation of the instance that it would not happen again actual behavior tf will plan and apply a action on the instance citing the same change reason steps to reproduce terraform apply you might want to tell me what additional information you need important factoids moved from to while other envs in this repo use instead
0
43,685
9,479,313,425
IssuesEvent
2019-04-20 07:09:53
postmanlabs/postman-app-support
https://api.github.com/repos/postmanlabs/postman-app-support
closed
Generated python.http code for PUT and DELETE requests not working
bug code-generator product/desktop-app
Version/App Information: 1. Postman Version: v4.10.7 2. App (Chrome app or Mac app): Mac app 3. OS details: 4. Is the Interceptor on and enabled in the app: yes 5. Did you encounter this recently, or has this bug always been there: recently 6. Expected behaviour: 7. Console logs (http://blog.getpostman.com/2014/01/27/enabling-chrome-developer-tools-inside-postman/ for the Chrome App, View->Toggle Dev Tools for the Mac app): 8. Screenshots (if applicable): <img width="1182" alt="screen shot 1" src="https://cloud.githubusercontent.com/assets/19791614/25566927/d6bc867e-2e00-11e7-8cea-8ea6c8d05fb2.png"> <img width="709" alt="screen shot 2" src="https://cloud.githubusercontent.com/assets/19791614/25566928/d6bfb556-2e00-11e7-91c6-334b6d26395d.png"> <img width="1167" alt="screen shot 3" src="https://cloud.githubusercontent.com/assets/19791614/25566929/d6c653b6-2e00-11e7-8a42-f5ac112b5de5.png"> Hello, I’ve been trying to run some of the generated scripts on Postman in Python so that I can manipulate them and make a few updates on Jive, by using the Jive REST API. The generated Python scripts for GET and POST requests seem to be working fine. However, the PUT and DELETE requests keep returning a similar response on IDLE, as seen in the screenshots below. For example, if I want to delete a document on Jive, I just send the following request using the Postman app, and this works. The document is successfully deleted from Jive. (Screenshot 1) The Python code generated to run this request is seen below: (Screenshot 2) However, an attempt to run this code using a Python IDE, results in the following: (Screenshot 3) The document is not deleted on Jive. This sort of output is produced upon executing the generated code for any sort of PUT or DELETE request. This is an urgent requirement for my team at the moment, and we’d be extremely gratuitous if the issue could be resolved at the earliest. Hoping for a swift response, Thanks and regards, Rahul Singh
1.0
Generated python.http code for PUT and DELETE requests not working - Version/App Information: 1. Postman Version: v4.10.7 2. App (Chrome app or Mac app): Mac app 3. OS details: 4. Is the Interceptor on and enabled in the app: yes 5. Did you encounter this recently, or has this bug always been there: recently 6. Expected behaviour: 7. Console logs (http://blog.getpostman.com/2014/01/27/enabling-chrome-developer-tools-inside-postman/ for the Chrome App, View->Toggle Dev Tools for the Mac app): 8. Screenshots (if applicable): <img width="1182" alt="screen shot 1" src="https://cloud.githubusercontent.com/assets/19791614/25566927/d6bc867e-2e00-11e7-8cea-8ea6c8d05fb2.png"> <img width="709" alt="screen shot 2" src="https://cloud.githubusercontent.com/assets/19791614/25566928/d6bfb556-2e00-11e7-91c6-334b6d26395d.png"> <img width="1167" alt="screen shot 3" src="https://cloud.githubusercontent.com/assets/19791614/25566929/d6c653b6-2e00-11e7-8a42-f5ac112b5de5.png"> Hello, I’ve been trying to run some of the generated scripts on Postman in Python so that I can manipulate them and make a few updates on Jive, by using the Jive REST API. The generated Python scripts for GET and POST requests seem to be working fine. However, the PUT and DELETE requests keep returning a similar response on IDLE, as seen in the screenshots below. For example, if I want to delete a document on Jive, I just send the following request using the Postman app, and this works. The document is successfully deleted from Jive. (Screenshot 1) The Python code generated to run this request is seen below: (Screenshot 2) However, an attempt to run this code using a Python IDE, results in the following: (Screenshot 3) The document is not deleted on Jive. This sort of output is produced upon executing the generated code for any sort of PUT or DELETE request. This is an urgent requirement for my team at the moment, and we’d be extremely gratuitous if the issue could be resolved at the earliest. Hoping for a swift response, Thanks and regards, Rahul Singh
non_priority
generated python http code for put and delete requests not working version app information postman version app chrome app or mac app mac app os details is the interceptor on and enabled in the app yes did you encounter this recently or has this bug always been there recently expected behaviour console logs for the chrome app view toggle dev tools for the mac app screenshots if applicable img width alt screen shot src img width alt screen shot src img width alt screen shot src hello i’ve been trying to run some of the generated scripts on postman in python so that i can manipulate them and make a few updates on jive by using the jive rest api the generated python scripts for get and post requests seem to be working fine however the put and delete requests keep returning a similar response on idle as seen in the screenshots below for example if i want to delete a document on jive i just send the following request using the postman app and this works the document is successfully deleted from jive screenshot the python code generated to run this request is seen below screenshot however an attempt to run this code using a python ide results in the following screenshot the document is not deleted on jive this sort of output is produced upon executing the generated code for any sort of put or delete request this is an urgent requirement for my team at the moment and we’d be extremely gratuitous if the issue could be resolved at the earliest hoping for a swift response thanks and regards rahul singh
0
386,997
26,710,061,089
IssuesEvent
2023-01-27 22:26:30
intel/policy-library-intel-aws
https://api.github.com/repos/intel/policy-library-intel-aws
closed
License and other markdown templates missing
documentation
# Issue License is missing from the repository along with the code of conduct, contributing, security, and notice markdown files. ## Ask Should these be added to all of the Sentinel policy libraries?
1.0
License and other markdown templates missing - # Issue License is missing from the repository along with the code of conduct, contributing, security, and notice markdown files. ## Ask Should these be added to all of the Sentinel policy libraries?
non_priority
license and other markdown templates missing issue license is missing from the repository along with the code of conduct contributing security and notice markdown files ask should these be added to all of the sentinel policy libraries
0
384,913
11,405,676,298
IssuesEvent
2020-01-31 12:41:32
CodeGra-de/CodeGra.de
https://api.github.com/repos/CodeGra-de/CodeGra.de
closed
Renaming groups requires reload on submissions list
bug frontend priority-1-normal
# Bug report Renaming groups requires a reload for the submission list to see the change. ## Expected Behavior The change to take effect without a reload. ## Actual Behavior No change...
1.0
Renaming groups requires reload on submissions list - # Bug report Renaming groups requires a reload for the submission list to see the change. ## Expected Behavior The change to take effect without a reload. ## Actual Behavior No change...
priority
renaming groups requires reload on submissions list bug report renaming groups requires a reload for the submission list to see the change expected behavior the change to take effect without a reload actual behavior no change
1
24,073
12,027,122,685
IssuesEvent
2020-04-12 17:01:02
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
My project's pricing tier is S0 but prediction api limit to 2 requests per second
Pri2 cognitive-services/svc custom-vision/subsvc cxp product-question triaged
Hi, when i access my customvision project setting in https://www.customvision.ai/projects/&lt;projectid&gt;/settings it show my project tier is S0. But when i call prediction API, it said request limit to 2 requests per second. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 45a084d7-acb6-9893-b6d1-6fef01e0bd07 * Version Independent ID: c9c8b4b2-d080-f8a8-85e2-24f9d66df8c3 * Content: [Limits and quotas - Custom Vision Service - Azure Cognitive Services](https://docs.microsoft.com/en-us/azure/cognitive-services/custom-vision-service/limits-and-quotas#feedback) * Content Source: [articles/cognitive-services/Custom-Vision-Service/limits-and-quotas.md](https://github.com/Microsoft/azure-docs/blob/master/articles/cognitive-services/Custom-Vision-Service/limits-and-quotas.md) * Service: **cognitive-services** * Sub-service: **custom-vision** * GitHub Login: @anrothMSFT * Microsoft Alias: **anroth**
1.0
My project's pricing tier is S0 but prediction api limit to 2 requests per second - Hi, when i access my customvision project setting in https://www.customvision.ai/projects/&lt;projectid&gt;/settings it show my project tier is S0. But when i call prediction API, it said request limit to 2 requests per second. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 45a084d7-acb6-9893-b6d1-6fef01e0bd07 * Version Independent ID: c9c8b4b2-d080-f8a8-85e2-24f9d66df8c3 * Content: [Limits and quotas - Custom Vision Service - Azure Cognitive Services](https://docs.microsoft.com/en-us/azure/cognitive-services/custom-vision-service/limits-and-quotas#feedback) * Content Source: [articles/cognitive-services/Custom-Vision-Service/limits-and-quotas.md](https://github.com/Microsoft/azure-docs/blob/master/articles/cognitive-services/Custom-Vision-Service/limits-and-quotas.md) * Service: **cognitive-services** * Sub-service: **custom-vision** * GitHub Login: @anrothMSFT * Microsoft Alias: **anroth**
non_priority
my project s pricing tier is but prediction api limit to requests per second hi when i access my customvision project setting in it show my project tier is but when i call prediction api it said request limit to requests per second document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service cognitive services sub service custom vision github login anrothmsft microsoft alias anroth
0
120,853
17,644,321,544
IssuesEvent
2021-08-20 02:12:20
Baneeishaque/Raindrop-Removal-With-Light-Field-Image-Using-Image-Inpainting
https://api.github.com/repos/Baneeishaque/Raindrop-Removal-With-Light-Field-Image-Using-Image-Inpainting
opened
CVE-2021-37635 (High) detected in tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl
security vulnerability
## CVE-2021-37635 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/92/2b/e3af15221da9ff323521565fa3324b0d7c7c5b1d7a8ca66984c8d59cb0ce/tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/92/2b/e3af15221da9ff323521565fa3324b0d7c7c5b1d7a8ca66984c8d59cb0ce/tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: Raindrop-Removal-With-Light-Field-Image-Using-Image-Inpainting/requirements.txt</p> <p>Path to vulnerable library: Raindrop-Removal-With-Light-Field-Image-Using-Image-Inpainting/requirements.txt</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an end-to-end open source platform for machine learning. In affected versions the implementation of sparse reduction operations in TensorFlow can trigger accesses outside of bounds of heap allocated data. The [implementation](https://github.com/tensorflow/tensorflow/blob/a1bc56203f21a5a4995311825ffaba7a670d7747/tensorflow/core/kernels/sparse_reduce_op.cc#L217-L228) fails to validate that each reduction group does not overflow and that each corresponding index does not point to outside the bounds of the input tensor. We have patched the issue in GitHub commit 87158f43f05f2720a374f3e6d22a7aaa3a33f750. The fix will be included in TensorFlow 2.6.0. We will also cherrypick this commit on TensorFlow 2.5.1, TensorFlow 2.4.3, and TensorFlow 2.3.4, as these are also affected and still in supported range. <p>Publish Date: 2021-08-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37635>CVE-2021-37635</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: Low - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-cgfm-62j4-v4rf">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-cgfm-62j4-v4rf</a></p> <p>Release Date: 2021-08-12</p> <p>Fix Resolution: tensorflow - 2.3.4, 2.4.3, 2.5.1, 2.6.0, tensorflow-cpu - 2.3.4, 2.4.3, 2.5.1, 2.6.0, tensorflow-gpu - 2.3.4, 2.4.3, 2.5.1, 2.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-37635 (High) detected in tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl - ## CVE-2021-37635 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/92/2b/e3af15221da9ff323521565fa3324b0d7c7c5b1d7a8ca66984c8d59cb0ce/tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/92/2b/e3af15221da9ff323521565fa3324b0d7c7c5b1d7a8ca66984c8d59cb0ce/tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: Raindrop-Removal-With-Light-Field-Image-Using-Image-Inpainting/requirements.txt</p> <p>Path to vulnerable library: Raindrop-Removal-With-Light-Field-Image-Using-Image-Inpainting/requirements.txt</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an end-to-end open source platform for machine learning. In affected versions the implementation of sparse reduction operations in TensorFlow can trigger accesses outside of bounds of heap allocated data. The [implementation](https://github.com/tensorflow/tensorflow/blob/a1bc56203f21a5a4995311825ffaba7a670d7747/tensorflow/core/kernels/sparse_reduce_op.cc#L217-L228) fails to validate that each reduction group does not overflow and that each corresponding index does not point to outside the bounds of the input tensor. We have patched the issue in GitHub commit 87158f43f05f2720a374f3e6d22a7aaa3a33f750. The fix will be included in TensorFlow 2.6.0. We will also cherrypick this commit on TensorFlow 2.5.1, TensorFlow 2.4.3, and TensorFlow 2.3.4, as these are also affected and still in supported range. <p>Publish Date: 2021-08-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37635>CVE-2021-37635</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: Low - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-cgfm-62j4-v4rf">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-cgfm-62j4-v4rf</a></p> <p>Release Date: 2021-08-12</p> <p>Fix Resolution: tensorflow - 2.3.4, 2.4.3, 2.5.1, 2.6.0, tensorflow-cpu - 2.3.4, 2.4.3, 2.5.1, 2.6.0, tensorflow-gpu - 2.3.4, 2.4.3, 2.5.1, 2.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in tensorflow whl cve high severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file raindrop removal with light field image using image inpainting requirements txt path to vulnerable library raindrop removal with light field image using image inpainting requirements txt dependency hierarchy x tensorflow whl vulnerable library found in base branch main vulnerability details tensorflow is an end to end open source platform for machine learning in affected versions the implementation of sparse reduction operations in tensorflow can trigger accesses outside of bounds of heap allocated data the fails to validate that each reduction group does not overflow and that each corresponding index does not point to outside the bounds of the input tensor we have patched the issue in github commit the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact low availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with whitesource
0
49,824
13,187,277,550
IssuesEvent
2020-08-13 02:54:22
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
Import error when trying to import MuonGun (Trac #2142)
Incomplete Migration Migrated from Trac cvmfs defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2142">https://code.icecube.wisc.edu/ticket/2142</a>, reported by thomas.kittler and owned by david.schultz</em></summary> <p> ```json { "status": "closed", "changetime": "2018-05-15T18:03:40", "description": "{{{\n\nIn [1]: from icecube import MuonGun\n---------------------------------------------------------------------------\nImportError Traceback (most recent call last)\n<ipython-input-1-4e878b76dfa2> in <module>()\n----> 1 from icecube import MuonGun\n\n/cvmfs/icecube.opensciencegrid.org/py2-v3/RHEL_7_x86_64/metaprojects/combo/stable/lib/icecube/MuonGun/__init__.py in <module>()\n 1 from icecube.load_pybindings import load_pybindings\n 2 import icecube.icetray, icecube.dataclasses, icecube.simclasses, icecube.phys_services# be nice and pull in our dependencies\n----> 3 load_pybindings(__name__,__path__)\n 4 \n 5 import inspect\n\n/cvmfs/icecube.opensciencegrid.org/py2-v3/RHEL_7_x86_64/metaprojects/combo/stable/lib/icecube/load_pybindings.py in load_pybindings(name, path)\n 56 import imp, sys\n 57 thismod = sys.modules[name]\n---> 58 m = imp.load_dynamic(name, path[0] + \".so\")\n 59 sys.modules[name] = thismod # Some python versions overwrite the Python\n 60 # module entry with the C++ one. We don't want\n\nImportError: /cvmfs/icecube.opensciencegrid.org/py2-v3/RHEL_7_x86_64/metaprojects/combo/stable/lib/icecube/MuonGun.so: undefined symbol: _ZN5boost6python5numpy10initializeEb\n\nIn [2]:\n\n}}}", "reporter": "thomas.kittler", "cc": "", "resolution": "fixed", "_ts": "1526407420536099", "component": "cvmfs", "summary": "Import error when trying to import MuonGun", "priority": "blocker", "keywords": "", "time": "2018-03-12T17:04:39", "milestone": "", "owner": "david.schultz", "type": "defect" } ``` </p> </details>
1.0
Import error when trying to import MuonGun (Trac #2142) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2142">https://code.icecube.wisc.edu/ticket/2142</a>, reported by thomas.kittler and owned by david.schultz</em></summary> <p> ```json { "status": "closed", "changetime": "2018-05-15T18:03:40", "description": "{{{\n\nIn [1]: from icecube import MuonGun\n---------------------------------------------------------------------------\nImportError Traceback (most recent call last)\n<ipython-input-1-4e878b76dfa2> in <module>()\n----> 1 from icecube import MuonGun\n\n/cvmfs/icecube.opensciencegrid.org/py2-v3/RHEL_7_x86_64/metaprojects/combo/stable/lib/icecube/MuonGun/__init__.py in <module>()\n 1 from icecube.load_pybindings import load_pybindings\n 2 import icecube.icetray, icecube.dataclasses, icecube.simclasses, icecube.phys_services# be nice and pull in our dependencies\n----> 3 load_pybindings(__name__,__path__)\n 4 \n 5 import inspect\n\n/cvmfs/icecube.opensciencegrid.org/py2-v3/RHEL_7_x86_64/metaprojects/combo/stable/lib/icecube/load_pybindings.py in load_pybindings(name, path)\n 56 import imp, sys\n 57 thismod = sys.modules[name]\n---> 58 m = imp.load_dynamic(name, path[0] + \".so\")\n 59 sys.modules[name] = thismod # Some python versions overwrite the Python\n 60 # module entry with the C++ one. We don't want\n\nImportError: /cvmfs/icecube.opensciencegrid.org/py2-v3/RHEL_7_x86_64/metaprojects/combo/stable/lib/icecube/MuonGun.so: undefined symbol: _ZN5boost6python5numpy10initializeEb\n\nIn [2]:\n\n}}}", "reporter": "thomas.kittler", "cc": "", "resolution": "fixed", "_ts": "1526407420536099", "component": "cvmfs", "summary": "Import error when trying to import MuonGun", "priority": "blocker", "keywords": "", "time": "2018-03-12T17:04:39", "milestone": "", "owner": "david.schultz", "type": "defect" } ``` </p> </details>
non_priority
import error when trying to import muongun trac migrated from json status closed changetime description n nin from icecube import muongun n nimporterror traceback most recent call last n in n from icecube import muongun n n cvmfs icecube opensciencegrid org rhel metaprojects combo stable lib icecube muongun init py in n from icecube load pybindings import load pybindings n import icecube icetray icecube dataclasses icecube simclasses icecube phys services be nice and pull in our dependencies n load pybindings name path n n import inspect n n cvmfs icecube opensciencegrid org rhel metaprojects combo stable lib icecube load pybindings py in load pybindings name path n import imp sys n thismod sys modules n m imp load dynamic name path so n sys modules thismod some python versions overwrite the python n module entry with the c one we don t want n nimporterror cvmfs icecube opensciencegrid org rhel metaprojects combo stable lib icecube muongun so undefined symbol n nin n n reporter thomas kittler cc resolution fixed ts component cvmfs summary import error when trying to import muongun priority blocker keywords time milestone owner david schultz type defect
0
710,332
24,414,734,616
IssuesEvent
2022-10-05 15:00:30
disorderedmaterials/dissolve
https://api.github.com/repos/disorderedmaterials/dissolve
opened
Epic / Fragments
Priority: Medium
### Focus Implement the ability to detect bound fragments within `Species` for the purposes of subsequent use as analysis sites or as targets for manipulation in `Configuration` generation. The relevant data boils down to a vector of relevant `Molecule` atoms and an associated `Site` or `OrientedSite`. ### Tasks - [ ] Enable returning of matched atoms from `NETADefinition::match()`. - [ ] Exclude previously-matched atoms from subsequent `NETADefinition::match()` calls (use selection in `Species`?, or operate on separate vector data?) - [ ] Develop `FragmentDefinition` and `Fragment` classes. `FragmentDefinition`s would be a companion to analysis sites, while `Fragment`s are the individual generated fragments from one `Species`/`Molecule`. - [ ] Incorporate into `SelectNode` as a new target type. ...
1.0
Epic / Fragments - ### Focus Implement the ability to detect bound fragments within `Species` for the purposes of subsequent use as analysis sites or as targets for manipulation in `Configuration` generation. The relevant data boils down to a vector of relevant `Molecule` atoms and an associated `Site` or `OrientedSite`. ### Tasks - [ ] Enable returning of matched atoms from `NETADefinition::match()`. - [ ] Exclude previously-matched atoms from subsequent `NETADefinition::match()` calls (use selection in `Species`?, or operate on separate vector data?) - [ ] Develop `FragmentDefinition` and `Fragment` classes. `FragmentDefinition`s would be a companion to analysis sites, while `Fragment`s are the individual generated fragments from one `Species`/`Molecule`. - [ ] Incorporate into `SelectNode` as a new target type. ...
priority
epic fragments focus implement the ability to detect bound fragments within species for the purposes of subsequent use as analysis sites or as targets for manipulation in configuration generation the relevant data boils down to a vector of relevant molecule atoms and an associated site or orientedsite tasks enable returning of matched atoms from netadefinition match exclude previously matched atoms from subsequent netadefinition match calls use selection in species or operate on separate vector data develop fragmentdefinition and fragment classes fragmentdefinition s would be a companion to analysis sites while fragment s are the individual generated fragments from one species molecule incorporate into selectnode as a new target type
1
59,585
14,422,021,757
IssuesEvent
2020-12-05 01:05:45
kenferrara/esri-leaflet
https://api.github.com/repos/kenferrara/esri-leaflet
opened
CVE-2019-10775 (High) detected in ecstatic-2.2.1.tgz
security vulnerability
## CVE-2019-10775 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ecstatic-2.2.1.tgz</b></p></summary> <p>A simple static file server middleware that works with both Express and Flatiron</p> <p>Library home page: <a href="https://registry.npmjs.org/ecstatic/-/ecstatic-2.2.1.tgz">https://registry.npmjs.org/ecstatic/-/ecstatic-2.2.1.tgz</a></p> <p>Path to dependency file: esri-leaflet/package.json</p> <p>Path to vulnerable library: esri-leaflet/node_modules/ecstatic/package.json</p> <p> Dependency Hierarchy: - http-server-0.10.0.tgz (Root Library) - :x: **ecstatic-2.2.1.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ecstatic have a denial of service vulnerability. Successful exploitation could lead to crash of an application. <p>Publish Date: 2020-01-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10775>CVE-2019-10775</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jfhbrook/node-ecstatic/tree/4.1.3">https://github.com/jfhbrook/node-ecstatic/tree/4.1.3</a></p> <p>Release Date: 2020-01-02</p> <p>Fix Resolution: 4.1.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"ecstatic","packageVersion":"2.2.1","isTransitiveDependency":true,"dependencyTree":"http-server:0.10.0;ecstatic:2.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.1.3"}],"vulnerabilityIdentifier":"CVE-2019-10775","vulnerabilityDetails":"ecstatic have a denial of service vulnerability. Successful exploitation could lead to crash of an application.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10775","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-10775 (High) detected in ecstatic-2.2.1.tgz - ## CVE-2019-10775 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ecstatic-2.2.1.tgz</b></p></summary> <p>A simple static file server middleware that works with both Express and Flatiron</p> <p>Library home page: <a href="https://registry.npmjs.org/ecstatic/-/ecstatic-2.2.1.tgz">https://registry.npmjs.org/ecstatic/-/ecstatic-2.2.1.tgz</a></p> <p>Path to dependency file: esri-leaflet/package.json</p> <p>Path to vulnerable library: esri-leaflet/node_modules/ecstatic/package.json</p> <p> Dependency Hierarchy: - http-server-0.10.0.tgz (Root Library) - :x: **ecstatic-2.2.1.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ecstatic have a denial of service vulnerability. Successful exploitation could lead to crash of an application. <p>Publish Date: 2020-01-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10775>CVE-2019-10775</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jfhbrook/node-ecstatic/tree/4.1.3">https://github.com/jfhbrook/node-ecstatic/tree/4.1.3</a></p> <p>Release Date: 2020-01-02</p> <p>Fix Resolution: 4.1.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"ecstatic","packageVersion":"2.2.1","isTransitiveDependency":true,"dependencyTree":"http-server:0.10.0;ecstatic:2.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.1.3"}],"vulnerabilityIdentifier":"CVE-2019-10775","vulnerabilityDetails":"ecstatic have a denial of service vulnerability. Successful exploitation could lead to crash of an application.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10775","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_priority
cve high detected in ecstatic tgz cve high severity vulnerability vulnerable library ecstatic tgz a simple static file server middleware that works with both express and flatiron library home page a href path to dependency file esri leaflet package json path to vulnerable library esri leaflet node modules ecstatic package json dependency hierarchy http server tgz root library x ecstatic tgz vulnerable library vulnerability details ecstatic have a denial of service vulnerability successful exploitation could lead to crash of an application publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails ecstatic have a denial of service vulnerability successful exploitation could lead to crash of an application vulnerabilityurl
0
260,984
27,785,039,028
IssuesEvent
2023-03-17 01:58:47
Ortal-GI/mithril.js
https://api.github.com/repos/Ortal-GI/mithril.js
opened
CVE-2023-28155 (Medium) detected in request-2.88.0.tgz
Mend: dependency security vulnerability
## CVE-2023-28155 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>request-2.88.0.tgz</b></p></summary> <p>Simplified HTTP request client.</p> <p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.88.0.tgz">https://registry.npmjs.org/request/-/request-2.88.0.tgz</a></p> <p> Dependency Hierarchy: - :x: **request-2.88.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Ortal-GI/mithril.js/commit/f177f06ccde92925d322fbb3829b5d1612fe8ef4">f177f06ccde92925d322fbb3829b5d1612fe8ef4</a></p> <p>Found in base branch: <b>next</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ** UNSUPPORTED WHEN ASSIGNED ** The Request package through 2.88.1 for Node.js allows a bypass of SSRF mitigations via an attacker-controller server that does a cross-protocol redirect (HTTP to HTTPS, or HTTPS to HTTP). NOTE: This vulnerability only affects products that are no longer supported by the maintainer. <p>Publish Date: 2023-03-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28155>CVE-2023-28155</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-28155 (Medium) detected in request-2.88.0.tgz - ## CVE-2023-28155 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>request-2.88.0.tgz</b></p></summary> <p>Simplified HTTP request client.</p> <p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.88.0.tgz">https://registry.npmjs.org/request/-/request-2.88.0.tgz</a></p> <p> Dependency Hierarchy: - :x: **request-2.88.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Ortal-GI/mithril.js/commit/f177f06ccde92925d322fbb3829b5d1612fe8ef4">f177f06ccde92925d322fbb3829b5d1612fe8ef4</a></p> <p>Found in base branch: <b>next</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ** UNSUPPORTED WHEN ASSIGNED ** The Request package through 2.88.1 for Node.js allows a bypass of SSRF mitigations via an attacker-controller server that does a cross-protocol redirect (HTTP to HTTPS, or HTTPS to HTTP). NOTE: This vulnerability only affects products that are no longer supported by the maintainer. <p>Publish Date: 2023-03-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28155>CVE-2023-28155</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in request tgz cve medium severity vulnerability vulnerable library request tgz simplified http request client library home page a href dependency hierarchy x request tgz vulnerable library found in head commit a href found in base branch next vulnerability details unsupported when assigned the request package through for node js allows a bypass of ssrf mitigations via an attacker controller server that does a cross protocol redirect http to https or https to http note this vulnerability only affects products that are no longer supported by the maintainer publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend
0
139,608
20,917,895,815
IssuesEvent
2022-03-24 14:53:14
microsoft/pyright
https://api.github.com/repos/microsoft/pyright
closed
False error when values of dictionary (which are functions) are assigned to variable
as designed
**Describe the bug** I am getting a false error when trying to assign a dictionary value (which is function) to a variable inside a class. The error message is ``` Argument of type "str | FunctionType" cannot be assigned to parameter "__k" of type "str" in function "__getitem__"  Type "str | FunctionType" cannot be assigned to type "str"    "FunctionType" is incompatible with "str" Pyright (reportGeneralTypeIssues) [55, 27] ``` The exact location are the lines ``` self.solver = METHOD[method] self.p = ORDER[method] ``` in the code snippet below. **To Reproduce** conda environment with - python = 3.10 - numpy pyright -> latest version **Expected behavior** There should be no errors. **Screenshots or Code** ```python import inspect import numpy as np from numpy.typing import NDArray from typing import Callable, Optional, Any, Union ODE = Union[ Callable[[float, NDArray[Any], Optional[Any]], Union[NDArray[Any], NDArray[np.float64]]], Callable[[float, NDArray[Any]], Union[NDArray[Any], NDArray[np.float64]]] ] def RKF45(): return 0 def RKF78(): return 1 METHOD = { "RKF45": RKF45, "RKF78": RKF78 } ORDER = { "CKRK45": 4, "DOP54": 5, "DVERK65": 6, "Euler": 1, "RALS4": 4, "RK4": 4, "RK42": 4, "RKF45": 4, "RKF78": 7 } class check: def __init__(self, derivative: ODE, h: float, tspan: NDArray[Any], IC: NDArray[np.float64], Atol: float, Rtol: float, method: str = "RKF78"): self.ode = derivative self.h = h self.tspan = tspan self.IC = IC self.Atol = Atol self.Rtol = Rtol # Check solver method and stepsize controller if method not in METHOD and not inspect.isfunction(method): raise ValueError("`method` must be one of {}".format(METHOD)) else: self.solver = METHOD[method] self.p = ORDER[method] # Test ------------------------------------------------------------------------ def test(t: float, y: NDArray[np.float64]) -> NDArray[Any]: return np.array([ y[1], -4*t*y[1] - (2 + 4*t**2)*y[0] ]) if __name__=="__main__": # Setup values tf = 20 # final time t = 0 # initial time ti = t x = np.array([1, 0.5]) # initial conditions of states h = 0.01 # initial stepsize p = 4 # order of the numerical integration atol = 1e-7 # absolute tolerance rtol = 1e-6 # relative tolerance ck = check(test, h, np.array([ti, tf]), x, atol, rtol) print(ck.solver) print(ck.p) ``` **VS Code extension or command-line** Running on command-line. Neovim with Native LSP on Pyright = 1.1.232
1.0
False error when values of dictionary (which are functions) are assigned to variable - **Describe the bug** I am getting a false error when trying to assign a dictionary value (which is function) to a variable inside a class. The error message is ``` Argument of type "str | FunctionType" cannot be assigned to parameter "__k" of type "str" in function "__getitem__"  Type "str | FunctionType" cannot be assigned to type "str"    "FunctionType" is incompatible with "str" Pyright (reportGeneralTypeIssues) [55, 27] ``` The exact location are the lines ``` self.solver = METHOD[method] self.p = ORDER[method] ``` in the code snippet below. **To Reproduce** conda environment with - python = 3.10 - numpy pyright -> latest version **Expected behavior** There should be no errors. **Screenshots or Code** ```python import inspect import numpy as np from numpy.typing import NDArray from typing import Callable, Optional, Any, Union ODE = Union[ Callable[[float, NDArray[Any], Optional[Any]], Union[NDArray[Any], NDArray[np.float64]]], Callable[[float, NDArray[Any]], Union[NDArray[Any], NDArray[np.float64]]] ] def RKF45(): return 0 def RKF78(): return 1 METHOD = { "RKF45": RKF45, "RKF78": RKF78 } ORDER = { "CKRK45": 4, "DOP54": 5, "DVERK65": 6, "Euler": 1, "RALS4": 4, "RK4": 4, "RK42": 4, "RKF45": 4, "RKF78": 7 } class check: def __init__(self, derivative: ODE, h: float, tspan: NDArray[Any], IC: NDArray[np.float64], Atol: float, Rtol: float, method: str = "RKF78"): self.ode = derivative self.h = h self.tspan = tspan self.IC = IC self.Atol = Atol self.Rtol = Rtol # Check solver method and stepsize controller if method not in METHOD and not inspect.isfunction(method): raise ValueError("`method` must be one of {}".format(METHOD)) else: self.solver = METHOD[method] self.p = ORDER[method] # Test ------------------------------------------------------------------------ def test(t: float, y: NDArray[np.float64]) -> NDArray[Any]: return np.array([ y[1], -4*t*y[1] - (2 + 4*t**2)*y[0] ]) if __name__=="__main__": # Setup values tf = 20 # final time t = 0 # initial time ti = t x = np.array([1, 0.5]) # initial conditions of states h = 0.01 # initial stepsize p = 4 # order of the numerical integration atol = 1e-7 # absolute tolerance rtol = 1e-6 # relative tolerance ck = check(test, h, np.array([ti, tf]), x, atol, rtol) print(ck.solver) print(ck.p) ``` **VS Code extension or command-line** Running on command-line. Neovim with Native LSP on Pyright = 1.1.232
non_priority
false error when values of dictionary which are functions are assigned to variable describe the bug i am getting a false error when trying to assign a dictionary value which is function to a variable inside a class the error message is argument of type str functiontype cannot be assigned to parameter k of type str in function getitem   type str functiontype cannot be assigned to type str      functiontype is incompatible with str pyright reportgeneraltypeissues the exact location are the lines self solver method self p order in the code snippet below to reproduce conda environment with python numpy pyright latest version expected behavior there should be no errors screenshots or code python import inspect import numpy as np from numpy typing import ndarray from typing import callable optional any union ode union callable optional union ndarray callable union ndarray def return def return method order euler class check def init self derivative ode h float tspan ndarray ic ndarray atol float rtol float method str self ode derivative self h h self tspan tspan self ic ic self atol atol self rtol rtol check solver method and stepsize controller if method not in method and not inspect isfunction method raise valueerror method must be one of format method else self solver method self p order test def test t float y ndarray ndarray return np array y t y t y if name main setup values tf final time t initial time ti t x np array initial conditions of states h initial stepsize p order of the numerical integration atol absolute tolerance rtol relative tolerance ck check test h np array x atol rtol print ck solver print ck p vs code extension or command line running on command line neovim with native lsp on pyright
0
675,694
23,101,802,254
IssuesEvent
2022-07-27 04:07:44
yugabyte/yugabyte-db
https://api.github.com/repos/yugabyte/yugabyte-db
closed
[docdb] TSAN race on QLTransactionTest_Simple
kind/bug area/docdb priority/medium
Jira Link: [DB-1914](https://yugabyte.atlassian.net/browse/DB-1914) > ybd tsan --cxx-test client_ql-transaction-test --gtest_filter QLTransactionTest.Simple -n 500 --tp 4 ```Write of size 8 at 0x7b70000ae4b8 by thread T179 (mutexes: write M51984): #0 std::__1::unique_ptr<yb::tablet::TransactionParticipant, std::__1::default_delete<yb::tablet::TransactionParticipant> >::reset(yb::tablet::TransactionParticipant*) /n/jenkins/thirdparty/yugabyte-thirdparty-2019-10-18T11_18_41/third party/installed/tsan/libcxx/include/c++/v1/memory:2632:20 (libtablet.so+0x13bdac) #1 std::__1::unique_ptr<yb::tablet::TransactionParticipant, std::__1::default_delete<yb::tablet::TransactionParticipant> >::~unique_ptr() /n/jenkins/thirdparty/yugabyte-thirdparty-2019-10-18T11_18_41/thirdparty/installed/tsan/libcxx/include/c++/v1/memory:2588 (libtablet.so+0x13bdac) #2 yb::tablet::Tablet::~Tablet() /net/centos-gcp-cloud-jenkins-slave-40c/share/jenkins/workspace/github-yugabyte-db-centos-master-clang-tsan/build/tsan-clang-dynamic-ninja/../../src/yb/tablet/tablet.cc:414 (libtablet.so+0x13bdac) Previous read of size 8 at 0x7b70000ae4b8 by main thread: #0 std::__1::unique_ptr<yb::tablet::TransactionParticipant, std::__1::default_delete<yb::tablet::TransactionParticipant> >::get() const /n/jenkins/thirdparty/yugabyte-thirdparty-2019-10-18T11_18_41/thirdparty/installed/tsan/libcxx/include/c++/v1/memory:2607:19 (libql-dml-test-base.so+0x417bd) #1 yb::tablet::Tablet::transaction_participant() const /net/centos-gcp-cloud-jenkins-slave-40c/share/jenkins/workspace/github-yugabyte-db-centos-master-clang-tsan/build/tsan-clang-dynamic-ninja/../../ent/src/yb/tablet/../../../../src/yb/tablet/tablet.h:453 (libql-dml-test-base.so+0x417bd) #2 yb::client::TransactionTestBase::CheckNoRunningTransactions() /net/centos-gcp-cloud-jenkins-slave-40c/share/jenkins/workspace/github-yugabyte-db-centos-master-clang-tsan/build/tsan-clang-dynamic-ninja/../../src/yb/client/txn-test-base.cc:304:42 (libql-dml-test-base.so+0x3f098)```
1.0
[docdb] TSAN race on QLTransactionTest_Simple - Jira Link: [DB-1914](https://yugabyte.atlassian.net/browse/DB-1914) > ybd tsan --cxx-test client_ql-transaction-test --gtest_filter QLTransactionTest.Simple -n 500 --tp 4 ```Write of size 8 at 0x7b70000ae4b8 by thread T179 (mutexes: write M51984): #0 std::__1::unique_ptr<yb::tablet::TransactionParticipant, std::__1::default_delete<yb::tablet::TransactionParticipant> >::reset(yb::tablet::TransactionParticipant*) /n/jenkins/thirdparty/yugabyte-thirdparty-2019-10-18T11_18_41/third party/installed/tsan/libcxx/include/c++/v1/memory:2632:20 (libtablet.so+0x13bdac) #1 std::__1::unique_ptr<yb::tablet::TransactionParticipant, std::__1::default_delete<yb::tablet::TransactionParticipant> >::~unique_ptr() /n/jenkins/thirdparty/yugabyte-thirdparty-2019-10-18T11_18_41/thirdparty/installed/tsan/libcxx/include/c++/v1/memory:2588 (libtablet.so+0x13bdac) #2 yb::tablet::Tablet::~Tablet() /net/centos-gcp-cloud-jenkins-slave-40c/share/jenkins/workspace/github-yugabyte-db-centos-master-clang-tsan/build/tsan-clang-dynamic-ninja/../../src/yb/tablet/tablet.cc:414 (libtablet.so+0x13bdac) Previous read of size 8 at 0x7b70000ae4b8 by main thread: #0 std::__1::unique_ptr<yb::tablet::TransactionParticipant, std::__1::default_delete<yb::tablet::TransactionParticipant> >::get() const /n/jenkins/thirdparty/yugabyte-thirdparty-2019-10-18T11_18_41/thirdparty/installed/tsan/libcxx/include/c++/v1/memory:2607:19 (libql-dml-test-base.so+0x417bd) #1 yb::tablet::Tablet::transaction_participant() const /net/centos-gcp-cloud-jenkins-slave-40c/share/jenkins/workspace/github-yugabyte-db-centos-master-clang-tsan/build/tsan-clang-dynamic-ninja/../../ent/src/yb/tablet/../../../../src/yb/tablet/tablet.h:453 (libql-dml-test-base.so+0x417bd) #2 yb::client::TransactionTestBase::CheckNoRunningTransactions() /net/centos-gcp-cloud-jenkins-slave-40c/share/jenkins/workspace/github-yugabyte-db-centos-master-clang-tsan/build/tsan-clang-dynamic-ninja/../../src/yb/client/txn-test-base.cc:304:42 (libql-dml-test-base.so+0x3f098)```
priority
tsan race on qltransactiontest simple jira link ybd tsan cxx test client ql transaction test gtest filter qltransactiontest simple n tp write of size at by thread mutexes write std unique ptr reset yb tablet transactionparticipant n jenkins thirdparty yugabyte thirdparty third party installed tsan libcxx include c memory libtablet so std unique ptr unique ptr n jenkins thirdparty yugabyte thirdparty thirdparty installed tsan libcxx include c memory libtablet so yb tablet tablet tablet net centos gcp cloud jenkins slave share jenkins workspace github yugabyte db centos master clang tsan build tsan clang dynamic ninja src yb tablet tablet cc libtablet so previous read of size at by main thread std unique ptr get const n jenkins thirdparty yugabyte thirdparty thirdparty installed tsan libcxx include c memory libql dml test base so yb tablet tablet transaction participant const net centos gcp cloud jenkins slave share jenkins workspace github yugabyte db centos master clang tsan build tsan clang dynamic ninja ent src yb tablet src yb tablet tablet h libql dml test base so yb client transactiontestbase checknorunningtransactions net centos gcp cloud jenkins slave share jenkins workspace github yugabyte db centos master clang tsan build tsan clang dynamic ninja src yb client txn test base cc libql dml test base so
1
132,888
28,378,874,041
IssuesEvent
2023-04-13 00:09:23
dtcxzyw/llvm-ci
https://api.github.com/repos/dtcxzyw/llvm-ci
closed
Regressions Report [rv64gc-Os-thinlto] April 6th 2023, 12:53:19 pm
regression codegen reasonable
## Metadata + Workflow URL: https://github.com/dtcxzyw/llvm-ci/actions/runs/4628651695 ## Change Logs from b02e2ed7ac71137430fc9fc8b32bed344f5e7fef to 846712b0cb133f0ebfaf5be41818183b4c86f0a1 [846712b0cb133f0ebfaf5be41818183b4c86f0a1](https://github.com/llvm/llvm-project/commit/846712b0cb133f0ebfaf5be41818183b4c86f0a1) [DAG] combineSelect - select(i1,vXi1,vXi1) - only cast &lt;X x i1&gt; constants to iX pre-legalization or if its a legal type [5348a25377652c5e32f9a85754da7231cb9bf9f6](https://github.com/llvm/llvm-project/commit/5348a25377652c5e32f9a85754da7231cb9bf9f6) [llvm-exegesis][AArch64] Add return statement at the end of the snippet correctly [ef1b22a20fd4d8c0cfea64089d4e732dd95dc265](https://github.com/llvm/llvm-project/commit/ef1b22a20fd4d8c0cfea64089d4e732dd95dc265) [clang][ClangScanDeps][test] Remove the file before overwriting it [f3d9abf1f87c308e9d7bd368b3ad54cb8f28ac4f](https://github.com/llvm/llvm-project/commit/f3d9abf1f87c308e9d7bd368b3ad54cb8f28ac4f) [X86][mem-fold] Use the generated memory folding table [63df77d90520a5df4d870c710c0c6eb961be8887](https://github.com/llvm/llvm-project/commit/63df77d90520a5df4d870c710c0c6eb961be8887) [Test] Add expensive-range-sharpening mode for test [eda14ebf6a43d9ada6a2be3d1b06b8b6036eb774](https://github.com/llvm/llvm-project/commit/eda14ebf6a43d9ada6a2be3d1b06b8b6036eb774) [X86][mem-fold] Refine the code in X86FoldTablesEmitter.cpp, NFCI [ffa0a2ecd4c0042fb0fd8fc5d84bb92ddd3608a3](https://github.com/llvm/llvm-project/commit/ffa0a2ecd4c0042fb0fd8fc5d84bb92ddd3608a3) [AArch64][SME] Fix LDR and STR asm parser [55bbcbf511b4df0dfe524b2d06fc561a939c6b16](https://github.com/llvm/llvm-project/commit/55bbcbf511b4df0dfe524b2d06fc561a939c6b16) [clang] Reset track of immediate function context when entering new function [93fb1bac674b07134f7de59ba22d3d5acf741637](https://github.com/llvm/llvm-project/commit/93fb1bac674b07134f7de59ba22d3d5acf741637) [X86] LowerVectorAllEqual - don&#x27;t attempt to match comparisons of float data [3d7242f05e6ebae394fe47039abb0568083209a0](https://github.com/llvm/llvm-project/commit/3d7242f05e6ebae394fe47039abb0568083209a0) Reapply &quot;[LSR] Preserve LCSSA when rewriting instruction with PHI user&quot; [65031c189bc818ee449294b2f02abc55a5f88691](https://github.com/llvm/llvm-project/commit/65031c189bc818ee449294b2f02abc55a5f88691) [SVE][InstCombine] Don&#x27;t convert calls to fp binop intrinsics to instructions when strictfp is required. [5f2145adf9e343e950935cac1c9030f8ca597912](https://github.com/llvm/llvm-project/commit/5f2145adf9e343e950935cac1c9030f8ca597912) [RISCV] Use non-strided load if VL=1 for optimized zero stride loads ## Regressions (Size) |Name|Baseline MD5|Current MD5|Baseline Size|Current Size|Ratio| |:--|:--:|:--:|--:|--:|--:| |MultiSource/Benchmarks/DOE-ProxyApps-C/CoMD/CoMD|269b1c47ea33f5534f9a0f3be174af4d|edaed94a3c5258ec350cc8d947e4fc50|19310|19354|1.002| |MultiSource/Benchmarks/ASCI_Purple/SMG2000/smg2000|f1a5fb56c822aa429dbeb6733f97e614|e3928712057816499c500a1e6523c809|97928|97960|1.000| ## Regressions (Time) |Name|Baseline MD5|Current MD5|Baseline Time|Current Time|Ratio| |:--|:--:|:--:|--:|--:|--:| |MultiSource/Applications/SIBsim4/SIBsim4|15e40ef2c1251c3bbacbc3264159768d|89708662d312cf2a47dd7a9d32778552|12.346281844|12.348638757|1.000| |MultiSource/Benchmarks/lzbench/lzbench|b9131a100696ff7fec4ad1064335fcc8|b66e586d31c4ce28400b7bc0be4931d3|501.72679052|501.770194473|1.000| |MultiSource/Benchmarks/MiBench/consumer-lame/consumer-lame|dd01c3bdebe9202c68d9f92c1adb9800|151fd1daae67c64b1e2710bad99205d5|0.981094664|0.981105753|1.000| |MultiSource/Applications/ClamAV/clamscan|cb5c5d8ba44eb3976f6694631eedf8b8|2c91fea70b70f14ced66384f84bd1274|0.487347923|0.487350167|1.000| |MultiSource/Benchmarks/mafft/pairlocalalign|02b88f8644043ec7de8bb814944514f6|50da3e48ccef6e598759bd5aa503a0de|107.356278764|107.356288014|1.000| ## Differences (Size) |Name|Baseline MD5|Current MD5|Baseline Size|Current Size|Ratio| |:--|:--:|:--:|--:|--:|--:| |MultiSource/Benchmarks/DOE-ProxyApps-C/CoMD/CoMD|269b1c47ea33f5534f9a0f3be174af4d|edaed94a3c5258ec350cc8d947e4fc50|19310|19354|1.002| |MultiSource/Benchmarks/ASCI_Purple/SMG2000/smg2000|f1a5fb56c822aa429dbeb6733f97e614|e3928712057816499c500a1e6523c809|97928|97960|1.000| |MultiSource/Benchmarks/DOE-ProxyApps-C++/miniFE/miniFE|20ea780f6aa79dbc5ae8e4f3027f1126|f5923a4d31356e70074c10e15bf5a195|45492|45500|1.000| |MultiSource/Benchmarks/mafft/pairlocalalign|02b88f8644043ec7de8bb814944514f6|50da3e48ccef6e598759bd5aa503a0de|83662|83670|1.000| |MultiSource/Benchmarks/MallocBench/espresso/espresso|d089af231e75239e7f90030729dd27c6|60f70f548e14bc04db6cda43082ed69b|107986|107994|1.000| |MultiSource/Applications/ClamAV/clamscan|cb5c5d8ba44eb3976f6694631eedf8b8|2c91fea70b70f14ced66384f84bd1274|326910|326918|1.000| |MultiSource/Applications/JM/lencod/lencod|2b7dc022fe55a70db58c5be34f20884f|59d3fed7a750d00b63835f921cb94d5b|375620|375624|1.000| |MultiSource/Benchmarks/FreeBench/neural/neural|6a492826d452277d08af97dd6dab59a2|a0d3cf3a30c942a1ee9743dc73b12012|2924|2924|1.000| |MultiSource/Benchmarks/MiBench/consumer-jpeg/consumer-jpeg|ba0447c4af043c96daaeb728e9a3b729|59f7eb335d753adf355bcab2bc94662e|49806|49806|1.000| |MultiSource/Benchmarks/MiBench/consumer-typeset/consumer-typeset|7a005af363107288dd10236bb560e754|6500b86b61801736c81e02c69041589b|343420|343420|1.000| |MultiSource/Benchmarks/mediabench/jpeg/jpeg-6a/cjpeg|06d1478ba757a150d7d4f923ac6107eb|770d2f105a4a31bc5da7059f125870d0|44698|44698|1.000| |MultiSource/Benchmarks/lzbench/lzbench|b9131a100696ff7fec4ad1064335fcc8|b66e586d31c4ce28400b7bc0be4931d3|1572792|1572760|1.000| |MultiSource/Benchmarks/MiBench/consumer-lame/consumer-lame|dd01c3bdebe9202c68d9f92c1adb9800|151fd1daae67c64b1e2710bad99205d5|75362|75342|1.000| |MultiSource/Applications/SIBsim4/SIBsim4|15e40ef2c1251c3bbacbc3264159768d|89708662d312cf2a47dd7a9d32778552|26692|26678|0.999| |GeoMeans|N/A|N/A|111683.795|111708.066|1.000| ## Differences (Time) |Name|Baseline MD5|Current MD5|Baseline Time|Current Time|Ratio| |:--|:--:|:--:|--:|--:|--:| |MultiSource/Applications/SIBsim4/SIBsim4|15e40ef2c1251c3bbacbc3264159768d|89708662d312cf2a47dd7a9d32778552|12.346281844|12.348638757|1.000| |MultiSource/Benchmarks/lzbench/lzbench|b9131a100696ff7fec4ad1064335fcc8|b66e586d31c4ce28400b7bc0be4931d3|501.72679052|501.770194473|1.000| |MultiSource/Benchmarks/MiBench/consumer-lame/consumer-lame|dd01c3bdebe9202c68d9f92c1adb9800|151fd1daae67c64b1e2710bad99205d5|0.981094664|0.981105753|1.000| |MultiSource/Applications/ClamAV/clamscan|cb5c5d8ba44eb3976f6694631eedf8b8|2c91fea70b70f14ced66384f84bd1274|0.487347923|0.487350167|1.000| |MultiSource/Benchmarks/mafft/pairlocalalign|02b88f8644043ec7de8bb814944514f6|50da3e48ccef6e598759bd5aa503a0de|107.356278764|107.356288014|1.000| |MultiSource/Benchmarks/MallocBench/espresso/espresso|d089af231e75239e7f90030729dd27c6|60f70f548e14bc04db6cda43082ed69b|2.165321788|2.165321788|1.000| |MultiSource/Benchmarks/MiBench/consumer-jpeg/consumer-jpeg|ba0447c4af043c96daaeb728e9a3b729|59f7eb335d753adf355bcab2bc94662e|0.024813697|0.024813697|1.000| |MultiSource/Applications/JM/lencod/lencod|2b7dc022fe55a70db58c5be34f20884f|59d3fed7a750d00b63835f921cb94d5b|42.437328252|42.437328029|1.000| |MultiSource/Benchmarks/DOE-ProxyApps-C/CoMD/CoMD|269b1c47ea33f5534f9a0f3be174af4d|edaed94a3c5258ec350cc8d947e4fc50|13.053597444|13.053596478|1.000| |MultiSource/Benchmarks/MiBench/consumer-typeset/consumer-typeset|7a005af363107288dd10236bb560e754|6500b86b61801736c81e02c69041589b|0.639010516|0.639010338|1.000| |MultiSource/Benchmarks/FreeBench/neural/neural|6a492826d452277d08af97dd6dab59a2|a0d3cf3a30c942a1ee9743dc73b12012|0.680771356|0.680766036|1.000| |MultiSource/Benchmarks/mediabench/jpeg/jpeg-6a/cjpeg|06d1478ba757a150d7d4f923ac6107eb|770d2f105a4a31bc5da7059f125870d0|0.01518234|0.01518144|1.000| |MultiSource/Benchmarks/DOE-ProxyApps-C++/miniFE/miniFE|20ea780f6aa79dbc5ae8e4f3027f1126|f5923a4d31356e70074c10e15bf5a195|27.688832037|27.674151902|0.999| |MultiSource/Benchmarks/ASCI_Purple/SMG2000/smg2000|f1a5fb56c822aa429dbeb6733f97e614|e3928712057816499c500a1e6523c809|11.575908837|11.374923672|0.983| |GeoMeans|N/A|N/A|5.180|5.172|0.999|
1.0
Regressions Report [rv64gc-Os-thinlto] April 6th 2023, 12:53:19 pm - ## Metadata + Workflow URL: https://github.com/dtcxzyw/llvm-ci/actions/runs/4628651695 ## Change Logs from b02e2ed7ac71137430fc9fc8b32bed344f5e7fef to 846712b0cb133f0ebfaf5be41818183b4c86f0a1 [846712b0cb133f0ebfaf5be41818183b4c86f0a1](https://github.com/llvm/llvm-project/commit/846712b0cb133f0ebfaf5be41818183b4c86f0a1) [DAG] combineSelect - select(i1,vXi1,vXi1) - only cast &lt;X x i1&gt; constants to iX pre-legalization or if its a legal type [5348a25377652c5e32f9a85754da7231cb9bf9f6](https://github.com/llvm/llvm-project/commit/5348a25377652c5e32f9a85754da7231cb9bf9f6) [llvm-exegesis][AArch64] Add return statement at the end of the snippet correctly [ef1b22a20fd4d8c0cfea64089d4e732dd95dc265](https://github.com/llvm/llvm-project/commit/ef1b22a20fd4d8c0cfea64089d4e732dd95dc265) [clang][ClangScanDeps][test] Remove the file before overwriting it [f3d9abf1f87c308e9d7bd368b3ad54cb8f28ac4f](https://github.com/llvm/llvm-project/commit/f3d9abf1f87c308e9d7bd368b3ad54cb8f28ac4f) [X86][mem-fold] Use the generated memory folding table [63df77d90520a5df4d870c710c0c6eb961be8887](https://github.com/llvm/llvm-project/commit/63df77d90520a5df4d870c710c0c6eb961be8887) [Test] Add expensive-range-sharpening mode for test [eda14ebf6a43d9ada6a2be3d1b06b8b6036eb774](https://github.com/llvm/llvm-project/commit/eda14ebf6a43d9ada6a2be3d1b06b8b6036eb774) [X86][mem-fold] Refine the code in X86FoldTablesEmitter.cpp, NFCI [ffa0a2ecd4c0042fb0fd8fc5d84bb92ddd3608a3](https://github.com/llvm/llvm-project/commit/ffa0a2ecd4c0042fb0fd8fc5d84bb92ddd3608a3) [AArch64][SME] Fix LDR and STR asm parser [55bbcbf511b4df0dfe524b2d06fc561a939c6b16](https://github.com/llvm/llvm-project/commit/55bbcbf511b4df0dfe524b2d06fc561a939c6b16) [clang] Reset track of immediate function context when entering new function [93fb1bac674b07134f7de59ba22d3d5acf741637](https://github.com/llvm/llvm-project/commit/93fb1bac674b07134f7de59ba22d3d5acf741637) [X86] LowerVectorAllEqual - don&#x27;t attempt to match comparisons of float data [3d7242f05e6ebae394fe47039abb0568083209a0](https://github.com/llvm/llvm-project/commit/3d7242f05e6ebae394fe47039abb0568083209a0) Reapply &quot;[LSR] Preserve LCSSA when rewriting instruction with PHI user&quot; [65031c189bc818ee449294b2f02abc55a5f88691](https://github.com/llvm/llvm-project/commit/65031c189bc818ee449294b2f02abc55a5f88691) [SVE][InstCombine] Don&#x27;t convert calls to fp binop intrinsics to instructions when strictfp is required. [5f2145adf9e343e950935cac1c9030f8ca597912](https://github.com/llvm/llvm-project/commit/5f2145adf9e343e950935cac1c9030f8ca597912) [RISCV] Use non-strided load if VL=1 for optimized zero stride loads ## Regressions (Size) |Name|Baseline MD5|Current MD5|Baseline Size|Current Size|Ratio| |:--|:--:|:--:|--:|--:|--:| |MultiSource/Benchmarks/DOE-ProxyApps-C/CoMD/CoMD|269b1c47ea33f5534f9a0f3be174af4d|edaed94a3c5258ec350cc8d947e4fc50|19310|19354|1.002| |MultiSource/Benchmarks/ASCI_Purple/SMG2000/smg2000|f1a5fb56c822aa429dbeb6733f97e614|e3928712057816499c500a1e6523c809|97928|97960|1.000| ## Regressions (Time) |Name|Baseline MD5|Current MD5|Baseline Time|Current Time|Ratio| |:--|:--:|:--:|--:|--:|--:| |MultiSource/Applications/SIBsim4/SIBsim4|15e40ef2c1251c3bbacbc3264159768d|89708662d312cf2a47dd7a9d32778552|12.346281844|12.348638757|1.000| |MultiSource/Benchmarks/lzbench/lzbench|b9131a100696ff7fec4ad1064335fcc8|b66e586d31c4ce28400b7bc0be4931d3|501.72679052|501.770194473|1.000| |MultiSource/Benchmarks/MiBench/consumer-lame/consumer-lame|dd01c3bdebe9202c68d9f92c1adb9800|151fd1daae67c64b1e2710bad99205d5|0.981094664|0.981105753|1.000| |MultiSource/Applications/ClamAV/clamscan|cb5c5d8ba44eb3976f6694631eedf8b8|2c91fea70b70f14ced66384f84bd1274|0.487347923|0.487350167|1.000| |MultiSource/Benchmarks/mafft/pairlocalalign|02b88f8644043ec7de8bb814944514f6|50da3e48ccef6e598759bd5aa503a0de|107.356278764|107.356288014|1.000| ## Differences (Size) |Name|Baseline MD5|Current MD5|Baseline Size|Current Size|Ratio| |:--|:--:|:--:|--:|--:|--:| |MultiSource/Benchmarks/DOE-ProxyApps-C/CoMD/CoMD|269b1c47ea33f5534f9a0f3be174af4d|edaed94a3c5258ec350cc8d947e4fc50|19310|19354|1.002| |MultiSource/Benchmarks/ASCI_Purple/SMG2000/smg2000|f1a5fb56c822aa429dbeb6733f97e614|e3928712057816499c500a1e6523c809|97928|97960|1.000| |MultiSource/Benchmarks/DOE-ProxyApps-C++/miniFE/miniFE|20ea780f6aa79dbc5ae8e4f3027f1126|f5923a4d31356e70074c10e15bf5a195|45492|45500|1.000| |MultiSource/Benchmarks/mafft/pairlocalalign|02b88f8644043ec7de8bb814944514f6|50da3e48ccef6e598759bd5aa503a0de|83662|83670|1.000| |MultiSource/Benchmarks/MallocBench/espresso/espresso|d089af231e75239e7f90030729dd27c6|60f70f548e14bc04db6cda43082ed69b|107986|107994|1.000| |MultiSource/Applications/ClamAV/clamscan|cb5c5d8ba44eb3976f6694631eedf8b8|2c91fea70b70f14ced66384f84bd1274|326910|326918|1.000| |MultiSource/Applications/JM/lencod/lencod|2b7dc022fe55a70db58c5be34f20884f|59d3fed7a750d00b63835f921cb94d5b|375620|375624|1.000| |MultiSource/Benchmarks/FreeBench/neural/neural|6a492826d452277d08af97dd6dab59a2|a0d3cf3a30c942a1ee9743dc73b12012|2924|2924|1.000| |MultiSource/Benchmarks/MiBench/consumer-jpeg/consumer-jpeg|ba0447c4af043c96daaeb728e9a3b729|59f7eb335d753adf355bcab2bc94662e|49806|49806|1.000| |MultiSource/Benchmarks/MiBench/consumer-typeset/consumer-typeset|7a005af363107288dd10236bb560e754|6500b86b61801736c81e02c69041589b|343420|343420|1.000| |MultiSource/Benchmarks/mediabench/jpeg/jpeg-6a/cjpeg|06d1478ba757a150d7d4f923ac6107eb|770d2f105a4a31bc5da7059f125870d0|44698|44698|1.000| |MultiSource/Benchmarks/lzbench/lzbench|b9131a100696ff7fec4ad1064335fcc8|b66e586d31c4ce28400b7bc0be4931d3|1572792|1572760|1.000| |MultiSource/Benchmarks/MiBench/consumer-lame/consumer-lame|dd01c3bdebe9202c68d9f92c1adb9800|151fd1daae67c64b1e2710bad99205d5|75362|75342|1.000| |MultiSource/Applications/SIBsim4/SIBsim4|15e40ef2c1251c3bbacbc3264159768d|89708662d312cf2a47dd7a9d32778552|26692|26678|0.999| |GeoMeans|N/A|N/A|111683.795|111708.066|1.000| ## Differences (Time) |Name|Baseline MD5|Current MD5|Baseline Time|Current Time|Ratio| |:--|:--:|:--:|--:|--:|--:| |MultiSource/Applications/SIBsim4/SIBsim4|15e40ef2c1251c3bbacbc3264159768d|89708662d312cf2a47dd7a9d32778552|12.346281844|12.348638757|1.000| |MultiSource/Benchmarks/lzbench/lzbench|b9131a100696ff7fec4ad1064335fcc8|b66e586d31c4ce28400b7bc0be4931d3|501.72679052|501.770194473|1.000| |MultiSource/Benchmarks/MiBench/consumer-lame/consumer-lame|dd01c3bdebe9202c68d9f92c1adb9800|151fd1daae67c64b1e2710bad99205d5|0.981094664|0.981105753|1.000| |MultiSource/Applications/ClamAV/clamscan|cb5c5d8ba44eb3976f6694631eedf8b8|2c91fea70b70f14ced66384f84bd1274|0.487347923|0.487350167|1.000| |MultiSource/Benchmarks/mafft/pairlocalalign|02b88f8644043ec7de8bb814944514f6|50da3e48ccef6e598759bd5aa503a0de|107.356278764|107.356288014|1.000| |MultiSource/Benchmarks/MallocBench/espresso/espresso|d089af231e75239e7f90030729dd27c6|60f70f548e14bc04db6cda43082ed69b|2.165321788|2.165321788|1.000| |MultiSource/Benchmarks/MiBench/consumer-jpeg/consumer-jpeg|ba0447c4af043c96daaeb728e9a3b729|59f7eb335d753adf355bcab2bc94662e|0.024813697|0.024813697|1.000| |MultiSource/Applications/JM/lencod/lencod|2b7dc022fe55a70db58c5be34f20884f|59d3fed7a750d00b63835f921cb94d5b|42.437328252|42.437328029|1.000| |MultiSource/Benchmarks/DOE-ProxyApps-C/CoMD/CoMD|269b1c47ea33f5534f9a0f3be174af4d|edaed94a3c5258ec350cc8d947e4fc50|13.053597444|13.053596478|1.000| |MultiSource/Benchmarks/MiBench/consumer-typeset/consumer-typeset|7a005af363107288dd10236bb560e754|6500b86b61801736c81e02c69041589b|0.639010516|0.639010338|1.000| |MultiSource/Benchmarks/FreeBench/neural/neural|6a492826d452277d08af97dd6dab59a2|a0d3cf3a30c942a1ee9743dc73b12012|0.680771356|0.680766036|1.000| |MultiSource/Benchmarks/mediabench/jpeg/jpeg-6a/cjpeg|06d1478ba757a150d7d4f923ac6107eb|770d2f105a4a31bc5da7059f125870d0|0.01518234|0.01518144|1.000| |MultiSource/Benchmarks/DOE-ProxyApps-C++/miniFE/miniFE|20ea780f6aa79dbc5ae8e4f3027f1126|f5923a4d31356e70074c10e15bf5a195|27.688832037|27.674151902|0.999| |MultiSource/Benchmarks/ASCI_Purple/SMG2000/smg2000|f1a5fb56c822aa429dbeb6733f97e614|e3928712057816499c500a1e6523c809|11.575908837|11.374923672|0.983| |GeoMeans|N/A|N/A|5.180|5.172|0.999|
non_priority
regressions report april pm metadata workflow url change logs from to combineselect select only cast lt x x gt constants to ix pre legalization or if its a legal type add return statement at the end of the snippet correctly remove the file before overwriting it use the generated memory folding table add expensive range sharpening mode for test refine the code in cpp nfci fix ldr and str asm parser reset track of immediate function context when entering new function lowervectorallequal don t attempt to match comparisons of float data reapply quot preserve lcssa when rewriting instruction with phi user quot don t convert calls to fp binop intrinsics to instructions when strictfp is required use non strided load if vl for optimized zero stride loads regressions size name baseline current baseline size current size ratio multisource benchmarks doe proxyapps c comd comd multisource benchmarks asci purple regressions time name baseline current baseline time current time ratio multisource applications multisource benchmarks lzbench lzbench multisource benchmarks mibench consumer lame consumer lame multisource applications clamav clamscan multisource benchmarks mafft pairlocalalign differences size name baseline current baseline size current size ratio multisource benchmarks doe proxyapps c comd comd multisource benchmarks asci purple multisource benchmarks doe proxyapps c minife minife multisource benchmarks mafft pairlocalalign multisource benchmarks mallocbench espresso espresso multisource applications clamav clamscan multisource applications jm lencod lencod multisource benchmarks freebench neural neural multisource benchmarks mibench consumer jpeg consumer jpeg multisource benchmarks mibench consumer typeset consumer typeset multisource benchmarks mediabench jpeg jpeg cjpeg multisource benchmarks lzbench lzbench multisource benchmarks mibench consumer lame consumer lame multisource applications geomeans n a n a differences time name baseline current baseline time current time ratio multisource applications multisource benchmarks lzbench lzbench multisource benchmarks mibench consumer lame consumer lame multisource applications clamav clamscan multisource benchmarks mafft pairlocalalign multisource benchmarks mallocbench espresso espresso multisource benchmarks mibench consumer jpeg consumer jpeg multisource applications jm lencod lencod multisource benchmarks doe proxyapps c comd comd multisource benchmarks mibench consumer typeset consumer typeset multisource benchmarks freebench neural neural multisource benchmarks mediabench jpeg jpeg cjpeg multisource benchmarks doe proxyapps c minife minife multisource benchmarks asci purple geomeans n a n a
0
23,119
21,048,236,133
IssuesEvent
2022-03-31 18:08:48
CH-Earth/summa
https://api.github.com/repos/CH-Earth/summa
opened
Add a warning if SUMMA performs a relative humidity correction?
enhancement usability
#### Feature Requests https://github.com/CH-Earth/summa/blob/fa9adf808229a45085defdc2bb8ef05836b9b3aa/build/source/engine/derivforce.f90#L276-L282 Might be helpful to users if this correction also prints a warning, so that it becomes easier to track what happens to the provided forcing data
True
Add a warning if SUMMA performs a relative humidity correction? - #### Feature Requests https://github.com/CH-Earth/summa/blob/fa9adf808229a45085defdc2bb8ef05836b9b3aa/build/source/engine/derivforce.f90#L276-L282 Might be helpful to users if this correction also prints a warning, so that it becomes easier to track what happens to the provided forcing data
non_priority
add a warning if summa performs a relative humidity correction feature requests might be helpful to users if this correction also prints a warning so that it becomes easier to track what happens to the provided forcing data
0
752,115
26,273,947,260
IssuesEvent
2023-01-06 19:53:30
getodk/central-backend
https://api.github.com/repos/getodk/central-backend
closed
Warn if new form definition has structural change
enhancement high-priority
Deleting a form field can add complications to a data analysis workflow, because the user may still need to access the data from the deleted field. This comes up in particular when a user tries to rename a field or moves a field into or out of a group, because Backend understands that as two operations: the deletion of the old field and the addition of a new field. In v1.4, we are adding the ability to include deleted fields in the submissions export. @yanokwa has suggested for a later release that we also warn the user if they upload a form definition that deletes a field. The text of the warning would point to the docs. The user experience would be similar to how warnings are currently shown for an XLSForm. One tricky case to consider is if an uploaded form definition generates XLSForm warnings and also deletes a field: how do we return both sets of warnings from Backend? Related Forum topic: https://forum.getodk.org/t/data-loss-scare-when-updating-to-a-new-form-version/39777
1.0
Warn if new form definition has structural change - Deleting a form field can add complications to a data analysis workflow, because the user may still need to access the data from the deleted field. This comes up in particular when a user tries to rename a field or moves a field into or out of a group, because Backend understands that as two operations: the deletion of the old field and the addition of a new field. In v1.4, we are adding the ability to include deleted fields in the submissions export. @yanokwa has suggested for a later release that we also warn the user if they upload a form definition that deletes a field. The text of the warning would point to the docs. The user experience would be similar to how warnings are currently shown for an XLSForm. One tricky case to consider is if an uploaded form definition generates XLSForm warnings and also deletes a field: how do we return both sets of warnings from Backend? Related Forum topic: https://forum.getodk.org/t/data-loss-scare-when-updating-to-a-new-form-version/39777
priority
warn if new form definition has structural change deleting a form field can add complications to a data analysis workflow because the user may still need to access the data from the deleted field this comes up in particular when a user tries to rename a field or moves a field into or out of a group because backend understands that as two operations the deletion of the old field and the addition of a new field in we are adding the ability to include deleted fields in the submissions export yanokwa has suggested for a later release that we also warn the user if they upload a form definition that deletes a field the text of the warning would point to the docs the user experience would be similar to how warnings are currently shown for an xlsform one tricky case to consider is if an uploaded form definition generates xlsform warnings and also deletes a field how do we return both sets of warnings from backend related forum topic
1
149,635
23,500,964,944
IssuesEvent
2022-08-18 08:23:18
mui/mui-toolpad
https://api.github.com/repos/mui/mui-toolpad
opened
Visually separate props form component Id and name
discussion design: ux
### Duplicates - [X] I have searched the existing issues ### Latest version - [X] I have tested the latest version ### Summary 💡 Currently, the right panel shows everything together. We should have some visual separation based on the type of items. Moreover, we should also separate UI props from data props. <img width="301" alt="Screenshot 2022-08-18 at 1 50 04 PM" src="https://user-images.githubusercontent.com/92228082/185347082-9ba6917c-5acb-4e71-8055-3b22e0b59382.png"> ### Examples 🌈 _No response_ ### Motivation 🔦 _No response_
1.0
Visually separate props form component Id and name - ### Duplicates - [X] I have searched the existing issues ### Latest version - [X] I have tested the latest version ### Summary 💡 Currently, the right panel shows everything together. We should have some visual separation based on the type of items. Moreover, we should also separate UI props from data props. <img width="301" alt="Screenshot 2022-08-18 at 1 50 04 PM" src="https://user-images.githubusercontent.com/92228082/185347082-9ba6917c-5acb-4e71-8055-3b22e0b59382.png"> ### Examples 🌈 _No response_ ### Motivation 🔦 _No response_
non_priority
visually separate props form component id and name duplicates i have searched the existing issues latest version i have tested the latest version summary 💡 currently the right panel shows everything together we should have some visual separation based on the type of items moreover we should also separate ui props from data props img width alt screenshot at pm src examples 🌈 no response motivation 🔦 no response
0
53,379
13,798,837,443
IssuesEvent
2020-10-10 03:33:52
tt9133github/jetty.project
https://api.github.com/repos/tt9133github/jetty.project
opened
CVE-2015-5237 (High) detected in protobuf-java-3.2.0.jar
security vulnerability
## CVE-2015-5237 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>protobuf-java-3.2.0.jar</b></p></summary> <p>Core Protocol Buffers library. Protocol Buffers are a way of encoding structured data in an efficient yet extensible format.</p> <p>Library home page: <a href="https://developers.google.com/protocol-buffers/">https://developers.google.com/protocol-buffers/</a></p> <p>Path to dependency file: jetty.project/jetty-gcloud/jetty-gcloud-session-manager/pom.xml</p> <p>Path to vulnerable library: 20201010033015_ZWAISP/downloadResource_XUFRLM/20201010033300/protobuf-java-3.2.0.jar</p> <p> Dependency Hierarchy: - :x: **protobuf-java-3.2.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/tt9133github/jetty.project/commit/5b64054edfd8f14ce518b950213113328da64915">5b64054edfd8f14ce518b950213113328da64915</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> protobuf allows remote authenticated attackers to cause a heap-based buffer overflow. <p>Publish Date: 2017-09-25 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-5237>CVE-2015-5237</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/protocolbuffers/protobuf/releases/tag/v3.4.0">https://github.com/protocolbuffers/protobuf/releases/tag/v3.4.0</a></p> <p>Release Date: 2017-09-25</p> <p>Fix Resolution: 3.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2015-5237 (High) detected in protobuf-java-3.2.0.jar - ## CVE-2015-5237 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>protobuf-java-3.2.0.jar</b></p></summary> <p>Core Protocol Buffers library. Protocol Buffers are a way of encoding structured data in an efficient yet extensible format.</p> <p>Library home page: <a href="https://developers.google.com/protocol-buffers/">https://developers.google.com/protocol-buffers/</a></p> <p>Path to dependency file: jetty.project/jetty-gcloud/jetty-gcloud-session-manager/pom.xml</p> <p>Path to vulnerable library: 20201010033015_ZWAISP/downloadResource_XUFRLM/20201010033300/protobuf-java-3.2.0.jar</p> <p> Dependency Hierarchy: - :x: **protobuf-java-3.2.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/tt9133github/jetty.project/commit/5b64054edfd8f14ce518b950213113328da64915">5b64054edfd8f14ce518b950213113328da64915</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> protobuf allows remote authenticated attackers to cause a heap-based buffer overflow. <p>Publish Date: 2017-09-25 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-5237>CVE-2015-5237</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/protocolbuffers/protobuf/releases/tag/v3.4.0">https://github.com/protocolbuffers/protobuf/releases/tag/v3.4.0</a></p> <p>Release Date: 2017-09-25</p> <p>Fix Resolution: 3.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in protobuf java jar cve high severity vulnerability vulnerable library protobuf java jar core protocol buffers library protocol buffers are a way of encoding structured data in an efficient yet extensible format library home page a href path to dependency file jetty project jetty gcloud jetty gcloud session manager pom xml path to vulnerable library zwaisp downloadresource xufrlm protobuf java jar dependency hierarchy x protobuf java jar vulnerable library found in head commit a href vulnerability details protobuf allows remote authenticated attackers to cause a heap based buffer overflow publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
343,553
10,332,474,492
IssuesEvent
2019-09-03 00:17:43
wyvernlang/wyvern
https://api.github.com/repos/wyvernlang/wyvern
closed
Types aren't handled properly in arrow types
bug priority: high
In the text-editor application, in [the textEditor module](https://github.com/wyvernlang/wyvern/blob/master/examples/text-editor/textEditor.wyv), make the type of the last module parameter more precise by substituting the `Dyn` type with the arrow type `Unit -> TextEditor`, i.e., change the textEditor module header, which is currently: ``` module def textEditor(java: Java, logger: Logger, createTextEditorInstance: Dyn): TextEditor ``` to be: ``` module def textEditor(java: Java, logger: Logger, createTextEditorInstance: Unit -> TextEditor): TextEditor ``` Run the text-editor application. _Expected behavior:_ No error is thrown. _Actual behavior:_ The following error is thrown: ``` Type TextEditor is not defined at location file wyvern/examples/text-editor/textEditor.wyv on line 1 column 85 ``` It's unclear whether this is the problem for all arrow types, but, at least, it manifests in module parameters.
1.0
Types aren't handled properly in arrow types - In the text-editor application, in [the textEditor module](https://github.com/wyvernlang/wyvern/blob/master/examples/text-editor/textEditor.wyv), make the type of the last module parameter more precise by substituting the `Dyn` type with the arrow type `Unit -> TextEditor`, i.e., change the textEditor module header, which is currently: ``` module def textEditor(java: Java, logger: Logger, createTextEditorInstance: Dyn): TextEditor ``` to be: ``` module def textEditor(java: Java, logger: Logger, createTextEditorInstance: Unit -> TextEditor): TextEditor ``` Run the text-editor application. _Expected behavior:_ No error is thrown. _Actual behavior:_ The following error is thrown: ``` Type TextEditor is not defined at location file wyvern/examples/text-editor/textEditor.wyv on line 1 column 85 ``` It's unclear whether this is the problem for all arrow types, but, at least, it manifests in module parameters.
priority
types aren t handled properly in arrow types in the text editor application in make the type of the last module parameter more precise by substituting the dyn type with the arrow type unit texteditor i e change the texteditor module header which is currently module def texteditor java java logger logger createtexteditorinstance dyn texteditor to be module def texteditor java java logger logger createtexteditorinstance unit texteditor texteditor run the text editor application expected behavior no error is thrown actual behavior the following error is thrown type texteditor is not defined at location file wyvern examples text editor texteditor wyv on line column it s unclear whether this is the problem for all arrow types but at least it manifests in module parameters
1
21,301
7,001,590,539
IssuesEvent
2017-12-18 10:48:10
minishift/minishift
https://api.github.com/repos/minishift/minishift
closed
Make dep installation as part of Makefile same as go-bindata
component/build kind/task priority/minor
We should add dep installation as part of Makefile, same way we are doing for `go-bindata`.
1.0
Make dep installation as part of Makefile same as go-bindata - We should add dep installation as part of Makefile, same way we are doing for `go-bindata`.
non_priority
make dep installation as part of makefile same as go bindata we should add dep installation as part of makefile same way we are doing for go bindata
0
451,612
13,039,283,092
IssuesEvent
2020-07-28 16:29:24
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
closed
GET https://127.0.0.1:9443/api/am/publisher/v1/subscriptions is not working as expected
Priority/Highest Type/Bug commitment
### Description: APIM 310 Retrieving all subscriptions for the user's APIs return a Bad request. GET https://127.0.0.1:9443/api/am/publisher/v1/subscriptions Response: **{"code":400,"message":"Bad Request","description":"Validation Error","moreInfo":"","error":[{"code":"400_subscriptionsGet.arg0","message":"subscriptionsGet.arg0: may not be null","description":null}]}** ### Steps to reproduce: - Create an API and subscribe it to an app - Use below curl command to retrieve all subscriptions **curl -X GET "https://localhost:9443/api/am/publisher/v1/subscriptions" -H "accept: application/json" -u admin:admin -k** or **curl -X GET "https://localhost:9443/api/am/publisher/v1/subscriptions" -H "accept: application/json" -H "Authorization: Bearer 1e99cbc4-cd8e-3fa3-ae6a-cb273ce48672" -k** (token bound with apim:subscription_view) - Observe above mentioned response
1.0
GET https://127.0.0.1:9443/api/am/publisher/v1/subscriptions is not working as expected - ### Description: APIM 310 Retrieving all subscriptions for the user's APIs return a Bad request. GET https://127.0.0.1:9443/api/am/publisher/v1/subscriptions Response: **{"code":400,"message":"Bad Request","description":"Validation Error","moreInfo":"","error":[{"code":"400_subscriptionsGet.arg0","message":"subscriptionsGet.arg0: may not be null","description":null}]}** ### Steps to reproduce: - Create an API and subscribe it to an app - Use below curl command to retrieve all subscriptions **curl -X GET "https://localhost:9443/api/am/publisher/v1/subscriptions" -H "accept: application/json" -u admin:admin -k** or **curl -X GET "https://localhost:9443/api/am/publisher/v1/subscriptions" -H "accept: application/json" -H "Authorization: Bearer 1e99cbc4-cd8e-3fa3-ae6a-cb273ce48672" -k** (token bound with apim:subscription_view) - Observe above mentioned response
priority
get is not working as expected description apim retrieving all subscriptions for the user s apis return a bad request get response code message bad request description validation error moreinfo error steps to reproduce create an api and subscribe it to an app use below curl command to retrieve all subscriptions curl x get h accept application json u admin admin k or curl x get h accept application json h authorization bearer k token bound with apim subscription view observe above mentioned response
1
748,295
26,116,100,811
IssuesEvent
2022-12-28 06:22:18
longhorn/longhorn
https://api.github.com/repos/longhorn/longhorn
closed
[BUG] Check if the replica size is matched with volume size before running backup
kind/bug area/engine priority/0 area/volume-backup-restore area/stability
## Describe the bug Current backup mechanism is randomly choosing a replica and then executing backup without any replica size check. When a volume fails to expand and leaves some replicas with a mismatched size, Longhorn will potentialy back up the wrong replica to remote storage and is not aware of the size mismatch ([code](https://github.com/longhorn/longhorn-manager/blob/0eec020db8d889ba37318d9a4717268cd2daca74/controller/backup_controller.go#L492)). To prevent from the issue, the size check and setting the wrong replica to ERR for triggering rebuilding will be implemented in this ticket. ## To Reproduce Reproduce steps 1. Create a 2 GiB volume with 3 replicas 2. Go to one of the instance-manager-r 3. Download `[grpc-client-cli](https://github.com/vadimi/grpc-client-cli)` and execute `grpc-client-cli --address ${replicas process's replica server address}` to connect to replica server e.g: `grpc-client-cli --address localhost:10000` 4. `ptypes.ReplicaService` > `ReplicaExpand` > `{"size": "4294967296"}` > enter 5. Create a tons of replica. The issue will be observed. ## Expected behavior A clear and concise description of what you expected to happen. ## Log or Support bundle If applicable, add the Longhorn managers' log or support bundle when the issue happens. You can generate a Support Bundle using the link at the footer of the Longhorn UI. ## Environment - Longhorn version: v1.2.x, v1.3.0 and v1.3.1 - Installation method (e.g. Rancher Catalog App/Helm/Kubectl): - Kubernetes distro (e.g. RKE/K3s/EKS/OpenShift) and version: - Number of management node in the cluster: - Number of worker node in the cluster: - Node config - OS type and version: - CPU per node: - Memory per node: - Disk type(e.g. SSD/NVMe): - Network bandwidth between the nodes: - Underlying Infrastructure (e.g. on AWS/GCE, EKS/GKE, VMWare/KVM, Baremetal): - Number of Longhorn volumes in the cluster: ## Additional context Add any other context about the problem here.
1.0
[BUG] Check if the replica size is matched with volume size before running backup - ## Describe the bug Current backup mechanism is randomly choosing a replica and then executing backup without any replica size check. When a volume fails to expand and leaves some replicas with a mismatched size, Longhorn will potentialy back up the wrong replica to remote storage and is not aware of the size mismatch ([code](https://github.com/longhorn/longhorn-manager/blob/0eec020db8d889ba37318d9a4717268cd2daca74/controller/backup_controller.go#L492)). To prevent from the issue, the size check and setting the wrong replica to ERR for triggering rebuilding will be implemented in this ticket. ## To Reproduce Reproduce steps 1. Create a 2 GiB volume with 3 replicas 2. Go to one of the instance-manager-r 3. Download `[grpc-client-cli](https://github.com/vadimi/grpc-client-cli)` and execute `grpc-client-cli --address ${replicas process's replica server address}` to connect to replica server e.g: `grpc-client-cli --address localhost:10000` 4. `ptypes.ReplicaService` > `ReplicaExpand` > `{"size": "4294967296"}` > enter 5. Create a tons of replica. The issue will be observed. ## Expected behavior A clear and concise description of what you expected to happen. ## Log or Support bundle If applicable, add the Longhorn managers' log or support bundle when the issue happens. You can generate a Support Bundle using the link at the footer of the Longhorn UI. ## Environment - Longhorn version: v1.2.x, v1.3.0 and v1.3.1 - Installation method (e.g. Rancher Catalog App/Helm/Kubectl): - Kubernetes distro (e.g. RKE/K3s/EKS/OpenShift) and version: - Number of management node in the cluster: - Number of worker node in the cluster: - Node config - OS type and version: - CPU per node: - Memory per node: - Disk type(e.g. SSD/NVMe): - Network bandwidth between the nodes: - Underlying Infrastructure (e.g. on AWS/GCE, EKS/GKE, VMWare/KVM, Baremetal): - Number of Longhorn volumes in the cluster: ## Additional context Add any other context about the problem here.
priority
check if the replica size is matched with volume size before running backup describe the bug current backup mechanism is randomly choosing a replica and then executing backup without any replica size check when a volume fails to expand and leaves some replicas with a mismatched size longhorn will potentialy back up the wrong replica to remote storage and is not aware of the size mismatch to prevent from the issue the size check and setting the wrong replica to err for triggering rebuilding will be implemented in this ticket to reproduce reproduce steps create a gib volume with replicas go to one of the instance manager r download and execute grpc client cli address replicas process s replica server address to connect to replica server e g grpc client cli address localhost ptypes replicaservice replicaexpand size enter create a tons of replica the issue will be observed expected behavior a clear and concise description of what you expected to happen log or support bundle if applicable add the longhorn managers log or support bundle when the issue happens you can generate a support bundle using the link at the footer of the longhorn ui environment longhorn version x and installation method e g rancher catalog app helm kubectl kubernetes distro e g rke eks openshift and version number of management node in the cluster number of worker node in the cluster node config os type and version cpu per node memory per node disk type e g ssd nvme network bandwidth between the nodes underlying infrastructure e g on aws gce eks gke vmware kvm baremetal number of longhorn volumes in the cluster additional context add any other context about the problem here
1
8,588
2,999,261,234
IssuesEvent
2015-07-23 18:08:44
luchanz/edilot-angular
https://api.github.com/repos/luchanz/edilot-angular
closed
Pantalla Home Distribuidor
Amend Test
Mati, En el home de Nazareno / Diego, se debe mostrar recuadrado la cantidad de pedidos pendientes, los pedidos que lleva cargados del mes (Estén validados o no) y el importe que debería cobrar.
1.0
Pantalla Home Distribuidor - Mati, En el home de Nazareno / Diego, se debe mostrar recuadrado la cantidad de pedidos pendientes, los pedidos que lleva cargados del mes (Estén validados o no) y el importe que debería cobrar.
non_priority
pantalla home distribuidor mati en el home de nazareno diego se debe mostrar recuadrado la cantidad de pedidos pendientes los pedidos que lleva cargados del mes estén validados o no y el importe que debería cobrar
0
682,811
23,358,538,184
IssuesEvent
2022-08-10 09:35:13
ansible-collections/servicenow.itsm
https://api.github.com/repos/ansible-collections/servicenow.itsm
closed
configuration_item_batch: return created and updated records
type/enhancement priority/medium
<!--- Verify first that your feature was not already discussed on GitHub --> <!--- Complete *all* sections as described, this form is processed automatically --> ##### SUMMARY <!--- Describe the new feature/improvement briefly below --> Return created and updated records on configuration_item_batch. ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME <!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure --> configuration_item_batch ##### ADDITIONAL INFORMATION <!--- Describe how the feature would be used, why it is needed and what it would solve --> Actually `configuration_item_batch` does not returns much useful information and consequently it is difficult to identify created or updated records to manipulate later. It is specially useful to have sys_ids at least. <!--- Paste example playbooks or commands between quotes below --> <!--- HINT: You can also paste gist.github.com links for larger files -->
1.0
configuration_item_batch: return created and updated records - <!--- Verify first that your feature was not already discussed on GitHub --> <!--- Complete *all* sections as described, this form is processed automatically --> ##### SUMMARY <!--- Describe the new feature/improvement briefly below --> Return created and updated records on configuration_item_batch. ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME <!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure --> configuration_item_batch ##### ADDITIONAL INFORMATION <!--- Describe how the feature would be used, why it is needed and what it would solve --> Actually `configuration_item_batch` does not returns much useful information and consequently it is difficult to identify created or updated records to manipulate later. It is specially useful to have sys_ids at least. <!--- Paste example playbooks or commands between quotes below --> <!--- HINT: You can also paste gist.github.com links for larger files -->
priority
configuration item batch return created and updated records summary return created and updated records on configuration item batch issue type feature idea component name configuration item batch additional information actually configuration item batch does not returns much useful information and consequently it is difficult to identify created or updated records to manipulate later it is specially useful to have sys ids at least
1
423,942
12,304,056,850
IssuesEvent
2020-05-11 19:48:43
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
samples/cfb/display flickers with SSD1306
area: Display bug priority: low
**Describe the bug** When running the example for the SSD1306, the display flickers between "012345" shown, to blank. The following code in the example sets the font to the largest available (in my case, height of 32) ``` for (int idx = 0; idx < 42; idx++) { if (cfb_get_font_size(dev, idx, &font_width, &font_height)) { break; } cfb_framebuffer_set_font(dev, idx); printf("font width %d, font height %d\n", font_width, font_height); } ``` In the console output, it prints: ``` x_res 128, y_res 32, ppt 8, rows 4, cols 128 ``` which I assume means: - `x_res` = x resolution, 128 pixels - `y_res` = y resolution, 32 pixels - `ppt` = pixels per tile, 8 pixels (tiles are vertical only, so it is a 1x8 pixel tile) - `rows` = number of tile rows, 4 (= y_res/ppt) - `cols` = number of tile cols, 128 However in the main loop it then runs a for loop for each *tile* row, not character row: ``` while (1) { for (int i = 0; i < rows; i++) { // <<<< rows is the number of vertical tiles, not character spaces cfb_framebuffer_clear(dev, false); if (cfb_print(dev, "0123456789mMgj!\\\"§$%&/()=", 0, i * ppt)) { printf("Failed to print a string\n"); continue; } cfb_framebuffer_finalize(dev); } } ``` Should it actually be something like: ``` while (1) { for (int i = 0; i < (display_height / font_height); i++) { // use the number of character rows cfb_framebuffer_clear(dev, false); if (cfb_print(dev, "0123456789mMgj!\\\"§$%&/()=", 0, i * font_height)) { printf("Failed to print a string\n"); continue; } cfb_framebuffer_finalize(dev); } } ``` Although this causes flickering if there is more than one character row (in my example, `font_height=16`), as the display will 1. draw the row=0 case, wrapping the extra characters onto the next row 2. erase the display, wiping rows 0 and 1 3. write to row 1 only 4. erase the display, wiping rows 0 and 1 So row 0 flickers between having text and being blank, and row 1 flickers between wrapped characters and written characters. I think the actual intention of this example is to do something like: ``` while (1) { cfb_framebuffer_clear(dev, false); for (int i = 0; i < (display_height / font_height); i++) { if (cfb_print(dev, "0123456789mMgj!\\\"§$%&/()=", 0, i * font_height)) { printf("Failed to print a string\n"); continue; } } cfb_framebuffer_finalize(dev); } ``` Which just prints that text on every line (or as much as possible). This seems much more sensible than the existing sample. I can create a PR for this last case, I'm just looking for some guidance as to what the sample is actually supposed to do. **To Reproduce** Steps to reproduce the behavior: 1. Run the sample with an SSD1306 **Expected behavior** The display should not flicker, or it should at least be documented that this is the expected behaviour (although I'm not sure how useful that is as an example). **Screenshots or console output** ``` *** Booting Zephyr OS build v2.1.99-ncs1 *** initialized SSD1306 font width 10, font height 16 font width 15, font height 24 font width 20, font height 32 x_res 128, y_res 32, ppt 8, rows 4, cols 128 [00:00:00.012,390] <dbg> cfb.cfb_framebuffer_init: number of fonts 3 ``` **Environment (please complete the following information):** - OS: Ubuntu 18.04 - Toolchain: gnuarmemb - Commit: ff214d637e
1.0
samples/cfb/display flickers with SSD1306 - **Describe the bug** When running the example for the SSD1306, the display flickers between "012345" shown, to blank. The following code in the example sets the font to the largest available (in my case, height of 32) ``` for (int idx = 0; idx < 42; idx++) { if (cfb_get_font_size(dev, idx, &font_width, &font_height)) { break; } cfb_framebuffer_set_font(dev, idx); printf("font width %d, font height %d\n", font_width, font_height); } ``` In the console output, it prints: ``` x_res 128, y_res 32, ppt 8, rows 4, cols 128 ``` which I assume means: - `x_res` = x resolution, 128 pixels - `y_res` = y resolution, 32 pixels - `ppt` = pixels per tile, 8 pixels (tiles are vertical only, so it is a 1x8 pixel tile) - `rows` = number of tile rows, 4 (= y_res/ppt) - `cols` = number of tile cols, 128 However in the main loop it then runs a for loop for each *tile* row, not character row: ``` while (1) { for (int i = 0; i < rows; i++) { // <<<< rows is the number of vertical tiles, not character spaces cfb_framebuffer_clear(dev, false); if (cfb_print(dev, "0123456789mMgj!\\\"§$%&/()=", 0, i * ppt)) { printf("Failed to print a string\n"); continue; } cfb_framebuffer_finalize(dev); } } ``` Should it actually be something like: ``` while (1) { for (int i = 0; i < (display_height / font_height); i++) { // use the number of character rows cfb_framebuffer_clear(dev, false); if (cfb_print(dev, "0123456789mMgj!\\\"§$%&/()=", 0, i * font_height)) { printf("Failed to print a string\n"); continue; } cfb_framebuffer_finalize(dev); } } ``` Although this causes flickering if there is more than one character row (in my example, `font_height=16`), as the display will 1. draw the row=0 case, wrapping the extra characters onto the next row 2. erase the display, wiping rows 0 and 1 3. write to row 1 only 4. erase the display, wiping rows 0 and 1 So row 0 flickers between having text and being blank, and row 1 flickers between wrapped characters and written characters. I think the actual intention of this example is to do something like: ``` while (1) { cfb_framebuffer_clear(dev, false); for (int i = 0; i < (display_height / font_height); i++) { if (cfb_print(dev, "0123456789mMgj!\\\"§$%&/()=", 0, i * font_height)) { printf("Failed to print a string\n"); continue; } } cfb_framebuffer_finalize(dev); } ``` Which just prints that text on every line (or as much as possible). This seems much more sensible than the existing sample. I can create a PR for this last case, I'm just looking for some guidance as to what the sample is actually supposed to do. **To Reproduce** Steps to reproduce the behavior: 1. Run the sample with an SSD1306 **Expected behavior** The display should not flicker, or it should at least be documented that this is the expected behaviour (although I'm not sure how useful that is as an example). **Screenshots or console output** ``` *** Booting Zephyr OS build v2.1.99-ncs1 *** initialized SSD1306 font width 10, font height 16 font width 15, font height 24 font width 20, font height 32 x_res 128, y_res 32, ppt 8, rows 4, cols 128 [00:00:00.012,390] <dbg> cfb.cfb_framebuffer_init: number of fonts 3 ``` **Environment (please complete the following information):** - OS: Ubuntu 18.04 - Toolchain: gnuarmemb - Commit: ff214d637e
priority
samples cfb display flickers with describe the bug when running the example for the the display flickers between shown to blank the following code in the example sets the font to the largest available in my case height of for int idx idx idx if cfb get font size dev idx font width font height break cfb framebuffer set font dev idx printf font width d font height d n font width font height in the console output it prints x res y res ppt rows cols which i assume means x res x resolution pixels y res y resolution pixels ppt pixels per tile pixels tiles are vertical only so it is a pixel tile rows number of tile rows y res ppt cols number of tile cols however in the main loop it then runs a for loop for each tile row not character row while for int i i rows i rows is the number of vertical tiles not character spaces cfb framebuffer clear dev false if cfb print dev § i ppt printf failed to print a string n continue cfb framebuffer finalize dev should it actually be something like while for int i i display height font height i use the number of character rows cfb framebuffer clear dev false if cfb print dev § i font height printf failed to print a string n continue cfb framebuffer finalize dev although this causes flickering if there is more than one character row in my example font height as the display will draw the row case wrapping the extra characters onto the next row erase the display wiping rows and write to row only erase the display wiping rows and so row flickers between having text and being blank and row flickers between wrapped characters and written characters i think the actual intention of this example is to do something like while cfb framebuffer clear dev false for int i i display height font height i if cfb print dev § i font height printf failed to print a string n continue cfb framebuffer finalize dev which just prints that text on every line or as much as possible this seems much more sensible than the existing sample i can create a pr for this last case i m just looking for some guidance as to what the sample is actually supposed to do to reproduce steps to reproduce the behavior run the sample with an expected behavior the display should not flicker or it should at least be documented that this is the expected behaviour although i m not sure how useful that is as an example screenshots or console output booting zephyr os build initialized font width font height font width font height font width font height x res y res ppt rows cols cfb cfb framebuffer init number of fonts environment please complete the following information os ubuntu toolchain gnuarmemb commit
1
105,522
4,237,377,961
IssuesEvent
2016-07-05 21:36:20
Microsoft/TypeScript
https://api.github.com/repos/Microsoft/TypeScript
closed
STOFL if narrowed variable is mentioned in 'case' condition expression
Bug High Priority
A very simple repro: ```ts function f(x: {a: "A", b} | {a: "C", e}) { switch (x.a) { case x: break; } } ``` that results in stack overflow: ``` C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript>node built\local\tsc.js hello.ts C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript\built\local\tsc.js:37776 throw e; ^ RangeError: Maximum call stack size exceeded at Object.hasOwnProperty (native) at Object.hasProperty (C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript\built\local\tsc.js:1060:31) at getPropertyOfUnionOrIntersectionType (C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript\built\local\tsc.js:17227:20) at getPropertyOfType (C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript\built\local\tsc.js:17255:24) at getTypeOfPropertyOfType (C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript\built\local\tsc.js:16003:24) at narrowTypeBySwitchOnDiscriminant (C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript\built\local\tsc.js:20364:32) at getTypeAtSwitchClause (C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript\built\local\tsc.js:20207:24) at getTypeAtFlowNode (C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript\built\local\tsc.js:20145:32) at getFlowTypeOfReference (C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript\built\local\tsc.js:20118:26) at checkIdentifier (C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript\built\local\tsc.js:20577:28) ```
1.0
STOFL if narrowed variable is mentioned in 'case' condition expression - A very simple repro: ```ts function f(x: {a: "A", b} | {a: "C", e}) { switch (x.a) { case x: break; } } ``` that results in stack overflow: ``` C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript>node built\local\tsc.js hello.ts C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript\built\local\tsc.js:37776 throw e; ^ RangeError: Maximum call stack size exceeded at Object.hasOwnProperty (native) at Object.hasProperty (C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript\built\local\tsc.js:1060:31) at getPropertyOfUnionOrIntersectionType (C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript\built\local\tsc.js:17227:20) at getPropertyOfType (C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript\built\local\tsc.js:17255:24) at getTypeOfPropertyOfType (C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript\built\local\tsc.js:16003:24) at narrowTypeBySwitchOnDiscriminant (C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript\built\local\tsc.js:20364:32) at getTypeAtSwitchClause (C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript\built\local\tsc.js:20207:24) at getTypeAtFlowNode (C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript\built\local\tsc.js:20145:32) at getFlowTypeOfReference (C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript\built\local\tsc.js:20118:26) at checkIdentifier (C:\Users\Anton.Lobov\Desktop\ts_new\TypeScript\built\local\tsc.js:20577:28) ```
priority
stofl if narrowed variable is mentioned in case condition expression a very simple repro ts function f x a a b a c e switch x a case x break that results in stack overflow c users anton lobov desktop ts new typescript node built local tsc js hello ts c users anton lobov desktop ts new typescript built local tsc js throw e rangeerror maximum call stack size exceeded at object hasownproperty native at object hasproperty c users anton lobov desktop ts new typescript built local tsc js at getpropertyofunionorintersectiontype c users anton lobov desktop ts new typescript built local tsc js at getpropertyoftype c users anton lobov desktop ts new typescript built local tsc js at gettypeofpropertyoftype c users anton lobov desktop ts new typescript built local tsc js at narrowtypebyswitchondiscriminant c users anton lobov desktop ts new typescript built local tsc js at gettypeatswitchclause c users anton lobov desktop ts new typescript built local tsc js at gettypeatflownode c users anton lobov desktop ts new typescript built local tsc js at getflowtypeofreference c users anton lobov desktop ts new typescript built local tsc js at checkidentifier c users anton lobov desktop ts new typescript built local tsc js
1
118,417
9,986,781,667
IssuesEvent
2019-07-10 20:00:51
mili-technologies/mili-shit-app
https://api.github.com/repos/mili-technologies/mili-shit-app
closed
Crashing on Gray Add button
bug ready for test
In order page - some items have different add button with quantity increase & decrease button. When click on a plus button to icrease the quantity app gets crashed.
1.0
Crashing on Gray Add button - In order page - some items have different add button with quantity increase & decrease button. When click on a plus button to icrease the quantity app gets crashed.
non_priority
crashing on gray add button in order page some items have different add button with quantity increase decrease button when click on a plus button to icrease the quantity app gets crashed
0
65,946
27,287,222,682
IssuesEvent
2023-02-23 14:19:42
docker-mailserver/docker-mailserver
https://api.github.com/repos/docker-mailserver/docker-mailserver
closed
[FR] Disable "AUTH LOGIN" for non-submission smtp port
priority/low service/postfix kind/improvement area/configuration (file) stale-bot/ignore
# Feature Request ## Context Currently is possible to connect to 25 port and send email after "AUTH LOGIN" command. I want to disallow any type of sending emails by connection to this port and use it only for receiving emails. Sending should be possible only by connecting to submission ports which can be protected by firewall. ### Describe the Solution you'd like Add environment variable to disable this and set it by default by security reasons ### Are you going to implement it? First I want to know maybe this issue can be handled by existing way. Maybe I missed something
1.0
[FR] Disable "AUTH LOGIN" for non-submission smtp port - # Feature Request ## Context Currently is possible to connect to 25 port and send email after "AUTH LOGIN" command. I want to disallow any type of sending emails by connection to this port and use it only for receiving emails. Sending should be possible only by connecting to submission ports which can be protected by firewall. ### Describe the Solution you'd like Add environment variable to disable this and set it by default by security reasons ### Are you going to implement it? First I want to know maybe this issue can be handled by existing way. Maybe I missed something
non_priority
disable auth login for non submission smtp port feature request context currently is possible to connect to port and send email after auth login command i want to disallow any type of sending emails by connection to this port and use it only for receiving emails sending should be possible only by connecting to submission ports which can be protected by firewall describe the solution you d like add environment variable to disable this and set it by default by security reasons are you going to implement it first i want to know maybe this issue can be handled by existing way maybe i missed something
0
716,006
24,618,029,957
IssuesEvent
2022-10-15 15:09:39
azerothcore/azerothcore-wotlk
https://api.github.com/repos/azerothcore/azerothcore-wotlk
closed
Horde Warsong Gulch gates dont despawn
PvP Confirmed ChromieCraft Generic Priority-Trivial
original issue https://github.com/chromiecraft/chromiecraft/issues/3721 ### Current Behaviour The gates that fall down when a battle starts on the horde side never despawn https://user-images.githubusercontent.com/11332559/194896032-1f13e6cd-9b56-4f97-ae0f-22f6003a5963.mp4 ### Expected Blizzlike Behaviour They should despawn a few seconds after the battle starts ### Source https://youtu.be/BPRXfYA5AqI?t=11 bit sped up, but they run through the gate and he looks back shortly afterwards. It already despawned at that time ### Steps to reproduce the problem .debug bg queue WSG ### Extra Notes _No response_ ### AC rev. hash/commit https://github.com/chromiecraft/azerothcore-wotlk/commit/0ef7f999db7284c2a7776b70c9ecf008d8551f16 ### Operating system Ubuntu 20.04 ### Custom changes or Modules [mod-ah-bot](https://github.com/azerothcore/mod-ah-bot) [mod-bg-item-reward](https://github.com/azerothcore/mod-bg-item-reward) [mod-cfbg](https://github.com/azerothcore/mod-cfbg) [mod-chat-transmitter](https://github.com/azerothcore/mod-chat-transmitter) [mod-chromie-xp](https://github.com/azerothcore/mod-chromie-xp) [mod-cta-switch](https://github.com/azerothcore/mod-cta-switch) [mod-desertion-warnings](https://github.com/azerothcore/mod-desertion-warnings) [mod-dmf-switch](https://github.com/azerothcore/mod-dmf-switch) [mod-duel-reset](https://github.com/azerothcore/mod-duel-reset) [mod-eluna](https://github.com/azerothcore/mod-eluna) [mod-ip-tracker](https://github.com/azerothcore/mod-ip-tracker) [mod-low-level-arena](https://github.com/azerothcore/mod-low-level-arena) [mod-low-level-rbg](https://github.com/azerothcore/mod-low-level-rbg) [mod-multi-client-check](https://github.com/azerothcore/mod-multi-client-check) [mod-progression-system](https://github.com/azerothcore/mod-progression-system) [mod-pvp-titles](https://github.com/azerothcore/mod-pvp-titles) [mod-pvpstats-announcer](https://github.com/azerothcore/mod-pvpstats-announcer) [mod-queue-list-cache](https://github.com/azerothcore/mod-queue-list-cache) [mod-rdf-expansion](https://github.com/azerothcore/mod-rdf-expansion) [mod-transmog](https://github.com/azerothcore/mod-transmog) [mod-weekend-xp](https://github.com/azerothcore/mod-weekend-xp) [lua-carbon-copy](https://github.com/55Honey/Acore_CarbonCopy) [lua-exchange-npc](https://github.com/55Honey/Acore_ExchangeNpc) [lua-custom-worldboss](https://github.com/55Honey/Acore_CustomWorldboss) [lua-level-up-reward](https://github.com/55Honey/Acore_LevelUpReward) [lua-recruit-a-friend](https://github.com/55Honey/Acore_RecruitAFriend) [lua-send-and-bind](https://github.com/55Honey/Acore_SendAndBind) [lua-temp-announcements](https://github.com/55Honey/Acore_TempAnnouncements) [lua-zonecheck](https://github.com/55Honey/acore_Zonecheck) [lua-zone-debuff](https://github.com/55Honey/Acore_ZoneDebuff)
1.0
Horde Warsong Gulch gates dont despawn - original issue https://github.com/chromiecraft/chromiecraft/issues/3721 ### Current Behaviour The gates that fall down when a battle starts on the horde side never despawn https://user-images.githubusercontent.com/11332559/194896032-1f13e6cd-9b56-4f97-ae0f-22f6003a5963.mp4 ### Expected Blizzlike Behaviour They should despawn a few seconds after the battle starts ### Source https://youtu.be/BPRXfYA5AqI?t=11 bit sped up, but they run through the gate and he looks back shortly afterwards. It already despawned at that time ### Steps to reproduce the problem .debug bg queue WSG ### Extra Notes _No response_ ### AC rev. hash/commit https://github.com/chromiecraft/azerothcore-wotlk/commit/0ef7f999db7284c2a7776b70c9ecf008d8551f16 ### Operating system Ubuntu 20.04 ### Custom changes or Modules [mod-ah-bot](https://github.com/azerothcore/mod-ah-bot) [mod-bg-item-reward](https://github.com/azerothcore/mod-bg-item-reward) [mod-cfbg](https://github.com/azerothcore/mod-cfbg) [mod-chat-transmitter](https://github.com/azerothcore/mod-chat-transmitter) [mod-chromie-xp](https://github.com/azerothcore/mod-chromie-xp) [mod-cta-switch](https://github.com/azerothcore/mod-cta-switch) [mod-desertion-warnings](https://github.com/azerothcore/mod-desertion-warnings) [mod-dmf-switch](https://github.com/azerothcore/mod-dmf-switch) [mod-duel-reset](https://github.com/azerothcore/mod-duel-reset) [mod-eluna](https://github.com/azerothcore/mod-eluna) [mod-ip-tracker](https://github.com/azerothcore/mod-ip-tracker) [mod-low-level-arena](https://github.com/azerothcore/mod-low-level-arena) [mod-low-level-rbg](https://github.com/azerothcore/mod-low-level-rbg) [mod-multi-client-check](https://github.com/azerothcore/mod-multi-client-check) [mod-progression-system](https://github.com/azerothcore/mod-progression-system) [mod-pvp-titles](https://github.com/azerothcore/mod-pvp-titles) [mod-pvpstats-announcer](https://github.com/azerothcore/mod-pvpstats-announcer) [mod-queue-list-cache](https://github.com/azerothcore/mod-queue-list-cache) [mod-rdf-expansion](https://github.com/azerothcore/mod-rdf-expansion) [mod-transmog](https://github.com/azerothcore/mod-transmog) [mod-weekend-xp](https://github.com/azerothcore/mod-weekend-xp) [lua-carbon-copy](https://github.com/55Honey/Acore_CarbonCopy) [lua-exchange-npc](https://github.com/55Honey/Acore_ExchangeNpc) [lua-custom-worldboss](https://github.com/55Honey/Acore_CustomWorldboss) [lua-level-up-reward](https://github.com/55Honey/Acore_LevelUpReward) [lua-recruit-a-friend](https://github.com/55Honey/Acore_RecruitAFriend) [lua-send-and-bind](https://github.com/55Honey/Acore_SendAndBind) [lua-temp-announcements](https://github.com/55Honey/Acore_TempAnnouncements) [lua-zonecheck](https://github.com/55Honey/acore_Zonecheck) [lua-zone-debuff](https://github.com/55Honey/Acore_ZoneDebuff)
priority
horde warsong gulch gates dont despawn original issue current behaviour the gates that fall down when a battle starts on the horde side never despawn expected blizzlike behaviour they should despawn a few seconds after the battle starts source bit sped up but they run through the gate and he looks back shortly afterwards it already despawned at that time steps to reproduce the problem debug bg queue wsg extra notes no response ac rev hash commit operating system ubuntu custom changes or modules
1
270,941
29,147,217,503
IssuesEvent
2023-05-18 04:54:27
momo-tong/org.springframework.boot-spring-boot-2.2.10.RELEASE
https://api.github.com/repos/momo-tong/org.springframework.boot-spring-boot-2.2.10.RELEASE
closed
spring-messaging-5.2.9.RELEASE.jar: 1 vulnerabilities (highest severity is: 6.5) - autoclosed
Mend: dependency security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-messaging-5.2.9.RELEASE.jar</b></p></summary> <p>Spring Messaging</p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-messaging/5.2.9.RELEASE/spring-messaging-5.2.9.RELEASE.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/momo-tong/org.springframework.boot-spring-boot-2.2.10.RELEASE/commit/e47b37552e432ad2d67525c45be46ca5af41326f">e47b37552e432ad2d67525c45be46ca5af41326f</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (spring-messaging version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2022-22971](https://www.mend.io/vulnerability-database/CVE-2022-22971) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.5 | spring-messaging-5.2.9.RELEASE.jar | Direct | 5.2.22.RELEASE | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2022-22971</summary> ### Vulnerable Library - <b>spring-messaging-5.2.9.RELEASE.jar</b></p> <p>Spring Messaging</p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-messaging/5.2.9.RELEASE/spring-messaging-5.2.9.RELEASE.jar</p> <p> Dependency Hierarchy: - :x: **spring-messaging-5.2.9.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/momo-tong/org.springframework.boot-spring-boot-2.2.10.RELEASE/commit/e47b37552e432ad2d67525c45be46ca5af41326f">e47b37552e432ad2d67525c45be46ca5af41326f</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In spring framework versions prior to 5.3.20+ , 5.2.22+ and old unsupported versions, application with a STOMP over WebSocket endpoint is vulnerable to a denial of service attack by an authenticated user. <p>Publish Date: 2022-05-12 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-22971>CVE-2022-22971</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2022-22971">https://tanzu.vmware.com/security/cve-2022-22971</a></p> <p>Release Date: 2022-05-12</p> <p>Fix Resolution: 5.2.22.RELEASE</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
True
spring-messaging-5.2.9.RELEASE.jar: 1 vulnerabilities (highest severity is: 6.5) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-messaging-5.2.9.RELEASE.jar</b></p></summary> <p>Spring Messaging</p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-messaging/5.2.9.RELEASE/spring-messaging-5.2.9.RELEASE.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/momo-tong/org.springframework.boot-spring-boot-2.2.10.RELEASE/commit/e47b37552e432ad2d67525c45be46ca5af41326f">e47b37552e432ad2d67525c45be46ca5af41326f</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (spring-messaging version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2022-22971](https://www.mend.io/vulnerability-database/CVE-2022-22971) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.5 | spring-messaging-5.2.9.RELEASE.jar | Direct | 5.2.22.RELEASE | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2022-22971</summary> ### Vulnerable Library - <b>spring-messaging-5.2.9.RELEASE.jar</b></p> <p>Spring Messaging</p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-messaging/5.2.9.RELEASE/spring-messaging-5.2.9.RELEASE.jar</p> <p> Dependency Hierarchy: - :x: **spring-messaging-5.2.9.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/momo-tong/org.springframework.boot-spring-boot-2.2.10.RELEASE/commit/e47b37552e432ad2d67525c45be46ca5af41326f">e47b37552e432ad2d67525c45be46ca5af41326f</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In spring framework versions prior to 5.3.20+ , 5.2.22+ and old unsupported versions, application with a STOMP over WebSocket endpoint is vulnerable to a denial of service attack by an authenticated user. <p>Publish Date: 2022-05-12 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-22971>CVE-2022-22971</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2022-22971">https://tanzu.vmware.com/security/cve-2022-22971</a></p> <p>Release Date: 2022-05-12</p> <p>Fix Resolution: 5.2.22.RELEASE</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
non_priority
spring messaging release jar vulnerabilities highest severity is autoclosed vulnerable library spring messaging release jar spring messaging path to dependency file pom xml path to vulnerable library home wss scanner repository org springframework spring messaging release spring messaging release jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in spring messaging version remediation available medium spring messaging release jar direct release details cve vulnerable library spring messaging release jar spring messaging path to dependency file pom xml path to vulnerable library home wss scanner repository org springframework spring messaging release spring messaging release jar dependency hierarchy x spring messaging release jar vulnerable library found in head commit a href found in base branch master vulnerability details in spring framework versions prior to and old unsupported versions application with a stomp over websocket endpoint is vulnerable to a denial of service attack by an authenticated user publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution release step up your open source security game with mend
0
475,922
13,728,175,563
IssuesEvent
2020-10-04 10:21:39
OS-Games-Official/SoulsOfTheForest
https://api.github.com/repos/OS-Games-Official/SoulsOfTheForest
closed
Player flies into the air if you walk into the tree stump/other obstacles
bug medium priority
Version: v0.0.3 To reproduce: Walk into any obstacle, the tree stump provides the greatest effect, and keep walking forwards even after you are over it. This sends you into the air, with larger objects sending you higher. No falling animation plays. Screenshot: ![Screenshot (946)](https://user-images.githubusercontent.com/59693282/95000658-a417a100-05ba-11eb-8aac-f4731b6ecbfd.png)
1.0
Player flies into the air if you walk into the tree stump/other obstacles - Version: v0.0.3 To reproduce: Walk into any obstacle, the tree stump provides the greatest effect, and keep walking forwards even after you are over it. This sends you into the air, with larger objects sending you higher. No falling animation plays. Screenshot: ![Screenshot (946)](https://user-images.githubusercontent.com/59693282/95000658-a417a100-05ba-11eb-8aac-f4731b6ecbfd.png)
priority
player flies into the air if you walk into the tree stump other obstacles version to reproduce walk into any obstacle the tree stump provides the greatest effect and keep walking forwards even after you are over it this sends you into the air with larger objects sending you higher no falling animation plays screenshot
1
120,416
4,788,867,062
IssuesEvent
2016-10-30 19:49:37
michaeljcalkins/rangersteve-ideas
https://api.github.com/repos/michaeljcalkins/rangersteve-ideas
closed
Leaderboard alignment updates
Priority: Medium Status: Accepted Time: > Week Type: Maintenance
- [x] Add kill streak to new column the leaderboard hud - [x] Fix the name alignment in the leaderboard modal
1.0
Leaderboard alignment updates - - [x] Add kill streak to new column the leaderboard hud - [x] Fix the name alignment in the leaderboard modal
priority
leaderboard alignment updates add kill streak to new column the leaderboard hud fix the name alignment in the leaderboard modal
1
276,784
30,543,839,790
IssuesEvent
2023-07-20 01:00:33
billmcchesney1/hadoop
https://api.github.com/repos/billmcchesney1/hadoop
opened
CVE-2022-25883 (Medium) detected in multiple libraries
Mend: dependency security vulnerability
## CVE-2022-25883 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>semver-4.3.6.tgz</b>, <b>semver-5.0.3.tgz</b>, <b>semver-5.3.0.tgz</b>, <b>semver-5.7.1.tgz</b></p></summary> <p> <details><summary><b>semver-4.3.6.tgz</b></p></summary> <p>The semantic version parser used by npm.</p> <p>Library home page: <a href="https://registry.npmjs.org/semver/-/semver-4.3.6.tgz">https://registry.npmjs.org/semver/-/semver-4.3.6.tgz</a></p> <p>Path to dependency file: /hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/package.json</p> <p>Path to vulnerable library: /hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/node_modules/ember-cli-dependency-checker/node_modules/semver/package.json,/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/node_modules/ember-cli-release/node_modules/semver/package.json,/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/node_modules/ember-cli/node_modules/semver/package.json</p> <p> Dependency Hierarchy: - ember-cli-dependency-checker-1.2.0.tgz (Root Library) - :x: **semver-4.3.6.tgz** (Vulnerable Library) </details> <details><summary><b>semver-5.0.3.tgz</b></p></summary> <p>The semantic version parser used by npm.</p> <p>Library home page: <a href="https://registry.npmjs.org/semver/-/semver-5.0.3.tgz">https://registry.npmjs.org/semver/-/semver-5.0.3.tgz</a></p> <p>Path to dependency file: /hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/package.json</p> <p>Path to vulnerable library: /hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/node_modules/npm/node_modules/semver/package.json</p> <p> Dependency Hierarchy: - ember-cli-1.13.14.tgz (Root Library) - npm-2.14.10.tgz - :x: **semver-5.0.3.tgz** (Vulnerable Library) </details> <details><summary><b>semver-5.3.0.tgz</b></p></summary> <p>The semantic version parser used by npm.</p> <p>Library home page: <a href="https://registry.npmjs.org/semver/-/semver-5.3.0.tgz">https://registry.npmjs.org/semver/-/semver-5.3.0.tgz</a></p> <p>Path to dependency file: /hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/package.json</p> <p>Path to vulnerable library: /hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/node_modules/node-gyp/node_modules/semver/package.json,/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-catalog/hadoop-yarn-applications-catalog-webapp/node_modules/semver/package.json</p> <p> Dependency Hierarchy: - ember-cli-sass-7.0.0.tgz (Root Library) - broccoli-sass-source-maps-2.2.0.tgz - node-sass-4.14.1.tgz - node-gyp-3.8.0.tgz - :x: **semver-5.3.0.tgz** (Vulnerable Library) </details> <details><summary><b>semver-5.7.1.tgz</b></p></summary> <p>The semantic version parser used by npm.</p> <p>Library home page: <a href="https://registry.npmjs.org/semver/-/semver-5.7.1.tgz">https://registry.npmjs.org/semver/-/semver-5.7.1.tgz</a></p> <p>Path to dependency file: /hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/package.json</p> <p>Path to vulnerable library: /hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/node_modules/semver/package.json</p> <p> Dependency Hierarchy: - bower-shrinkwrap-resolver-ext-0.1.0.tgz (Root Library) - :x: **semver-5.7.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/hadoop/commit/6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a">6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a</a></p> <p>Found in base branch: <b>trunk</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Versions of the package semver before 7.5.2 are vulnerable to Regular Expression Denial of Service (ReDoS) via the function new Range, when untrusted user data is provided as a range. <p>Publish Date: 2023-06-21 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25883>CVE-2022-25883</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2023-06-21</p> <p>Fix Resolution: semver - 7.5.2</p> </p> </details> <p></p>
True
CVE-2022-25883 (Medium) detected in multiple libraries - ## CVE-2022-25883 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>semver-4.3.6.tgz</b>, <b>semver-5.0.3.tgz</b>, <b>semver-5.3.0.tgz</b>, <b>semver-5.7.1.tgz</b></p></summary> <p> <details><summary><b>semver-4.3.6.tgz</b></p></summary> <p>The semantic version parser used by npm.</p> <p>Library home page: <a href="https://registry.npmjs.org/semver/-/semver-4.3.6.tgz">https://registry.npmjs.org/semver/-/semver-4.3.6.tgz</a></p> <p>Path to dependency file: /hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/package.json</p> <p>Path to vulnerable library: /hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/node_modules/ember-cli-dependency-checker/node_modules/semver/package.json,/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/node_modules/ember-cli-release/node_modules/semver/package.json,/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/node_modules/ember-cli/node_modules/semver/package.json</p> <p> Dependency Hierarchy: - ember-cli-dependency-checker-1.2.0.tgz (Root Library) - :x: **semver-4.3.6.tgz** (Vulnerable Library) </details> <details><summary><b>semver-5.0.3.tgz</b></p></summary> <p>The semantic version parser used by npm.</p> <p>Library home page: <a href="https://registry.npmjs.org/semver/-/semver-5.0.3.tgz">https://registry.npmjs.org/semver/-/semver-5.0.3.tgz</a></p> <p>Path to dependency file: /hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/package.json</p> <p>Path to vulnerable library: /hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/node_modules/npm/node_modules/semver/package.json</p> <p> Dependency Hierarchy: - ember-cli-1.13.14.tgz (Root Library) - npm-2.14.10.tgz - :x: **semver-5.0.3.tgz** (Vulnerable Library) </details> <details><summary><b>semver-5.3.0.tgz</b></p></summary> <p>The semantic version parser used by npm.</p> <p>Library home page: <a href="https://registry.npmjs.org/semver/-/semver-5.3.0.tgz">https://registry.npmjs.org/semver/-/semver-5.3.0.tgz</a></p> <p>Path to dependency file: /hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/package.json</p> <p>Path to vulnerable library: /hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/node_modules/node-gyp/node_modules/semver/package.json,/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-catalog/hadoop-yarn-applications-catalog-webapp/node_modules/semver/package.json</p> <p> Dependency Hierarchy: - ember-cli-sass-7.0.0.tgz (Root Library) - broccoli-sass-source-maps-2.2.0.tgz - node-sass-4.14.1.tgz - node-gyp-3.8.0.tgz - :x: **semver-5.3.0.tgz** (Vulnerable Library) </details> <details><summary><b>semver-5.7.1.tgz</b></p></summary> <p>The semantic version parser used by npm.</p> <p>Library home page: <a href="https://registry.npmjs.org/semver/-/semver-5.7.1.tgz">https://registry.npmjs.org/semver/-/semver-5.7.1.tgz</a></p> <p>Path to dependency file: /hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/package.json</p> <p>Path to vulnerable library: /hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/node_modules/semver/package.json</p> <p> Dependency Hierarchy: - bower-shrinkwrap-resolver-ext-0.1.0.tgz (Root Library) - :x: **semver-5.7.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/hadoop/commit/6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a">6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a</a></p> <p>Found in base branch: <b>trunk</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Versions of the package semver before 7.5.2 are vulnerable to Regular Expression Denial of Service (ReDoS) via the function new Range, when untrusted user data is provided as a range. <p>Publish Date: 2023-06-21 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25883>CVE-2022-25883</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2023-06-21</p> <p>Fix Resolution: semver - 7.5.2</p> </p> </details> <p></p>
non_priority
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries semver tgz semver tgz semver tgz semver tgz semver tgz the semantic version parser used by npm library home page a href path to dependency file hadoop yarn project hadoop yarn hadoop yarn ui src main webapp package json path to vulnerable library hadoop yarn project hadoop yarn hadoop yarn ui src main webapp node modules ember cli dependency checker node modules semver package json hadoop yarn project hadoop yarn hadoop yarn ui src main webapp node modules ember cli release node modules semver package json hadoop yarn project hadoop yarn hadoop yarn ui src main webapp node modules ember cli node modules semver package json dependency hierarchy ember cli dependency checker tgz root library x semver tgz vulnerable library semver tgz the semantic version parser used by npm library home page a href path to dependency file hadoop yarn project hadoop yarn hadoop yarn ui src main webapp package json path to vulnerable library hadoop yarn project hadoop yarn hadoop yarn ui src main webapp node modules npm node modules semver package json dependency hierarchy ember cli tgz root library npm tgz x semver tgz vulnerable library semver tgz the semantic version parser used by npm library home page a href path to dependency file hadoop yarn project hadoop yarn hadoop yarn ui src main webapp package json path to vulnerable library hadoop yarn project hadoop yarn hadoop yarn ui src main webapp node modules node gyp node modules semver package json hadoop yarn project hadoop yarn hadoop yarn applications hadoop yarn applications catalog hadoop yarn applications catalog webapp node modules semver package json dependency hierarchy ember cli sass tgz root library broccoli sass source maps tgz node sass tgz node gyp tgz x semver tgz vulnerable library semver tgz the semantic version parser used by npm library home page a href path to dependency file hadoop yarn project hadoop yarn hadoop yarn ui src main webapp package json path to vulnerable library hadoop yarn project hadoop yarn hadoop yarn ui src main webapp node modules semver package json dependency hierarchy bower shrinkwrap resolver ext tgz root library x semver tgz vulnerable library found in head commit a href found in base branch trunk vulnerability details versions of the package semver before are vulnerable to regular expression denial of service redos via the function new range when untrusted user data is provided as a range publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version release date fix resolution semver
0
590,417
17,777,413,860
IssuesEvent
2021-08-30 21:11:13
o3de/o3de
https://api.github.com/repos/o3de/o3de
opened
deleting a UI canvas from script via “destroy entity and child” caused the editor to crash
kind/bug needs-triage sig/content status/backlog triage/accepted priority/critical feature/lyshine WF2
[Migrated from JIRA LYN-4248] – We needed to use Ui Canvas Unload instead. we should not crash the editor.
1.0
deleting a UI canvas from script via “destroy entity and child” caused the editor to crash - [Migrated from JIRA LYN-4248] – We needed to use Ui Canvas Unload instead. we should not crash the editor.
priority
deleting a ui canvas from script via “destroy entity and child” caused the editor to crash – we needed to use ui canvas unload instead we should not crash the editor
1
66,731
16,685,231,865
IssuesEvent
2021-06-08 07:19:43
spring-projects/spring-security
https://api.github.com/repos/spring-projects/spring-security
reopened
Build failures due to Timestamp precision
in: build type: bug
Tests relying on HSQLDB 2.5.1 are failing because of https://sourceforge.net/p/hsqldb/bugs/1585/ We will temporarily change our assertions on Timestamp values to be more lenient (to the millisecond).
1.0
Build failures due to Timestamp precision - Tests relying on HSQLDB 2.5.1 are failing because of https://sourceforge.net/p/hsqldb/bugs/1585/ We will temporarily change our assertions on Timestamp values to be more lenient (to the millisecond).
non_priority
build failures due to timestamp precision tests relying on hsqldb are failing because of we will temporarily change our assertions on timestamp values to be more lenient to the millisecond
0
477,625
13,765,483,149
IssuesEvent
2020-10-07 13:28:17
zeoflow/material-elements
https://api.github.com/repos/zeoflow/material-elements
closed
SquareComponent<T>
@feature @priority-low
**Is your feature request related to a problem? Please describe.** Create a square view using a class annotator We also happily accept [pull requests](https://github.com/zeoflow/material-elements/pulls).
1.0
SquareComponent<T> - **Is your feature request related to a problem? Please describe.** Create a square view using a class annotator We also happily accept [pull requests](https://github.com/zeoflow/material-elements/pulls).
priority
squarecomponent is your feature request related to a problem please describe create a square view using a class annotator we also happily accept
1
20,337
29,868,713,002
IssuesEvent
2023-06-20 06:57:18
jg-rp/liquid
https://api.github.com/repos/jg-rp/liquid
closed
The `map` filter should flatten its input sequence
incompatibility
The reference implementation of the `map` filter uses [`InputIterator`](https://github.com/Shopify/liquid/blob/master/lib/liquid/standardfilters.rb#L932) on its input value, causing the input to be flattened if it's a nested array, or be a single element array if it's a hash. We have similar behaviour for other filters, using our `sequence_filter` decorator, but we do not apply it to `map`.
True
The `map` filter should flatten its input sequence - The reference implementation of the `map` filter uses [`InputIterator`](https://github.com/Shopify/liquid/blob/master/lib/liquid/standardfilters.rb#L932) on its input value, causing the input to be flattened if it's a nested array, or be a single element array if it's a hash. We have similar behaviour for other filters, using our `sequence_filter` decorator, but we do not apply it to `map`.
non_priority
the map filter should flatten its input sequence the reference implementation of the map filter uses on its input value causing the input to be flattened if it s a nested array or be a single element array if it s a hash we have similar behaviour for other filters using our sequence filter decorator but we do not apply it to map
0
95,439
3,951,556,912
IssuesEvent
2016-04-29 02:07:18
Captianrock/android_PV
https://api.github.com/repos/Captianrock/android_PV
closed
Move sdks into app
Low Priority New Feature
We are using sdk 23, so just move that sdk inside project so user doesn't need to provide them. In the future, we can give the user the option to select the sdk.
1.0
Move sdks into app - We are using sdk 23, so just move that sdk inside project so user doesn't need to provide them. In the future, we can give the user the option to select the sdk.
priority
move sdks into app we are using sdk so just move that sdk inside project so user doesn t need to provide them in the future we can give the user the option to select the sdk
1
454,965
13,109,806,693
IssuesEvent
2020-08-04 19:24:25
CDH-Studio/UpSkill
https://api.github.com/repos/CDH-Studio/UpSkill
closed
Check i18n files for missing and unused keys
Front-end Low Priority enhancement
**Note:** This would be a nice to have to give us peace of mind knowing that the app is translated The following could be done in a GitHub action or just in a script in package.json - To make sure everything is in both languages - in the i18n JSON files, they should both have the same keys (simple javascript) - detect hardcoded strings in the app (could extend eslint https://www.npmjs.com/package/eslint-plugin-react-intl) - To clean up the files - we could automatically remove unused keys by searching the whole frontend project to see if they are used (simple javascript) - we could sort the whole file (lodash/javascript) Would prevent issues like #417 #378 #211 #151
1.0
Check i18n files for missing and unused keys - **Note:** This would be a nice to have to give us peace of mind knowing that the app is translated The following could be done in a GitHub action or just in a script in package.json - To make sure everything is in both languages - in the i18n JSON files, they should both have the same keys (simple javascript) - detect hardcoded strings in the app (could extend eslint https://www.npmjs.com/package/eslint-plugin-react-intl) - To clean up the files - we could automatically remove unused keys by searching the whole frontend project to see if they are used (simple javascript) - we could sort the whole file (lodash/javascript) Would prevent issues like #417 #378 #211 #151
priority
check files for missing and unused keys note this would be a nice to have to give us peace of mind knowing that the app is translated the following could be done in a github action or just in a script in package json to make sure everything is in both languages in the json files they should both have the same keys simple javascript detect hardcoded strings in the app could extend eslint to clean up the files we could automatically remove unused keys by searching the whole frontend project to see if they are used simple javascript we could sort the whole file lodash javascript would prevent issues like
1
9,081
24,036,582,812
IssuesEvent
2022-09-15 19:48:13
MicrosoftDocs/architecture-center
https://api.github.com/repos/MicrosoftDocs/architecture-center
closed
Broken nav link "Secure access to IoT apps with Azure AD"
assigned-to-author triaged architecture-center/svc reference-architecture/subsvc Pri2
Left nav broken link: Secure access to IoT apps with Azure AD Used to point to this page (found on web archive): https://web.archive.org/web/20200825221210/https://docs.microsoft.com/en-us/azure/architecture/example-scenario/iot-aad/iot-aad Please fix link (to original source on docs.microsoft.com obviously not web archive). [Enter feedback here] --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: cf328b1f-6bc5-dff2-619f-82b95030c3cf * Version Independent ID: 4c7bd8b7-5ebb-5b1e-010b-2acd7fdcec0f * Content: [Azure IoT reference architecture - Azure Reference Architectures](https://docs.microsoft.com/en-us/azure/architecture/reference-architectures/iot) * Content Source: [docs/reference-architectures/iot.yml](https://github.com/microsoftdocs/architecture-center/blob/master/docs/reference-architectures/iot.yml) * Service: **architecture-center** * Sub-service: **reference-architecture** * GitHub Login: @doodlemania2 * Microsoft Alias: **pnp**
2.0
Broken nav link "Secure access to IoT apps with Azure AD" - Left nav broken link: Secure access to IoT apps with Azure AD Used to point to this page (found on web archive): https://web.archive.org/web/20200825221210/https://docs.microsoft.com/en-us/azure/architecture/example-scenario/iot-aad/iot-aad Please fix link (to original source on docs.microsoft.com obviously not web archive). [Enter feedback here] --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: cf328b1f-6bc5-dff2-619f-82b95030c3cf * Version Independent ID: 4c7bd8b7-5ebb-5b1e-010b-2acd7fdcec0f * Content: [Azure IoT reference architecture - Azure Reference Architectures](https://docs.microsoft.com/en-us/azure/architecture/reference-architectures/iot) * Content Source: [docs/reference-architectures/iot.yml](https://github.com/microsoftdocs/architecture-center/blob/master/docs/reference-architectures/iot.yml) * Service: **architecture-center** * Sub-service: **reference-architecture** * GitHub Login: @doodlemania2 * Microsoft Alias: **pnp**
non_priority
broken nav link secure access to iot apps with azure ad left nav broken link secure access to iot apps with azure ad used to point to this page found on web archive please fix link to original source on docs microsoft com obviously not web archive document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service architecture center sub service reference architecture github login microsoft alias pnp
0
274,251
23,824,201,820
IssuesEvent
2022-09-05 13:40:29
londonmedicallaboratory/api-sdk
https://api.github.com/repos/londonmedicallaboratory/api-sdk
closed
Product banner
Ready for test
**Description** On current lml you can create banner that will be displayed on every product page. Banner has id, text and enabled properties. **Screenshots** ![image](https://user-images.githubusercontent.com/34042955/187892874-dec81b34-8310-45c4-8263-45fcd54593e5.png) **URL** https://www.londonmedicallaboratory.com/product/prostate-profile
1.0
Product banner - **Description** On current lml you can create banner that will be displayed on every product page. Banner has id, text and enabled properties. **Screenshots** ![image](https://user-images.githubusercontent.com/34042955/187892874-dec81b34-8310-45c4-8263-45fcd54593e5.png) **URL** https://www.londonmedicallaboratory.com/product/prostate-profile
non_priority
product banner description on current lml you can create banner that will be displayed on every product page banner has id text and enabled properties screenshots url
0
2,076
3,269,008,111
IssuesEvent
2015-10-23 14:30:40
lucasb/leafbird
https://api.github.com/repos/lucasb/leafbird
closed
Performance of loops
performance
Opening an issue for us to verify the code the actual need of using `` `for (key in args)` `` because if we can replace regressive basic loops we would have a gain of considerable performace, about 99%. See the example that [Jsperf](http://jsperf.com/fastest-array-loops-in-javascript/417)
True
Performance of loops - Opening an issue for us to verify the code the actual need of using `` `for (key in args)` `` because if we can replace regressive basic loops we would have a gain of considerable performace, about 99%. See the example that [Jsperf](http://jsperf.com/fastest-array-loops-in-javascript/417)
non_priority
performance of loops opening an issue for us to verify the code the actual need of using for key in args because if we can replace regressive basic loops we would have a gain of considerable performace about see the example that
0
49,337
20,738,884,369
IssuesEvent
2022-03-14 15:55:12
ClickHouse/ClickHouse
https://api.github.com/repos/ClickHouse/ClickHouse
closed
why doesn't my ngrambf_v1 index work well
question comp-skipidx v20.1 support-services
### 1.my clickhouse version is 20.1.6.30 ### 2.table ddl CREATE TABLE default.ip_test2 ( `found_time` UInt32, `d_ip` String, INDEX d_ip_bl_idx d_ip TYPE bloom_filter(0.001) GRANULARITY 1, INDEX d_ip_idx2 d_ip TYPE ngrambf_v1(3, 315119, 6, 0) GRANULARITY 1 ) ENGINE = MergeTree() PARTITION BY toRelativeHourNum(toDateTime(found_time)) ORDER BY found_time SETTINGS index_granularity = 8192; ### 3.the data count is 10000000. the sample data like below ┌─found_time─┬─d_ip────────────────────────────────────┐ │ 1592591894 │ 3200:1010:1100:0100:0110:1110:1100:0001 │ │ 1592591894 │ 8010:1100:1010:0100:0000:1010:1010:2031 │ │ 1592591894 │ 102.171.195.13 │ │ 1592591894 │ d110:1100:0100:1000:0100:0010:1100:1310 │ │ 1592591894 │ 3110:0010:0000:0110:1100:0110:1100:0012 │ │ 1592591894 │ a300:0100:1110:1000:0000:0010:1110:0000 │ │ 1592591894 │ 2200:1000:0000:0100:1110:0100:1000:3131 │ │ 1592591894 │ 5210:0110:0100:0100:1000:1000:1000:0001 │ │ 1592591894 │ 3001:1000:0110:0000:1010:0110:0100:2320 │ │ 1592591894 │ 116.184.47.2 │ └────────────┴─────────────────────────────────────────┘ ### 4.my query sql:select * from default.ip_test2 where d_ip like '%171.195%' limit 10;the below is query log.My question is why doesn't d_ip_idx2 drop 0 granules?are the prameters i give to the ngrambf_v1 wrong? >The query result is 10 rows,if every result is in a index_granularity 8192,10 rows locates in 81920 rows,but the query shows that it scan 400million rows! > [bd-server-19] 2020.07.24 09:07:18.203411 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> InterpreterSelectQuery: MergeTreeWhereOptimizer: condition "d_ip LIKE '%171.195%'" moved to PREWHERE [bd-server-19] 2020.07.24 09:07:18.203639 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Trace> ContextAccess (default): Access granted: SELECT(found_time, d_ip) ON default.ip_test2 [bd-server-19] 2020.07.24 09:07:18.203816 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Key condition: unknown [bd-server-19] 2020.07.24 09:07:18.203847 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): MinMax index condition: unknown [bd-server-19] 2020.07.24 09:07:18.219786 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.248175 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.281673 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.306771 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.330602 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.345376 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.361686 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.377417 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.384960 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.392819 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.401857 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.409763 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.418204 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.426288 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.434560 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.441925 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.457873 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.457909 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Selected 17 parts by date, 17 parts by key, 1230 marks to read from 17 ranges [bd-server-19] 2020.07.24 09:07:18.458151 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Trace> default.ip_test2 (SelectExecutor): Reading approx. 10076160 rows with 28 streams the resultis : ┌─found_time─┬─d_ip───────────┐ │ 1592601429 │ 46.171.195.198 │ └────────────┴────────────────┘ ┌─found_time─┬─d_ip───────────┐ │ 1592633446 │ 90.180.171.195 │ └────────────┴────────────────┘ ┌─found_time─┬─d_ip──────────┐ │ 1592613116 │ 122.171.195.5 │ └────────────┴───────────────┘ ┌─found_time─┬─d_ip─────────┐ │ 1592593924 │ 76.171.195.1 │ └────────────┴──────────────┘ ┌─found_time─┬─d_ip───────────┐ │ 1592623000 │ 184.171.195.78 │ └────────────┴────────────────┘ ┌─found_time─┬─d_ip─────────┐ │ 1592597161 │ 17.171.195.3 │ └────────────┴──────────────┘ ┌─found_time─┬─d_ip───────────┐ │ 1592643661 │ 171.195.229.10 │ │ 1592643661 │ 213.171.195.8 │ └────────────┴────────────────┘ ┌─found_time─┬─d_ip──────────┐ │ 1592601429 │ 171.195.231.7 │ └────────────┴───────────────┘ ┌─found_time─┬─d_ip───────────┐ │ 1592625726 │ 78.218.171.195 │ └────────────┴────────────────┘ **_10 rows in set. Elapsed: 0.334 sec. Processed 4.00 million rows,_** 139.73 MB (11.97 million rows/s., 418.41 MB/s.)
1.0
why doesn't my ngrambf_v1 index work well - ### 1.my clickhouse version is 20.1.6.30 ### 2.table ddl CREATE TABLE default.ip_test2 ( `found_time` UInt32, `d_ip` String, INDEX d_ip_bl_idx d_ip TYPE bloom_filter(0.001) GRANULARITY 1, INDEX d_ip_idx2 d_ip TYPE ngrambf_v1(3, 315119, 6, 0) GRANULARITY 1 ) ENGINE = MergeTree() PARTITION BY toRelativeHourNum(toDateTime(found_time)) ORDER BY found_time SETTINGS index_granularity = 8192; ### 3.the data count is 10000000. the sample data like below ┌─found_time─┬─d_ip────────────────────────────────────┐ │ 1592591894 │ 3200:1010:1100:0100:0110:1110:1100:0001 │ │ 1592591894 │ 8010:1100:1010:0100:0000:1010:1010:2031 │ │ 1592591894 │ 102.171.195.13 │ │ 1592591894 │ d110:1100:0100:1000:0100:0010:1100:1310 │ │ 1592591894 │ 3110:0010:0000:0110:1100:0110:1100:0012 │ │ 1592591894 │ a300:0100:1110:1000:0000:0010:1110:0000 │ │ 1592591894 │ 2200:1000:0000:0100:1110:0100:1000:3131 │ │ 1592591894 │ 5210:0110:0100:0100:1000:1000:1000:0001 │ │ 1592591894 │ 3001:1000:0110:0000:1010:0110:0100:2320 │ │ 1592591894 │ 116.184.47.2 │ └────────────┴─────────────────────────────────────────┘ ### 4.my query sql:select * from default.ip_test2 where d_ip like '%171.195%' limit 10;the below is query log.My question is why doesn't d_ip_idx2 drop 0 granules?are the prameters i give to the ngrambf_v1 wrong? >The query result is 10 rows,if every result is in a index_granularity 8192,10 rows locates in 81920 rows,but the query shows that it scan 400million rows! > [bd-server-19] 2020.07.24 09:07:18.203411 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> InterpreterSelectQuery: MergeTreeWhereOptimizer: condition "d_ip LIKE '%171.195%'" moved to PREWHERE [bd-server-19] 2020.07.24 09:07:18.203639 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Trace> ContextAccess (default): Access granted: SELECT(found_time, d_ip) ON default.ip_test2 [bd-server-19] 2020.07.24 09:07:18.203816 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Key condition: unknown [bd-server-19] 2020.07.24 09:07:18.203847 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): MinMax index condition: unknown [bd-server-19] 2020.07.24 09:07:18.219786 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.248175 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.281673 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.306771 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.330602 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.345376 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.361686 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.377417 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.384960 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.392819 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.401857 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.409763 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.418204 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.426288 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.434560 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.441925 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.457873 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Index `d_ip_idx2` has dropped 0 granules. [bd-server-19] 2020.07.24 09:07:18.457909 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Debug> default.ip_test2 (SelectExecutor): Selected 17 parts by date, 17 parts by key, 1230 marks to read from 17 ranges [bd-server-19] 2020.07.24 09:07:18.458151 [ 127961 ] {b5d1900f-6870-4c13-8383-03de0b0f126e} <Trace> default.ip_test2 (SelectExecutor): Reading approx. 10076160 rows with 28 streams the resultis : ┌─found_time─┬─d_ip───────────┐ │ 1592601429 │ 46.171.195.198 │ └────────────┴────────────────┘ ┌─found_time─┬─d_ip───────────┐ │ 1592633446 │ 90.180.171.195 │ └────────────┴────────────────┘ ┌─found_time─┬─d_ip──────────┐ │ 1592613116 │ 122.171.195.5 │ └────────────┴───────────────┘ ┌─found_time─┬─d_ip─────────┐ │ 1592593924 │ 76.171.195.1 │ └────────────┴──────────────┘ ┌─found_time─┬─d_ip───────────┐ │ 1592623000 │ 184.171.195.78 │ └────────────┴────────────────┘ ┌─found_time─┬─d_ip─────────┐ │ 1592597161 │ 17.171.195.3 │ └────────────┴──────────────┘ ┌─found_time─┬─d_ip───────────┐ │ 1592643661 │ 171.195.229.10 │ │ 1592643661 │ 213.171.195.8 │ └────────────┴────────────────┘ ┌─found_time─┬─d_ip──────────┐ │ 1592601429 │ 171.195.231.7 │ └────────────┴───────────────┘ ┌─found_time─┬─d_ip───────────┐ │ 1592625726 │ 78.218.171.195 │ └────────────┴────────────────┘ **_10 rows in set. Elapsed: 0.334 sec. Processed 4.00 million rows,_** 139.73 MB (11.97 million rows/s., 418.41 MB/s.)
non_priority
why doesn t my ngrambf index work well my clickhouse version is table ddl create table default ip found time d ip string index d ip bl idx d ip type bloom filter granularity index d ip d ip type ngrambf granularity engine mergetree partition by torelativehournum todatetime found time order by found time settings index granularity the data count is the sample data like below ┌─found time─┬─d ip────────────────────────────────────┐ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ │ └────────────┴─────────────────────────────────────────┘ my query sql select from default ip where d ip like limit the below is query log my question is why doesn t d ip drop granules are the prameters i give to the ngrambf wrong the query result is rows if every result is in a index granularity rows locates in rows but the query shows that it scan rows interpreterselectquery mergetreewhereoptimizer condition d ip like moved to prewhere contextaccess default access granted select found time d ip on default ip default ip selectexecutor key condition unknown default ip selectexecutor minmax index condition unknown default ip selectexecutor index d ip has dropped granules default ip selectexecutor index d ip has dropped granules default ip selectexecutor index d ip has dropped granules default ip selectexecutor index d ip has dropped granules default ip selectexecutor index d ip has dropped granules default ip selectexecutor index d ip has dropped granules default ip selectexecutor index d ip has dropped granules default ip selectexecutor index d ip has dropped granules default ip selectexecutor index d ip has dropped granules default ip selectexecutor index d ip has dropped granules default ip selectexecutor index d ip has dropped granules default ip selectexecutor index d ip has dropped granules default ip selectexecutor index d ip has dropped granules default ip selectexecutor index d ip has dropped granules default ip selectexecutor index d ip has dropped granules default ip selectexecutor index d ip has dropped granules default ip selectexecutor index d ip has dropped granules default ip selectexecutor selected parts by date parts by key marks to read from ranges default ip selectexecutor reading approx rows with streams the resultis ┌─found time─┬─d ip───────────┐ │ │ │ └────────────┴────────────────┘ ┌─found time─┬─d ip───────────┐ │ │ │ └────────────┴────────────────┘ ┌─found time─┬─d ip──────────┐ │ │ │ └────────────┴───────────────┘ ┌─found time─┬─d ip─────────┐ │ │ │ └────────────┴──────────────┘ ┌─found time─┬─d ip───────────┐ │ │ │ └────────────┴────────────────┘ ┌─found time─┬─d ip─────────┐ │ │ │ └────────────┴──────────────┘ ┌─found time─┬─d ip───────────┐ │ │ │ │ │ │ └────────────┴────────────────┘ ┌─found time─┬─d ip──────────┐ │ │ │ └────────────┴───────────────┘ ┌─found time─┬─d ip───────────┐ │ │ │ └────────────┴────────────────┘ rows in set elapsed sec processed million rows mb million rows s mb s
0
574,693
17,023,859,472
IssuesEvent
2021-07-03 04:13:29
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
[amenity-points] render historic=monument
Component: mapnik Priority: minor Resolution: duplicate Type: enhancement
**[Submitted to the original trac issue database at 8.07pm, Friday, 12th April 2013]** With the default Mapnik style in OSM, the tag historic=monument renders the name as text (with no icon) when applied to an area: http://www.openstreetmap.org/browse/way/216272330 But nothing is rendered when it is applied to a node: http://www.openstreetmap.org/browse/node/2210726836 http://www.openstreetmap.org/browse/node/2210726849 It doesn't make sense, shouldn't it have the same behaviour? I think it should behave like memorials, perhaps with its own icon.
1.0
[amenity-points] render historic=monument - **[Submitted to the original trac issue database at 8.07pm, Friday, 12th April 2013]** With the default Mapnik style in OSM, the tag historic=monument renders the name as text (with no icon) when applied to an area: http://www.openstreetmap.org/browse/way/216272330 But nothing is rendered when it is applied to a node: http://www.openstreetmap.org/browse/node/2210726836 http://www.openstreetmap.org/browse/node/2210726849 It doesn't make sense, shouldn't it have the same behaviour? I think it should behave like memorials, perhaps with its own icon.
priority
render historic monument with the default mapnik style in osm the tag historic monument renders the name as text with no icon when applied to an area but nothing is rendered when it is applied to a node it doesn t make sense shouldn t it have the same behaviour i think it should behave like memorials perhaps with its own icon
1
723,307
24,892,138,956
IssuesEvent
2022-10-28 12:57:24
DDMAL/Rodan
https://api.github.com/repos/DDMAL/Rodan
opened
Workflow on rodan "incomplete" for no apparent reason (production)
Priority: LOW bug Rodan Client
I built a small workflow in production to test certain jobs, but the status on the top right said that the "setup is incomplete." I can't click on that notification, so I can't get any more information about what the problem could be. Rodan won't allow me to run the workflow, because it's incomplete, but nothing I do completes it. I deleted all the jobs and rebuilt the exact same workflow and it was valid again. Photo evidence: <img width="1438" alt="Incomplete?" src="https://user-images.githubusercontent.com/83373378/198592478-588d801d-130f-46cb-a47a-619be2f3bc6b.png"> <img width="1440" alt="Valid" src="https://user-images.githubusercontent.com/83373378/198592521-479ba05e-5404-40dd-abbb-02a90fbb9e0b.png"> I have a possible theory: I also get the "incomplete" notification when I have jobs in my workflow that aren't connected to any other jobs. Once I connect them, the "incomplete" changes to "valid." Also, when I add more jobs to a workflow, they all appear in the same spot, overlapping each other. It's therefore possible that some jobs are hidden behind others, so they're triggering an "incomplete" notification, but I can't see them, so I can't fix it.
1.0
Workflow on rodan "incomplete" for no apparent reason (production) - I built a small workflow in production to test certain jobs, but the status on the top right said that the "setup is incomplete." I can't click on that notification, so I can't get any more information about what the problem could be. Rodan won't allow me to run the workflow, because it's incomplete, but nothing I do completes it. I deleted all the jobs and rebuilt the exact same workflow and it was valid again. Photo evidence: <img width="1438" alt="Incomplete?" src="https://user-images.githubusercontent.com/83373378/198592478-588d801d-130f-46cb-a47a-619be2f3bc6b.png"> <img width="1440" alt="Valid" src="https://user-images.githubusercontent.com/83373378/198592521-479ba05e-5404-40dd-abbb-02a90fbb9e0b.png"> I have a possible theory: I also get the "incomplete" notification when I have jobs in my workflow that aren't connected to any other jobs. Once I connect them, the "incomplete" changes to "valid." Also, when I add more jobs to a workflow, they all appear in the same spot, overlapping each other. It's therefore possible that some jobs are hidden behind others, so they're triggering an "incomplete" notification, but I can't see them, so I can't fix it.
priority
workflow on rodan incomplete for no apparent reason production i built a small workflow in production to test certain jobs but the status on the top right said that the setup is incomplete i can t click on that notification so i can t get any more information about what the problem could be rodan won t allow me to run the workflow because it s incomplete but nothing i do completes it i deleted all the jobs and rebuilt the exact same workflow and it was valid again photo evidence img width alt incomplete src img width alt valid src i have a possible theory i also get the incomplete notification when i have jobs in my workflow that aren t connected to any other jobs once i connect them the incomplete changes to valid also when i add more jobs to a workflow they all appear in the same spot overlapping each other it s therefore possible that some jobs are hidden behind others so they re triggering an incomplete notification but i can t see them so i can t fix it
1
100,991
30,841,731,340
IssuesEvent
2023-08-02 11:07:51
PaddlePaddle/Paddle
https://api.github.com/repos/PaddlePaddle/Paddle
closed
在美国的linux服务器上进行试验 python -v结果 :3.7.3 64位python pip install paddlepaddle -i https://mirror.baidu.com/pypi/simple pip install paddlepaddle 等结果 报错为 ERROR: Could not find a version that satisfies the requirement paddlepaddle (from versions: none) ERROR: No matching
status/following-up status/new-issue type/build status/close
### 问题描述 Issue Description 在美国的linux服务器上进行试验 python -v结果 :3.7.3 64位python pip install paddlepaddle -i https://mirror.baidu.com/pypi/simple pip install paddlepaddle 等结果 报错为 ERROR: Could not find a version that satisfies the requirement paddlepaddle (from versions: none) ERROR: No matching distribution found for paddlepaddle 感谢工作人员~ ### 版本&环境信息 Version & Environment Information 在美国的linux服务器上进行试验 python -v结果 :3.7.3 64位python pip install paddlepaddle -i https://mirror.baidu.com/pypi/simple pip install paddlepaddle 等结果 报错为 ERROR: Could not find a version that satisfies the requirement paddlepaddle (from versions: none) ERROR: No matching
1.0
在美国的linux服务器上进行试验 python -v结果 :3.7.3 64位python pip install paddlepaddle -i https://mirror.baidu.com/pypi/simple pip install paddlepaddle 等结果 报错为 ERROR: Could not find a version that satisfies the requirement paddlepaddle (from versions: none) ERROR: No matching - ### 问题描述 Issue Description 在美国的linux服务器上进行试验 python -v结果 :3.7.3 64位python pip install paddlepaddle -i https://mirror.baidu.com/pypi/simple pip install paddlepaddle 等结果 报错为 ERROR: Could not find a version that satisfies the requirement paddlepaddle (from versions: none) ERROR: No matching distribution found for paddlepaddle 感谢工作人员~ ### 版本&环境信息 Version & Environment Information 在美国的linux服务器上进行试验 python -v结果 :3.7.3 64位python pip install paddlepaddle -i https://mirror.baidu.com/pypi/simple pip install paddlepaddle 等结果 报错为 ERROR: Could not find a version that satisfies the requirement paddlepaddle (from versions: none) ERROR: No matching
non_priority
在美国的linux服务器上进行试验 python v结果 : pip install paddlepaddle i pip install paddlepaddle 等结果 报错为 error could not find a version that satisfies the requirement paddlepaddle from versions none error no matching 问题描述 issue description 在美国的linux服务器上进行试验 python v结果 : pip install paddlepaddle i pip install paddlepaddle 等结果 报错为 error could not find a version that satisfies the requirement paddlepaddle from versions none error no matching distribution found for paddlepaddle 感谢工作人员 版本 环境信息 version environment information 在美国的linux服务器上进行试验 python v结果 : pip install paddlepaddle i pip install paddlepaddle 等结果 报错为 error could not find a version that satisfies the requirement paddlepaddle from versions none error no matching
0
124,901
12,241,467,788
IssuesEvent
2020-05-05 04:10:17
celery/celery
https://api.github.com/repos/celery/celery
opened
Is MongoDB suported as result backend?
Category: Documentation Issue Type: Bug Report
The [documentation](https://docs.celeryproject.org/en/stable/userguide/configuration.html#task-result-backend-settings) about result backend settings does not shows MongoDB as a suported option. The options available are: * rpc * database (SQLAlchemy) * redis * cache * cassandra * elasticsearch * ironcache * couchbase * arangodb * couchdb * cosmosdbsql (experimental) * filesystem * consul * azureblockblob * s3
1.0
Is MongoDB suported as result backend? - The [documentation](https://docs.celeryproject.org/en/stable/userguide/configuration.html#task-result-backend-settings) about result backend settings does not shows MongoDB as a suported option. The options available are: * rpc * database (SQLAlchemy) * redis * cache * cassandra * elasticsearch * ironcache * couchbase * arangodb * couchdb * cosmosdbsql (experimental) * filesystem * consul * azureblockblob * s3
non_priority
is mongodb suported as result backend the about result backend settings does not shows mongodb as a suported option the options available are rpc database sqlalchemy redis cache cassandra elasticsearch ironcache couchbase arangodb couchdb cosmosdbsql experimental filesystem consul azureblockblob
0
153,191
19,702,832,759
IssuesEvent
2022-01-12 18:23:41
harrinry/bazel
https://api.github.com/repos/harrinry/bazel
opened
CVE-2016-9843 (High) detected in zlibv1.2.8
security vulnerability
## CVE-2016-9843 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>zlibv1.2.8</b></p></summary> <p> <p>A massively spiffy yet delicately unobtrusive compression library.</p> <p>Library home page: <a href=https://github.com/cyanskies/zlib.git>https://github.com/cyanskies/zlib.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/harrinry/bazel/commit/c258fdc57bed0b623184b81d1c7127a55e698fdd">c258fdc57bed0b623184b81d1c7127a55e698fdd</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/crc32.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The crc32_big function in crc32.c in zlib 1.2.8 might allow context-dependent attackers to have unspecified impact via vectors involving big-endian CRC calculation. <p>Publish Date: 2017-05-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-9843>CVE-2016-9843</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-9843">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-9843</a></p> <p>Release Date: 2017-05-23</p> <p>Fix Resolution: v1.2.9</p> </p> </details> <p></p>
True
CVE-2016-9843 (High) detected in zlibv1.2.8 - ## CVE-2016-9843 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>zlibv1.2.8</b></p></summary> <p> <p>A massively spiffy yet delicately unobtrusive compression library.</p> <p>Library home page: <a href=https://github.com/cyanskies/zlib.git>https://github.com/cyanskies/zlib.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/harrinry/bazel/commit/c258fdc57bed0b623184b81d1c7127a55e698fdd">c258fdc57bed0b623184b81d1c7127a55e698fdd</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/crc32.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The crc32_big function in crc32.c in zlib 1.2.8 might allow context-dependent attackers to have unspecified impact via vectors involving big-endian CRC calculation. <p>Publish Date: 2017-05-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-9843>CVE-2016-9843</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-9843">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-9843</a></p> <p>Release Date: 2017-05-23</p> <p>Fix Resolution: v1.2.9</p> </p> </details> <p></p>
non_priority
cve high detected in cve high severity vulnerability vulnerable library a massively spiffy yet delicately unobtrusive compression library library home page a href found in head commit a href found in base branch master vulnerable source files c vulnerability details the big function in c in zlib might allow context dependent attackers to have unspecified impact via vectors involving big endian crc calculation publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
0
254,538
21,792,261,209
IssuesEvent
2022-05-15 04:27:41
milvus-io/milvus
https://api.github.com/repos/milvus-io/milvus
closed
[Bug]: [benchmark][standalone] RpcError in the first search: _MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded
kind/bug needs-triage stale test/benchmark
### Is there an existing issue for this? - [X] I have searched the existing issues ### Environment ```markdown - Milvus version: master-20220407-bb7a076 - Deployment mode(standalone or cluster): standalone - SDK version(e.g. pymilvus v2.0.0rc2): 2.1.0.dev18 - OS(Ubuntu or CentOS): - CPU/Memory: - GPU: - Others: ``` ### Current Behavior argo task: benchmark-cron-1649347200 ### scene 1 test yaml: client-configmap:client-acc-sift-ivf-flat,instanceId:1,server-configmap:server-single-8c16m client pod: benchmark-cron-1649347200-1665149048 client log: ``` [2022-04-07 16:06:47,745] [ DEBUG] - End insert, start flush (milvus_benchmark.runners.accuracy:256) [2022-04-07 16:06:50,776] [ DEBUG] - Milvus flush run in 3.0303s (milvus_benchmark.client:53) [2022-04-07 16:06:50,776] [ DEBUG] - End flush (milvus_benchmark.runners.accuracy:258) [2022-04-07 16:06:50,780] [ DEBUG] - Row count: 1000000 in collection: <sift_128_euclidean> (milvus_benchmark.client:424) [2022-04-07 16:06:50,781] [ INFO] - Table: sift_128_euclidean, row count: 1000000 (milvus_benchmark.runners.accuracy:260) [2022-04-07 16:06:50,784] [ INFO] - None (milvus_benchmark.client:293) [2022-04-07 16:06:50,785] [ INFO] - Drop index: sift_128_euclidean (milvus_benchmark.client:304) [2022-04-07 16:06:50,787] [ INFO] - Re-create index: sift_128_euclidean (milvus_benchmark.runners.accuracy:265) [2022-04-07 16:06:50,788] [ INFO] - Building index start, collection_name: sift_128_euclidean, index_type: IVF_FLAT, metric_type: L2 (milvus_benchmark.client:276) [2022-04-07 16:06:50,788] [ INFO] - {'nlist': 1024} (milvus_benchmark.client:278) [2022-04-07 16:06:50,789] [ DEBUG] - collection: sift_128_euclidean Index params: {'index_type': 'IVF_FLAT', 'metric_type': 'L2', 'params': {'nlist': 1024}} (milvus_benchmark.client:284) [2022-04-07 16:08:28,280] [ DEBUG] - Building index done, collection_name: sift_128_euclidean, response: Status(code=0, message='') (milvus_benchmark.client:286) [2022-04-07 16:08:28,282] [ DEBUG] - Milvus create_index run in 97.4946s (milvus_benchmark.client:53) [2022-04-07 16:08:28,286] [ INFO] - None (milvus_benchmark.client:293) [2022-04-07 16:08:28,288] [ INFO] - {'index_type': 'flat', 'metric_type': None, 'index_param': None} (milvus_benchmark.runners.accuracy:267) [2022-04-07 16:08:28,290] [ INFO] - Start load collection: sift_128_euclidean (milvus_benchmark.runners.accuracy:268) [2022-04-07 16:08:28,315] [ DEBUG] - Milvus load_collection run in 0.0226s (milvus_benchmark.client:53) [2022-04-07 16:08:28,315] [ INFO] - End load collection: sift_128_euclidean (milvus_benchmark.runners.accuracy:271) [2022-04-07 16:08:28,359] [ INFO] - Start run case (milvus_benchmark.main:88) [2022-04-07 16:13:28,604] [ ERROR] - grpc RpcError: [_execute_search_requests], <_MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded>, <Time:{'RPC start': '2022-04-07 16:08:28.602915', 'gRPC error': '2022-04-07 16:13:28.604619'}> (pymilvus.decorators:81) [2022-04-07 16:13:28,605] [ ERROR] - grpc RpcError: [search], <_MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded>, <Time:{'RPC start': '2022-04-07 16:08:28.363162', 'gRPC error': '2022-04-07 16:13:28.605807'}> (pymilvus.decorators:81) [2022-04-07 16:13:28,610] [ ERROR] - Traceback (most recent call last): File "main.py", line 95, in run_suite result = runner.run_case(case_metric, **case) File "/src/milvus_benchmark/runners/accuracy.py", line 282, in run_case guarantee_timestamp=case_param["guarantee_timestamp"]) File "/src/milvus_benchmark/client.py", line 49, in wrapper result = func(*args, **kwargs) File "/src/milvus_benchmark/client.py", line 324, in query result = self._milvus.search(tmp_collection_name, **params) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 835, in search partition_names, output_fields, timeout, round_decimal, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 49, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 41, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 21, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 458, in search return self._execute_search_requests(requests, timeout, **_kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 405, in _execute_search_requests raise pre_err File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 393, in _execute_search_requests response = ft.result() File "/usr/local/lib/python3.6/site-packages/grpc/_channel.py", line 744, in result raise self grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "{"created":"@1649348008.603559154","description":"Error received from peer ipv4:10.96.163.148:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}" > (milvus_benchmark.main:98) [2022-04-07 16:13:28,611] [ INFO] - None (milvus_benchmark.main:99) [2022-04-07 16:13:28,611] [ DEBUG] - {'type': 'ann_accuracy', 'value': {}} (milvus_benchmark.main:107) [2022-04-07 16:13:28,612] [ DEBUG] - {'_version': '0.1', '_type': 'case', 'run_id': 1649353300, 'mode': 'local', 'server': <milvus_benchmark.metrics.models.server.Server object at 0x7f7d6c645f98>, 'hardware': <milvus_benchmark.metrics.models.hardware.Hardware object at 0x7f7d6c645eb8>, 'env': <milvus_benchmark.metrics.models.env.Env object at 0x7f7d6c645dd8>, 'status': 'RUN_FAILED', 'err_message': '<_MultiThreadedRendezvous of RPC that terminated with:\n\tstatus = StatusCode.DEADLINE_EXCEEDED\n\tdetails = "Deadline Exceeded"\n\tdebug_error_string = "{"created":"@1649348008.603559154","description":"Error received from peer ipv4:10.96.163.148:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}"\n>\nTraceback (most recent call last):\n File "main.py", line 95, in run_suite\n result = runner.run_case(case_metric, **case)\n File "/src/milvus_benchmark/runners/accuracy.py", line 282, in run_case\n guarantee_timestamp=case_param["guarantee_timestamp"])\n File "/src/milvus_benchmark/client.py", line 49, in wrapper\n result = func(*args, **kwargs)\n File "/src/milvus_benchmark/client.py", line 324, in query\n result = self._milvus.search(tmp_collection_name, **params)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 835, in search\n partition_names, output_fields, timeout, round_decimal, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 49, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 41, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler\n return func(*args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 21, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 458, in search\n return self._execute_search_requests(requests, timeout, **_kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler\n return func(*args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 405, in _execute_search_requests\n raise pre_err\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 393, in _execute_search_requests\n response = ft.result()\n File "/usr/local/lib/python3.6/site-packages/grpc/_channel.py", line 744, in result\n raise self\ngrpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:\n\tstatus = StatusCode.DEADLINE_EXCEEDED\n\tdetails = "Deadline Exceeded"\n\tdebug_error_string = "{"created":"@1649348008.603559154","description":"Error received from peer ipv4:10.96.163.148:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}"\n>\n', 'collection': {'dimension': 128, 'metric_type': 'l2', 'dataset_name': 'sift_128_euclidean', 'shards_num': None}, 'index': {'index_type': 'ivf_flat', 'index_param': {'nlist': 1024}}, 'search': {'nq': 10000, 'topk': 10, 'search_param': {'nprobe': 1}, 'filter': [], 'guarantee_timestamp': None}, 'run_params': None, 'metrics': {'type': 'ann_accuracy', 'value': {}}, 'datetime': '2022-04-07 16:03:21.708543', 'type': 'metric'} (milvus_benchmark.metric.api:29) [2022-04-07 16:13:29,812] [ DEBUG] - Milvus query run in 1.1615s (milvus_benchmark.client:53) ``` server: ``` NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES benchmark-cron-1649347200-1-etcd-0 1/1 Running 0 3h57m 10.97.17.227 qa-node014.zilliz.local <none> <none> benchmark-cron-1649347200-1-milvus-standalone-5cb44fc7c7-vfzwm 1/1 Running 0 3h57m 10.97.20.232 qa-node018.zilliz.local <none> <none> benchmark-cron-1649347200-1-minio-76c9c4c444-dmqp5 1/1 Running 0 3h57m 10.97.12.19 qa-node015.zilliz.local <none> <none> ``` ### scene 2 test yaml: client-configmap:client-acc-sift-flat,instanceId:2,server-configmap:server-single-8c16m client pod: benchmark-cron-1649347200-543944498 client log: ``` [2022-04-07 16:06:20,390] [ DEBUG] - End flush (milvus_benchmark.runners.accuracy:258) [2022-04-07 16:06:20,394] [ DEBUG] - Row count: 1000000 in collection: <sift_128_euclidean> (milvus_benchmark.client:424) [2022-04-07 16:06:20,395] [ INFO] - Table: sift_128_euclidean, row count: 1000000 (milvus_benchmark.runners.accuracy:260) [2022-04-07 16:06:20,402] [ INFO] - None (milvus_benchmark.client:293) [2022-04-07 16:06:20,402] [ INFO] - Drop index: sift_128_euclidean (milvus_benchmark.client:304) [2022-04-07 16:06:20,405] [ INFO] - Re-create index: sift_128_euclidean (milvus_benchmark.runners.accuracy:265) [2022-04-07 16:06:20,406] [ INFO] - Building index start, collection_name: sift_128_euclidean, index_type: FLAT, metric_type: L2 (milvus_benchmark.client:276) [2022-04-07 16:06:20,406] [ INFO] - {'nlist': 1024} (milvus_benchmark.client:278) [2022-04-07 16:06:20,407] [ DEBUG] - collection: sift_128_euclidean Index params: {'index_type': 'FLAT', 'metric_type': 'L2', 'params': {'nlist': 1024}} (milvus_benchmark.client:284) [2022-04-07 16:06:20,411] [ DEBUG] - Building index done, collection_name: sift_128_euclidean, response: Status(code=0, message='Warning: It is not necessary to build index with index_type: FLAT') (milvus_benchmark.client:286) [2022-04-07 16:06:20,411] [ DEBUG] - Milvus create_index run in 0.0058s (milvus_benchmark.client:53) [2022-04-07 16:06:20,414] [ INFO] - None (milvus_benchmark.client:293) [2022-04-07 16:06:20,414] [ INFO] - {'index_type': 'flat', 'metric_type': None, 'index_param': None} (milvus_benchmark.runners.accuracy:267) [2022-04-07 16:06:20,415] [ INFO] - Start load collection: sift_128_euclidean (milvus_benchmark.runners.accuracy:268) [2022-04-07 16:06:20,431] [ DEBUG] - Milvus load_collection run in 0.0152s (milvus_benchmark.client:53) [2022-04-07 16:06:20,432] [ INFO] - End load collection: sift_128_euclidean (milvus_benchmark.runners.accuracy:271) [2022-04-07 16:06:20,481] [ INFO] - Start run case (milvus_benchmark.main:88) [2022-04-07 16:11:20,775] [ ERROR] - grpc RpcError: [_execute_search_requests], <_MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded>, <Time:{'RPC start': '2022-04-07 16:06:20.773926', 'gRPC error': '2022-04-07 16:11:20.775437'}> (pymilvus.decorators:81) [2022-04-07 16:11:20,778] [ ERROR] - grpc RpcError: [search], <_MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded>, <Time:{'RPC start': '2022-04-07 16:06:20.482406', 'gRPC error': '2022-04-07 16:11:20.778149'}> (pymilvus.decorators:81) [2022-04-07 16:11:20,784] [ ERROR] - Traceback (most recent call last): File "main.py", line 95, in run_suite result = runner.run_case(case_metric, **case) File "/src/milvus_benchmark/runners/accuracy.py", line 282, in run_case guarantee_timestamp=case_param["guarantee_timestamp"]) File "/src/milvus_benchmark/client.py", line 49, in wrapper result = func(*args, **kwargs) File "/src/milvus_benchmark/client.py", line 324, in query result = self._milvus.search(tmp_collection_name, **params) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 835, in search partition_names, output_fields, timeout, round_decimal, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 49, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 41, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 21, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 458, in search return self._execute_search_requests(requests, timeout, **_kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 405, in _execute_search_requests raise pre_err File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 393, in _execute_search_requests response = ft.result() File "/usr/local/lib/python3.6/site-packages/grpc/_channel.py", line 744, in result raise self grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "{"created":"@1649347880.774488333","description":"Error received from peer ipv4:10.96.124.71:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}" > (milvus_benchmark.main:98) [2022-04-07 16:11:20,786] [ INFO] - None (milvus_benchmark.main:99) [2022-04-07 16:11:20,788] [ DEBUG] - {'type': 'ann_accuracy', 'value': {}} (milvus_benchmark.main:107) [2022-04-07 16:11:20,791] [ DEBUG] - {'_version': '0.1', '_type': 'case', 'run_id': 1649348428, 'mode': 'local', 'server': <milvus_benchmark.metrics.models.server.Server object at 0x7f17f8867f98>, 'hardware': <milvus_benchmark.metrics.models.hardware.Hardware object at 0x7f17f8867eb8>, 'env': <milvus_benchmark.metrics.models.env.Env object at 0x7f17f8867dd8>, 'status': 'RUN_FAILED', 'err_message': '<_MultiThreadedRendezvous of RPC that terminated with:\n\tstatus = StatusCode.DEADLINE_EXCEEDED\n\tdetails = "Deadline Exceeded"\n\tdebug_error_string = "{"created":"@1649347880.774488333","description":"Error received from peer ipv4:10.96.124.71:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}"\n>\nTraceback (most recent call last):\n File "main.py", line 95, in run_suite\n result = runner.run_case(case_metric, **case)\n File "/src/milvus_benchmark/runners/accuracy.py", line 282, in run_case\n guarantee_timestamp=case_param["guarantee_timestamp"])\n File "/src/milvus_benchmark/client.py", line 49, in wrapper\n result = func(*args, **kwargs)\n File "/src/milvus_benchmark/client.py", line 324, in query\n result = self._milvus.search(tmp_collection_name, **params)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 835, in search\n partition_names, output_fields, timeout, round_decimal, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 49, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 41, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler\n return func(*args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 21, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 458, in search\n return self._execute_search_requests(requests, timeout, **_kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler\n return func(*args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 405, in _execute_search_requests\n raise pre_err\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 393, in _execute_search_requests\n response = ft.result()\n File "/usr/local/lib/python3.6/site-packages/grpc/_channel.py", line 744, in result\n raise self\ngrpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:\n\tstatus = StatusCode.DEADLINE_EXCEEDED\n\tdetails = "Deadline Exceeded"\n\tdebug_error_string = "{"created":"@1649347880.774488333","description":"Error received from peer ipv4:10.96.124.71:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}"\n>\n', 'collection': {'dimension': 128, 'metric_type': 'l2', 'dataset_name': 'sift_128_euclidean', 'shards_num': None}, 'index': {'index_type': 'flat', 'index_param': {'nlist': 1024}}, 'search': {'nq': 10000, 'topk': 10, 'search_param': {'nprobe': 1}, 'filter': [], 'guarantee_timestamp': None}, 'run_params': None, 'metrics': {'type': 'ann_accuracy', 'value': {}}, 'datetime': '2022-04-07 16:03:27.458830', 'type': 'metric'} (milvus_benchmark.metric.api:29) ``` server: ``` NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES benchmark-cron-1649347200-2-etcd-0 1/1 Running 0 3h57m 10.97.17.222 qa-node014.zilliz.local <none> <none> benchmark-cron-1649347200-2-milvus-standalone-7ddfc7fbfc-ndmfd 1/1 Running 0 3h57m 10.97.10.18 qa-node008.zilliz.local <none> <none> benchmark-cron-1649347200-2-minio-f9dccfb9c-b5fkw 1/1 Running 0 3h57m 10.97.16.67 qa-node013.zilliz.local <none> <none> ``` ### scene 3 test yaml: client-configmap:client-acc-sift-ivf-sq8,instanceId:3,server-configmap:server-single-8c16m client pod: benchmark-cron-1649347200-2815052319 client log: ``` [2022-04-07 16:05:19,601] [ DEBUG] - Row count: 1000000 in collection: <sift_128_euclidean> (milvus_benchmark.client:424) [2022-04-07 16:05:19,655] [ INFO] - Table: sift_128_euclidean, row count: 1000000 (milvus_benchmark.runners.accuracy:260) [2022-04-07 16:05:19,659] [ INFO] - None (milvus_benchmark.client:293) [2022-04-07 16:05:19,661] [ INFO] - Drop index: sift_128_euclidean (milvus_benchmark.client:304) [2022-04-07 16:05:19,664] [ INFO] - Re-create index: sift_128_euclidean (milvus_benchmark.runners.accuracy:265) [2022-04-07 16:05:19,751] [ INFO] - Building index start, collection_name: sift_128_euclidean, index_type: IVF_SQ8, metric_type: L2 (milvus_benchmark.client:276) [2022-04-07 16:05:19,752] [ INFO] - {'nlist': 1024} (milvus_benchmark.client:278) [2022-04-07 16:05:19,753] [ DEBUG] - collection: sift_128_euclidean Index params: {'index_type': 'IVF_SQ8', 'metric_type': 'L2', 'params': {'nlist': 1024}} (milvus_benchmark.client:284) [2022-04-07 16:06:07,539] [ DEBUG] - Building index done, collection_name: sift_128_euclidean, response: Status(code=0, message='') (milvus_benchmark.client:286) [2022-04-07 16:06:07,549] [ DEBUG] - Milvus create_index run in 47.7976s (milvus_benchmark.client:53) [2022-04-07 16:06:07,551] [ INFO] - None (milvus_benchmark.client:293) [2022-04-07 16:06:07,552] [ INFO] - {'index_type': 'flat', 'metric_type': None, 'index_param': None} (milvus_benchmark.runners.accuracy:267) [2022-04-07 16:06:07,552] [ INFO] - Start load collection: sift_128_euclidean (milvus_benchmark.runners.accuracy:268) [2022-04-07 16:06:07,571] [ DEBUG] - Milvus load_collection run in 0.0178s (milvus_benchmark.client:53) [2022-04-07 16:06:07,572] [ INFO] - End load collection: sift_128_euclidean (milvus_benchmark.runners.accuracy:271) [2022-04-07 16:06:07,617] [ INFO] - Start run case (milvus_benchmark.main:88) [2022-04-07 16:11:07,855] [ ERROR] - grpc RpcError: [_execute_search_requests], <_MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded>, <Time:{'RPC start': '2022-04-07 16:06:07.853407', 'gRPC error': '2022-04-07 16:11:07.855146'}> (pymilvus.decorators:81) [2022-04-07 16:11:07,856] [ ERROR] - grpc RpcError: [search], <_MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded>, <Time:{'RPC start': '2022-04-07 16:06:07.618333', 'gRPC error': '2022-04-07 16:11:07.856616'}> (pymilvus.decorators:81) [2022-04-07 16:11:07,861] [ ERROR] - Traceback (most recent call last): File "main.py", line 95, in run_suite result = runner.run_case(case_metric, **case) File "/src/milvus_benchmark/runners/accuracy.py", line 282, in run_case guarantee_timestamp=case_param["guarantee_timestamp"]) File "/src/milvus_benchmark/client.py", line 49, in wrapper result = func(*args, **kwargs) File "/src/milvus_benchmark/client.py", line 324, in query result = self._milvus.search(tmp_collection_name, **params) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 835, in search partition_names, output_fields, timeout, round_decimal, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 49, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 41, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 21, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 458, in search return self._execute_search_requests(requests, timeout, **_kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 405, in _execute_search_requests raise pre_err File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 393, in _execute_search_requests response = ft.result() File "/usr/local/lib/python3.6/site-packages/grpc/_channel.py", line 744, in result raise self grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "{"created":"@1649347867.853764926","description":"Error received from peer ipv4:10.96.240.114:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}" > (milvus_benchmark.main:98) [2022-04-07 16:11:07,862] [ INFO] - None (milvus_benchmark.main:99) [2022-04-07 16:11:07,862] [ DEBUG] - {'type': 'ann_accuracy', 'value': {}} (milvus_benchmark.main:107) [2022-04-07 16:11:07,863] [ DEBUG] - {'_version': '0.1', '_type': 'case', 'run_id': 1649355398, 'mode': 'local', 'server': <milvus_benchmark.metrics.models.server.Server object at 0x7fd3bc52ff98>, 'hardware': <milvus_benchmark.metrics.models.hardware.Hardware object at 0x7fd3bc52feb8>, 'env': <milvus_benchmark.metrics.models.env.Env object at 0x7fd3bc52fdd8>, 'status': 'RUN_FAILED', 'err_message': '<_MultiThreadedRendezvous of RPC that terminated with:\n\tstatus = StatusCode.DEADLINE_EXCEEDED\n\tdetails = "Deadline Exceeded"\n\tdebug_error_string = "{"created":"@1649347867.853764926","description":"Error received from peer ipv4:10.96.240.114:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}"\n>\nTraceback (most recent call last):\n File "main.py", line 95, in run_suite\n result = runner.run_case(case_metric, **case)\n File "/src/milvus_benchmark/runners/accuracy.py", line 282, in run_case\n guarantee_timestamp=case_param["guarantee_timestamp"])\n File "/src/milvus_benchmark/client.py", line 49, in wrapper\n result = func(*args, **kwargs)\n File "/src/milvus_benchmark/client.py", line 324, in query\n result = self._milvus.search(tmp_collection_name, **params)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 835, in search\n partition_names, output_fields, timeout, round_decimal, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 49, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 41, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler\n return func(*args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 21, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 458, in search\n return self._execute_search_requests(requests, timeout, **_kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler\n return func(*args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 405, in _execute_search_requests\n raise pre_err\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 393, in _execute_search_requests\n response = ft.result()\n File "/usr/local/lib/python3.6/site-packages/grpc/_channel.py", line 744, in result\n raise self\ngrpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:\n\tstatus = StatusCode.DEADLINE_EXCEEDED\n\tdetails = "Deadline Exceeded"\n\tdebug_error_string = "{"created":"@1649347867.853764926","description":"Error received from peer ipv4:10.96.240.114:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}"\n>\n', 'collection': {'dimension': 128, 'metric_type': 'l2', 'dataset_name': 'sift_128_euclidean', 'shards_num': None}, 'index': {'index_type': 'ivf_sq8', 'index_param': {'nlist': 1024}}, 'search': {'nq': 10000, 'topk': 10, 'search_param': {'nprobe': 1}, 'filter': [], 'guarantee_timestamp': None}, 'run_params': None, 'metrics': {'type': 'ann_accuracy', 'value': {}}, 'datetime': '2022-04-07 16:02:58.136148', 'type': 'metric'} (milvus_benchmark.metric.api:29) [2022-04-07 16:11:08,791] [ DEBUG] - Milvus query run in 0.8903s (milvus_benchmark.client:53) ``` server: ``` NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES benchmark-cron-1649347200-3-etcd-0 1/1 Running 0 3h57m 10.97.16.69 qa-node013.zilliz.local <none> <none> benchmark-cron-1649347200-3-milvus-standalone-5bdd8bb54f-8h4sh 1/1 Running 0 3h57m 10.97.17.208 qa-node014.zilliz.local <none> <none> benchmark-cron-1649347200-3-minio-6d59c84445-4x2nt 1/1 Running 0 3h57m 10.97.12.13 qa-node015.zilliz.local <none> <none> ``` ### scene 4 test yaml: client-configmap:client-acc-sift-hnsw,instanceId:4,server-configmap:server-single-8c16m client pod: benchmark-cron-1649347200-1294725969 client log: ``` [2022-04-07 16:04:54,864] [ DEBUG] - Row count: 1000000 in collection: <sift_128_euclidean> (milvus_benchmark.client:424) [2022-04-07 16:04:54,864] [ INFO] - Table: sift_128_euclidean, row count: 1000000 (milvus_benchmark.runners.accuracy:260) [2022-04-07 16:04:54,866] [ INFO] - None (milvus_benchmark.client:293) [2022-04-07 16:04:54,866] [ INFO] - Drop index: sift_128_euclidean (milvus_benchmark.client:304) [2022-04-07 16:04:54,868] [ INFO] - Re-create index: sift_128_euclidean (milvus_benchmark.runners.accuracy:265) [2022-04-07 16:04:54,869] [ INFO] - Building index start, collection_name: sift_128_euclidean, index_type: HNSW, metric_type: L2 (milvus_benchmark.client:276) [2022-04-07 16:04:54,869] [ INFO] - {'M': 16, 'efConstruction': 500} (milvus_benchmark.client:278) [2022-04-07 16:04:54,870] [ DEBUG] - collection: sift_128_euclidean Index params: {'index_type': 'HNSW', 'metric_type': 'L2', 'params': {'M': 16, 'efConstruction': 500}} (milvus_benchmark.client:284) [2022-04-07 16:09:18,805] [ DEBUG] - Building index done, collection_name: sift_128_euclidean, response: Status(code=0, message='') (milvus_benchmark.client:286) [2022-04-07 16:09:18,806] [ DEBUG] - Milvus create_index run in 263.937s (milvus_benchmark.client:53) [2022-04-07 16:09:18,808] [ INFO] - None (milvus_benchmark.client:293) [2022-04-07 16:09:18,808] [ INFO] - {'index_type': 'flat', 'metric_type': None, 'index_param': None} (milvus_benchmark.runners.accuracy:267) [2022-04-07 16:09:18,809] [ INFO] - Start load collection: sift_128_euclidean (milvus_benchmark.runners.accuracy:268) [2022-04-07 16:09:18,826] [ DEBUG] - Milvus load_collection run in 0.0176s (milvus_benchmark.client:53) [2022-04-07 16:09:18,827] [ INFO] - End load collection: sift_128_euclidean (milvus_benchmark.runners.accuracy:271) [2022-04-07 16:09:18,829] [ INFO] - Start run case (milvus_benchmark.main:88) [2022-04-07 16:14:19,007] [ ERROR] - grpc RpcError: [_execute_search_requests], <_MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded>, <Time:{'RPC start': '2022-04-07 16:09:19.005930', 'gRPC error': '2022-04-07 16:14:19.007620'}> (pymilvus.decorators:81) [2022-04-07 16:14:19,008] [ ERROR] - grpc RpcError: [search], <_MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded>, <Time:{'RPC start': '2022-04-07 16:09:18.829597', 'gRPC error': '2022-04-07 16:14:19.008847'}> (pymilvus.decorators:81) [2022-04-07 16:14:19,011] [ ERROR] - Traceback (most recent call last): File "main.py", line 95, in run_suite result = runner.run_case(case_metric, **case) File "/src/milvus_benchmark/runners/accuracy.py", line 282, in run_case guarantee_timestamp=case_param["guarantee_timestamp"]) File "/src/milvus_benchmark/client.py", line 49, in wrapper result = func(*args, **kwargs) File "/src/milvus_benchmark/client.py", line 324, in query result = self._milvus.search(tmp_collection_name, **params) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 835, in search partition_names, output_fields, timeout, round_decimal, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 49, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 41, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 21, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 458, in search return self._execute_search_requests(requests, timeout, **_kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 405, in _execute_search_requests raise pre_err File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 393, in _execute_search_requests response = ft.result() File "/usr/local/lib/python3.6/site-packages/grpc/_channel.py", line 744, in result raise self grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "{"created":"@1649348059.006772396","description":"Error received from peer ipv4:10.96.221.3:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}" > (milvus_benchmark.main:98) [2022-04-07 16:14:19,012] [ INFO] - None (milvus_benchmark.main:99) [2022-04-07 16:14:19,012] [ DEBUG] - {'type': 'ann_accuracy', 'value': {}} (milvus_benchmark.main:107) [2022-04-07 16:14:19,012] [ DEBUG] - {'_version': '0.1', '_type': 'case', 'run_id': 1649348560, 'mode': 'local', 'server': <milvus_benchmark.metrics.models.server.Server object at 0x7f2b4ca0efd0>, 'hardware': <milvus_benchmark.metrics.models.hardware.Hardware object at 0x7f2b4ca0eef0>, 'env': <milvus_benchmark.metrics.models.env.Env object at 0x7f2b4ca0ee10>, 'status': 'RUN_FAILED', 'err_message': '<_MultiThreadedRendezvous of RPC that terminated with:\n\tstatus = StatusCode.DEADLINE_EXCEEDED\n\tdetails = "Deadline Exceeded"\n\tdebug_error_string = "{"created":"@1649348059.006772396","description":"Error received from peer ipv4:10.96.221.3:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}"\n>\nTraceback (most recent call last):\n File "main.py", line 95, in run_suite\n result = runner.run_case(case_metric, **case)\n File "/src/milvus_benchmark/runners/accuracy.py", line 282, in run_case\n guarantee_timestamp=case_param["guarantee_timestamp"])\n File "/src/milvus_benchmark/client.py", line 49, in wrapper\n result = func(*args, **kwargs)\n File "/src/milvus_benchmark/client.py", line 324, in query\n result = self._milvus.search(tmp_collection_name, **params)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 835, in search\n partition_names, output_fields, timeout, round_decimal, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 49, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 41, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler\n return func(*args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 21, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 458, in search\n return self._execute_search_requests(requests, timeout, **_kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler\n return func(*args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 405, in _execute_search_requests\n raise pre_err\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 393, in _execute_search_requests\n response = ft.result()\n File "/usr/local/lib/python3.6/site-packages/grpc/_channel.py", line 744, in result\n raise self\ngrpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:\n\tstatus = StatusCode.DEADLINE_EXCEEDED\n\tdetails = "Deadline Exceeded"\n\tdebug_error_string = "{"created":"@1649348059.006772396","description":"Error received from peer ipv4:10.96.221.3:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}"\n>\n', 'collection': {'dimension': 128, 'metric_type': 'l2', 'dataset_name': 'sift_128_euclidean', 'shards_num': None}, 'index': {'index_type': 'hnsw', 'index_param': {'M': 16, 'efConstruction': 500}}, 'search': {'nq': 10000, 'topk': 10, 'search_param': {'ef': 16}, 'filter': [], 'guarantee_timestamp': None}, 'run_params': None, 'metrics': {'type': 'ann_accuracy', 'value': {}}, 'datetime': '2022-04-07 16:02:45.863579', 'type': 'metric'} (milvus_benchmark.metric.api:29) [2022-04-07 16:14:19,874] [ DEBUG] - Milvus query run in 0.8345s (milvus_benchmark.client:53) ``` server: ``` NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES benchmark-cron-1649347200-4-etcd-0 1/1 Running 0 3h57m 10.97.17.214 qa-node014.zilliz.local <none> <none> benchmark-cron-1649347200-4-milvus-standalone-65559bf9c4-v65m5 1/1 Running 0 3h57m 10.97.19.74 qa-node016.zilliz.local <none> <none> benchmark-cron-1649347200-4-minio-9f49dbd6f-22ppq 1/1 Running 0 3h57m 10.97.19.72 qa-node016.zilliz.local <none> <none> ``` ### scene 5 test yaml: client-configmap:client-search-sift50m-ivf-flat-2048,instanceId:6,server-configmap:server-single-16c64m client pod: benchmark-cron-1649347200-4136737022 client log: ``` [2022-04-07 17:28:25,859] [ DEBUG] - Row count: 50000000 in collection: <sift_50m_128_l2> (milvus_benchmark.client:424) [2022-04-07 17:28:25,860] [ INFO] - 50000000 (milvus_benchmark.runners.search:280) [2022-04-07 17:28:25,860] [ INFO] - Start load collection (milvus_benchmark.runners.search:281) [2022-04-07 17:28:25,984] [ DEBUG] - Milvus load_collection run in 0.1234s (milvus_benchmark.client:53) [2022-04-07 17:28:25,985] [ DEBUG] - {'load_time': 0.1242} (milvus_benchmark.runners.search:284) [2022-04-07 17:28:25,986] [ INFO] - Start run case (milvus_benchmark.main:88) [2022-04-07 17:28:25,986] [ DEBUG] - Start run query, run 1 of 30 (milvus_benchmark.runners.search:294) [2022-04-07 17:28:25,987] [ INFO] - {'nq': 1, 'topk': 1, 'search_param': {'nprobe': 8}, 'filter': [], 'guarantee_timestamp': None} (milvus_benchmark.runners.search:295) [2022-04-07 17:33:25,996] [ ERROR] - grpc RpcError: [_execute_search_requests], <_MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded>, <Time:{'RPC start': '2022-04-07 17:28:25.994302', 'gRPC error': '2022-04-07 17:33:25.995836'}> (pymilvus.decorators:81) [2022-04-07 17:33:26,002] [ ERROR] - grpc RpcError: [search], <_MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded>, <Time:{'RPC start': '2022-04-07 17:28:25.988694', 'gRPC error': '2022-04-07 17:33:26.002716'}> (pymilvus.decorators:81) [2022-04-07 17:33:26,009] [ ERROR] - Traceback (most recent call last): File "main.py", line 95, in run_suite result = runner.run_case(case_metric, **case) File "/src/milvus_benchmark/runners/search.py", line 298, in run_case guarantee_timestamp=case_param["guarantee_timestamp"]) File "/src/milvus_benchmark/client.py", line 49, in wrapper result = func(*args, **kwargs) File "/src/milvus_benchmark/client.py", line 324, in query result = self._milvus.search(tmp_collection_name, **params) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 835, in search partition_names, output_fields, timeout, round_decimal, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 49, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 41, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 21, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 458, in search return self._execute_search_requests(requests, timeout, **_kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 405, in _execute_search_requests raise pre_err File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 393, in _execute_search_requests response = ft.result() File "/usr/local/lib/python3.6/site-packages/grpc/_channel.py", line 744, in result raise self grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "{"created":"@1649352805.994898509","description":"Error received from peer ipv4:10.96.105.127:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}" > (milvus_benchmark.main:98) [2022-04-07 17:33:26,011] [ INFO] - None (milvus_benchmark.main:99) [2022-04-07 17:33:26,012] [ DEBUG] - {'type': 'insert_search_performance', 'value': {}} (milvus_benchmark.main:107) [2022-04-07 17:33:26,012] [ DEBUG] - {'_version': '0.1', '_type': 'case', 'run_id': 1649352312, 'mode': 'local', 'server': <milvus_benchmark.metrics.models.server.Server object at 0x7efdbc2c22e8>, 'hardware': <milvus_benchmark.metrics.models.hardware.Hardware object at 0x7efdbc2c26a0>, 'env': <milvus_benchmark.metrics.models.env.Env object at 0x7efdbc2c2940>, 'status': 'RUN_FAILED', 'err_message': '<_MultiThreadedRendezvous of RPC that terminated with:\n\tstatus = StatusCode.DEADLINE_EXCEEDED\n\tdetails = "Deadline Exceeded"\n\tdebug_error_string = "{"created":"@1649352805.994898509","description":"Error received from peer ipv4:10.96.105.127:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}"\n>\nTraceback (most recent call last):\n File "main.py", line 95, in run_suite\n result = runner.run_case(case_metric, **case)\n File "/src/milvus_benchmark/runners/search.py", line 298, in run_case\n guarantee_timestamp=case_param["guarantee_timestamp"])\n File "/src/milvus_benchmark/client.py", line 49, in wrapper\n result = func(*args, **kwargs)\n File "/src/milvus_benchmark/client.py", line 324, in query\n result = self._milvus.search(tmp_collection_name, **params)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 835, in search\n partition_names, output_fields, timeout, round_decimal, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 49, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 41, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler\n return func(*args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 21, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 458, in search\n return self._execute_search_requests(requests, timeout, **_kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler\n return func(*args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 405, in _execute_search_requests\n raise pre_err\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 393, in _execute_search_requests\n response = ft.result()\n File "/usr/local/lib/python3.6/site-packages/grpc/_channel.py", line 744, in result\n raise self\ngrpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:\n\tstatus = StatusCode.DEADLINE_EXCEEDED\n\tdetails = "Deadline Exceeded"\n\tdebug_error_string = "{"created":"@1649352805.994898509","description":"Error received from peer ipv4:10.96.105.127:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}"\n>\n', 'collection': {'dimension': 128, 'metric_type': 'l2', 'dataset_name': 'sift_50m_128_l2', 'fields': None, 'shards_num': None}, 'index': {'index_type': 'ivf_flat', 'index_param': {'nlist': 2048}}, 'search': {'nq': 1, 'topk': 1, 'search_param': {'nprobe': 8}, 'filter': [], 'guarantee_timestamp': None}, 'run_params': None, 'metrics': {'type': 'insert_search_performance', 'value': {}}, 'datetime': '2022-04-07 16:02:57.589034', 'type': 'metric'} (milvus_benchmark.metric.api:29) ``` server: ``` NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES benchmark-cron-1649347200-6-etcd-0 1/1 Running 0 3h57m 10.97.17.213 qa-node014.zilliz.local <none> <none> benchmark-cron-1649347200-6-milvus-standalone-76ddb48b6-2dh8b 1/1 Running 0 3h57m 10.97.20.230 qa-node018.zilliz.local <none> <none> benchmark-cron-1649347200-6-minio-6d7c6d494c-d8hpm 1/1 Running 0 3h57m 10.97.19.73 qa-node016.zilliz.local <none> <none> ``` ### Expected Behavior _No response_ ### Steps To Reproduce ```markdown scene 1、2、3、4: template: ann_accuracy steps: 1、create collection 2、insert vectors 3、flush collection 4、build index 5、load collection 6、search 《- raise error scene 15: template: insert_search_performance steps: 1、create collection 2、build index 3、insert vectors 4、flush collection 5、build index again 6、load collection 7、search 《- raise error ``` ### Anything else? _No response_
1.0
[Bug]: [benchmark][standalone] RpcError in the first search: _MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Environment ```markdown - Milvus version: master-20220407-bb7a076 - Deployment mode(standalone or cluster): standalone - SDK version(e.g. pymilvus v2.0.0rc2): 2.1.0.dev18 - OS(Ubuntu or CentOS): - CPU/Memory: - GPU: - Others: ``` ### Current Behavior argo task: benchmark-cron-1649347200 ### scene 1 test yaml: client-configmap:client-acc-sift-ivf-flat,instanceId:1,server-configmap:server-single-8c16m client pod: benchmark-cron-1649347200-1665149048 client log: ``` [2022-04-07 16:06:47,745] [ DEBUG] - End insert, start flush (milvus_benchmark.runners.accuracy:256) [2022-04-07 16:06:50,776] [ DEBUG] - Milvus flush run in 3.0303s (milvus_benchmark.client:53) [2022-04-07 16:06:50,776] [ DEBUG] - End flush (milvus_benchmark.runners.accuracy:258) [2022-04-07 16:06:50,780] [ DEBUG] - Row count: 1000000 in collection: <sift_128_euclidean> (milvus_benchmark.client:424) [2022-04-07 16:06:50,781] [ INFO] - Table: sift_128_euclidean, row count: 1000000 (milvus_benchmark.runners.accuracy:260) [2022-04-07 16:06:50,784] [ INFO] - None (milvus_benchmark.client:293) [2022-04-07 16:06:50,785] [ INFO] - Drop index: sift_128_euclidean (milvus_benchmark.client:304) [2022-04-07 16:06:50,787] [ INFO] - Re-create index: sift_128_euclidean (milvus_benchmark.runners.accuracy:265) [2022-04-07 16:06:50,788] [ INFO] - Building index start, collection_name: sift_128_euclidean, index_type: IVF_FLAT, metric_type: L2 (milvus_benchmark.client:276) [2022-04-07 16:06:50,788] [ INFO] - {'nlist': 1024} (milvus_benchmark.client:278) [2022-04-07 16:06:50,789] [ DEBUG] - collection: sift_128_euclidean Index params: {'index_type': 'IVF_FLAT', 'metric_type': 'L2', 'params': {'nlist': 1024}} (milvus_benchmark.client:284) [2022-04-07 16:08:28,280] [ DEBUG] - Building index done, collection_name: sift_128_euclidean, response: Status(code=0, message='') (milvus_benchmark.client:286) [2022-04-07 16:08:28,282] [ DEBUG] - Milvus create_index run in 97.4946s (milvus_benchmark.client:53) [2022-04-07 16:08:28,286] [ INFO] - None (milvus_benchmark.client:293) [2022-04-07 16:08:28,288] [ INFO] - {'index_type': 'flat', 'metric_type': None, 'index_param': None} (milvus_benchmark.runners.accuracy:267) [2022-04-07 16:08:28,290] [ INFO] - Start load collection: sift_128_euclidean (milvus_benchmark.runners.accuracy:268) [2022-04-07 16:08:28,315] [ DEBUG] - Milvus load_collection run in 0.0226s (milvus_benchmark.client:53) [2022-04-07 16:08:28,315] [ INFO] - End load collection: sift_128_euclidean (milvus_benchmark.runners.accuracy:271) [2022-04-07 16:08:28,359] [ INFO] - Start run case (milvus_benchmark.main:88) [2022-04-07 16:13:28,604] [ ERROR] - grpc RpcError: [_execute_search_requests], <_MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded>, <Time:{'RPC start': '2022-04-07 16:08:28.602915', 'gRPC error': '2022-04-07 16:13:28.604619'}> (pymilvus.decorators:81) [2022-04-07 16:13:28,605] [ ERROR] - grpc RpcError: [search], <_MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded>, <Time:{'RPC start': '2022-04-07 16:08:28.363162', 'gRPC error': '2022-04-07 16:13:28.605807'}> (pymilvus.decorators:81) [2022-04-07 16:13:28,610] [ ERROR] - Traceback (most recent call last): File "main.py", line 95, in run_suite result = runner.run_case(case_metric, **case) File "/src/milvus_benchmark/runners/accuracy.py", line 282, in run_case guarantee_timestamp=case_param["guarantee_timestamp"]) File "/src/milvus_benchmark/client.py", line 49, in wrapper result = func(*args, **kwargs) File "/src/milvus_benchmark/client.py", line 324, in query result = self._milvus.search(tmp_collection_name, **params) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 835, in search partition_names, output_fields, timeout, round_decimal, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 49, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 41, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 21, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 458, in search return self._execute_search_requests(requests, timeout, **_kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 405, in _execute_search_requests raise pre_err File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 393, in _execute_search_requests response = ft.result() File "/usr/local/lib/python3.6/site-packages/grpc/_channel.py", line 744, in result raise self grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "{"created":"@1649348008.603559154","description":"Error received from peer ipv4:10.96.163.148:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}" > (milvus_benchmark.main:98) [2022-04-07 16:13:28,611] [ INFO] - None (milvus_benchmark.main:99) [2022-04-07 16:13:28,611] [ DEBUG] - {'type': 'ann_accuracy', 'value': {}} (milvus_benchmark.main:107) [2022-04-07 16:13:28,612] [ DEBUG] - {'_version': '0.1', '_type': 'case', 'run_id': 1649353300, 'mode': 'local', 'server': <milvus_benchmark.metrics.models.server.Server object at 0x7f7d6c645f98>, 'hardware': <milvus_benchmark.metrics.models.hardware.Hardware object at 0x7f7d6c645eb8>, 'env': <milvus_benchmark.metrics.models.env.Env object at 0x7f7d6c645dd8>, 'status': 'RUN_FAILED', 'err_message': '<_MultiThreadedRendezvous of RPC that terminated with:\n\tstatus = StatusCode.DEADLINE_EXCEEDED\n\tdetails = "Deadline Exceeded"\n\tdebug_error_string = "{"created":"@1649348008.603559154","description":"Error received from peer ipv4:10.96.163.148:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}"\n>\nTraceback (most recent call last):\n File "main.py", line 95, in run_suite\n result = runner.run_case(case_metric, **case)\n File "/src/milvus_benchmark/runners/accuracy.py", line 282, in run_case\n guarantee_timestamp=case_param["guarantee_timestamp"])\n File "/src/milvus_benchmark/client.py", line 49, in wrapper\n result = func(*args, **kwargs)\n File "/src/milvus_benchmark/client.py", line 324, in query\n result = self._milvus.search(tmp_collection_name, **params)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 835, in search\n partition_names, output_fields, timeout, round_decimal, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 49, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 41, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler\n return func(*args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 21, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 458, in search\n return self._execute_search_requests(requests, timeout, **_kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler\n return func(*args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 405, in _execute_search_requests\n raise pre_err\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 393, in _execute_search_requests\n response = ft.result()\n File "/usr/local/lib/python3.6/site-packages/grpc/_channel.py", line 744, in result\n raise self\ngrpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:\n\tstatus = StatusCode.DEADLINE_EXCEEDED\n\tdetails = "Deadline Exceeded"\n\tdebug_error_string = "{"created":"@1649348008.603559154","description":"Error received from peer ipv4:10.96.163.148:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}"\n>\n', 'collection': {'dimension': 128, 'metric_type': 'l2', 'dataset_name': 'sift_128_euclidean', 'shards_num': None}, 'index': {'index_type': 'ivf_flat', 'index_param': {'nlist': 1024}}, 'search': {'nq': 10000, 'topk': 10, 'search_param': {'nprobe': 1}, 'filter': [], 'guarantee_timestamp': None}, 'run_params': None, 'metrics': {'type': 'ann_accuracy', 'value': {}}, 'datetime': '2022-04-07 16:03:21.708543', 'type': 'metric'} (milvus_benchmark.metric.api:29) [2022-04-07 16:13:29,812] [ DEBUG] - Milvus query run in 1.1615s (milvus_benchmark.client:53) ``` server: ``` NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES benchmark-cron-1649347200-1-etcd-0 1/1 Running 0 3h57m 10.97.17.227 qa-node014.zilliz.local <none> <none> benchmark-cron-1649347200-1-milvus-standalone-5cb44fc7c7-vfzwm 1/1 Running 0 3h57m 10.97.20.232 qa-node018.zilliz.local <none> <none> benchmark-cron-1649347200-1-minio-76c9c4c444-dmqp5 1/1 Running 0 3h57m 10.97.12.19 qa-node015.zilliz.local <none> <none> ``` ### scene 2 test yaml: client-configmap:client-acc-sift-flat,instanceId:2,server-configmap:server-single-8c16m client pod: benchmark-cron-1649347200-543944498 client log: ``` [2022-04-07 16:06:20,390] [ DEBUG] - End flush (milvus_benchmark.runners.accuracy:258) [2022-04-07 16:06:20,394] [ DEBUG] - Row count: 1000000 in collection: <sift_128_euclidean> (milvus_benchmark.client:424) [2022-04-07 16:06:20,395] [ INFO] - Table: sift_128_euclidean, row count: 1000000 (milvus_benchmark.runners.accuracy:260) [2022-04-07 16:06:20,402] [ INFO] - None (milvus_benchmark.client:293) [2022-04-07 16:06:20,402] [ INFO] - Drop index: sift_128_euclidean (milvus_benchmark.client:304) [2022-04-07 16:06:20,405] [ INFO] - Re-create index: sift_128_euclidean (milvus_benchmark.runners.accuracy:265) [2022-04-07 16:06:20,406] [ INFO] - Building index start, collection_name: sift_128_euclidean, index_type: FLAT, metric_type: L2 (milvus_benchmark.client:276) [2022-04-07 16:06:20,406] [ INFO] - {'nlist': 1024} (milvus_benchmark.client:278) [2022-04-07 16:06:20,407] [ DEBUG] - collection: sift_128_euclidean Index params: {'index_type': 'FLAT', 'metric_type': 'L2', 'params': {'nlist': 1024}} (milvus_benchmark.client:284) [2022-04-07 16:06:20,411] [ DEBUG] - Building index done, collection_name: sift_128_euclidean, response: Status(code=0, message='Warning: It is not necessary to build index with index_type: FLAT') (milvus_benchmark.client:286) [2022-04-07 16:06:20,411] [ DEBUG] - Milvus create_index run in 0.0058s (milvus_benchmark.client:53) [2022-04-07 16:06:20,414] [ INFO] - None (milvus_benchmark.client:293) [2022-04-07 16:06:20,414] [ INFO] - {'index_type': 'flat', 'metric_type': None, 'index_param': None} (milvus_benchmark.runners.accuracy:267) [2022-04-07 16:06:20,415] [ INFO] - Start load collection: sift_128_euclidean (milvus_benchmark.runners.accuracy:268) [2022-04-07 16:06:20,431] [ DEBUG] - Milvus load_collection run in 0.0152s (milvus_benchmark.client:53) [2022-04-07 16:06:20,432] [ INFO] - End load collection: sift_128_euclidean (milvus_benchmark.runners.accuracy:271) [2022-04-07 16:06:20,481] [ INFO] - Start run case (milvus_benchmark.main:88) [2022-04-07 16:11:20,775] [ ERROR] - grpc RpcError: [_execute_search_requests], <_MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded>, <Time:{'RPC start': '2022-04-07 16:06:20.773926', 'gRPC error': '2022-04-07 16:11:20.775437'}> (pymilvus.decorators:81) [2022-04-07 16:11:20,778] [ ERROR] - grpc RpcError: [search], <_MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded>, <Time:{'RPC start': '2022-04-07 16:06:20.482406', 'gRPC error': '2022-04-07 16:11:20.778149'}> (pymilvus.decorators:81) [2022-04-07 16:11:20,784] [ ERROR] - Traceback (most recent call last): File "main.py", line 95, in run_suite result = runner.run_case(case_metric, **case) File "/src/milvus_benchmark/runners/accuracy.py", line 282, in run_case guarantee_timestamp=case_param["guarantee_timestamp"]) File "/src/milvus_benchmark/client.py", line 49, in wrapper result = func(*args, **kwargs) File "/src/milvus_benchmark/client.py", line 324, in query result = self._milvus.search(tmp_collection_name, **params) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 835, in search partition_names, output_fields, timeout, round_decimal, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 49, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 41, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 21, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 458, in search return self._execute_search_requests(requests, timeout, **_kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 405, in _execute_search_requests raise pre_err File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 393, in _execute_search_requests response = ft.result() File "/usr/local/lib/python3.6/site-packages/grpc/_channel.py", line 744, in result raise self grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "{"created":"@1649347880.774488333","description":"Error received from peer ipv4:10.96.124.71:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}" > (milvus_benchmark.main:98) [2022-04-07 16:11:20,786] [ INFO] - None (milvus_benchmark.main:99) [2022-04-07 16:11:20,788] [ DEBUG] - {'type': 'ann_accuracy', 'value': {}} (milvus_benchmark.main:107) [2022-04-07 16:11:20,791] [ DEBUG] - {'_version': '0.1', '_type': 'case', 'run_id': 1649348428, 'mode': 'local', 'server': <milvus_benchmark.metrics.models.server.Server object at 0x7f17f8867f98>, 'hardware': <milvus_benchmark.metrics.models.hardware.Hardware object at 0x7f17f8867eb8>, 'env': <milvus_benchmark.metrics.models.env.Env object at 0x7f17f8867dd8>, 'status': 'RUN_FAILED', 'err_message': '<_MultiThreadedRendezvous of RPC that terminated with:\n\tstatus = StatusCode.DEADLINE_EXCEEDED\n\tdetails = "Deadline Exceeded"\n\tdebug_error_string = "{"created":"@1649347880.774488333","description":"Error received from peer ipv4:10.96.124.71:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}"\n>\nTraceback (most recent call last):\n File "main.py", line 95, in run_suite\n result = runner.run_case(case_metric, **case)\n File "/src/milvus_benchmark/runners/accuracy.py", line 282, in run_case\n guarantee_timestamp=case_param["guarantee_timestamp"])\n File "/src/milvus_benchmark/client.py", line 49, in wrapper\n result = func(*args, **kwargs)\n File "/src/milvus_benchmark/client.py", line 324, in query\n result = self._milvus.search(tmp_collection_name, **params)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 835, in search\n partition_names, output_fields, timeout, round_decimal, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 49, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 41, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler\n return func(*args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 21, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 458, in search\n return self._execute_search_requests(requests, timeout, **_kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler\n return func(*args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 405, in _execute_search_requests\n raise pre_err\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 393, in _execute_search_requests\n response = ft.result()\n File "/usr/local/lib/python3.6/site-packages/grpc/_channel.py", line 744, in result\n raise self\ngrpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:\n\tstatus = StatusCode.DEADLINE_EXCEEDED\n\tdetails = "Deadline Exceeded"\n\tdebug_error_string = "{"created":"@1649347880.774488333","description":"Error received from peer ipv4:10.96.124.71:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}"\n>\n', 'collection': {'dimension': 128, 'metric_type': 'l2', 'dataset_name': 'sift_128_euclidean', 'shards_num': None}, 'index': {'index_type': 'flat', 'index_param': {'nlist': 1024}}, 'search': {'nq': 10000, 'topk': 10, 'search_param': {'nprobe': 1}, 'filter': [], 'guarantee_timestamp': None}, 'run_params': None, 'metrics': {'type': 'ann_accuracy', 'value': {}}, 'datetime': '2022-04-07 16:03:27.458830', 'type': 'metric'} (milvus_benchmark.metric.api:29) ``` server: ``` NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES benchmark-cron-1649347200-2-etcd-0 1/1 Running 0 3h57m 10.97.17.222 qa-node014.zilliz.local <none> <none> benchmark-cron-1649347200-2-milvus-standalone-7ddfc7fbfc-ndmfd 1/1 Running 0 3h57m 10.97.10.18 qa-node008.zilliz.local <none> <none> benchmark-cron-1649347200-2-minio-f9dccfb9c-b5fkw 1/1 Running 0 3h57m 10.97.16.67 qa-node013.zilliz.local <none> <none> ``` ### scene 3 test yaml: client-configmap:client-acc-sift-ivf-sq8,instanceId:3,server-configmap:server-single-8c16m client pod: benchmark-cron-1649347200-2815052319 client log: ``` [2022-04-07 16:05:19,601] [ DEBUG] - Row count: 1000000 in collection: <sift_128_euclidean> (milvus_benchmark.client:424) [2022-04-07 16:05:19,655] [ INFO] - Table: sift_128_euclidean, row count: 1000000 (milvus_benchmark.runners.accuracy:260) [2022-04-07 16:05:19,659] [ INFO] - None (milvus_benchmark.client:293) [2022-04-07 16:05:19,661] [ INFO] - Drop index: sift_128_euclidean (milvus_benchmark.client:304) [2022-04-07 16:05:19,664] [ INFO] - Re-create index: sift_128_euclidean (milvus_benchmark.runners.accuracy:265) [2022-04-07 16:05:19,751] [ INFO] - Building index start, collection_name: sift_128_euclidean, index_type: IVF_SQ8, metric_type: L2 (milvus_benchmark.client:276) [2022-04-07 16:05:19,752] [ INFO] - {'nlist': 1024} (milvus_benchmark.client:278) [2022-04-07 16:05:19,753] [ DEBUG] - collection: sift_128_euclidean Index params: {'index_type': 'IVF_SQ8', 'metric_type': 'L2', 'params': {'nlist': 1024}} (milvus_benchmark.client:284) [2022-04-07 16:06:07,539] [ DEBUG] - Building index done, collection_name: sift_128_euclidean, response: Status(code=0, message='') (milvus_benchmark.client:286) [2022-04-07 16:06:07,549] [ DEBUG] - Milvus create_index run in 47.7976s (milvus_benchmark.client:53) [2022-04-07 16:06:07,551] [ INFO] - None (milvus_benchmark.client:293) [2022-04-07 16:06:07,552] [ INFO] - {'index_type': 'flat', 'metric_type': None, 'index_param': None} (milvus_benchmark.runners.accuracy:267) [2022-04-07 16:06:07,552] [ INFO] - Start load collection: sift_128_euclidean (milvus_benchmark.runners.accuracy:268) [2022-04-07 16:06:07,571] [ DEBUG] - Milvus load_collection run in 0.0178s (milvus_benchmark.client:53) [2022-04-07 16:06:07,572] [ INFO] - End load collection: sift_128_euclidean (milvus_benchmark.runners.accuracy:271) [2022-04-07 16:06:07,617] [ INFO] - Start run case (milvus_benchmark.main:88) [2022-04-07 16:11:07,855] [ ERROR] - grpc RpcError: [_execute_search_requests], <_MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded>, <Time:{'RPC start': '2022-04-07 16:06:07.853407', 'gRPC error': '2022-04-07 16:11:07.855146'}> (pymilvus.decorators:81) [2022-04-07 16:11:07,856] [ ERROR] - grpc RpcError: [search], <_MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded>, <Time:{'RPC start': '2022-04-07 16:06:07.618333', 'gRPC error': '2022-04-07 16:11:07.856616'}> (pymilvus.decorators:81) [2022-04-07 16:11:07,861] [ ERROR] - Traceback (most recent call last): File "main.py", line 95, in run_suite result = runner.run_case(case_metric, **case) File "/src/milvus_benchmark/runners/accuracy.py", line 282, in run_case guarantee_timestamp=case_param["guarantee_timestamp"]) File "/src/milvus_benchmark/client.py", line 49, in wrapper result = func(*args, **kwargs) File "/src/milvus_benchmark/client.py", line 324, in query result = self._milvus.search(tmp_collection_name, **params) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 835, in search partition_names, output_fields, timeout, round_decimal, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 49, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 41, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 21, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 458, in search return self._execute_search_requests(requests, timeout, **_kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 405, in _execute_search_requests raise pre_err File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 393, in _execute_search_requests response = ft.result() File "/usr/local/lib/python3.6/site-packages/grpc/_channel.py", line 744, in result raise self grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "{"created":"@1649347867.853764926","description":"Error received from peer ipv4:10.96.240.114:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}" > (milvus_benchmark.main:98) [2022-04-07 16:11:07,862] [ INFO] - None (milvus_benchmark.main:99) [2022-04-07 16:11:07,862] [ DEBUG] - {'type': 'ann_accuracy', 'value': {}} (milvus_benchmark.main:107) [2022-04-07 16:11:07,863] [ DEBUG] - {'_version': '0.1', '_type': 'case', 'run_id': 1649355398, 'mode': 'local', 'server': <milvus_benchmark.metrics.models.server.Server object at 0x7fd3bc52ff98>, 'hardware': <milvus_benchmark.metrics.models.hardware.Hardware object at 0x7fd3bc52feb8>, 'env': <milvus_benchmark.metrics.models.env.Env object at 0x7fd3bc52fdd8>, 'status': 'RUN_FAILED', 'err_message': '<_MultiThreadedRendezvous of RPC that terminated with:\n\tstatus = StatusCode.DEADLINE_EXCEEDED\n\tdetails = "Deadline Exceeded"\n\tdebug_error_string = "{"created":"@1649347867.853764926","description":"Error received from peer ipv4:10.96.240.114:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}"\n>\nTraceback (most recent call last):\n File "main.py", line 95, in run_suite\n result = runner.run_case(case_metric, **case)\n File "/src/milvus_benchmark/runners/accuracy.py", line 282, in run_case\n guarantee_timestamp=case_param["guarantee_timestamp"])\n File "/src/milvus_benchmark/client.py", line 49, in wrapper\n result = func(*args, **kwargs)\n File "/src/milvus_benchmark/client.py", line 324, in query\n result = self._milvus.search(tmp_collection_name, **params)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 835, in search\n partition_names, output_fields, timeout, round_decimal, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 49, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 41, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler\n return func(*args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 21, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 458, in search\n return self._execute_search_requests(requests, timeout, **_kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler\n return func(*args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 405, in _execute_search_requests\n raise pre_err\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 393, in _execute_search_requests\n response = ft.result()\n File "/usr/local/lib/python3.6/site-packages/grpc/_channel.py", line 744, in result\n raise self\ngrpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:\n\tstatus = StatusCode.DEADLINE_EXCEEDED\n\tdetails = "Deadline Exceeded"\n\tdebug_error_string = "{"created":"@1649347867.853764926","description":"Error received from peer ipv4:10.96.240.114:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}"\n>\n', 'collection': {'dimension': 128, 'metric_type': 'l2', 'dataset_name': 'sift_128_euclidean', 'shards_num': None}, 'index': {'index_type': 'ivf_sq8', 'index_param': {'nlist': 1024}}, 'search': {'nq': 10000, 'topk': 10, 'search_param': {'nprobe': 1}, 'filter': [], 'guarantee_timestamp': None}, 'run_params': None, 'metrics': {'type': 'ann_accuracy', 'value': {}}, 'datetime': '2022-04-07 16:02:58.136148', 'type': 'metric'} (milvus_benchmark.metric.api:29) [2022-04-07 16:11:08,791] [ DEBUG] - Milvus query run in 0.8903s (milvus_benchmark.client:53) ``` server: ``` NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES benchmark-cron-1649347200-3-etcd-0 1/1 Running 0 3h57m 10.97.16.69 qa-node013.zilliz.local <none> <none> benchmark-cron-1649347200-3-milvus-standalone-5bdd8bb54f-8h4sh 1/1 Running 0 3h57m 10.97.17.208 qa-node014.zilliz.local <none> <none> benchmark-cron-1649347200-3-minio-6d59c84445-4x2nt 1/1 Running 0 3h57m 10.97.12.13 qa-node015.zilliz.local <none> <none> ``` ### scene 4 test yaml: client-configmap:client-acc-sift-hnsw,instanceId:4,server-configmap:server-single-8c16m client pod: benchmark-cron-1649347200-1294725969 client log: ``` [2022-04-07 16:04:54,864] [ DEBUG] - Row count: 1000000 in collection: <sift_128_euclidean> (milvus_benchmark.client:424) [2022-04-07 16:04:54,864] [ INFO] - Table: sift_128_euclidean, row count: 1000000 (milvus_benchmark.runners.accuracy:260) [2022-04-07 16:04:54,866] [ INFO] - None (milvus_benchmark.client:293) [2022-04-07 16:04:54,866] [ INFO] - Drop index: sift_128_euclidean (milvus_benchmark.client:304) [2022-04-07 16:04:54,868] [ INFO] - Re-create index: sift_128_euclidean (milvus_benchmark.runners.accuracy:265) [2022-04-07 16:04:54,869] [ INFO] - Building index start, collection_name: sift_128_euclidean, index_type: HNSW, metric_type: L2 (milvus_benchmark.client:276) [2022-04-07 16:04:54,869] [ INFO] - {'M': 16, 'efConstruction': 500} (milvus_benchmark.client:278) [2022-04-07 16:04:54,870] [ DEBUG] - collection: sift_128_euclidean Index params: {'index_type': 'HNSW', 'metric_type': 'L2', 'params': {'M': 16, 'efConstruction': 500}} (milvus_benchmark.client:284) [2022-04-07 16:09:18,805] [ DEBUG] - Building index done, collection_name: sift_128_euclidean, response: Status(code=0, message='') (milvus_benchmark.client:286) [2022-04-07 16:09:18,806] [ DEBUG] - Milvus create_index run in 263.937s (milvus_benchmark.client:53) [2022-04-07 16:09:18,808] [ INFO] - None (milvus_benchmark.client:293) [2022-04-07 16:09:18,808] [ INFO] - {'index_type': 'flat', 'metric_type': None, 'index_param': None} (milvus_benchmark.runners.accuracy:267) [2022-04-07 16:09:18,809] [ INFO] - Start load collection: sift_128_euclidean (milvus_benchmark.runners.accuracy:268) [2022-04-07 16:09:18,826] [ DEBUG] - Milvus load_collection run in 0.0176s (milvus_benchmark.client:53) [2022-04-07 16:09:18,827] [ INFO] - End load collection: sift_128_euclidean (milvus_benchmark.runners.accuracy:271) [2022-04-07 16:09:18,829] [ INFO] - Start run case (milvus_benchmark.main:88) [2022-04-07 16:14:19,007] [ ERROR] - grpc RpcError: [_execute_search_requests], <_MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded>, <Time:{'RPC start': '2022-04-07 16:09:19.005930', 'gRPC error': '2022-04-07 16:14:19.007620'}> (pymilvus.decorators:81) [2022-04-07 16:14:19,008] [ ERROR] - grpc RpcError: [search], <_MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded>, <Time:{'RPC start': '2022-04-07 16:09:18.829597', 'gRPC error': '2022-04-07 16:14:19.008847'}> (pymilvus.decorators:81) [2022-04-07 16:14:19,011] [ ERROR] - Traceback (most recent call last): File "main.py", line 95, in run_suite result = runner.run_case(case_metric, **case) File "/src/milvus_benchmark/runners/accuracy.py", line 282, in run_case guarantee_timestamp=case_param["guarantee_timestamp"]) File "/src/milvus_benchmark/client.py", line 49, in wrapper result = func(*args, **kwargs) File "/src/milvus_benchmark/client.py", line 324, in query result = self._milvus.search(tmp_collection_name, **params) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 835, in search partition_names, output_fields, timeout, round_decimal, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 49, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 41, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 21, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 458, in search return self._execute_search_requests(requests, timeout, **_kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 405, in _execute_search_requests raise pre_err File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 393, in _execute_search_requests response = ft.result() File "/usr/local/lib/python3.6/site-packages/grpc/_channel.py", line 744, in result raise self grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "{"created":"@1649348059.006772396","description":"Error received from peer ipv4:10.96.221.3:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}" > (milvus_benchmark.main:98) [2022-04-07 16:14:19,012] [ INFO] - None (milvus_benchmark.main:99) [2022-04-07 16:14:19,012] [ DEBUG] - {'type': 'ann_accuracy', 'value': {}} (milvus_benchmark.main:107) [2022-04-07 16:14:19,012] [ DEBUG] - {'_version': '0.1', '_type': 'case', 'run_id': 1649348560, 'mode': 'local', 'server': <milvus_benchmark.metrics.models.server.Server object at 0x7f2b4ca0efd0>, 'hardware': <milvus_benchmark.metrics.models.hardware.Hardware object at 0x7f2b4ca0eef0>, 'env': <milvus_benchmark.metrics.models.env.Env object at 0x7f2b4ca0ee10>, 'status': 'RUN_FAILED', 'err_message': '<_MultiThreadedRendezvous of RPC that terminated with:\n\tstatus = StatusCode.DEADLINE_EXCEEDED\n\tdetails = "Deadline Exceeded"\n\tdebug_error_string = "{"created":"@1649348059.006772396","description":"Error received from peer ipv4:10.96.221.3:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}"\n>\nTraceback (most recent call last):\n File "main.py", line 95, in run_suite\n result = runner.run_case(case_metric, **case)\n File "/src/milvus_benchmark/runners/accuracy.py", line 282, in run_case\n guarantee_timestamp=case_param["guarantee_timestamp"])\n File "/src/milvus_benchmark/client.py", line 49, in wrapper\n result = func(*args, **kwargs)\n File "/src/milvus_benchmark/client.py", line 324, in query\n result = self._milvus.search(tmp_collection_name, **params)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 835, in search\n partition_names, output_fields, timeout, round_decimal, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 49, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 41, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler\n return func(*args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 21, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 458, in search\n return self._execute_search_requests(requests, timeout, **_kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler\n return func(*args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 405, in _execute_search_requests\n raise pre_err\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 393, in _execute_search_requests\n response = ft.result()\n File "/usr/local/lib/python3.6/site-packages/grpc/_channel.py", line 744, in result\n raise self\ngrpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:\n\tstatus = StatusCode.DEADLINE_EXCEEDED\n\tdetails = "Deadline Exceeded"\n\tdebug_error_string = "{"created":"@1649348059.006772396","description":"Error received from peer ipv4:10.96.221.3:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}"\n>\n', 'collection': {'dimension': 128, 'metric_type': 'l2', 'dataset_name': 'sift_128_euclidean', 'shards_num': None}, 'index': {'index_type': 'hnsw', 'index_param': {'M': 16, 'efConstruction': 500}}, 'search': {'nq': 10000, 'topk': 10, 'search_param': {'ef': 16}, 'filter': [], 'guarantee_timestamp': None}, 'run_params': None, 'metrics': {'type': 'ann_accuracy', 'value': {}}, 'datetime': '2022-04-07 16:02:45.863579', 'type': 'metric'} (milvus_benchmark.metric.api:29) [2022-04-07 16:14:19,874] [ DEBUG] - Milvus query run in 0.8345s (milvus_benchmark.client:53) ``` server: ``` NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES benchmark-cron-1649347200-4-etcd-0 1/1 Running 0 3h57m 10.97.17.214 qa-node014.zilliz.local <none> <none> benchmark-cron-1649347200-4-milvus-standalone-65559bf9c4-v65m5 1/1 Running 0 3h57m 10.97.19.74 qa-node016.zilliz.local <none> <none> benchmark-cron-1649347200-4-minio-9f49dbd6f-22ppq 1/1 Running 0 3h57m 10.97.19.72 qa-node016.zilliz.local <none> <none> ``` ### scene 5 test yaml: client-configmap:client-search-sift50m-ivf-flat-2048,instanceId:6,server-configmap:server-single-16c64m client pod: benchmark-cron-1649347200-4136737022 client log: ``` [2022-04-07 17:28:25,859] [ DEBUG] - Row count: 50000000 in collection: <sift_50m_128_l2> (milvus_benchmark.client:424) [2022-04-07 17:28:25,860] [ INFO] - 50000000 (milvus_benchmark.runners.search:280) [2022-04-07 17:28:25,860] [ INFO] - Start load collection (milvus_benchmark.runners.search:281) [2022-04-07 17:28:25,984] [ DEBUG] - Milvus load_collection run in 0.1234s (milvus_benchmark.client:53) [2022-04-07 17:28:25,985] [ DEBUG] - {'load_time': 0.1242} (milvus_benchmark.runners.search:284) [2022-04-07 17:28:25,986] [ INFO] - Start run case (milvus_benchmark.main:88) [2022-04-07 17:28:25,986] [ DEBUG] - Start run query, run 1 of 30 (milvus_benchmark.runners.search:294) [2022-04-07 17:28:25,987] [ INFO] - {'nq': 1, 'topk': 1, 'search_param': {'nprobe': 8}, 'filter': [], 'guarantee_timestamp': None} (milvus_benchmark.runners.search:295) [2022-04-07 17:33:25,996] [ ERROR] - grpc RpcError: [_execute_search_requests], <_MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded>, <Time:{'RPC start': '2022-04-07 17:28:25.994302', 'gRPC error': '2022-04-07 17:33:25.995836'}> (pymilvus.decorators:81) [2022-04-07 17:33:26,002] [ ERROR] - grpc RpcError: [search], <_MultiThreadedRendezvous: StatusCode.DEADLINE_EXCEEDED, Deadline Exceeded>, <Time:{'RPC start': '2022-04-07 17:28:25.988694', 'gRPC error': '2022-04-07 17:33:26.002716'}> (pymilvus.decorators:81) [2022-04-07 17:33:26,009] [ ERROR] - Traceback (most recent call last): File "main.py", line 95, in run_suite result = runner.run_case(case_metric, **case) File "/src/milvus_benchmark/runners/search.py", line 298, in run_case guarantee_timestamp=case_param["guarantee_timestamp"]) File "/src/milvus_benchmark/client.py", line 49, in wrapper result = func(*args, **kwargs) File "/src/milvus_benchmark/client.py", line 324, in query result = self._milvus.search(tmp_collection_name, **params) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 835, in search partition_names, output_fields, timeout, round_decimal, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 49, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 41, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 21, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 458, in search return self._execute_search_requests(requests, timeout, **_kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 405, in _execute_search_requests raise pre_err File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 393, in _execute_search_requests response = ft.result() File "/usr/local/lib/python3.6/site-packages/grpc/_channel.py", line 744, in result raise self grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "{"created":"@1649352805.994898509","description":"Error received from peer ipv4:10.96.105.127:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}" > (milvus_benchmark.main:98) [2022-04-07 17:33:26,011] [ INFO] - None (milvus_benchmark.main:99) [2022-04-07 17:33:26,012] [ DEBUG] - {'type': 'insert_search_performance', 'value': {}} (milvus_benchmark.main:107) [2022-04-07 17:33:26,012] [ DEBUG] - {'_version': '0.1', '_type': 'case', 'run_id': 1649352312, 'mode': 'local', 'server': <milvus_benchmark.metrics.models.server.Server object at 0x7efdbc2c22e8>, 'hardware': <milvus_benchmark.metrics.models.hardware.Hardware object at 0x7efdbc2c26a0>, 'env': <milvus_benchmark.metrics.models.env.Env object at 0x7efdbc2c2940>, 'status': 'RUN_FAILED', 'err_message': '<_MultiThreadedRendezvous of RPC that terminated with:\n\tstatus = StatusCode.DEADLINE_EXCEEDED\n\tdetails = "Deadline Exceeded"\n\tdebug_error_string = "{"created":"@1649352805.994898509","description":"Error received from peer ipv4:10.96.105.127:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}"\n>\nTraceback (most recent call last):\n File "main.py", line 95, in run_suite\n result = runner.run_case(case_metric, **case)\n File "/src/milvus_benchmark/runners/search.py", line 298, in run_case\n guarantee_timestamp=case_param["guarantee_timestamp"])\n File "/src/milvus_benchmark/client.py", line 49, in wrapper\n result = func(*args, **kwargs)\n File "/src/milvus_benchmark/client.py", line 324, in query\n result = self._milvus.search(tmp_collection_name, **params)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 835, in search\n partition_names, output_fields, timeout, round_decimal, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 49, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 41, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler\n return func(*args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 21, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 458, in search\n return self._execute_search_requests(requests, timeout, **_kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 82, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/decorators.py", line 70, in handler\n return func(*args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 405, in _execute_search_requests\n raise pre_err\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 393, in _execute_search_requests\n response = ft.result()\n File "/usr/local/lib/python3.6/site-packages/grpc/_channel.py", line 744, in result\n raise self\ngrpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:\n\tstatus = StatusCode.DEADLINE_EXCEEDED\n\tdetails = "Deadline Exceeded"\n\tdebug_error_string = "{"created":"@1649352805.994898509","description":"Error received from peer ipv4:10.96.105.127:19530","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Deadline Exceeded","grpc_status":4}"\n>\n', 'collection': {'dimension': 128, 'metric_type': 'l2', 'dataset_name': 'sift_50m_128_l2', 'fields': None, 'shards_num': None}, 'index': {'index_type': 'ivf_flat', 'index_param': {'nlist': 2048}}, 'search': {'nq': 1, 'topk': 1, 'search_param': {'nprobe': 8}, 'filter': [], 'guarantee_timestamp': None}, 'run_params': None, 'metrics': {'type': 'insert_search_performance', 'value': {}}, 'datetime': '2022-04-07 16:02:57.589034', 'type': 'metric'} (milvus_benchmark.metric.api:29) ``` server: ``` NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES benchmark-cron-1649347200-6-etcd-0 1/1 Running 0 3h57m 10.97.17.213 qa-node014.zilliz.local <none> <none> benchmark-cron-1649347200-6-milvus-standalone-76ddb48b6-2dh8b 1/1 Running 0 3h57m 10.97.20.230 qa-node018.zilliz.local <none> <none> benchmark-cron-1649347200-6-minio-6d7c6d494c-d8hpm 1/1 Running 0 3h57m 10.97.19.73 qa-node016.zilliz.local <none> <none> ``` ### Expected Behavior _No response_ ### Steps To Reproduce ```markdown scene 1、2、3、4: template: ann_accuracy steps: 1、create collection 2、insert vectors 3、flush collection 4、build index 5、load collection 6、search 《- raise error scene 15: template: insert_search_performance steps: 1、create collection 2、build index 3、insert vectors 4、flush collection 5、build index again 6、load collection 7、search 《- raise error ``` ### Anything else? _No response_
non_priority
rpcerror in the first search multithreadedrendezvous statuscode deadline exceeded deadline exceeded is there an existing issue for this i have searched the existing issues environment markdown milvus version master deployment mode standalone or cluster standalone sdk version e g pymilvus os ubuntu or centos cpu memory gpu others current behavior argo task benchmark cron scene test yaml client configmap client acc sift ivf flat instanceid server configmap server single client pod benchmark cron client log end insert start flush milvus benchmark runners accuracy milvus flush run in milvus benchmark client end flush milvus benchmark runners accuracy row count in collection milvus benchmark client table sift euclidean row count milvus benchmark runners accuracy none milvus benchmark client drop index sift euclidean milvus benchmark client re create index sift euclidean milvus benchmark runners accuracy building index start collection name sift euclidean index type ivf flat metric type milvus benchmark client nlist milvus benchmark client collection sift euclidean index params index type ivf flat metric type params nlist milvus benchmark client building index done collection name sift euclidean response status code message milvus benchmark client milvus create index run in milvus benchmark client none milvus benchmark client index type flat metric type none index param none milvus benchmark runners accuracy start load collection sift euclidean milvus benchmark runners accuracy milvus load collection run in milvus benchmark client end load collection sift euclidean milvus benchmark runners accuracy start run case milvus benchmark main grpc rpcerror pymilvus decorators grpc rpcerror pymilvus decorators traceback most recent call last file main py line in run suite result runner run case case metric case file src milvus benchmark runners accuracy py line in run case guarantee timestamp case param file src milvus benchmark client py line in wrapper result func args kwargs file src milvus benchmark client py line in query result self milvus search tmp collection name params file usr local lib site packages pymilvus client stub py line in search partition names output fields timeout round decimal kwargs file usr local lib site packages pymilvus decorators py line in handler raise e file usr local lib site packages pymilvus decorators py line in handler return func self args kwargs file usr local lib site packages pymilvus decorators py line in handler raise e file usr local lib site packages pymilvus decorators py line in handler return func args kwargs file usr local lib site packages pymilvus decorators py line in handler return func self args kwargs file usr local lib site packages pymilvus client grpc handler py line in search return self execute search requests requests timeout kwargs file usr local lib site packages pymilvus decorators py line in handler raise e file usr local lib site packages pymilvus decorators py line in handler return func args kwargs file usr local lib site packages pymilvus client grpc handler py line in execute search requests raise pre err file usr local lib site packages pymilvus client grpc handler py line in execute search requests response ft result file usr local lib site packages grpc channel py line in result raise self grpc channel multithreadedrendezvous multithreadedrendezvous of rpc that terminated with status statuscode deadline exceeded details deadline exceeded debug error string created description error received from peer file src core lib surface call cc file line grpc message deadline exceeded grpc status milvus benchmark main none milvus benchmark main type ann accuracy value milvus benchmark main version type case run id mode local server hardware env status run failed err message ntraceback most recent call last n file main py line in run suite n result runner run case case metric case n file src milvus benchmark runners accuracy py line in run case n guarantee timestamp case param n file src milvus benchmark client py line in wrapper n result func args kwargs n file src milvus benchmark client py line in query n result self milvus search tmp collection name params n file usr local lib site packages pymilvus client stub py line in search n partition names output fields timeout round decimal kwargs n file usr local lib site packages pymilvus decorators py line in handler n raise e n file usr local lib site packages pymilvus decorators py line in handler n return func self args kwargs n file usr local lib site packages pymilvus decorators py line in handler n raise e n file usr local lib site packages pymilvus decorators py line in handler n return func args kwargs n file usr local lib site packages pymilvus decorators py line in handler n return func self args kwargs n file usr local lib site packages pymilvus client grpc handler py line in search n return self execute search requests requests timeout kwargs n file usr local lib site packages pymilvus decorators py line in handler n raise e n file usr local lib site packages pymilvus decorators py line in handler n return func args kwargs n file usr local lib site packages pymilvus client grpc handler py line in execute search requests n raise pre err n file usr local lib site packages pymilvus client grpc handler py line in execute search requests n response ft result n file usr local lib site packages grpc channel py line in result n raise self ngrpc channel multithreadedrendezvous n collection dimension metric type dataset name sift euclidean shards num none index index type ivf flat index param nlist search nq topk search param nprobe filter guarantee timestamp none run params none metrics type ann accuracy value datetime type metric milvus benchmark metric api milvus query run in milvus benchmark client server name ready status restarts age ip node nominated node readiness gates benchmark cron etcd running qa zilliz local benchmark cron milvus standalone vfzwm running qa zilliz local benchmark cron minio running qa zilliz local scene test yaml client configmap client acc sift flat instanceid server configmap server single client pod benchmark cron client log end flush milvus benchmark runners accuracy row count in collection milvus benchmark client table sift euclidean row count milvus benchmark runners accuracy none milvus benchmark client drop index sift euclidean milvus benchmark client re create index sift euclidean milvus benchmark runners accuracy building index start collection name sift euclidean index type flat metric type milvus benchmark client nlist milvus benchmark client collection sift euclidean index params index type flat metric type params nlist milvus benchmark client building index done collection name sift euclidean response status code message warning it is not necessary to build index with index type flat milvus benchmark client milvus create index run in milvus benchmark client none milvus benchmark client index type flat metric type none index param none milvus benchmark runners accuracy start load collection sift euclidean milvus benchmark runners accuracy milvus load collection run in milvus benchmark client end load collection sift euclidean milvus benchmark runners accuracy start run case milvus benchmark main grpc rpcerror pymilvus decorators grpc rpcerror pymilvus decorators traceback most recent call last file main py line in run suite result runner run case case metric case file src milvus benchmark runners accuracy py line in run case guarantee timestamp case param file src milvus benchmark client py line in wrapper result func args kwargs file src milvus benchmark client py line in query result self milvus search tmp collection name params file usr local lib site packages pymilvus client stub py line in search partition names output fields timeout round decimal kwargs file usr local lib site packages pymilvus decorators py line in handler raise e file usr local lib site packages pymilvus decorators py line in handler return func self args kwargs file usr local lib site packages pymilvus decorators py line in handler raise e file usr local lib site packages pymilvus decorators py line in handler return func args kwargs file usr local lib site packages pymilvus decorators py line in handler return func self args kwargs file usr local lib site packages pymilvus client grpc handler py line in search return self execute search requests requests timeout kwargs file usr local lib site packages pymilvus decorators py line in handler raise e file usr local lib site packages pymilvus decorators py line in handler return func args kwargs file usr local lib site packages pymilvus client grpc handler py line in execute search requests raise pre err file usr local lib site packages pymilvus client grpc handler py line in execute search requests response ft result file usr local lib site packages grpc channel py line in result raise self grpc channel multithreadedrendezvous multithreadedrendezvous of rpc that terminated with status statuscode deadline exceeded details deadline exceeded debug error string created description error received from peer file src core lib surface call cc file line grpc message deadline exceeded grpc status milvus benchmark main none milvus benchmark main type ann accuracy value milvus benchmark main version type case run id mode local server hardware env status run failed err message ntraceback most recent call last n file main py line in run suite n result runner run case case metric case n file src milvus benchmark runners accuracy py line in run case n guarantee timestamp case param n file src milvus benchmark client py line in wrapper n result func args kwargs n file src milvus benchmark client py line in query n result self milvus search tmp collection name params n file usr local lib site packages pymilvus client stub py line in search n partition names output fields timeout round decimal kwargs n file usr local lib site packages pymilvus decorators py line in handler n raise e n file usr local lib site packages pymilvus decorators py line in handler n return func self args kwargs n file usr local lib site packages pymilvus decorators py line in handler n raise e n file usr local lib site packages pymilvus decorators py line in handler n return func args kwargs n file usr local lib site packages pymilvus decorators py line in handler n return func self args kwargs n file usr local lib site packages pymilvus client grpc handler py line in search n return self execute search requests requests timeout kwargs n file usr local lib site packages pymilvus decorators py line in handler n raise e n file usr local lib site packages pymilvus decorators py line in handler n return func args kwargs n file usr local lib site packages pymilvus client grpc handler py line in execute search requests n raise pre err n file usr local lib site packages pymilvus client grpc handler py line in execute search requests n response ft result n file usr local lib site packages grpc channel py line in result n raise self ngrpc channel multithreadedrendezvous n collection dimension metric type dataset name sift euclidean shards num none index index type flat index param nlist search nq topk search param nprobe filter guarantee timestamp none run params none metrics type ann accuracy value datetime type metric milvus benchmark metric api server name ready status restarts age ip node nominated node readiness gates benchmark cron etcd running qa zilliz local benchmark cron milvus standalone ndmfd running qa zilliz local benchmark cron minio running qa zilliz local scene test yaml client configmap client acc sift ivf instanceid server configmap server single client pod benchmark cron client log row count in collection milvus benchmark client table sift euclidean row count milvus benchmark runners accuracy none milvus benchmark client drop index sift euclidean milvus benchmark client re create index sift euclidean milvus benchmark runners accuracy building index start collection name sift euclidean index type ivf metric type milvus benchmark client nlist milvus benchmark client collection sift euclidean index params index type ivf metric type params nlist milvus benchmark client building index done collection name sift euclidean response status code message milvus benchmark client milvus create index run in milvus benchmark client none milvus benchmark client index type flat metric type none index param none milvus benchmark runners accuracy start load collection sift euclidean milvus benchmark runners accuracy milvus load collection run in milvus benchmark client end load collection sift euclidean milvus benchmark runners accuracy start run case milvus benchmark main grpc rpcerror pymilvus decorators grpc rpcerror pymilvus decorators traceback most recent call last file main py line in run suite result runner run case case metric case file src milvus benchmark runners accuracy py line in run case guarantee timestamp case param file src milvus benchmark client py line in wrapper result func args kwargs file src milvus benchmark client py line in query result self milvus search tmp collection name params file usr local lib site packages pymilvus client stub py line in search partition names output fields timeout round decimal kwargs file usr local lib site packages pymilvus decorators py line in handler raise e file usr local lib site packages pymilvus decorators py line in handler return func self args kwargs file usr local lib site packages pymilvus decorators py line in handler raise e file usr local lib site packages pymilvus decorators py line in handler return func args kwargs file usr local lib site packages pymilvus decorators py line in handler return func self args kwargs file usr local lib site packages pymilvus client grpc handler py line in search return self execute search requests requests timeout kwargs file usr local lib site packages pymilvus decorators py line in handler raise e file usr local lib site packages pymilvus decorators py line in handler return func args kwargs file usr local lib site packages pymilvus client grpc handler py line in execute search requests raise pre err file usr local lib site packages pymilvus client grpc handler py line in execute search requests response ft result file usr local lib site packages grpc channel py line in result raise self grpc channel multithreadedrendezvous multithreadedrendezvous of rpc that terminated with status statuscode deadline exceeded details deadline exceeded debug error string created description error received from peer file src core lib surface call cc file line grpc message deadline exceeded grpc status milvus benchmark main none milvus benchmark main type ann accuracy value milvus benchmark main version type case run id mode local server hardware env status run failed err message ntraceback most recent call last n file main py line in run suite n result runner run case case metric case n file src milvus benchmark runners accuracy py line in run case n guarantee timestamp case param n file src milvus benchmark client py line in wrapper n result func args kwargs n file src milvus benchmark client py line in query n result self milvus search tmp collection name params n file usr local lib site packages pymilvus client stub py line in search n partition names output fields timeout round decimal kwargs n file usr local lib site packages pymilvus decorators py line in handler n raise e n file usr local lib site packages pymilvus decorators py line in handler n return func self args kwargs n file usr local lib site packages pymilvus decorators py line in handler n raise e n file usr local lib site packages pymilvus decorators py line in handler n return func args kwargs n file usr local lib site packages pymilvus decorators py line in handler n return func self args kwargs n file usr local lib site packages pymilvus client grpc handler py line in search n return self execute search requests requests timeout kwargs n file usr local lib site packages pymilvus decorators py line in handler n raise e n file usr local lib site packages pymilvus decorators py line in handler n return func args kwargs n file usr local lib site packages pymilvus client grpc handler py line in execute search requests n raise pre err n file usr local lib site packages pymilvus client grpc handler py line in execute search requests n response ft result n file usr local lib site packages grpc channel py line in result n raise self ngrpc channel multithreadedrendezvous n collection dimension metric type dataset name sift euclidean shards num none index index type ivf index param nlist search nq topk search param nprobe filter guarantee timestamp none run params none metrics type ann accuracy value datetime type metric milvus benchmark metric api milvus query run in milvus benchmark client server name ready status restarts age ip node nominated node readiness gates benchmark cron etcd running qa zilliz local benchmark cron milvus standalone running qa zilliz local benchmark cron minio running qa zilliz local scene test yaml client configmap client acc sift hnsw instanceid server configmap server single client pod benchmark cron client log row count in collection milvus benchmark client table sift euclidean row count milvus benchmark runners accuracy none milvus benchmark client drop index sift euclidean milvus benchmark client re create index sift euclidean milvus benchmark runners accuracy building index start collection name sift euclidean index type hnsw metric type milvus benchmark client m efconstruction milvus benchmark client collection sift euclidean index params index type hnsw metric type params m efconstruction milvus benchmark client building index done collection name sift euclidean response status code message milvus benchmark client milvus create index run in milvus benchmark client none milvus benchmark client index type flat metric type none index param none milvus benchmark runners accuracy start load collection sift euclidean milvus benchmark runners accuracy milvus load collection run in milvus benchmark client end load collection sift euclidean milvus benchmark runners accuracy start run case milvus benchmark main grpc rpcerror pymilvus decorators grpc rpcerror pymilvus decorators traceback most recent call last file main py line in run suite result runner run case case metric case file src milvus benchmark runners accuracy py line in run case guarantee timestamp case param file src milvus benchmark client py line in wrapper result func args kwargs file src milvus benchmark client py line in query result self milvus search tmp collection name params file usr local lib site packages pymilvus client stub py line in search partition names output fields timeout round decimal kwargs file usr local lib site packages pymilvus decorators py line in handler raise e file usr local lib site packages pymilvus decorators py line in handler return func self args kwargs file usr local lib site packages pymilvus decorators py line in handler raise e file usr local lib site packages pymilvus decorators py line in handler return func args kwargs file usr local lib site packages pymilvus decorators py line in handler return func self args kwargs file usr local lib site packages pymilvus client grpc handler py line in search return self execute search requests requests timeout kwargs file usr local lib site packages pymilvus decorators py line in handler raise e file usr local lib site packages pymilvus decorators py line in handler return func args kwargs file usr local lib site packages pymilvus client grpc handler py line in execute search requests raise pre err file usr local lib site packages pymilvus client grpc handler py line in execute search requests response ft result file usr local lib site packages grpc channel py line in result raise self grpc channel multithreadedrendezvous multithreadedrendezvous of rpc that terminated with status statuscode deadline exceeded details deadline exceeded debug error string created description error received from peer file src core lib surface call cc file line grpc message deadline exceeded grpc status milvus benchmark main none milvus benchmark main type ann accuracy value milvus benchmark main version type case run id mode local server hardware env status run failed err message ntraceback most recent call last n file main py line in run suite n result runner run case case metric case n file src milvus benchmark runners accuracy py line in run case n guarantee timestamp case param n file src milvus benchmark client py line in wrapper n result func args kwargs n file src milvus benchmark client py line in query n result self milvus search tmp collection name params n file usr local lib site packages pymilvus client stub py line in search n partition names output fields timeout round decimal kwargs n file usr local lib site packages pymilvus decorators py line in handler n raise e n file usr local lib site packages pymilvus decorators py line in handler n return func self args kwargs n file usr local lib site packages pymilvus decorators py line in handler n raise e n file usr local lib site packages pymilvus decorators py line in handler n return func args kwargs n file usr local lib site packages pymilvus decorators py line in handler n return func self args kwargs n file usr local lib site packages pymilvus client grpc handler py line in search n return self execute search requests requests timeout kwargs n file usr local lib site packages pymilvus decorators py line in handler n raise e n file usr local lib site packages pymilvus decorators py line in handler n return func args kwargs n file usr local lib site packages pymilvus client grpc handler py line in execute search requests n raise pre err n file usr local lib site packages pymilvus client grpc handler py line in execute search requests n response ft result n file usr local lib site packages grpc channel py line in result n raise self ngrpc channel multithreadedrendezvous n collection dimension metric type dataset name sift euclidean shards num none index index type hnsw index param m efconstruction search nq topk search param ef filter guarantee timestamp none run params none metrics type ann accuracy value datetime type metric milvus benchmark metric api milvus query run in milvus benchmark client server name ready status restarts age ip node nominated node readiness gates benchmark cron etcd running qa zilliz local benchmark cron milvus standalone running qa zilliz local benchmark cron minio running qa zilliz local scene test yaml client configmap client search ivf flat instanceid server configmap server single client pod benchmark cron client log row count in collection milvus benchmark client milvus benchmark runners search start load collection milvus benchmark runners search milvus load collection run in milvus benchmark client load time milvus benchmark runners search start run case milvus benchmark main start run query run of milvus benchmark runners search nq topk search param nprobe filter guarantee timestamp none milvus benchmark runners search grpc rpcerror pymilvus decorators grpc rpcerror pymilvus decorators traceback most recent call last file main py line in run suite result runner run case case metric case file src milvus benchmark runners search py line in run case guarantee timestamp case param file src milvus benchmark client py line in wrapper result func args kwargs file src milvus benchmark client py line in query result self milvus search tmp collection name params file usr local lib site packages pymilvus client stub py line in search partition names output fields timeout round decimal kwargs file usr local lib site packages pymilvus decorators py line in handler raise e file usr local lib site packages pymilvus decorators py line in handler return func self args kwargs file usr local lib site packages pymilvus decorators py line in handler raise e file usr local lib site packages pymilvus decorators py line in handler return func args kwargs file usr local lib site packages pymilvus decorators py line in handler return func self args kwargs file usr local lib site packages pymilvus client grpc handler py line in search return self execute search requests requests timeout kwargs file usr local lib site packages pymilvus decorators py line in handler raise e file usr local lib site packages pymilvus decorators py line in handler return func args kwargs file usr local lib site packages pymilvus client grpc handler py line in execute search requests raise pre err file usr local lib site packages pymilvus client grpc handler py line in execute search requests response ft result file usr local lib site packages grpc channel py line in result raise self grpc channel multithreadedrendezvous multithreadedrendezvous of rpc that terminated with status statuscode deadline exceeded details deadline exceeded debug error string created description error received from peer file src core lib surface call cc file line grpc message deadline exceeded grpc status milvus benchmark main none milvus benchmark main type insert search performance value milvus benchmark main version type case run id mode local server hardware env status run failed err message ntraceback most recent call last n file main py line in run suite n result runner run case case metric case n file src milvus benchmark runners search py line in run case n guarantee timestamp case param n file src milvus benchmark client py line in wrapper n result func args kwargs n file src milvus benchmark client py line in query n result self milvus search tmp collection name params n file usr local lib site packages pymilvus client stub py line in search n partition names output fields timeout round decimal kwargs n file usr local lib site packages pymilvus decorators py line in handler n raise e n file usr local lib site packages pymilvus decorators py line in handler n return func self args kwargs n file usr local lib site packages pymilvus decorators py line in handler n raise e n file usr local lib site packages pymilvus decorators py line in handler n return func args kwargs n file usr local lib site packages pymilvus decorators py line in handler n return func self args kwargs n file usr local lib site packages pymilvus client grpc handler py line in search n return self execute search requests requests timeout kwargs n file usr local lib site packages pymilvus decorators py line in handler n raise e n file usr local lib site packages pymilvus decorators py line in handler n return func args kwargs n file usr local lib site packages pymilvus client grpc handler py line in execute search requests n raise pre err n file usr local lib site packages pymilvus client grpc handler py line in execute search requests n response ft result n file usr local lib site packages grpc channel py line in result n raise self ngrpc channel multithreadedrendezvous n collection dimension metric type dataset name sift fields none shards num none index index type ivf flat index param nlist search nq topk search param nprobe filter guarantee timestamp none run params none metrics type insert search performance value datetime type metric milvus benchmark metric api server name ready status restarts age ip node nominated node readiness gates benchmark cron etcd running qa zilliz local benchmark cron milvus standalone running qa zilliz local benchmark cron minio running qa zilliz local expected behavior no response steps to reproduce markdown scene 、 、 、 template ann accuracy steps 、create collection 、insert vectors 、flush collection 、build index 、load collection 、search 《 raise error scene template insert search performance steps 、create collection 、build index 、insert vectors 、flush collection 、build index again 、load collection 、search 《 raise error anything else no response
0
254,279
8,072,150,187
IssuesEvent
2018-08-06 15:09:25
InfiniteFlightAirportEditing/Navigation
https://api.github.com/repos/InfiniteFlightAirportEditing/Navigation
opened
RKSO
ILS Priority ILS Error
# Summary of Navigation issue: (ILS, Waypoints, etc.) Runway added # Runway(s) with ILS: (If any) 09L,27R # References: (Official sources/charts) http://www.ut.ac.kr/cmm/fms/FileDown.do;jsessionid=F40rEV2kMe2yMLa2aeLPmqd2xnEECA3J1HRH8JRkqPX5THuZMml8yn8YHwzwk9MY.homepage_servlet_engine1?atchFileId=FILE_000000000106643&fileSn=2
1.0
RKSO - # Summary of Navigation issue: (ILS, Waypoints, etc.) Runway added # Runway(s) with ILS: (If any) 09L,27R # References: (Official sources/charts) http://www.ut.ac.kr/cmm/fms/FileDown.do;jsessionid=F40rEV2kMe2yMLa2aeLPmqd2xnEECA3J1HRH8JRkqPX5THuZMml8yn8YHwzwk9MY.homepage_servlet_engine1?atchFileId=FILE_000000000106643&fileSn=2
priority
rkso summary of navigation issue ils waypoints etc runway added runway s with ils if any references official sources charts
1
270,523
28,962,283,142
IssuesEvent
2023-05-10 04:19:55
nidhi7598/external_curl_AOSP10_r33
https://api.github.com/repos/nidhi7598/external_curl_AOSP10_r33
opened
CVE-2021-22926 (High) detected in curlcurl-7_64_1
Mend: dependency security vulnerability
## CVE-2021-22926 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>curlcurl-7_64_1</b></p></summary> <p> <p>A command line tool and library for transferring data with URL syntax, supporting HTTP, HTTPS, FTP, FTPS, GOPHER, TFTP, SCP, SFTP, SMB, TELNET, DICT, LDAP, LDAPS, FILE, IMAP, SMTP, POP3, RTSP and RTMP. libcurl offers a myriad of powerful features</p> <p>Library home page: <a href=https://github.com/curl/curl.git>https://github.com/curl/curl.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/external_curl_AOSP10_r33/commit/481a49fc7dbc30e43cd670ab40fa6cca41715464">481a49fc7dbc30e43cd670ab40fa6cca41715464</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/lib/vtls/sectransp.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> libcurl-using applications can ask for a specific client certificate to be used in a transfer. This is done with the `CURLOPT_SSLCERT` option (`--cert` with the command line tool).When libcurl is built to use the macOS native TLS library Secure Transport, an application can ask for the client certificate by name or with a file name - using the same option. If the name exists as a file, it will be used instead of by name.If the appliction runs with a current working directory that is writable by other users (like `/tmp`), a malicious user can create a file name with the same name as the app wants to use by name, and thereby trick the application to use the file based cert instead of the one referred to by name making libcurl send the wrong client certificate in the TLS connection handshake. <p>Publish Date: 2021-08-05 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-22926>CVE-2021-22926</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://curl.se/docs/CVE-2021-22926.html">https://curl.se/docs/CVE-2021-22926.html</a></p> <p>Release Date: 2021-08-05</p> <p>Fix Resolution: curl-7_78_0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-22926 (High) detected in curlcurl-7_64_1 - ## CVE-2021-22926 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>curlcurl-7_64_1</b></p></summary> <p> <p>A command line tool and library for transferring data with URL syntax, supporting HTTP, HTTPS, FTP, FTPS, GOPHER, TFTP, SCP, SFTP, SMB, TELNET, DICT, LDAP, LDAPS, FILE, IMAP, SMTP, POP3, RTSP and RTMP. libcurl offers a myriad of powerful features</p> <p>Library home page: <a href=https://github.com/curl/curl.git>https://github.com/curl/curl.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/external_curl_AOSP10_r33/commit/481a49fc7dbc30e43cd670ab40fa6cca41715464">481a49fc7dbc30e43cd670ab40fa6cca41715464</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/lib/vtls/sectransp.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> libcurl-using applications can ask for a specific client certificate to be used in a transfer. This is done with the `CURLOPT_SSLCERT` option (`--cert` with the command line tool).When libcurl is built to use the macOS native TLS library Secure Transport, an application can ask for the client certificate by name or with a file name - using the same option. If the name exists as a file, it will be used instead of by name.If the appliction runs with a current working directory that is writable by other users (like `/tmp`), a malicious user can create a file name with the same name as the app wants to use by name, and thereby trick the application to use the file based cert instead of the one referred to by name making libcurl send the wrong client certificate in the TLS connection handshake. <p>Publish Date: 2021-08-05 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-22926>CVE-2021-22926</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://curl.se/docs/CVE-2021-22926.html">https://curl.se/docs/CVE-2021-22926.html</a></p> <p>Release Date: 2021-08-05</p> <p>Fix Resolution: curl-7_78_0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in curlcurl cve high severity vulnerability vulnerable library curlcurl a command line tool and library for transferring data with url syntax supporting http https ftp ftps gopher tftp scp sftp smb telnet dict ldap ldaps file imap smtp rtsp and rtmp libcurl offers a myriad of powerful features library home page a href found in head commit a href found in base branch main vulnerable source files lib vtls sectransp c vulnerability details libcurl using applications can ask for a specific client certificate to be used in a transfer this is done with the curlopt sslcert option cert with the command line tool when libcurl is built to use the macos native tls library secure transport an application can ask for the client certificate by name or with a file name using the same option if the name exists as a file it will be used instead of by name if the appliction runs with a current working directory that is writable by other users like tmp a malicious user can create a file name with the same name as the app wants to use by name and thereby trick the application to use the file based cert instead of the one referred to by name making libcurl send the wrong client certificate in the tls connection handshake publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution curl step up your open source security game with mend
0